Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

Database over 2GB in MongoDB

We've got a file-based program we want to convert to use a document database, specifically MongoDB. Problem is, MongoDB is limited to 2GB on 32-bit machines (according to http://www.mongodb.org/display/DOCS/FAQ#FAQ-Whatarethe32bitlimitations%3F), and a lot of our users will have over 2GB of data. Is there a way to have MongoDB use more than one file somehow?

I thought perhaps I could implement sharding on a single machine, meaning I'd run more than one mongod on the same machine and they'd somehow communicate. Could that work?

like image 544
configurator Avatar asked Jun 07 '10 10:06

configurator


People also ask

How much data can be stored in a MongoDB database?

The maximum size an individual document can be in MongoDB is 16MB with a nested depth of 100 levels. Edit: There is no max size for an individual MongoDB database.

How much is 1gb in MongoDB?

Once the 1 GB of free data usage has been consumed, MongoDB bills Charts data transfers at a fixed rate of $1.00 per GB.

Can MongoDB store large data?

Large objects, or "files", are easily stored in MongoDB. It is no problem to store 100MB videos in the database. This has a number of advantages over files stored in a file system. Unlike a file system, the database will have no problem dealing with millions of objects.

Can MongoDB handle petabytes of data?

Over a third of the Fortune 100 and many of the most successful and innovative web companies rely on MongoDB. They've grown from single server deployments to clusters with over 1,000 nodes, delivering millions of operations per second on over 100 billion documents and petabytes of data.


1 Answers

The only way to have more than 2GB on a single node is to run multiple mongod processes. So sharding is one option (like you said) or doing some manual partitioning across processes.

like image 90
mdirolf Avatar answered Sep 24 '22 15:09

mdirolf