[01:39:49] <patforg> I get this error when ever I try to do something "Can't take a write lock while out of disk space" even when deleting to make free space
[01:40:51] <patforg> I thought my sharding was working... turns out gridfs doesn't shard by default
[01:41:08] <patforg> eventhough I enable sharding on my DB
[01:46:26] <patforg> any ideas on what should one do?
[03:30:08] <vrim> is there some metric collection in the mongo java driver - something like application success/error count?
[04:08:12] <therealkasey> anyone know if it's possible to stop an index building operation?
[04:08:37] <therealkasey> I set up a unique index on a field in a collection that already had duplicates before I started the operation
[04:32:14] <dstorrs> I've built a job manager using Mongo as the datastore.
[04:32:26] <dstorrs> it's for a harvesting system that writes data about YouTube videos.
[04:32:41] <ron> well, just because people are around, doesn't mean they're going to answer ;)
[04:32:46] <dstorrs> I'm thinking about breaking the "job control" collections out into a separate mongod from the data
[04:33:15] <dstorrs> but I'm not sure that's a good idea. This post (http://stackoverflow.com/questions/9203418/mongodb-sharding-on-one-machine) has Derick saying that mongod's will fight for all RAM.
[04:33:35] <dstorrs> however, the jobs collections are a very small set (a few dozen megs at most), so
[04:33:52] <dstorrs> I'm wondering if that's really "all ram on box" or "all ram needed for dataset"
[04:33:58] <ron> why use a different mongod for it?
[04:34:19] <dstorrs> because we are currently bottlenecking on the write lock.
[04:34:41] <dstorrs> this is based on extensive profiling / experimentation.
[04:40:40] <dstorrs> these are feeds that we collected but have never had the disk space to use until now.
[04:41:46] <dstorrs> I reused the harvester architecture to pull them out of CloudFiles, untar / split / queue up harvest jobs / process them. It works so well that it ends up with a bazillion harvest jobs and it chokes out.
[04:42:13] <dstorrs> I've added a usleep(125,000) and that helps a lot, but that's insane.
[06:47:40] <edmund> hi. i have a question about upsert with modifiers.. what i would really like to do is 'set a path through the matched object' .. e.g. if i do soemthing like upsert {id:1} {$set: {a: {$set: {b: 5}}}, then {id:5, a:{c:7}} becomes {id:5, a:{b:5,c:7}}
[06:47:54] <edmund> from poking around a bit it seems like this isnt possible. or am i doing something horribly wrong?
[06:48:30] <edmund> it's okay i just went to freenode $mongodb
[07:16:04] <Mortah> somethng like... infinite loop: each time, pick a random user and do the update/queries they need... repeat
[07:16:30] <Mortah> maybe add some bias to the probability for each user to get 'heavy usage' / 'low usage'
[07:16:35] <circlicious> ok i will try it out, i hope it works well, else my heart will break :P already made most of the transition fo this part of app from mysql to mongodb
[07:49:09] <tknz> can someone help me with this. I'm honestly not sure what I'm doing but I'm trying to group by a keyfunction (the year part of the date field). http://pastebin.com/TegGyTDS
[09:37:36] <woozly> but I ensure indexes and other..
[09:54:58] <IAD> woozly: be sure, what indexes are correct for your find operations. and use profiler http://www.mongodb.org/display/DOCS/Database+Profiler#DatabaseProfiler-Viewslowoperationsonly
[09:58:30] <tomoyuki28jp> Does MongoDB have a function like mysql's enum type?
[10:01:02] <woozly> I see while quering, mongod CPU% usage 99-100%
[10:01:12] <woozly> what it can be? data stored = 5gb
[10:02:01] <kali> woozly: look for long running queries in db.currentOp()
[10:03:31] <circlicious> how do you manage things like indexing ? do you just store all indexing queries in a txt file and then simply execute each of them in mongo shell of server(production)?
[12:00:10] <fredix> when I try to use mongo::fromjson compil fail on this error : /home/fred/Dropbox/code/nodecast/ncw/externals/mongodb/db/mongomutex.h:235: error: call of overloaded 'msgasserted(int, std::basic_string<char>)' is ambiguous
[13:17:35] <jpadilla> hey guys, quick question. I have a map loading a couple of markers, I also have that data on Mongo with the geospatial index. I'm querying using within box, but the results I'm getting are not "complete". In the map I'm seeing at least 5, in the mongo query I'm being returned one.
[13:32:39] <jpadilla> NodeX: I'm using the boundaries of the whole map. This is the markers I'm seeing on the map, and only one result of the query on the side https://img.skitch.com/20120724-twq64b3k9i34s5qqc51f3qtmwd.jpg
[14:53:15] <kylefaris> I did something like this: var mongo = require('mongojs').connect('localhost', ['clients']); mongo.clients.save({foo:'bar'});
[14:54:19] <kylefaris> and it works great. I can even restart nodejs and mongodb and the data is persistent (as expected). But, if I open the mongodb console in the command line, the 'clients' collection isn't even listed as a option to use.
[14:59:24] <kchodorow_> kylefaris: are you switching to the correct db in the shell
[15:19:46] <kchodorow_> system collections are used by mongo
[15:19:59] <kchodorow_> admin, local, and config dbs are important
[15:20:35] <kchodorow_> so don't drop any of those
[15:34:33] <kylefaris> kchodorow: thanks! But, let's just say I did delete the `local` database... what would be the consequences? would I need to re-install mongodb or just reacreate the database?
[15:35:04] <kylefaris> I don't have anything I'm worried about losing at this moment
[15:35:51] <kali> kylefaris: as long as you don't use Replica Set, you can afford to ditch local
[17:35:23] <Almindor> hey, what can cause connection timeout resets from mongodb? We had a big import script (reads data from M$ SQL and puts them in mongoDB) and sometimes after a random amount of time we got connection resets
[17:35:46] <Almindor> usually when they started it was like a "buffer" was filled and they kept creeping up after a few 100 documents
[17:41:13] <Almindor> I get this with pymongo: AutoReconnect: could not connect to our.server:port timed out
[18:18:02] <addisonj> anyone have any guides on doing a migration of a production mongodb instance to a new server? I am not using replica sets on the old instance but plan on doing it on the new cluster, would adding the old host to the cluster make sense? and then just promote one of the new boxes to master?
[20:47:05] <chubz> how can i restrict the output of db.isMaster() to just return a certain field?
[20:54:12] <kali> chubz: what are you trying to do ?
[21:31:11] <dstorrs> http://pastie.org/4326428 describes a data structure and a query. What index should I set to make the query efficient? I'm not clear on how to set an index into an array
[21:32:28] <dstorrs> is it just db.ensureIndex({ 'pages.locked_until' : -1, 'pages.action' : 1, 'pages.owner' : 1 })
[22:01:17] <kchodorow_> dstorrs: that's how you index fields in an array, but i'm not sure how efficient you can really make that query
[22:01:39] <kchodorow_> $ne and $exists can't use an index
[22:01:58] <kchodorow_> and i think you mean pages.owner_host, not pages.owner, right?
[22:03:35] <dstorrs> kchodorow_: no, I meant owner.
[22:03:46] <dstorrs> owner_host means "any worker on this machine"
[22:04:01] <dstorrs> owner means "this process on this machine"
[22:04:12] <kchodorow_> you generally don't need to index fields you're not querying on
[22:04:32] <dstorrs> ok...but isn't an update effectively a query?
[22:04:51] <dstorrs> "find an element that matches X, and change it"
[22:05:19] <dstorrs> or is update always a table scan?
[22:06:08] <kchodorow_> the second doc is a modifier
[22:06:14] <kchodorow_> that operates on whatever was found
[22:06:35] <kchodorow_> build indexes to optimize the query
[22:07:01] <dstorrs> right. that's what I'm asking about.
[22:08:51] <kchodorow_> the query part can use an index to find the document(s) to change
[22:08:57] <dstorrs> as to the $ne issue...ok, if the index can get me to a subset of documents that match up until we consider their 'owner_host', that would be a win
[22:09:00] <kchodorow_> but the update always has to touch the document itself+indexes
[22:12:14] <dstorrs> that's...complicated. there are 50 workers hitting the DB simultatenously. each one locks one page on one job (decrementing np in the process), processes it to completion, and then removes it
[22:20:57] <dstorrs> errr, with the explain, you mean this? > db.jobs_harvest_Profile.find({"pages.locked_until":1,"pages.owner_host":1}).explain()
[22:21:07] <dstorrs> or should that be the query params?
[22:21:22] <dstorrs> I think the latter, but I'm not familiar with explain yet.
[23:11:40] <ccasey> hello all, i'm new to mongo and i'm having some issues with mongoexport. i'm trying to export a collection and i want the date to be formatted in the iso string format. from what i've read on stack overflow it's not possible…thought i'd check here too to see if anyone knows of any other options/workarounds
[23:31:53] <krz> anyway to install mongodb 2.2 in homebrew?