[00:02:24] <hadees> sirious: yeah but I don't think that is all there is. I think actually if limit it by everything less than created_at it would work most of the time. The problem is if multiple comments are created at the same time.
[02:12:59] <sam___> keep getting this - assertion: 9997 auth failed: { errmsg: "auth fails", ok: 0.0 } - when i try mongodump even with the right credential, any idea?
[02:25:13] <aep> hm if i query for arbitrary garbage like {'abcdef' : 1234' } which is not contained in any document, i get all documents. is that supposed to work this way?
[05:21:42] <jwilliams_> does mongodb's mapreduce perform full table (collection) scan? or if within map function, using indexed field will let map function perform more effectively?
[05:47:06] <timah> hey guys… looking for some opinions regarding the best (reliable, efficient, fast, etc.) method for mass transforming about 30 million documents from a single collection to many (about 40) collections… i wrote a very slick mongodb shell script that serves in such capacity, however i can only seem to get between 2-5k documents read/transformed/written every few seconds or so.
[05:51:05] <vr_> it appears that GridFS doesn't support TTL.
[10:19:27] <Guest_1448> "is there a better/faster way to do this: http://stackoverflow.com/a/9388189 with the new aggregation framework?"
[10:35:18] <kali> Guest_1448: well, it would be pretty standard stuff with the AF: you can extract components of a datetime with http://docs.mongodb.org/manual/reference/aggregation/#date-operators, then group
[11:38:38] <xcat> If I have a logging collection that expires via a TTL date field and a type field which I want to filter messages by, how many indexes would I need and on which fields?
[12:12:30] <NodeX> xcat : one index for the TTL and one for the "type" field
[12:15:28] <xcat> NodeX: if I also want to sort by insert order (either ID or date) do I need to create a compound index on one of those fields?
[12:16:35] <NodeX> you get a free index on insert order ... order with _id
[12:16:55] <NodeX> _id -1 for desc (last first) or no sort for asc
[12:17:25] <NodeX> there is no need to order on a date in my opinion
[13:36:23] <NodeX> quick question regarding the _id index.... if I do somehting like this ... {_id:ObjectId("1234"),uid:"5678"} with an index on UID is this as efficient as just looking up on _id ?
[13:37:34] <NodeX> or do I need a secondary compound index on _id,uid
[15:12:23] <JoeyJoeJo> I need to performance test how long it takes to insert my data (tens of millions of documents). What's a good way of doing that? I was thinking about writing a python script and just doing finish time - start time. Does mongo have anything built in to handle that?
[16:32:31] <timah> every md5 gets replaced with its corresponding mysql pkid.
[16:33:47] <timah> so publisher gets moved out of the schema altogether and its hashes mysql pkid value becomes part of the collection name… so… 1 collection per publisher… e.g. p.25.e
[16:34:07] <timah> p = publisher, 25 = id, e = event.
[16:55:39] <timah> i mean… i was even going to go as far as using mongoexport with --query (per publisher) and --fields (only the fields i need) and dumping each publisher in its own file during that export… then using something super simple like php fgets for buffer read on large files and have php just swap all the values out and basically just transform the exports into imports.
[16:56:58] <timah> but if you think that having php (or anything else, really) do the work will speed it up a bit… it surely wouldn't take too long to port it over.
[17:50:42] <JoeyJoeJo> Based on this output from the sh.status() command, can anyone tell if my database is using all four of my servers? It looks to me like it's only using one server. http://pastebin.com/zKK4Rvy2
[17:51:44] <kali> for now everything is in one single shard, there is not enough data
[17:52:48] <JoeyJoeJo> How much is enough? I'm running my insertion script now and I'm up to 1.5 million documents out of 180+ million
[17:55:02] <JoeyJoeJo> Is there any way to force mongo to always write to all shards? Right now I seem to be inserting about 5000 documents per second on the one server. Couldn't I quadruple that if all servers were being used?
[17:56:27] <kali> JoeyJoeJo: you know, this is documented extensively
[18:52:14] <MatheusOl> (because _id is unique, only)
[20:27:17] <owen1> is there a way to see logs of all my replica set?
[20:31:20] <w3pm> why are there different functions to connect - connect() and rs_connect() in the Erlang driver for example.. i thought replica sets were supposed to be transparent to the app?
[21:26:02] <SEJeff_work> I've got a record that has a key named "roles" with a value like ["super_user", "secret-sauce"]
[21:26:24] <SEJeff_work> How can I do a query to return all users with a "super_user" role?
[21:26:56] <SEJeff_work> I thought something like this might work, but it doesn't: db.users.find({"roles": {"$in": ["super_users"]}})
[21:33:19] <SEJeff_work> This is what I was after: db.users.find({"roles": {"$in": ["super-users"]}}, {login: 1, _id: 0})
[22:45:38] <Ironhand> hello; is anyone able to give me a rough idea of the current status of MongoDB on ARM CPU's? the information I've found suggests some people have done some serious work on it, but with very few recent signs of activity
[22:47:13] <tjadc> Ironhand: I am unable to give you personal experience info
[22:47:21] <tjadc> but if you'd like I can do a cross search for you
[22:47:30] <tjadc> - do the search and see what I find
[22:47:41] <tjadc> Don't know how useful it'd be :P
[22:49:29] <Ironhand> well, any information would be useful, just hope you won't waste too much time on whatever a cross search would actually amount to
[22:52:00] <Ironhand> after a lot of consideration I decided I wanted to migrate a personal Python-based project from SQL to MongoDB, read up on various libraries, got a taste for the whole thing, then when I finally decided to actually install stuff it turned out running MongoDB on my QNAP NAS running Debian wasn't quite that easy...
[22:52:43] <Ironhand> I'm kind of hoping the popularity of the Raspberry Pi might have sparked some new interest in MongoDB on non-x86
[23:02:37] <owen1> is there a way to see logs of all my replica set?
[23:10:46] <tjadc> Ironhand: and by cross search I just mean searching it up as a second person, I probably search differently and may stumble upon something you haven't
[23:11:02] <tjadc> ehershey: so far have waited 1.5 weeks
[23:11:13] <midinerd> I'm looking at http://mongoosejs.com/docs/index.html <--- where it starts with //getting-started.js (where exactly is this taking place, a javascript file located <somewhere>, within npm, in mongoose?)
[23:13:56] <midinerd> ehershey: Thanks! So I include node.js from somewhere and then continue on with javascript documentation as they've got shown there?
[23:14:05] <ehershey> I wish I'd ordered a raspberry pi a year ago
[23:14:59] <tjadc> Ironhand: You've obviously taken a look @ skrabban/mongo-nonx86 - what are your thoughts around it :?
[23:15:03] <ehershey> midinerd: you might want to step through some basic node.js stuff before diving into mongoose imho
[23:15:26] <tjadc> ehershey: well, if you ordered recently you get the newly upgraded pi :)
[23:16:05] <midinerd> I don't want to... but I will.
[23:16:30] <midinerd> ehershey: btw you never said "yes" or "no"
[23:18:27] <w3pm> hey guys, why is there rs_connect and connect? my understanding was replica sets would be invisible to the app
[23:18:40] <w3pm> but it seems like i need to know whether i have a replica set or not
[23:20:35] <midinerd> mongoosejs answered my question foiist
[23:20:36] <midinerd> [17:18:30] <aheckmann> midinerd: once you have node installed you can execute javascript files as shown in getting-started: node someScript.js
[23:20:36] <midinerd> [17:18:50] <aheckmann> midinerd: make sure mongo is running as well
[23:28:03] <ehershey> midinerd: yes or no to which?
[23:29:00] <ehershey> w3pm: I think that the replica set layout can be somewhat transparent to the app but the app needs to know that it's connecting to a replica set vs a single instance
[23:30:20] <ehershey> for example what if you were building an administrative tool that needed to connect to an individual instance that is also a member of a replica set
[23:30:32] <ehershey> you would need a way to specify a single instance and know for sure you are connecting to that instance