[00:24:25] <mkmkmk> if mongos dies on another box and wont come back, i'm going to be dead in the water
[01:54:13] <Kryten001> Hi, here http://www.mongodb.org/display/DOCS/Object+IDs they say that an _id must be unique for a given db cluster, does that mean that two objects in the same db but different collection can not have the same _id ?
[04:34:07] <dstorrs> I have a collection that stores statistics. When I add new stats, I would like them to include incremental data (current stat - previous value). Is a mapReduce the best option here, do you think?
[04:34:24] <dstorrs> There's about 10M records per harvest
[04:34:54] <dstorrs> Doing it one by one is certainly out, but is there an alternative to M/R that I haven't considered?
[14:09:02] <mitsuhiko> it's neither reliable nor easy to use nor well designed. but compared to many other nonsql solutions it has some merits
[14:09:03] <Defusal> which brings me to the point where i need to start deciding whether continuing to invest in mongo for projects is worth it in the long run
[14:09:09] <Defusal> or if i should start comparing alternatives
[14:15:16] <Defusal> while there are some downsides to the query interface, for the most part, the schemaless nature does streamline development, especially for prototypes
[14:15:29] <Defusal> and all my projects at this point in time start as prototypes
[14:15:43] <Defusal> so any steamlining is more than welcome
[14:16:19] <Defusal> i cringe when i have to make a single change to a project using a SQL database just because of the effort it takes to update the schema
[14:16:33] <Defusal> where as a mongo project the change is made in seconds
[14:17:01] <mitsuhiko> Defusal: the problem with mongodb is not mongodb itself as much as the fact that everybody thinks he needs to use it
[14:22:45] <Defusal> you obviously need a replica on a seperate machine
[14:23:06] <Defusal> point is, is it really necessary?
[14:24:23] <Defusal> a replica server would obviously be better protection than backups every few hours, sure, but how often is there gonna really be a data loss issue that will require recovery?
[14:24:29] <mitsuhiko> Defusal: if you use mongodb you need replica sets
[14:24:37] <mitsuhiko> it's one of the constraints of the design. that's how you run it
[14:25:04] <Defusal> then you could run a replica set on the same machine
[14:25:24] <Defusal> that would be protection against internal mongo corruption
[14:25:25] <kali> not if you're ready to loose all writes between two backups
[14:26:05] <Defusal> kali, you mean in the case of hardware failure?
[14:26:08] <mitsuhiko> Defusal: mongodb is just fine for what it is. but you have to take it with all it's warts and bugs :)
[14:26:20] <mitsuhiko> i think the argument for mongodb is that 10gen is very, very active and cares abuot improving their product
[14:26:36] <Defusal> mitsuhiko, well then again, please point me to an article that provides details on this
[14:35:54] <mitsuhiko> Defusal: because you need to introduce a mongo controller, mongo routers and the balancer will spend a long time moving your data over :)
[18:35:07] <skot> with what I've seen with shared server running mongodb, and the mmap os behavior, I can imagine it would not go well using that model with varnish
[18:36:07] <kali> well, it make sense to have something like a haproxy or an nginx, as they are not memory greedy
[18:36:21] <kali> but you don't want to share a VM between varnish and mongodb :)
[18:53:52] <jpnance> this is a really pedestrian question
[18:54:28] <jpnance> let's say i have a collection of documents such as: { "field1": 10, "field2": 5 }
[18:54:45] <jpnance> what's the typical way to query that collection for all documents which have field1 > field2?
[18:55:18] <wereHamster> which is slow, because it can't use an index. If you want to use an index, make a third field, field1minusfield2, and index that. and then run a query on that
[18:55:38] <jpnance> wow, i somehow have completely missed that $where is a thing
[18:55:54] <jpnance> yeah, both of those make sense