PMXBOT Log file Viewer

Help | Karma | Search:

#mongodb logs for Tuesday the 31st of July, 2012

(Back to #mongodb overview) (Back to channel listing) (Animate logs)
[00:02:12] <awestwell> ya
[00:02:21] <awestwell> its in the class path
[00:02:32] <awestwell> using spring data
[00:02:40] <awestwell> so it does not make sense
[00:10:13] <cgriego> What's up with the deb package server? Took 8 minutes to download the file.
[00:20:12] <skot> It is on S3...
[00:24:55] <dstorrs> ah yes, Amazon. "Here, have these awesome cloud services! You get to choose your mandatory catastrophic failure mode: too slow or too volatile!"
[00:31:56] <cgriego> downloads-distro.mongodb.org doesn't act like it's on S3. Claims Apache 2
[00:32:58] <cgriego> http://downloads-distro.mongodb.org/repo/
[01:23:31] <Frozenlock> Is there a way to return every other item in a collection? (or every other other other.... item)
[01:59:51] <sx_> hallo. new to mongodb so this is probably easy for you guys. trying to upsert a document. if uniqueid exists, just need to $push to a field. if doesn't exist, need to add entire object plus the $push to it.
[02:11:14] <dstorrs> sx_: so...you've got 3 cases, I think? document doesn't exist, exists with uniqueid, exists w/o uniqueid .
[02:11:17] <dstorrs> right?
[02:11:48] <sx_> dstorrs: even easier. either id exists or it doesnt
[02:12:02] <dstorrs> to the room -- splitvector is taking 400-500 ms and happening relatively often. is there any way to speed it up? does indexing help?
[02:12:24] <dstorrs> sx_: so you know the document exists ?
[02:12:56] <dstorrs> if so, why are you doing upsert?
[02:13:10] <sx_> dstorrs: no, i don't know if it exists. let me rephrase the question
[02:13:39] <dstorrs> maybe easier if you back up and expalin the overall situation, actually. what's the end goal at a business level?
[02:14:48] <sx_> dstorrs: so I am querying an API, and retrieving a JSON object. if this object's id is already saved in my db, then I just want to $push some values to two array fields. if the object ID is not in my db (i.e. new document), then I want to save the entire object, and also $push to it.
[02:15:27] <sx_> i can easily do this with one separate find() and $push, but I am thinking that one update/upsert command may be able to do it all
[02:16:24] <sx_> the belabored way: z=find({id:x}). if (z) {$push) else {save+$push)
[02:16:49] <topriddy> hello people
[02:17:06] <dstorrs> sx_: race condition there, unfortunately
[02:17:39] <sx_> i'd nest it of course
[02:17:47] <topriddy> say I want to have an entity have an attribute that stores someone's country. does having a list of country and doing foreign key linking usual in mongodb too?
[02:17:57] <sx_> but thinking one upset command could do it?
[02:18:15] <topriddy> also, how would i do a mass load of list of countries save writing the code manually myself.
[02:19:02] <dstorrs> topriddy: just put the country in the document where you want it.
[02:19:12] <dstorrs> sx_: this is actually trickier than it seems.
[02:19:25] <topriddy> dstorrs: i dont get.
[02:19:57] <topriddy> dstorrs: you mean i should just store country as is in the entity. what happens to normalization etc?
[02:20:03] <dstorrs> if you do this : db.coll.update({ _id : 'foo', blah : 1 }, { $push : ... }, {upsert : 1 })
[02:20:30] <dstorrs> topriddy: this is a NoSQL system. denormalization is not just normal, it's preferred.
[02:21:14] <dstorrs> sx_: the problem with that is that if it only matches the document if all of _id and blah match
[02:21:38] <dstorrs> if the doc doesn't exist, it will create it with both in place, which is good.
[02:21:48] <dstorrs> but you see the issue?
[02:22:14] <dstorrs> and if you leave out the non_id fields, then they aren't available to be saved in the upsert
[02:22:33] <sx_> yup
[02:22:44] <sx_> i guess it'll be two commands
[02:22:54] <dstorrs> well, then you have a race conditon.
[02:23:03] <sx_> not if nested
[02:23:09] <dstorrs> show me.
[02:23:37] <topriddy> dstorrs: what then do you do to avoid problems with denormalization? :S
[02:23:53] <dstorrs> topriddy: what problems? be specific.
[02:24:50] <dstorrs> topriddy: this isn't a SQL database. It works on different principles, and if you try to force it to act like SQL you are giving up most of the power and making your own life harder.
[02:25:10] <sx_> db.widgets.find({ id:x }, function(result) { if (result) $push({ a:b }) else db.widgets.save(obj._extend({ a:b }) });
[02:25:43] <topriddy> dstorrs: i have a country entity (pardon my use of sql terms) with say president name. What happens when the President changes and i have a denormalize database?
[02:26:30] <topriddy> SQL teaches me that if i have a foreign-key and well normalized db i just need to change this info in one place. I dont know what i'll get from nosql, but then it doesnt look "so-good" already
[02:26:38] <dstorrs> sx_: so far as I know, that's not legal syntax.
[02:26:40] <topriddy> dstorrs: what do you advise?
[02:27:00] <dstorrs> do you mean db.widgets.find({id:x}).forEach(function...) ?
[02:27:03] <sx_> dstorrs: yeah. the $push will actually be an update command
[02:27:16] <sx_> maybe -- gotta run but thanks for help
[02:27:22] <dstorrs> you still have a race.
[02:27:28] <dstorrs> what if your JS yields in the middle.
[02:27:34] <dstorrs> anyway, see you
[02:28:12] <topriddy> dstorrs: you didnt answeR?
[02:28:26] <dstorrs> because I was talking to sx_. be patient.
[02:28:38] <dstorrs> topriddy: use the right tool for the job. Judge each tool for the job it's intended for.
[02:29:00] <dstorrs> SQL is intended for implementing constraints on your data.
[02:29:32] <topriddy> dstorrs: i only chose Mongo cos of scalability. surely twitter is using NoSQL. Do i have to lose all constraints by using nosql?
[02:29:33] <dstorrs> used correctly, it lets you maintain a clean data store relatively easily, at the expense of making it very difficult to scale horizontally
[02:30:15] <topriddy> dstorrs: am asking sincere questions here. if you have a good link to point me too, i dont mind doing some more reading
[02:30:16] <dstorrs> constraints get moved to the app layer. Or, more commonly, to the ORM layer.
[02:30:47] <topriddy> dstorrs: i'm using java and Morphia
[02:30:54] <dstorrs> In general, yes, you lose all constraints.
[02:32:10] <dstorrs> as to good links...
[02:32:20] <dstorrs> hm. don't have one off the top of my head. one sec.
[02:32:31] <topriddy> dstorrs: but twitter somehow still avoids people using same "username", guaranteeing uniqueness. Also, really the trivial Country example. Data would CHANGE in real life. I cant possibly walk through all entities and update them cos of the denormalized data?
[02:32:43] <dstorrs> oh, that.
[02:32:45] <topriddy> dstorrs: okay i'll wait. thank you in advance.
[02:32:51] <dstorrs> Yes, uniqueness can be guaranteed.
[02:33:15] <dstorrs> db.coll.ensureIndex({username : 1 }, { unique : true })
[02:35:16] <topriddy> dstorrs: while waiting the link, i am having a need to store a user picture. considering setting up another entity just for that. (this is based on my SQL background/school of thought)
[02:35:16] <dstorrs> this is short and a bit trite, but all good points: http://facility9.com/2010/09/five-reasons-to-use-nosql/
[02:36:51] <dstorrs> topriddy: schema design still matters in NoSQL, and I don't know your app. but my users collection has attributes like this: name, age, thumbnail_url, canon_name, display_name
[02:37:53] <dstorrs> I like to have a canonical version of the name that makes for guaranteed deduplication and sort ordering, and then a 'display_name' which is how it was actually entered.
[02:40:01] <dstorrs> oh, here. read this: http://www.linuxjournal.com/article/10770
[02:40:16] <dstorrs> good discussion of CAP and how it applies to (No)SQL
[02:40:28] <dstorrs> topriddy: ^^
[02:40:49] <dstorrs> note that it's from 2010, so slightly dated as to the specifics about various dbs
[02:41:07] <topriddy> okay.
[02:41:43] <topriddy> dstorrs: well, i wont want the time-lag and redundancy from getting picture objects everytime i ask for a user. thats why am thinking of seperating it.,
[02:42:01] <dstorrs> oh, and one other thing which you probably already know -- don't store actual images in your DB. not only does it bloat everything, but you risk blowing the 16M-per-document limit
[02:42:34] <dstorrs> if you really want to do that, I would suggest using Mongo's GridFS
[02:42:54] <dstorrs> possibly with memcache (because it implements LRU semantics)
[02:43:34] <dstorrs> that said, think about how many users you have and whether or not the time to deep-link a picture is really going to matter.
[02:46:34] <topriddy> dstorrs: alright. maybe i can store in same entity afterall.
[02:47:17] <dstorrs> *shrug* up to you. like I said, I don't know your app or schema. I'm just making a suggestion, because I think it will make your life easier.
[02:56:20] <topriddy> dstorrs: yeah. thanks. :)
[03:17:06] <tsinnema> hello
[03:17:33] <tsinnema> is it possible to create a sparse index based not on the existence of a field but on the value in a field?
[03:18:30] <tsinnema> for example if you have things with category: "ubiquitousCategory" and category: "obscureCategory"
[03:18:49] <tsinnema> you'd only have an index for category: "obscureCategory"
[03:30:58] <skot> you want a filtered index and mongodb doesn't that feature yet.
[03:31:27] <dstorrs> skot: ooc, is it on the roadmap?
[03:31:32] <skot> https://jira.mongodb.org/browse/SERVER-785
[03:31:46] <hdm> still getting weird bson corruption - but with odd results. mongodump/mongoexport + mongorestore/mongoimport both work for this data, but doing a map-reduce triggers a invalid BSON error (randomish)
[03:32:00] <hdm> most queries still seem to work too, just not map-reduce
[03:37:56] <skot> Can you post an copy of the error to gist/pastebin?
[03:38:02] <skot> also, what version is this?
[03:39:32] <hdm> hit the problem consistently between 2.1.x, 2.2.0rc, and 2.0.6
[03:39:54] <hdm> let me paste the error in a second - ive been loading/reloading the data (takes 48 hours+ each time) trying to figure out if its something with my system
[03:40:32] <hdm> it seems to be triggered by certain content in the string field (binary bits)
[03:40:49] <hdm> at least, it doesnt trigger until i load a ton of http responses in
[03:44:00] <tsinnema> thanks for the reply skot!
[03:44:50] <hdm> waiting on the m/r to trigger it again
[03:54:16] <jwilliams> is it possible that bulk insert slow down due to too many servers are doing the same thing (around 50 servers are doing bulk insert at the same time) ?
[03:55:00] <hdm> jwilliams: use iostat -mh 1
[03:55:09] <hdm> see if you are hitting your disk or cpu limits
[03:55:32] <hdm> if 100% cpu and the disk isnt maxed on read or write, then its likely the cpu overhead of your indexes
[03:56:05] <hdm> if the disk is maxed on write (or IOPS), then you need faster storage
[03:56:28] <hdm> (indexes can cause *reads* to be bottleneck on inserts too)
[03:57:49] <jwilliams> the indexes was dropped.
[03:58:28] <jwilliams> memory doesn't use too much around 40% at shard
[03:58:38] <jwilliams> but cpu goes up to 100%
[03:59:21] <jwilliams> it looks like the data (in netin) doesn't actually come as max as previous it would have.
[04:02:43] <jwilliams> argh id filed has index.
[04:03:13] <jwilliams> is it possible to drop id filed index? issuing drop index seems not dropping id field index.
[04:05:15] <hdm> i think so
[04:10:15] <jwilliams> what is the right way to split _id field index?
[04:10:38] <jwilliams> search mongodb website, seems only mention dropIndex
[04:20:08] <wereHamster> jwilliams: split the index? what should that do?
[04:24:05] <jwilliams> wereHamster: not split the index, but want to drop _id field index
[04:24:18] <jwilliams> but just notice the doc mentions that _id field index can't be dropped.
[04:24:34] <jwilliams> http://www.mongodb.org/display/DOCS/Indexes#Indexes-The%5CidIndex
[04:27:01] <hdm> ah right, you can change how id is calculated, but not drop it
[06:01:25] <hdm> yuck - assertion: 13106 nextSafe(): { $err: "Invalid ns [*.*]", code: 16256 }
[06:01:41] <hdm> can no longer do mongodump
[06:04:09] <hdm> -d dbase seems to work, still, yuck
[07:01:55] <bullfrog3000> hi all, i am using pymongo, and am running a commond db.test.update({}, {$set, {"a":5}}). It seems without safe enabled, it returns right away. Is there no way to block until the call is done, without paying the safe=True performance cost? It appears significant on my dataset.
[07:14:37] <bullfrog3000> anyone around?
[07:15:52] <ron> I am. can't help you, but I'm around. just thought you'd like to know.
[07:16:52] <bullfrog3000> haha, thanks ron
[07:17:15] <bullfrog3000> in place updates appear to take at least 5x of a insert =)
[07:17:29] <bullfrog3000> trying to learn more… super-new
[07:17:33] <bullfrog3000> to mongo
[07:31:58] <jwilliams> is it possible for mongod to make use of multiple cpu?
[07:32:19] <jwilliams> checking with mpstat -P All 1 shows that only 1 cpu is used.
[07:38:56] <[AD]Turbo> hi there
[07:41:20] <ranman> jwilliams: what version of mongo?
[07:44:49] <jwilliams> 2.0.1
[07:45:19] <jwilliams> 64bits
[07:54:18] <bullfrog3000> jwilliams: you mean for writes?
[07:59:06] <ron> where's remnov? why is he such a lazy ass?
[08:28:27] <diegok> kchodorow_: the perl driver isn't setting the slave okay I think. It fails to query a slave even with slave_okay(1) :-/
[08:47:17] <jwilliams> bullfrog3000: yes.
[08:48:32] <jwilliams> we've shard the server, etc. but when monitor (mpstat -P All 1) the activities, noticing large amont of time only 1 cpu is not idle.
[08:52:17] <ron> Derick: Now that you guys are more active in the channel (and kudos for that), you may want to start using an IRC bouncer :)
[08:54:18] <Derick> nah, I was just checking some client settings in irssi
[08:54:33] <Derick> irssi runs on my server so shouldn't go down unless I do it myself
[08:54:44] <ron> that's true.
[09:31:20] <vak> hi all
[09:34:58] <Rozza> hi
[09:35:23] <vak> oops, I've got an unexpected performance artifact with pymongo. Although using of "fields=" in my case makes I/O bandwidth 3-times lower (as expected) it makes also the whole time of looping through the collection about 3 times longer... how come?..
[09:44:13] <vak> ok it is blows up badly in pypy, not in cpython
[10:11:46] <new2nosql> hi guys, running a repair ... getting a FAILED_TO_UNCOMPRESS any ideas
[11:16:48] <remonvv> Anyone aware of changes to findAndModify behaviour between 2.0 and 2.1? Unit tests of our system our failing on 2.1 that are okay on 2.0. Unspecified assertion error in find_and_modify.cpp:140
[11:18:07] <kali> remonvv: when you have found what has changed, i'm interested
[11:18:35] <remonvv> Trying 2.2.0-rc0 now, may have been a bug that was fixed.
[11:19:06] <kali> i'm in the laborious process of re-applying an adhoc patch on 2.2.0-rc...
[11:20:03] <remonvv> Jep, it's "broken" in 2.2.0 as well.
[11:20:04] <remonvv> Crap.
[11:20:41] <remonvv> command failed [command failed [findandmodify] { "serverUsed" : "127.0.0.1:27017" , "errmsg" : "exception: assertion src\\mongo\\db\\commands\\find_and_modify.cpp:146" , "code" : 0 , "ok" : 0.0}
[11:21:17] <remonvv> Seems unique to performing the findAndModify from a driver. Shell seems okay. Might be driver incompatability.
[11:25:14] <remonvv> Anyone using Java?
[11:25:22] <remonvv> If so, please try and reproduce if you have the time.
[11:59:18] <remonvv> Okay, so this findAndModify fails since 2.1+ : db.test.findAndModify({query:{stringField:1}, update:{$set:{stringField:2}}, upsert:true, new:true})
[12:00:10] <remonvv> Assuming it has something to do with querying on the same field I'm updating with new=true/upsert=true
[12:00:31] <kali> remonvv: with the shell too ?
[12:01:32] <NodeX> is it to do with upserts not being able to $set / update the field you query on
[12:01:37] <NodeX> ?
[12:05:17] <awestwell007> hey all
[12:06:23] <remonvv> kali, yes
[12:06:39] <awestwell007> I am trying to create a mapreduce in mongo and am getting empty values back from the reduce part. I have a collection with a id and a count I am trying to get all the values based on the id vale (views or counts)
[12:06:47] <awestwell007> here is the code
[12:06:48] <awestwell007> http://pastebin.com/KRBmEqeC
[12:06:55] <remonvv> NodeX, yes. It fails if you findAndModify with the query and the update using the same field and new=true and upsert=true. All other permutations of that command are okay.
[12:07:13] <remonvv> https://jira.mongodb.org/browse/SERVER-6659
[12:08:10] <remonvv> Bit of a pain. My unit tests for my mapping library is failing now.
[12:08:54] <kali> awestwell007: can you show us a couple documents ?
[12:09:19] <awestwell007> sure
[12:14:11] <awestwell007> example -> http://pastebin.com/GHsi4pth
[12:14:56] <awestwell007> So basically there could be the same primaryTopic more then once
[12:15:31] <awestwell007> I took the topic out in the example
[12:15:39] <awestwell007> but it is there in production
[12:17:33] <awestwell007> here is a better sample :) http://pastebin.com/1mFdaqmd
[12:19:38] <awestwell007> kali: Basically I want to get back { topic : "TopicA", Count: 1, Views: 2456 }
[12:19:47] <awestwell007> based on the sample data
[12:20:13] <awestwell007> ie aggregate by topic count and views based on the id
[12:20:21] <kali> lemme check have a look
[12:39:29] <awestwell007> thanks
[12:43:48] <kali> awestwell007: sorry, been interrupted. I can't see anything obviously wrong with what you're doing
[12:44:09] <awestwell007> so if I try and do this
[12:44:22] <awestwell007> function(v) {
[12:44:22] <awestwell007> print(v['count'])
[12:44:23] <kali> awestwell007: the only bit suspicious is the use of forEach. it might be worth trying with a stupid for(...)
[12:44:49] <awestwell007> in my map reduce it outputs the string "[object bson_object]"
[12:45:07] <awestwell007> not the value
[12:45:54] <kali> try printjson
[12:47:10] <awestwell007> outputs this
[12:47:12] <awestwell007> NumberLong(4)
[12:47:12] <awestwell007> 0
[12:47:12] <awestwell007> NumberLong(2)
[12:47:12] <awestwell007> 0
[12:47:39] <awestwell007> what are the zeros?
[13:05:40] <remonvv> Is it possible to perform rs.initiate from a driver?
[13:09:48] <remonvv> Never mind, got it
[13:24:14] <Samuel__CTX> I have a question. In PHP I can find devices based on some fiend: $devices = $conn_devices->find(array("key" => "value" ));, however when I try to do this on an _id this doesn't work: $devices = $conn_devices->find(array("_id" => "id" ));
[13:24:32] <Samuel__CTX> how can I get a document when I have its ID?
[13:24:50] <wereHamster> Samuel__CTX: "_id" => ObjectId(id)
[13:24:57] <wereHamster> you need to cast the id to its correct type
[13:28:33] <souza> Good morning all
[13:28:38] <Samuel__CTX> wereHamster: thanks
[13:30:38] <souza> guys i have a high level problem, i must to create a database to a HUGE data structure, my question is: I must "cut" this huge environment in a lot of collections or put all into a one collection
[13:30:56] <wereHamster> souza: you must nothing.
[13:31:13] <souza> wereHamster: sorry?
[13:31:38] <souza> wereHamster: i don't understand :/
[13:31:49] <wereHamster> me neither.
[13:32:11] <wereHamster> there is a limit of 16MB per document and about 16k collections. That's all. Anything else is up to you.
[13:33:31] <souza> putting this data in a XML i got around 10 levels in tags, and this 10 levels repeat some times, this is a hard question to explain
[13:33:32] <souza> :/
[13:34:11] <wereHamster> try harder.
[13:37:30] <souza> if you have a Big environment and must to put it in MongoDB, you will create some collections and share ObjectID's, or put all into one table.
[13:40:08] <wereHamster> depends
[13:40:30] <wereHamster> what do you mean by 'share objectids'?
[13:41:05] <ron> remonvv: dude! are you going to be in IBC?
[13:41:37] <souza> well, the ObjectID must to create relationships among the tables.
[13:42:00] <wereHamster> you can, yes. What does that have to do with your problem?
[13:42:16] <wereHamster> souza: go read http://www.mongodb.org/display/DOCS/Schema+Design
[13:44:06] <souza> wereHamster: Ok, it's a bussiness problem, i must to try get the best model to mongo
[13:47:04] <wereHamster> you must try to read the documentation and resources available.
[14:28:00] <Rrjois> hi, is this the place to ask Qs about spring-mongo?
[14:29:27] <souza> Rrjois: maybe, what's your question!? =D
[14:29:43] <remonvv> ron, what? where?
[14:34:08] <Rrjois> souza: I want to do case insensitive string match. I tried regex. but it returns the value even if its a substring. I tries "is" but that is case sensitive here is my code http://pastebin.com/4rAEBdPK
[14:34:44] <Rrjois> souza: please help
[14:34:54] <beholder> mongodump'ing to a full file system doesn't quite fail in the way I expected. It's pretty silent about the whole issue :(
[14:39:54] <souza> Rrjois: let me understand
[14:39:55] <souza> ...
[14:40:23] <Rrjois> souza: you want me to explain once?
[14:41:53] <souza> i think that i understand you want to create a query that get the data from database, that doesn't cares about the case of the string
[14:44:12] <remonvv> Rrjois, is that Morphia?
[14:47:13] <souza> Rrjois: i didn't found any documentation mentiong this, post this question in mongodb e-mail list, maybe help
[14:47:19] <ron> remonvv: http://www.ibc.org/
[14:47:35] <Rrjois> souza: yes. the problem is suppose i hve two values in my collection. say "bill status" and "Bill" . if I search bill, It returns be bill status as bill status(may be cause its found first). I want it to match the full string "Bill status"
[14:48:03] <Rrjois> remonvv: I dint get you.
[14:55:30] <remonvv> ron, oh that IBC ;) Maybe, probably not.
[14:56:39] <ron> remonvv: my company has a stand there, so if you do go, you should visit. unfortunately, I won't be there. I'm staying to hold down the fort.
[14:57:09] <remonvv> ron, I will. Why aren't they sending you?
[14:57:30] <ron> remonvv: "I'm staying to hold down the fort." :)
[14:58:42] <NodeX> ron is staying to learn PHP
[14:58:44] <ron> remonvv: gonna overview the R&D while the big guys are away. should be fairly exciting there though, it's basically the company's debut.
[15:02:16] <hjb> funny fact of the day: when that ff addon is installed: http://www.bolwin.com/software/snb.shtml it breaks access to the html admin interface of mongod
[15:02:26] <hjb> took me two hours to realize
[15:02:45] <hjb> btw. IE8 isn't able to access it at all
[15:02:55] <SisterArrow> Hiya! I updated mongodb with a apt-get update and restartet mongodb and restarted the MMS-agent on a machine, lets say machine2. But now(after 15 minutes wait) the MMS-webpage only reports machine1 and machine3 but not machine2 :o
[15:03:11] <SisterArrow> The debug output from the agent just says starting watching machine1 and machine3, but not itself.
[15:03:21] <SisterArrow> Anyone know why that might be?
[15:10:23] <SisterArrow> On the machine thats gone away(machine2) I updated from 2.0.4 to 2.0.6..
[15:14:59] <remonvv> ron, cool. I might go but I'm not that interested in the conference itself and I have to keep an eye on my annual conference budget ;)
[15:16:18] <ron> remonvv: hehe :) well, so far I didn't get to travel from work (though surprisingly, two of my team members do).
[15:17:25] <remonvv> ron, you're doing it wrong ;) Shame you're not coming to Amsterdam. Would be good to have a coffee.
[15:18:07] <SisterArrow> Ok, manually adding the server in MMS solved the issue.
[15:18:14] <ron> remonvv: don't be silly. I don't drink coffee. but yeah, had I come, I would let you know so we could meet. maybe to next year's IBC.
[15:18:47] <remonvv> Who knows ;)
[15:18:56] <remonvv> Let me check if anyone else is going here.
[15:19:18] <remonvv> Hm, just my boss I think.
[15:19:34] <remonvv> Oh, ha, and I'm on holiday in Spain then
[15:24:10] <ron> hehe
[16:55:56] <jwilliams> http://www.mongodb.org/display/DOCS/Inserting#Inserting-Bulkinserts says continuous on error is implied in shard environment and can not be disabled. does user need to explicit to enable it?
[16:56:08] <jwilliams> or just use DBcollection.insert would work?
[16:56:28] <jwilliams> mongodb 2.0.1, java driver 2.6.5
[18:32:23] <y3di> how do you generally go about doing backups?
[18:32:54] <y3di> I want to backup my db before I push my new changes to production
[18:33:19] <y3di> i only have one db running one one ec2 instance (the same one running my web app)
[18:33:54] <vsmatck> http://www.mongodb.org/display/DOCS/Backups
[19:02:24] <rnickb> i'm using scopeddbconnection and my program is crashing when it terminates. is there something i have to do to properly destroy the database connection pool?
[19:46:13] <y3di> vsmatck: I can't really figure out which option would be best, im worried im going to mess something up
[19:46:33] <cgriego> downloads-distro.mongodb.org is still running slow today :(
[19:51:50] <y3di> would it be possible to do a db restore from a simple log
[19:59:52] <y3di> how can i test if ./mongodump successfully stored all the data?
[20:00:20] <ron> you can import it to a new database?
[20:42:58] <linsys> y3di: Also mongodump doesn't restore indexes..
[20:43:34] <linsys> well backup indexes
[20:53:04] <y3di> linsys: it doesnt backup indexes?
[21:04:24] <linsys> no
[21:06:18] <linsys> Oh maybe it does now.. it didn't in thepast
[21:09:59] <reckoner> yooo
[21:11:40] <reckoner> is there a way to refresh the db schema?
[21:12:30] <reckoner> some old fields in my collection seem to persist, trying to get rid of them
[21:18:44] <y3di> linsys: i could still go back and readd indexes right?
[21:18:57] <y3di> and if i don't have any indexes it shouldnt matter
[21:57:21] <ribo> does mongo use more disk space propotional to the available disk space?
[21:58:25] <deoxxa> ribo: no..?
[21:58:28] <ribo> or is it making 20 2G local.# files to pre-empt a recovery
[21:58:39] <ribo> just upgrading my datastore on one RS node
[21:58:56] <ribo> as soon as I started mongodb, bam 60G consumed
[22:01:14] <ribo> that appears to be what it's doing
[22:01:17] <ribo> preallocating disk
[22:01:37] <ribo> the local.# are getting replaced with mydatabase.#
[22:02:25] <adiabatic> Weird default. players.0, .1, and .ns are 64, 128, and 16 MB apiecce
[22:02:33] <adiabatic> and these are minima
[22:13:08] <mmlac> How do I store tags best? As an embedded array or as a separate model that is linked to? Use-cases are: get the distinct tag names, display every model that has a specific tag
[22:19:19] <trbs> mmlac, for a similar use case i use a list in the document, then my application has a set() of all distinct tag names which if a unique tag gets added gets updated (if your tag list is big i guess you could put that in a collection as well) both my set and distinct tag names are small enough for this to be no problem.. and i am fine with the set() that gets checked and updated via celery (cron-alike jobs) every once in a while
[22:30:27] <mmlac> ic. How do you find the objects that contain a certain tag?