PMXBOT Log file Viewer

Help | Karma | Search:

#mongodb logs for Tuesday the 11th of September, 2012

(Back to #mongodb overview) (Back to channel listing) (Animate logs)
[02:38:04] <termite> anyone have any experience with geonear in node.js
[02:38:48] <termite> I am trying to use dbconn.command(dropQuery, handleDropPoints);
[02:39:07] <termite> but I get null result
[05:28:37] <Null_Route> Hi Guys!
[05:29:20] <Spaceghostc2c> There may be females here.
[05:29:37] <Null_Route> oh.
[05:29:50] <Null_Route> but I treat women equally
[05:29:56] <Null_Route> ...and don't make special cases for them.
[05:30:14] <Null_Route> *cough*
[05:30:22] <Null_Route> Hey peoples!
[05:30:40] <Null_Route> Anyway -
[05:31:29] <Null_Route> I've remove some hosts from a replica set. When I use listshards on a mongos instance, I still see the old hosts.
[05:32:12] <Null_Route> I've tried flushRouterConfig
[07:17:30] <gigo1980> hi can anyone tell me how the MongoPool works on php ?
[07:20:12] <Gargoyle> gigo1980: Automatically?
[07:22:30] <gigo1980> no i dont know what the pooling do ?
[07:23:00] <gigo1980> so if i read correctly the "perstit" connectionions are not longer available
[07:23:35] <gigo1980> on my application stack i have one mongos node on each application server
[07:24:00] <Gargoyle> I'm pretty sure they are still available. They are just automatically handled for you.
[07:24:42] <gigo1980> and at the moment each call to the webapp creates an connection to the mongorouter ...
[07:25:20] <gigo1980> so is there an way to make an pool that the webapp handels about (40) persistens connections
[07:25:30] <gigo1980> is this that, what the mongoPool does ?
[07:27:38] <Gargoyle> I guess that depends on the setup you have. I'm sure with apache + php the persistent connections and pooling is all handled automatically
[07:28:19] <LambdaDusk> I have some trouble with proper database design for mongo... if I have several users, each of them can have several blogs... should be the blogs and the posts always be their own collections? I have heard mongo is about redundancy everywhere
[07:28:36] <gigo1980> yesterday i had a problem with ulimit of my operating systems …
[07:29:03] <gigo1980> so there are to many file locks and this depends on the connections that the application makes to the mongos
[07:29:34] <Gargoyle> gigo1980: I don't think that is *just* about connections.
[07:29:53] <Gargoyle> gigo1980: What version are you running?
[07:30:06] <gigo1980> 2.2.x
[07:31:25] <[AD]Turbo> hi there
[07:31:27] <gigo1980> the mongos creates for each connection from the webapp, an connection the shards behind
[07:33:04] <gigo1980> follwing command gives me 866 established connections: lsof | grep mongos | grep TCP | wc -l
[07:33:57] <gigo1980> but in realtime there not so much connections 52 gives me : lsof | grep mongos | grep TCP | grep localhost | wc -l
[07:34:44] <gigo1980> and the 52 can rise … so what i want is, to set from the application side an pool limit
[07:56:50] <Rozza> gigo1980: what driver are you using?
[08:13:22] <kali> rolling out 2.2 !
[08:13:24] <kali> wooo hoo
[08:15:21] <NodeX> 2.2 rocks
[09:58:50] <gigo1980> hi, i will remove an shard. but the state is daining ongoing, and no chunks will be moved
[10:04:00] <Gargoyle> Derick: Driver is holding strong. pretty consistent 200ish connections to each of the 3 nodes in the RS. No segfaults spotted in apache, and delivered 25,000+ pageviews already this morning. :)
[10:07:14] <Derick> woo!
[10:07:26] <Derick> i'm making slight changes, but that should be for standalone servers only
[10:08:50] <Gargoyle> if my munin traffic graphs are correct, PRIMARY is pumping out 70Mbps, and the Secondaries 30Mbps each.
[10:14:22] <NodeX> what's the app Gargoyle ?
[10:14:35] <Gargoyle> http://thebestof.co.uk
[10:15:07] <NodeX> re-launch ?
[10:15:15] <Gargoyle> yup
[10:15:40] <NodeX> nice job
[10:16:03] <NodeX> how are you handling the area lookups out if interest ?
[10:16:07] <NodeX> of interest*
[10:16:31] <Gargoyle> can't claim much credit for the design itself. But it's been a nightmare fitting the "business logic" into that design.
[10:16:48] <Gargoyle> Internally there are 500+ locations that we search by name.
[10:17:03] <NodeX> ah ok
[10:17:17] <Gargoyle> if that fails, we use google to attempt to find the lat long for what you typed, and then find the closest match from the 500.
[10:17:27] <NodeX> it's one of my things that I do well - matching
[10:17:30] <NodeX> matching areas
[10:18:07] <Gargoyle> Although, we seem to have broken something.
[10:18:22] <Gargoyle> Maybe google have cut off another service
[10:18:23] <NodeX> I was going to say I found 2 bugs for you lol
[10:18:42] <Gargoyle> They dropped the weather API we were using last week!
[10:18:50] <NodeX> :/
[10:18:52] <Gargoyle> just the 2 ;)
[10:18:54] <Gargoyle> :P
[10:19:27] <NodeX> if you like I can give you my database to help you out, it's made up of 2million or so places in the UK including every postcode and placename (towns/villages)
[10:19:46] <NodeX> I spent 2 years creating it and fine tuning it
[10:19:56] <Gargoyle> nice!
[10:20:38] <Gargoyle> I might take you up on that. At the moment, the focus is on franchisees promoting their area - and not so much the geo-coding
[10:20:47] <Gargoyle> but that will come in the future.
[10:21:06] <lupisak> NodeX: what's the license on that db (esp. post codes)?
[10:21:08] <NodeX> http://www.theukjobsite.co.uk/seek <--- type in the area to see it in action - it suggests and also looks up on search
[10:21:15] <NodeX> lupisak : no liscence
[10:21:50] <lupisak> you mean you personally collected the postcodes, or is it from the openpostcode project?
[10:22:10] <Derick> Gargoyle: yr.no has a free whether API - it's also awesome
[10:22:24] <NodeX> lupisak various sources
[10:22:29] <Gargoyle> Derick: Oooh! Thanks.
[10:22:34] <NodeX> none of them being the PAF
[10:22:43] <Zelest> Is it safe to blindly upgrade mongodb from 2.0.3 to 2.2 ?
[10:22:47] <lupisak> NodeX: a lot of people are interested in postcodes, but the post office have copyrighted them, preventing free use (bastards!)
[10:22:47] <Derick> NodeX: are they all open and free, and perhaps useable for OSM?
[10:22:51] <NodeX> Zelest : I did
[10:22:58] <Zelest> Is that a yes? ;-)
[10:23:02] <Mmike> Hi. How do I run --repair option? I had a server crash, and now when I start server I get 'mongo.lock file exists...'. It is stated that I run mongod wih --repair option, but it just spits the same error in the log file
[10:23:13] <Mmike> i'm doing this: mongod --config /etc/mongod.conf --repair
[10:23:21] <lupisak> NodeX: ok, none from PAF - good to hear
[10:23:43] <Zelest> Mmike, check the log.. if the repair is fine and all, remove the lockfile.
[10:23:50] <NodeX> Derick : I am willing to open source my data, but in general UK postcode databases are made from the PAF which carries a licence charge from the Royal Mail (UK Post people)
[10:23:51] <Zelest> (someone tell me if I'm wrong here though)
[10:24:20] <solars> hey, is there a java object mapper much like mongoid for ruby?
[10:24:20] <BurtyB> think I had to remove the lock file manually to do the repair
[10:24:35] <NodeX> I geocoded (added a lat/long pair) to every single postcode, place, placename and geohashed it all too
[10:24:38] <Gargoyle> NodeX: Derick: Isn't some of that incorporated into the OS open data?
[10:24:45] <lupisak> NodeX: I can probably tell you if that's possible if you tell us where the data is from - which various sources?
[10:24:57] <Derick> Gargoyle: no, PAF is special :-(
[10:25:12] <NodeX> lupisak, anyone can make it
[10:25:17] <BurtyB> PAF has "special" pricing too
[10:25:19] <Derick> NodeX: yes, I know PAF is expensive and stupid
[10:25:20] <NodeX> a bit of regex and a few loops ;)
[10:25:40] <NodeX> creating the database isn't the hard part, geo coding it is
[10:25:54] <lupisak> NodeX: ah, so you mean you looped through your own address database, and ripped the postcodes?
[10:25:59] <Gargoyle> NodeX: Shame didn't meet you a few years back
[10:26:01] <Mmike> Zelest, thnx, seems to be all ok :)
[10:26:09] <NodeX> lupisak : no not exactly but not far off
[10:26:15] <Gargoyle> Had a full geogoded PAF database
[10:26:21] <Gargoyle> *coded
[10:26:31] <Derick> Gargoyle: From RoyalMail?
[10:26:38] <NodeX> Gargoyle : did you get it from PAF ? because I released one into the wild a few years back
[10:26:57] <Zelest> Mmike, Perhaps the repair should remove the lock-file once its done.. but I think I had a similar issue when I expienced the same thing.
[10:27:00] <Derick> also: http://freethepostcode.org
[10:27:01] <Gargoyle> Derick: NodeX: Yeah, and I think geocoded by a third party
[10:27:05] <NodeX> At the time I geocoded the 28million PAF database for a company
[10:27:16] <NodeX> company didn;'t pay me, data went out to the world ;)
[10:27:25] <NodeX> it was probably mine that you had!!
[10:27:28] <Derick> not sure how shady that is for OSM really
[10:27:33] <lupisak> NodeX: sorry to bug you on this, but could you explain exactly how you dreived the data? If done the right way, this data could be interesting for the projects mentioned.
[10:27:43] <NodeX> lupisak : I cant sorry
[10:27:53] <Mmike> Zelest, yes, that's what I was expecting, that the lock file would be removed by repair
[10:28:15] <Mmike> I am now running db.repairDatabase, just to make sure
[10:28:15] <NodeX> it's the basis of alot of my current contracts and retainers so I keep it close to my chest
[10:28:18] <lupisak> NodeX: too bad, it can't be used in open source projects then.
[10:28:43] <NodeX> oh well
[10:28:49] <lupisak> :)
[10:28:50] <NodeX> no skin off my nose ;)
[10:30:54] <Gargoyle> Derick: I'm going to have to learn to speak norwegian!
[10:31:37] <NodeX> Gargoyle : be careful with reverse geocoding with google, after a few requests they throttle like mad
[10:31:45] <NodeX> after a few thousand they cut you off
[10:32:02] <Gargoyle> We've probably been cut off then! :/
[10:32:17] <nkuttler> how can i order a distinct() query? i need an ordered list of embedded objects, or should i not use distinct()?
[10:32:21] <NodeX> perhaps add a caching layer if you dont already have one
[10:33:14] <Gargoyle> NodeX: We only look in google when the place name doesn't match.
[10:33:28] <NodeX> do you then add it to your databasE?
[10:33:37] <Gargoyle> nope.
[10:33:50] <NodeX> nkuttler : you might want to look at the aggregation framework
[10:34:23] <Gargoyle> But a full gazzateer is on the future list anyway. but we also have to rollout to .au in the new year.
[10:34:47] <nkuttler> NodeX: will do
[10:34:55] <NodeX> geonames have a full gazeteer but it needs work
[10:36:49] <lupisak> regarding post codes. Seems the Ordnance Survey has released their data under an open license: http://www.freepostcodes.org.uk/
[10:37:15] <Derick> but not open enough for PAF
[10:38:31] <NodeX> I removed the addresses from mine
[10:39:38] <NodeX> that data ^^ is enough to reverse geocode and create a postcode database of every postcode in the UK
[10:46:37] <lupisak> <@Derick> but not open enough for PAF -- what do you mean?
[10:47:19] <NodeX> I think he means for an address lookup
[10:47:26] <sysdef> hello. i read there is no version for android. there is (still) no version?
[11:13:41] <gigo1980> does anybody know this situation ?
[11:13:42] <gigo1980> https://jira.mongodb.org/browse/SERVER-7008
[11:18:29] <PDani> hi
[11:20:55] <PDani> I have a replicated environment with two members and an arbiter. I'm experiencing periodical slowdowns when running relatively intensively insert queries. I'm using 2.2.0. I investigated this issue a lot, I tried to turn off journaling, and I set syncdelay to 0 (for testing purposes), but it didn't help. I'm trying to find out what could cause this issue.
[11:22:15] <kali> PDani: have you checked for preallocation messages in the log ?
[11:22:34] <kali> PDani: what file system your data is on ?
[11:24:06] <PDani> At every slowdown, I see this in the log: http://goo.gl/RnDzr
[11:24:34] <PDani> and xfs
[11:25:17] <kali> ok, not preallocation, then
[11:25:36] <PDani> it happens approximately in every 30 seconds
[11:56:02] <Zelest> Derick, http://www.httpdb.org/rawr :-)
[12:16:59] <Derick> Zelest: :-)
[12:17:45] <Zelest> had the crawler dead for a while, hence the huge backlog :P
[12:18:44] <Zelest> it's far from optimized though, seeing the machine only got 1gb of ram
[12:18:48] <Zelest> need to sort that out somehow
[12:21:08] <BurtyB> heh Zelest I've started to add that kind of data to my db too.. tho you've got a head start!
[12:22:17] <NodeX> Zelest : have you got any aggregate stats yet
[12:22:34] <NodeX> I'd be interested to see the proportion per country of web servers
[12:22:37] <Zelest> Nah, I've sort of let it be for a few months
[12:22:46] <Zelest> Might pick it up again soonish :)
[12:22:54] <Zelest> Generate some sexy stats
[12:23:21] <Zelest> What bothers me the most is the performance..
[12:23:28] <Zelest> Takes ages to do queries in that amount of data.
[12:23:41] <NodeX> how big are the databases?
[12:23:57] <Zelest> how do I know? db.stats() ?
[12:24:02] <NodeX> yer
[12:24:23] <Zelest> "dataSize" : 5359165384,
[12:24:59] <Zelest> 4.43GB index data :P
[12:25:07] <Zelest> 1GB of ram in the machine
[12:25:16] <Zelest> I will most likely move it to my laptop when I generate the stats.
[12:25:17] <NodeX> I'd happily provide some resources for aggregations
[12:25:26] <NodeX> I have a few high powered servers
[12:25:39] <Zelest> Hmms
[12:25:55] <NodeX> I can at least aggregate down into months for you, then it will take the load off
[12:26:52] <Zelest> Thanks for the offer.. my main concern though is about not sharing the data..
[12:27:11] <Zelest> Not that I personally care, but I've got a few mails where people have been quite picky who has access to the data. :P
[12:27:35] <NodeX> no probs
[12:28:09] <Zelest> I barely know how to extract the aggregated data once it's all sorted though
[12:28:16] <Zelest> But that's a future issue :)
[12:28:31] <Zelest> My Mac has enough hardware to handle it just fine I hope
[12:28:37] <BurtyB> Zelest, do you only have 12M domains to search?
[12:28:54] <Zelest> BurtyB, No domains, only IP.
[12:29:04] <Zelest> That's somewhat the goal of it.
[12:29:09] <NodeX> I have a huge domain list
[12:29:11] <BurtyB> Zeelot, ah
[12:29:22] <NodeX> 140M domains and counting
[12:29:23] <Zelest> Seeing sites like netcraft do their research based upon domains.
[12:29:27] <Zelest> wow
[12:29:38] <BurtyB> NodeX, that including subdomains?
[12:29:38] <NodeX> all in mongo lol
[12:29:45] <NodeX> no, just root
[12:29:54] <NodeX> foo.com, bar.com
[12:30:35] <NodeX> 140,829,065 documents LOL
[12:31:07] <BurtyB> NodeX, ah, grabbing the zonefiles probably gets you that or more
[12:31:16] <NodeX> I already have them
[12:31:19] <NodeX> this is them plus more
[12:32:07] <NodeX> :/
[12:32:16] <NodeX> one of my clients is a massive domaineer
[12:32:28] <NodeX> he has me doing all manner of crap with domains
[13:27:42] <yhpark> Use a DBRef when you need to embed documents from multiple collections in documents from one collection. DBRefs also provide a common format and type to represent these relationships among documents.
[13:27:51] <yhpark> from http://docs.mongodb.org/manual/applications/database-references/
[13:29:21] <yhpark> hard to understand
[13:30:27] <yhpark> if (documents from one collection) embed (documents from multiple collections) : Use DBRef ?
[13:32:10] <yhpark> hmm..
[13:40:25] <gigo1980> "errmsg" : "MR post processing failed: { errmsg: \"exception: could not initialize cursor across all shards because : socket exception [CONNECT_ERROR] for set12/mongo07.luan.local:10022,mongo08.luan....\", code: 14827, ok: 0.0 }"
[13:40:35] <gigo1980> but this shard does not longer exists
[13:40:44] <gigo1980> also restart the config servers ...
[13:40:48] <gigo1980> and mongo router
[13:40:53] <gigo1980> NEED HELP!
[14:46:54] <W0rmDrink> Hi, I have a system consisting of n computers - request for each user can reach each system - but one request may be processed per user at any one time in the system overall - is there a name for mechanisms used to ensure this - I normally refer to it just as syncronization but its slightly vauge
[14:49:51] <sysdef> is there a way to get mongodb support? a mailing list or something like that? or an official support channel?
[14:50:53] <meghan> https://groups.google.com/group/mongodb-user is the best place
[14:51:55] <sysdef> looks like i don't have access to that page
[14:51:56] <W0rmDrink> further - to achive this we typically do something like : try_lock: db.collection.update( { userId : ..., ownerId : NONE }, { $set : { ownerId : ?self? } }, true, false );
[14:51:56] <W0rmDrink> is_locked: db.collection.count( { userId : ..., ownerId : ?self? } ) != 0;
[14:51:56] <W0rmDrink> unlock: db.collection.delete( { userId : ..., ownerId : ?self? } );
[14:51:56] <W0rmDrink> - and I would like to know if anybody has better Ideas
[14:52:58] <W0rmDrink> and if this is not best place to ask - where should I ask
[14:55:53] <landstalker> sysdef: 10gen provide paid support plans
[14:56:56] <Derick> sysdef: you need to joind the google groups before you can access it - but we (10gen) also provide paid support plans
[14:57:16] <Derick> sysdef: or you can just ask here
[14:58:42] <abourget> what's the primary shard used by mongos if sharding is not enabled ? is it the first server listed as 'configdb' in mongos.conf ?
[14:59:27] <sysdef> i asked teh questio two times and dien't got a response yet. that's why i asked for an other way
[14:59:31] <sysdef> i read there is no version for android. there is (still) no version?
[14:59:43] <landstalker> client or server?
[14:59:50] <sysdef> server
[14:59:58] <abourget> the 'shard0000' in the 'shards' collection in the config db ?
[15:00:00] <sysdef> actually i'm using BDB
[15:03:24] <landstalker> you want to runb MongoDB server on Android?
[15:03:54] <ron> haha!
[15:06:07] <Derick> abourget: it's the server where you ran "enable sharding"
[15:06:44] <Derick> sysdef: MongoDB isn't made to run on an embedded device really.
[15:06:56] <abourget> Derick, but where will mongos store the data, if it,s configured to deal with 3 shards, but none of them have been enabled with "enable sharding" ?
[15:07:35] <Derick> Will mongos even be able to talk to any shard in that case? I don't think it would.
[15:07:45] <Derick> I don't have a definitive answer on that one though
[15:07:48] <sysdef> Derick: ok, thank you. i'll stay with BerkeleyDB
[15:11:35] <abourget> Derick, ok.. I just saw, the shards do have 'shardsvr' == true
[15:11:55] <abourget> Derick, can we connect directly to a server that has `shardsvr` == true ? and store data there ? or *must* be accessed through mongos ?
[15:12:12] <Derick> you don't have to, but you really should
[15:13:00] <Derick> anyway, you you can find out what your primary shard is too
[15:13:02] <Derick> let me find out how
[15:14:09] <Derick> > use admin
[15:14:10] <Derick> > db.printShardingStatus();
[15:22:26] <abourget> ok thanks :)
[15:22:42] <abourget> mustn't `shardsvr = true` be applied to all shards ?
[15:22:53] <Derick> if you want to add them to the shard cluster, yes
[15:27:46] <P-Alex> hi all :)
[15:28:07] <P-Alex> it's possible to count all distinct values on a collection?
[15:28:34] <NodeX> distinct();p
[15:28:41] <P-Alex> distinct().count()
[15:28:42] <P-Alex> ?
[15:28:49] <NodeX> oh count them all
[15:29:27] <P-Alex> but i need to count only the distinct number o values
[15:29:42] <NodeX> you'll have to do it appside
[15:30:14] <NodeX> you can use length
[15:30:27] <NodeX> distinct('foo').length;
[15:30:53] <NodeX> I never thought to use length till now lol
[15:31:53] <P-Alex> :D ok it works thanx XD
[15:32:07] <NodeX> ;)
[15:49:57] <NodeX> anyone know if i can do an $inc on a sub document that doesnt exist yet?
[15:51:14] <NodeX> I want to do something like db.foo.update({ymd:20120911,type:'geohash'},{$inc:{"foo.myhash":1}});
[15:52:03] <jsaacmk> http://pastebin.com/5JJUA3Uf is there a more concise way for me to do the query at the bottom of the post (ideally with mongodb directives). Also, why does the result print twice when I run it in the mongo CLI?
[15:59:12] <nostalgeek> Hi all, I am fairly new to Mongo. I think I've got a replicas between 2 mongod working (ports 27018). I also have 2 mongos routers (ports 27017). Now, if I understand right, when I connect to the mongos (27017) and issue slaveOk() this allow routing of find() queries to my secondary. How can I tell if the query was served by the secondary or by the master?
[16:27:39] <astropirate> I am working on a CMS. Documents can have child documents. There is no limit how deep the tree structure can be. Should I reference the child documents or embed them? Also what would be the way to get the child documents indexed too based their "slug" property
[16:29:50] <astropirate> Any advice would be very much appreciated
[16:31:29] <NodeX> I dont embed if that;s help
[16:31:35] <Samurai> @astro from personal experience i would reference them, it will make querying much easier
[16:32:26] <NodeX> easy querying does not always trump performance
[16:32:41] <NodeX> infact it seldom does
[16:33:35] <NodeX> astropirate: the general rule of thumb is whatever suits your data
[16:35:14] <astropirate> hmm
[16:36:09] <Samurai> i have a quesion.. is there a way to return the whole collection from a disctinct command match?
[16:36:10] <astropirate> If I embed, I can't index them
[16:36:24] <NodeX> why can't you index embeded >?/
[16:36:44] <NodeX> Samurai : whole collection or whole document?
[16:37:01] <Samurai> whole document*
[16:37:07] <astropirate> NodeX, because I have no Idea how deep the embedded document is
[16:37:15] <NodeX> no because distinct in an aggregate of values
[16:37:27] <NodeX> it's not any one single document
[16:37:35] <Samurai> i see
[16:37:50] <NodeX> you want aggration framework if you need to return things like thaty
[16:38:12] <NodeX> astropirate : what does that matter?
[16:38:32] <Samurai> @NodeX thanks going to google it
[16:38:37] <NodeX> ;)
[16:39:16] <NodeX> astropirate : what will be queried more in your CMS?
[16:40:23] <astropirate> NodeX, what do you mean
[16:40:27] <astropirate> i didnt get the question
[16:40:47] <NodeX> I cant explain it any differently
[16:40:52] <NodeX> that's as simple as it gets
[16:43:25] <astropirate> NodeX, what will be queried more? more what of what two options?
[16:49:53] <NodeX> it's your app, you tell me /
[16:53:00] <nostalgeek> answer to my previous question: I enabled profiling on the DB, and confirmed my read requests are logged on my secondary, and any inserts are logged on the primary!
[17:58:20] <camonz> is there a way to coerce all figures returned in a map/reduce to integers?
[17:59:12] <camonz> I'm already wrapping every figure in each operation with a parseInt(...)
[18:00:49] <skot> you can use NumberInt(…)
[18:01:05] <skot> which will create a bson int, not a javascript number
[18:01:35] <skot> or NumberLong for a 64bit int
[18:15:55] <camonz> skot: thanks
[19:01:40] <iyp> Anyone have advice on dynamically generating json-schema compliant docs on homogeneous(ish) collections?
[19:16:09] <crudson> iyp: that's going to come from one of your application layers somewhere, and depending on what language you are using. There's nothing currently in mongodb to associate a schema with a collection.
[19:17:46] <iyp> crudson: Gotchya. Do you know if theres anything in the json-schema spec that allows for summarizing populated data? Or is that another problem entirely?
[19:19:29] <crudson> iyp: I have no idea, sorry. That would be a question for the json-schema folks.
[19:20:33] <iyp> crudson: Righto. Thanks anyways
[19:42:52] <iyp> crudson: Got any insight on data warehousing with mongodb?
[19:44:45] <iyp> I intend to keep a cache of the data sources with something like Hbase and then transform that data into mongodb collections to sit under a web interface for exploration.
[19:45:16] <iyp> See anything blatantly wrong with that? (input welcomed from anyone)
[20:43:16] <Rhaven> hi all, i have just added a shard on a running shard cluster, since that i have got some high latency... and i got some error like that
[20:44:18] <Rhaven> http://pastebin.com/wyWah8bi
[20:45:44] <_m> Have you checked here: https://groups.google.com/forum/?fromgroups=#!topic/mongodb-user/lm7Ht1yCpeQ
[20:46:48] <_m> There's a lot of information on the error messages you linked available via google. I suggest turning there first. Apologies in advance if you've already exhausted that route.
[21:08:26] <tomlikestorock> anyone here know how to connect to a replica set in mongohub over an ssh tunnel successfull?
[21:11:37] <tomlikestorock> successfully*
[21:19:44] <R-66Y> is there a way to ask for keys as constructed strings
[21:19:47] <R-66Y> for instance
[21:19:50] <R-66Y> var a = "week";
[21:19:57] <R-66Y> db.games.find({a: 1});
[21:20:18] <R-66Y> should be equivalent to: db.games.find({"week": 1});
[21:24:07] <Neptu> somethat that used pymongo how i can do db.lala.insert(..) or db.lele.insert() dynamically?
[21:36:21] <R-66Y> using $where got me there, i suppose
[21:36:24] <R-66Y> so like
[21:36:27] <R-66Y> var a = "week";
[21:36:42] <R-66Y> db.games.find({ "$where": a + " == 1" });
[21:36:49] <R-66Y> seems a little dirty but it works
[21:37:28] <Gaddel> hello, i'm brand new to mongodb. if i have a collection of documents and i want a string from one of the fields to the be primary key, is my best option just duplicating the field value for "_id" as well?
[21:37:33] <Gaddel> to be the*
[21:39:19] <kchodorow> R-66Y: q = {}; q[a] = 1; db.games.find(q)
[21:40:58] <kchodorow> Gaddel: just use the string as the _id field
[21:41:48] <crudson> Gaddel: you can just use it as _id providing it's unique and logically appropriate. Downside is that it may not be obvious to consumers of that document what that field means without further knowledge.
[21:42:03] <Gaddel> crudson: right, what was my concern. in this case the field is an IP address.
[21:42:50] <Gaddel> which isn't naturally very similar to an "ID" so to speak. but i could probably store it there and then re-assign it to "IP" when presenting the document
[21:47:34] <crudson> Gaddel: if your collection is 'ip_address_stats' or something, then it's obvious what the _id means, otherwise I personally see benefit in making it obvious what attributes mean. Just my opinion.
[22:22:40] <fg3> I can search for my-field true -- how can I search for my-field: false -- or does not exist
[22:37:20] <_m> fg3: Check the $exists documentation. http://bit.ly/xpiM0V
[22:53:31] <fg3> found it
[22:53:35] <fg3> thanks