[11:29:42] <kees_> just create a db with 150 documents, see how big that is, take that number, do it times 1000, look how big that number is, do it times 1000 again and there is your answer ;)
[11:32:15] <Mickelodeonion> so, there is no way to pre-calculate
[13:00:57] <hallucynogenyc> Hey guys, trying to get started with mongo
[13:01:49] <hallucynogenyc> installed via homebrew in mac, however the mongod process exits with status 2 (bad options according to the docs) but without writting anything to the logs
[13:02:01] <hallucynogenyc> have noidea how to debug this any further
[13:03:04] <joannac> hallucynogenyc: what ptions are you starting it with
[13:15:03] <remonvv> Hi guys, we have a weird issue; we have a database that takes 6.4Gb filesize for 58Mb storage size (according to db.stats()). This margin is inexplicably huge. The data are simply 2 field tuples, no growing documents.
[13:16:49] <remonvv> @joannac: Certainly not to that extent I think. Shouldn't it start reusing at some point? I know usePowerOf2Sizes helps but I was under the impression it reuses file extents at some point.
[13:19:22] <joannac> remonvv: yes it will, but free lists are populated by buckets.
[13:19:59] <joannac> remonvv: and we don't exhaustively search because that leads to performance problems. So we give up if we haven't found an appropriately sized extent and create a new one.
[13:20:37] <joannac> we give after after looking at X extents in the free list, I mean
[13:20:43] <remonvv> @joannac: Okay, but that seems to happen pretty much always.
[13:21:12] <remonvv> @joannac: So what would you suggest we do? Short term to correct this state and long term to avoid it happening again.
[13:21:50] <hallucynogenyc> SIGINT 2 Term Interrupt from keyboard
[13:22:00] <hallucynogenyc> What? I mean look at this
[13:23:40] <hallucynogenyc> how can it clain I stopped it?
[13:23:43] <joannac> you'll need to repair it... or maybe compact... i can't remember off the top of my head
[13:24:20] <remonvv> @joannac: I'll look it up. Thanks. I'll assume there are various JIRA tickets open for this being improved already?
[13:26:13] <remonvv> Hm, can't run compact on mongos apparently and both compact and repair lock the database.
[14:31:56] <hallucynogenyc> ok joannac I figured it out, the brew installation was using a wrong path for the mongod executable, it was missing that mongod is inside /bin
[14:32:11] <hallucynogenyc> why was this showing up as a keyboard interrupt I have no idea
[14:44:52] <hallucynogenyc> this is the pull request to fix it
[16:39:06] <kali> you can't have .anything_variable, this is not javascript
[16:43:47] <bryanlarsen> I'm trying to add a member to a replica set. I followed http://docs.mongodb.org/manual/tutorial/expand-replica-set/. Now rs.conf() on primary shows a stateStr of "UNKNOWN" for the new member and rs.conf() on the new member shows a startupStatus of 4, and an errmsg of ""can't currently get local.system.replset config from self or any seed (EMPTYUNREACHABLE)""
[16:44:12] <bryanlarsen> The database is large and recovery will take a while, but I don't think that it's recovering. Any pointers?
[16:49:32] <kali> can the two hosts speak to each other ?
[16:50:15] <bryanlarsen> yes, I used telnet to make sure that the ports are visible from each machine
[16:51:26] <bryanlarsen> I also was able to successfully move a configsvr onto the new machine
[16:51:46] <kali> no trivial mistake on hostname, port number ? because it really sound like something like that
[16:56:08] <bryanlarsen> in my mongodb.log on the new server, I get a bunch of "connection accepted" messages from all my mongos servers
[16:57:39] <bryanlarsen> the log says "trying to contact". I double checked that I was able to telnet to the host/port listed for all of those, and that works
[17:02:32] <bryanlarsen> now it's showing "initial sync cloning db: studio_production". It'll probably be like that for a day or so. Thanks for your help kali, much appreciated
[17:11:57] <Nodex> in all seriousness, just build one yourself, take a day or so
[17:12:14] <edsiper> thats not my point, i need to find people using HTTP interfaces to query mongo
[17:12:30] <edsiper> someone using something like this: http://docs.mongodb.org/ecosystem/tools/http-interfaces/
[17:13:06] <Nodex> I doubt you will find people because it's trivial to build one yourself. Why rely on someone else's blated code when it can be streamlined to your app
[17:13:50] <edsiper> why do yo assume that i need someone code ?.
[17:14:08] <edsiper> pointless discussion, well, if anyone is using some rest interface please ping me
[17:14:26] <Nodex> again, you will find it HARD to find someone becuase PEOPLE BUILD THIER OWN
[17:15:21] <edsiper> well, "build their own" and "use some existent HTTP interface as the listed above" applies...just need to contact people using/creating HTTP interfaces to query mongo
[17:26:36] <morsik> kali: ok, anyway… continuing, does mongodb could be good for big data things? as i read on mongodb page, it's already used for things like that, also i found that some company is using mongo for monitoring data
[17:28:56] <morsik> this looks promising. http://www.mongodb.com/presentations/humongous-data-server-density-approaching-1-billion-documents-mongodb + https://blog.serverdensity.com/server-density-stats-in-april-2011/
[17:47:10] <kali> morsik: mongodb is for big data, no doubt about that. but not at all for batch processing (where hadoop can help)
[17:48:15] <kali> morsik: for monitoring, as you're likely to manipulate time series all over the place, I personally think the document model is not a good fit. That said, there are blog entries by a mongodb staffer stating otherwise.
[17:52:12] <morsik> kali: well, we monitor our web pages. download times, http return codes, all html (when error occured). also mail checks (so different structure)
[18:05:52] <Aartsie> a hosting company asks me to work for them and what my price is by hour but i have no idea what the markt is at the moment, anyone a suggestion ?
[18:06:31] <morsik> depends on country, region in that country, what you should to there, etc. etc.
[18:57:36] <rkgarcia> kali: what are you doing here? >_>
[18:58:21] <def-lkb> Hi, I am a complete beginner with mongo. I think my setup is suboptimal… So far I have used only one instance, but I think I am reaching performance limit.
[18:59:08] <def-lkb> I bought a dedicated box for mongo (8 cores), only one is used. Queries often timeout if there are concurrent writes.
[18:59:38] <kali> rkgarcia: trying to avoid one more "mongodb sucks" blog post six months from today :)
[19:00:34] <def-lkb> (only one core). What is the common setup if I want to use multiple cores?
[19:02:07] <kali> def-lkb: you may have rush into buying this box. writes have a per-db lock. you can improves thing by splitting your data across more than one database, or sharding.
[19:02:19] <kali> def-lkb: how many writes/s are we talking about ?
[19:11:17] <def-lkb> kali: the common workload is no write at all for around ~10 minutes, then around 60000 documents are written in a burst, nothing for 10 minutes, and so on
[19:12:38] <kali> mmmm... what kind of write ? insert ?
[19:13:01] <def-lkb> kali: maybe explaining the use can helps: there is small core dataset (around ~1GB), which are processed by external tools, generating a few gigabytes at a time after a few minutes.
[19:14:23] <def-lkb> only on first iteration. then, its mostly updates
[21:05:51] <McSorley> Hey #mongodb, looking for some advice on importing some large datasets into Mongo. I have a number of large CSV files ranging in size from 15-20GB that i would like to load in, just curious to see if anyone can offer any advice/strategies for dealing with some large data files..?