[08:53:40] <KekSi> i'm having trouble with the mongodb java driver - MongoCredential.createMongoX509Credential in particular -- somehow it doesn't seem to read the correct certificate from the keystore i put in - can anyone help me with that?
[09:04:35] <maasan> mongodb is using lot of memory. How do i limit the memory?
[10:21:22] <Har1ekin> not exactly, there is ONE special field that must not (under no circumstance) be null or our hardware brakes
[10:22:51] <Har1ekin> my boss just want's to make sure (in addition to the check our main application is running before inserting into mongo) that the db throws back any inserts/updates, that would turn this field into null or empty
[10:23:51] <Har1ekin> like good old sql does... but mongo would be better for our product and i can't convince him without this "feature"
[10:27:48] <arussel> Har1ekin: no, this is an application level feature, mongo don't do this
[12:52:21] <mitereiter> so isn't it suposed to be moved to the shard specified by the tagrange?
[14:00:25] <fxmulder> so one of my mongod's won't start, I can start it standalone but starting in the replica set fails with Wed May 20 07:50:14.594 Invalid access at address: 0x73ac2e8504e8 from thread: repl writer worker 1
[14:22:14] <mitereiter> so isn't it suposed to be moved to the shard specified by the tagrange?
[14:22:23] <aps> Hi. I'm new to mongodb and have a doubt. Is it advisable to "have the backup replica servers in one shard serving as the other shard primary (replica of primary of other shard)"?
[14:22:46] <aps> like it's mentioned here - https://github.com/ansible/ansible-examples/tree/master/mongodb#deploy-the-cluster
[14:47:26] <dorongutman> anyways - connecting to the primary and trying to run the monitoring agent manually (/var/lib/mongodb-mms-automation/mongodb-mms-monitoring-agent-3.3.0.183-1.linux_x86_64/mongodb-mms-monitoring-agent ) prompts the following error:
[16:27:35] <jkli> so in what kind of field is it used for
[16:27:45] <jkli> where it excels compared to like mysql
[16:28:10] <StephenLynx> systems where you have a high load to handle, where you don't need to stitch data together dynamically.
[16:28:29] <StephenLynx> gridfs also makes it excellent for storing files.
[16:28:31] <fewknow> jkli: it is different and used for different patterns than mysql
[16:29:15] <fewknow> i use a combination of relational, mongodb, elastic search and hadoop
[16:29:22] <StephenLynx> also from what I heard, its easier to build shards with mongo
[16:29:41] <fewknow> StephenLynx: that is an understatement compared to mysql
[16:30:13] <StephenLynx> never built a cluster with neither so I try to not make bold statements related to that :v
[16:31:49] <fewknow> StephenLynx: understand...i have built with both...
[16:31:49] <StephenLynx> so in general, don't expect mongo to be neither a full replacement for relational dbs nor useless compared to them. databases are one of those things you just can't have a one-size-fits-all
[17:17:41] <StephenLynx> to be fair, _id has this SPECHIAL rule with projection.
[17:17:51] <StephenLynx> by default it is informed and you have to ask for it to not be informed.
[17:18:15] <StephenLynx> how mongo handles _id as a god damn special snowflake is the one thing that really bothers me with mongo
[19:43:10] <deathanchor> pymongo question: created a collection object and running a 20K finds on it (exhausting the cursor each time), basically is going to be the fastest method?
[19:48:50] <deathanchor> ack, now I know why it is slow
[19:49:58] <deathanchor> my query uses an index which doesn't have the field value I want.
[20:10:55] <Progz> hello, I am new in the mongo world. And I am wondering, is it healthy to have 400 GB of data in one signle mongo server in one database with 4 collection ( 1 : 5 GB, 2 : 10 GB, 3 : 40 GB, 4: 350 GO)
[20:15:21] <GothAlice> Progz: Nothing particularly unusual about that, depending entirely on your data.
[20:16:25] <GothAlice> Progz: I find http://www.devsmash.com/blog/mongodb-ad-hoc-analytics-aggregation-framework to be a useful article not just for if you're doing analytics or aggregation, but because it clearly demonstrates the performance and storage impact of different ways to store the same data. Might be interesting for you to see how design can change storage space.
[20:16:40] <Progz> We have a lot of trouble with our data model and with aggregate taking more than 10 seconde.
[20:17:54] <GothAlice> Then that link may be doubly useful for you. Pre-aggregation lets you organize your data (or a copy of the data) in a way that can give constant-time performance. I.e. to "query a week's worth of data", given 15,000 documents in that week, it'll clearly be faster to sum up 7 daily total documents instead of going through all 15,000 original documents.
[20:19:00] <GothAlice> (Esp. if the number of documents being evaluated varies based on usage; your queries then have variable performance. Not so much with pre-aggregation; if instead there were 15 *million* documents in that week, the pre-aggregate would still only sum the 7 daily documents. :D
[21:33:47] <pocketprotector> anyway to set this before the system boots? /sys/kernel/mm/transparent_hugepage/defrag
[21:34:40] <pocketprotector> the kernel parameter does not fix it
[21:54:52] <hackel> I've got a collection with a composite, sparse, unique index, but keep getting "duplicate key" errors when trying to insert a record without one of the indexed fields. Any ideas?
[23:10:42] <joannac> shaggy_surfer: did you have a question?
[23:18:45] <shaggy_surfer> joannac: Thanks, I was just testing that I could actually get a response since registering.
[23:26:06] <Antiarc> Is there a place to discuss the design decisions of the mongo driver? Some of the 2.0 changes are really pretty incomprehensible and frustrating. A single global logger? One database per connection?
[23:27:35] <Antiarc> Topology detection doesn't seem to properly determine that a single-node replica set is in replica set mode
[23:50:24] <joannac> Antiarc: is there something special you need from the replica set mode?
[23:51:07] <Antiarc> I'm tailing the oplogs to index documents into Elasticsearch, and need to be able to tell if the node I've connected to is a mongos router, a mongod node in RS mode, or a mongod node not in RS mode.
[23:51:35] <Antiarc> Easy to do via various status commands, but it seems like the topology detection should probably do what it suggests on the tin
[23:52:12] <Antiarc> (important, since non-RS nodes won't have an oplog!)
[23:52:17] <joannac> Antiarc: even connecting with the replica set option?
[23:52:49] <Antiarc> It works if you explicitly force connect: :replica_set. But I don't know ahead of time that the target is an RS. I can make that configurable, but the client should be able to figure that out on its own.
[23:53:24] <joannac> if we enabled that, then you would never be able to connect to a specific node in a replica set
[23:54:32] <Antiarc> Sorry, how do you mean? You could force-connect with direct mode if you wanted a single node.
[23:55:44] <Antiarc> I can distinguish between a node that is part of an RS or not via the replSetGetStatus command (and the exception raised if it's not part of an RS), without having to know anything about the target cluster. Just seems odd that the client wouldn't do something similar.
[23:57:20] <Boomtime> Antiarc: again, as joannac says, if we did that then you couldn't connect directly
[23:57:48] <Boomtime> the idea of a replica-set connection is for you to supply multiple hosts in the connection string
[23:58:14] <Boomtime> if you supply only one host then you're not really achieving good resilency to failure since your client can;'t connect if that one host is down
[23:58:44] <Boomtime> thus, there is no getting around the fact that the client must know ahead of time a rough idea of what it's connecting to
[23:59:14] <Boomtime> if it expects a replica-set it will name multiple hosts, to ensure that if any one of them is down it can still contact the remaining members
[23:59:47] <Antiarc> Yup. And if you pass an array with a single entry (ie, RS mode connections), it'll say that you're in single node mode.