[02:36:28] <jiffe> so I'm trying to run db.runCommand({cleanupOrphaned: "mail.message_data"}) and I'm getting "errmsg" : "Error encountered while deleting range: nsmail.message_data from { message_identifier ... cause by: :: caused by :: 10334 BSONObj size: 1542706610 (0x5BF3D5B2) is invalid. Size must be between 0 and 16793600(16MB)
[03:49:42] <Waheedi> is there a way to make a client interact with one specific node only. that node is obviously part of a replicaset
[03:55:55] <Waheedi> I wish i had two primaries :D
[04:04:37] <Boomtime> @Waheedi: yes, you can create a direct connection - every driver supports this, but if the node is secondary you won't be able to write to it (only read)
[04:10:24] <Waheedi> maybe that was the case only in 2.4.1
[04:10:27] <Boomtime> ok; it sounds like you have multiple sites and want to constrain reads occurring in one site to members in that site - is that a fair description?
[04:10:35] <Waheedi> but have not tried the 3.x yet
[04:10:53] <Waheedi> Boomtime: yeah might be close
[04:31:40] <jiffe> when I try to start a couple nodes I'm getting [repl writer worker 15] writer worker caught exception: :: caused by :: 10334 BSONObj size: 1542706610 (0x5BF3D5B2) is invalid
[04:31:47] <jiffe> is there a way I can just skip that?
[04:34:42] <jiffe> this used to happen with mysql where replication would fail due to a problem with a record and you could set the skip counter, replication would skip the record and go on its merry way
[04:35:46] <Waheedi> jiffe: i used to find these records and delete them from primary
[04:36:25] <Waheedi> maybe take a backup of that record
[04:38:14] <jiffe> so when I start up these nodes they're not trying to read this out of a local log, its pulling all the replication data from the master?
[04:39:35] <Waheedi> your question has two parts :)
[09:28:27] <mementomori> 2) MongoDB can not enforce unique indexes on a sharded collection other than the shard key itself, or indexes prefixed by the shard key.
[09:30:26] <mementomori> here I asnwered False because it could be possibile that data is stored on different shards and mongo can't enforce uniqueness
[09:30:59] <mementomori> 3) Any update that does not contain the shard key will be sent to all shards
[09:34:49] <kali> for 2) if two values are identical and contains the shard key, then the shardkey is identical and the values go to the same shard, so mongodb can ensure they are unique
[09:36:07] <mementomori> kali: 2) also says "or indexes prefixed by the shard key."
[09:36:41] <mementomori> and that could cause data to be in different shards
[14:07:21] <jiffe> when I try to start a couple nodes I'm getting [repl writer worker 15] writer worker caught exception: :: caused by :: 10334 BSONObj size: 1542706610 (0x5BF3D5B2) is invalid, is there a way to skip it?
[14:43:19] <jiffe> so do slaves copy objects from the master's oplog to the local oplog before applying?
[15:07:21] <jiffe> I searched for db.oplog.rs.find({h: 8979387659673550781}) on both the primary and secondary and it is not found in either, so where is this coming from?
[16:43:46] <jiffe> ok so I found the oplog record in the master
[16:44:55] <jiffe> there doesn't seem to be anything I can do about it, I can't remove it from the oplog nor can I remove the referenced record from the recovering node
[16:46:24] <kali> jiffe: i would fix the database on the master (by discarding or reducing the document too big) and then resync
[16:47:52] <jiffe> kali: resyncing these boxes takes about a week and this seems to happen frequently so I'm looking for a quicker route
[16:48:03] <Zelest> how can I decide what fields I wish to return using the ruby driver?
[16:52:41] <Zelest> oh, nevermind! .projection() is what I was looking for. :)
[16:54:58] <jiffe> maybe I can modify the master oplog record to delete a record that doesn't exist instead of the one its having a problem with
[18:01:42] <jiffe> "errmsg" : "failing update: objects in a capped ns cannot grow"
[19:06:25] <Zelest> does the field names take up diskspace? e.g, if I have a LOT of documents, will I save diskspace if I use shorter field names or does wiredtiger handle this automagically?