move id only when necessary RUBY-88
This commit is contained in:
parent
f6e505e511
commit
e075d25a91
|
@ -139,7 +139,7 @@ static int cmp_char(const void* a, const void* b) {
|
||||||
return *(char*)a - *(char*)b;
|
return *(char*)a - *(char*)b;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void write_doc(buffer_t buffer, VALUE hash, VALUE check_keys);
|
static void write_doc(buffer_t buffer, VALUE hash, VALUE check_keys, VALUE move_id);
|
||||||
static int write_element(VALUE key, VALUE value, VALUE extra);
|
static int write_element(VALUE key, VALUE value, VALUE extra);
|
||||||
static VALUE elements_to_hash(const char* buffer, int max);
|
static VALUE elements_to_hash(const char* buffer, int max);
|
||||||
|
|
||||||
|
@ -147,6 +147,10 @@ static VALUE pack_extra(buffer_t buffer, VALUE check_keys) {
|
||||||
return rb_ary_new3(2, LL2NUM((long long)buffer), check_keys);
|
return rb_ary_new3(2, LL2NUM((long long)buffer), check_keys);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static VALUE pack_triple(buffer_t buffer, VALUE check_keys, int allow_id) {
|
||||||
|
return rb_ary_new3(3, LL2NUM((long long)buffer), check_keys, allow_id);
|
||||||
|
}
|
||||||
|
|
||||||
static void write_name_and_type(buffer_t buffer, VALUE name, char type) {
|
static void write_name_and_type(buffer_t buffer, VALUE name, char type) {
|
||||||
SAFE_WRITE(buffer, &type, 1);
|
SAFE_WRITE(buffer, &type, 1);
|
||||||
name = TO_UTF8(name);
|
name = TO_UTF8(name);
|
||||||
|
@ -236,7 +240,7 @@ static int write_element_allow_id(VALUE key, VALUE value, VALUE extra, int allow
|
||||||
case T_HASH:
|
case T_HASH:
|
||||||
{
|
{
|
||||||
write_name_and_type(buffer, key, 0x03);
|
write_name_and_type(buffer, key, 0x03);
|
||||||
write_doc(buffer, value, check_keys);
|
write_doc(buffer, value, check_keys, Qfalse);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
case T_ARRAY:
|
case T_ARRAY:
|
||||||
|
@ -289,7 +293,7 @@ static int write_element_allow_id(VALUE key, VALUE value, VALUE extra, int allow
|
||||||
SAFE_WRITE(buffer, (char*)&length, 4);
|
SAFE_WRITE(buffer, (char*)&length, 4);
|
||||||
SAFE_WRITE(buffer, RSTRING_PTR(value), length - 1);
|
SAFE_WRITE(buffer, RSTRING_PTR(value), length - 1);
|
||||||
SAFE_WRITE(buffer, &zero, 1);
|
SAFE_WRITE(buffer, &zero, 1);
|
||||||
write_doc(buffer, rb_funcall(value, rb_intern("scope"), 0), Qfalse);
|
write_doc(buffer, rb_funcall(value, rb_intern("scope"), 0), Qfalse, Qfalse);
|
||||||
|
|
||||||
total_length = buffer_get_position(buffer) - start_position;
|
total_length = buffer_get_position(buffer) - start_position;
|
||||||
SAFE_WRITE_AT_POS(buffer, length_location, (const char*)&total_length, 4);
|
SAFE_WRITE_AT_POS(buffer, length_location, (const char*)&total_length, 4);
|
||||||
|
@ -463,10 +467,11 @@ static int write_element(VALUE key, VALUE value, VALUE extra) {
|
||||||
return write_element_allow_id(key, value, extra, 0);
|
return write_element_allow_id(key, value, extra, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void write_doc(buffer_t buffer, VALUE hash, VALUE check_keys) {
|
static void write_doc(buffer_t buffer, VALUE hash, VALUE check_keys, VALUE move_id) {
|
||||||
buffer_position start_position = buffer_get_position(buffer);
|
buffer_position start_position = buffer_get_position(buffer);
|
||||||
buffer_position length_location = buffer_save_space(buffer, 4);
|
buffer_position length_location = buffer_save_space(buffer, 4);
|
||||||
buffer_position length;
|
buffer_position length;
|
||||||
|
int allow_id;
|
||||||
VALUE id_str = rb_str_new2("_id");
|
VALUE id_str = rb_str_new2("_id");
|
||||||
VALUE id_sym = ID2SYM(rb_intern("_id"));
|
VALUE id_sym = ID2SYM(rb_intern("_id"));
|
||||||
|
|
||||||
|
@ -474,6 +479,9 @@ static void write_doc(buffer_t buffer, VALUE hash, VALUE check_keys) {
|
||||||
rb_raise(rb_eNoMemError, "failed to allocate memory in buffer.c");
|
rb_raise(rb_eNoMemError, "failed to allocate memory in buffer.c");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// write '_id' first if move_id is true
|
||||||
|
if(move_id == Qtrue) {
|
||||||
|
allow_id = 0;
|
||||||
if (rb_funcall(hash, rb_intern("has_key?"), 1, id_str) == Qtrue) {
|
if (rb_funcall(hash, rb_intern("has_key?"), 1, id_str) == Qtrue) {
|
||||||
VALUE id = rb_hash_aref(hash, id_str);
|
VALUE id = rb_hash_aref(hash, id_str);
|
||||||
write_element_allow_id(id_str, id, pack_extra(buffer, check_keys), 1);
|
write_element_allow_id(id_str, id, pack_extra(buffer, check_keys), 1);
|
||||||
|
@ -481,6 +489,14 @@ static void write_doc(buffer_t buffer, VALUE hash, VALUE check_keys) {
|
||||||
VALUE id = rb_hash_aref(hash, id_sym);
|
VALUE id = rb_hash_aref(hash, id_sym);
|
||||||
write_element_allow_id(id_sym, id, pack_extra(buffer, check_keys), 1);
|
write_element_allow_id(id_sym, id, pack_extra(buffer, check_keys), 1);
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
allow_id = 1;
|
||||||
|
if ((rb_funcall(hash, rb_intern("has_key?"), 1, id_str) == Qtrue) &&
|
||||||
|
(rb_funcall(hash, rb_intern("has_key?"), 1, id_sym) == Qtrue)) {
|
||||||
|
VALUE obj = rb_hash_delete(hash, id_str);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// we have to check for an OrderedHash and handle that specially
|
// we have to check for an OrderedHash and handle that specially
|
||||||
if (strcmp(rb_class2name(RBASIC(hash)->klass), "OrderedHash") == 0) {
|
if (strcmp(rb_class2name(RBASIC(hash)->klass), "OrderedHash") == 0) {
|
||||||
|
@ -490,10 +506,10 @@ static void write_doc(buffer_t buffer, VALUE hash, VALUE check_keys) {
|
||||||
VALUE key = RARRAY_PTR(keys)[i];
|
VALUE key = RARRAY_PTR(keys)[i];
|
||||||
VALUE value = rb_hash_aref(hash, key);
|
VALUE value = rb_hash_aref(hash, key);
|
||||||
|
|
||||||
write_element(key, value, pack_extra(buffer, check_keys));
|
write_element_allow_id(key, value, pack_extra(buffer, check_keys), allow_id);
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
rb_hash_foreach(hash, write_element, pack_extra(buffer, check_keys));
|
rb_hash_foreach(hash, write_element_allow_id, pack_triple(buffer, check_keys, allow_id));
|
||||||
}
|
}
|
||||||
|
|
||||||
// write null byte and fill in length
|
// write null byte and fill in length
|
||||||
|
@ -509,14 +525,14 @@ static void write_doc(buffer_t buffer, VALUE hash, VALUE check_keys) {
|
||||||
SAFE_WRITE_AT_POS(buffer, length_location, (const char*)&length, 4);
|
SAFE_WRITE_AT_POS(buffer, length_location, (const char*)&length, 4);
|
||||||
}
|
}
|
||||||
|
|
||||||
static VALUE method_serialize(VALUE self, VALUE doc, VALUE check_keys) {
|
static VALUE method_serialize(VALUE self, VALUE doc, VALUE check_keys, VALUE move_id) {
|
||||||
VALUE result;
|
VALUE result;
|
||||||
buffer_t buffer = buffer_new();
|
buffer_t buffer = buffer_new();
|
||||||
if (buffer == NULL) {
|
if (buffer == NULL) {
|
||||||
rb_raise(rb_eNoMemError, "failed to allocate memory in buffer.c");
|
rb_raise(rb_eNoMemError, "failed to allocate memory in buffer.c");
|
||||||
}
|
}
|
||||||
|
|
||||||
write_doc(buffer, doc, check_keys);
|
write_doc(buffer, doc, check_keys, move_id);
|
||||||
|
|
||||||
result = rb_str_new(buffer_get_buffer(buffer), buffer_get_position(buffer));
|
result = rb_str_new(buffer_get_buffer(buffer), buffer_get_position(buffer));
|
||||||
if (buffer_free(buffer) != 0) {
|
if (buffer_free(buffer) != 0) {
|
||||||
|
@ -872,7 +888,7 @@ void Init_cbson() {
|
||||||
CBson = rb_define_module("CBson");
|
CBson = rb_define_module("CBson");
|
||||||
ext_version = rb_str_new2(VERSION);
|
ext_version = rb_str_new2(VERSION);
|
||||||
rb_define_const(CBson, "VERSION", ext_version);
|
rb_define_const(CBson, "VERSION", ext_version);
|
||||||
rb_define_module_function(CBson, "serialize", method_serialize, 2);
|
rb_define_module_function(CBson, "serialize", method_serialize, 3);
|
||||||
rb_define_module_function(CBson, "deserialize", method_deserialize, 1);
|
rb_define_module_function(CBson, "deserialize", method_deserialize, 1);
|
||||||
|
|
||||||
rb_require("digest/md5");
|
rb_require("digest/md5");
|
||||||
|
|
|
@ -126,6 +126,8 @@ module Mongo
|
||||||
#
|
#
|
||||||
# @raise [RuntimeError]
|
# @raise [RuntimeError]
|
||||||
# if given unknown options
|
# if given unknown options
|
||||||
|
#
|
||||||
|
# @see Official, <a href="http://dochub.mongodb.org/find" name="find-instance_method" rel="find">Official Docs</a>
|
||||||
def find(selector={}, opts={})
|
def find(selector={}, opts={})
|
||||||
fields = opts.delete(:fields)
|
fields = opts.delete(:fields)
|
||||||
fields = ["_id"] if fields && fields.empty?
|
fields = ["_id"] if fields && fields.empty?
|
||||||
|
@ -198,6 +200,8 @@ module Mongo
|
||||||
# If true, check that the save succeeded. OperationFailure
|
# If true, check that the save succeeded. OperationFailure
|
||||||
# will be raised on an error. Note that a safe check requires an extra
|
# will be raised on an error. Note that a safe check requires an extra
|
||||||
# round-trip to the database.
|
# round-trip to the database.
|
||||||
|
#
|
||||||
|
# @see Official, <a href="http://dochub.mongodb.org/save" name="save-instance_method" rel="find">Official Docs</a>
|
||||||
def save(doc, options={})
|
def save(doc, options={})
|
||||||
if doc.has_key?(:_id) || doc.has_key?('_id')
|
if doc.has_key?(:_id) || doc.has_key?('_id')
|
||||||
id = doc[:_id] || doc['_id']
|
id = doc[:_id] || doc['_id']
|
||||||
|
@ -294,7 +298,7 @@ module Mongo
|
||||||
update_options += 2 if options[:multi]
|
update_options += 2 if options[:multi]
|
||||||
message.put_int(update_options)
|
message.put_int(update_options)
|
||||||
message.put_array(BSON.serialize(selector, false).to_a)
|
message.put_array(BSON.serialize(selector, false).to_a)
|
||||||
message.put_array(BSON.serialize(document, false).to_a)
|
message.put_array(BSON.serialize(document, false, true).to_a)
|
||||||
if options[:safe]
|
if options[:safe]
|
||||||
@connection.send_message_with_safe_check(Mongo::Constants::OP_UPDATE, message, @db.name,
|
@connection.send_message_with_safe_check(Mongo::Constants::OP_UPDATE, message, @db.name,
|
||||||
"db.#{@name}.update(#{selector.inspect}, #{document.inspect})")
|
"db.#{@name}.update(#{selector.inspect}, #{document.inspect})")
|
||||||
|
@ -609,7 +613,7 @@ EOS
|
||||||
# Initial byte is 0.
|
# Initial byte is 0.
|
||||||
message = ByteBuffer.new([0, 0, 0, 0])
|
message = ByteBuffer.new([0, 0, 0, 0])
|
||||||
BSON_RUBY.serialize_cstr(message, "#{@db.name}.#{collection_name}")
|
BSON_RUBY.serialize_cstr(message, "#{@db.name}.#{collection_name}")
|
||||||
documents.each { |doc| message.put_array(BSON.serialize(doc, check_keys).to_a) }
|
documents.each { |doc| message.put_array(BSON.serialize(doc, check_keys, true).to_a) }
|
||||||
if safe
|
if safe
|
||||||
@connection.send_message_with_safe_check(Mongo::Constants::OP_INSERT, message, @db.name,
|
@connection.send_message_with_safe_check(Mongo::Constants::OP_INSERT, message, @db.name,
|
||||||
"db.#{collection_name}.insert(#{documents.inspect})")
|
"db.#{collection_name}.insert(#{documents.inspect})")
|
||||||
|
|
|
@ -1,8 +1,8 @@
|
||||||
# A thin wrapper for the CBson class
|
# A thin wrapper for the CBson class
|
||||||
class BSON_C
|
class BSON_C
|
||||||
|
|
||||||
def self.serialize(obj, check_keys=false)
|
def self.serialize(obj, check_keys=false, move_id=false)
|
||||||
ByteBuffer.new(CBson.serialize(obj, check_keys))
|
ByteBuffer.new(CBson.serialize(obj, check_keys, move_id))
|
||||||
end
|
end
|
||||||
|
|
||||||
def self.deserialize(buf=nil)
|
def self.deserialize(buf=nil)
|
||||||
|
|
|
@ -87,15 +87,15 @@ class BSON_RUBY
|
||||||
|
|
||||||
# Serializes an object.
|
# Serializes an object.
|
||||||
# Implemented to ensure an API compatible with BSON extension.
|
# Implemented to ensure an API compatible with BSON extension.
|
||||||
def self.serialize(obj, check_keys=false)
|
def self.serialize(obj, check_keys=false, move_id=false)
|
||||||
new.serialize(obj, check_keys)
|
new.serialize(obj, check_keys, move_id)
|
||||||
end
|
end
|
||||||
|
|
||||||
def self.deserialize(buf=nil)
|
def self.deserialize(buf=nil)
|
||||||
new.deserialize(buf)
|
new.deserialize(buf)
|
||||||
end
|
end
|
||||||
|
|
||||||
def serialize(obj, check_keys=false)
|
def serialize(obj, check_keys=false, move_id=false)
|
||||||
raise "Document is null" unless obj
|
raise "Document is null" unless obj
|
||||||
|
|
||||||
@buf.rewind
|
@buf.rewind
|
||||||
|
@ -103,13 +103,19 @@ class BSON_RUBY
|
||||||
@buf.put_int(0)
|
@buf.put_int(0)
|
||||||
|
|
||||||
# Write key/value pairs. Always write _id first if it exists.
|
# Write key/value pairs. Always write _id first if it exists.
|
||||||
|
if move_id
|
||||||
if obj.has_key? '_id'
|
if obj.has_key? '_id'
|
||||||
serialize_key_value('_id', obj['_id'], check_keys)
|
serialize_key_value('_id', obj['_id'], check_keys)
|
||||||
elsif obj.has_key? :_id
|
elsif obj.has_key? :_id
|
||||||
serialize_key_value('_id', obj[:_id], check_keys)
|
serialize_key_value('_id', obj[:_id], check_keys)
|
||||||
end
|
end
|
||||||
|
|
||||||
obj.each {|k, v| serialize_key_value(k, v, check_keys) unless k == '_id' || k == :_id }
|
obj.each {|k, v| serialize_key_value(k, v, check_keys) unless k == '_id' || k == :_id }
|
||||||
|
else
|
||||||
|
if obj.has_key?('_id') && obj.has_key?(:_id)
|
||||||
|
obj.delete(:_id)
|
||||||
|
end
|
||||||
|
obj.each {|k, v| serialize_key_value(k, v, check_keys) }
|
||||||
|
end
|
||||||
|
|
||||||
serialize_eoo_element(@buf)
|
serialize_eoo_element(@buf)
|
||||||
if @buf.size > 4 * 1024 * 1024
|
if @buf.size > 4 * 1024 * 1024
|
||||||
|
|
|
@ -263,19 +263,19 @@ class BSONTest < Test::Unit::TestCase
|
||||||
val = OrderedHash.new
|
val = OrderedHash.new
|
||||||
val['not_id'] = 1
|
val['not_id'] = 1
|
||||||
val['_id'] = 2
|
val['_id'] = 2
|
||||||
roundtrip = BSON.deserialize(BSON.serialize(val).to_a)
|
roundtrip = BSON.deserialize(BSON.serialize(val, false, true).to_a)
|
||||||
assert_kind_of OrderedHash, roundtrip
|
assert_kind_of OrderedHash, roundtrip
|
||||||
assert_equal '_id', roundtrip.keys.first
|
assert_equal '_id', roundtrip.keys.first
|
||||||
|
|
||||||
val = {'a' => 'foo', 'b' => 'bar', :_id => 42, 'z' => 'hello'}
|
val = {'a' => 'foo', 'b' => 'bar', :_id => 42, 'z' => 'hello'}
|
||||||
roundtrip = BSON.deserialize(BSON.serialize(val).to_a)
|
roundtrip = BSON.deserialize(BSON.serialize(val, false, true).to_a)
|
||||||
assert_kind_of OrderedHash, roundtrip
|
assert_kind_of OrderedHash, roundtrip
|
||||||
assert_equal '_id', roundtrip.keys.first
|
assert_equal '_id', roundtrip.keys.first
|
||||||
end
|
end
|
||||||
|
|
||||||
def test_nil_id
|
def test_nil_id
|
||||||
doc = {"_id" => nil}
|
doc = {"_id" => nil}
|
||||||
assert_equal doc, BSON.deserialize(bson = BSON.serialize(doc).to_a)
|
assert_equal doc, BSON.deserialize(bson = BSON.serialize(doc, false, true).to_a)
|
||||||
end
|
end
|
||||||
|
|
||||||
def test_timestamp
|
def test_timestamp
|
||||||
|
@ -394,4 +394,35 @@ class BSONTest < Test::Unit::TestCase
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
|
def test_move_id
|
||||||
|
a = OrderedHash.new
|
||||||
|
a['text'] = 'abc'
|
||||||
|
a['key'] = 'abc'
|
||||||
|
a['_id'] = 1
|
||||||
|
|
||||||
|
assert_equal ")\000\000\000\020_id\000\001\000\000\000\002text" +
|
||||||
|
"\000\004\000\000\000abc\000\002key\000\004\000\000\000abc\000\000",
|
||||||
|
BSON.serialize(a, false, true).to_s
|
||||||
|
assert_equal ")\000\000\000\002text\000\004\000\000\000abc\000\002key" +
|
||||||
|
"\000\004\000\000\000abc\000\020_id\000\001\000\000\000\000",
|
||||||
|
BSON.serialize(a, false, false).to_s
|
||||||
|
end
|
||||||
|
|
||||||
|
def test_move_id_with_nested_doc
|
||||||
|
b = OrderedHash.new
|
||||||
|
b['text'] = 'abc'
|
||||||
|
b['_id'] = 2
|
||||||
|
c = OrderedHash.new
|
||||||
|
c['text'] = 'abc'
|
||||||
|
c['hash'] = b
|
||||||
|
c['_id'] = 3
|
||||||
|
assert_equal ">\000\000\000\020_id\000\003\000\000\000\002text" +
|
||||||
|
"\000\004\000\000\000abc\000\003hash\000\034\000\000" +
|
||||||
|
"\000\002text\000\004\000\000\000abc\000\020_id\000\002\000\000\000\000\000",
|
||||||
|
BSON.serialize(c, false, true).to_s
|
||||||
|
assert_equal ">\000\000\000\002text\000\004\000\000\000abc\000\003hash" +
|
||||||
|
"\000\034\000\000\000\002text\000\004\000\000\000abc\000\020_id" +
|
||||||
|
"\000\002\000\000\000\000\020_id\000\003\000\000\000\000",
|
||||||
|
BSON.serialize(c, false, false).to_s
|
||||||
|
end
|
||||||
end
|
end
|
||||||
|
|
Loading…
Reference in New Issue