master-slave
主的配置文件 /Users/olifer/middle/mongo/master-slave/master/mongod.conf
bind_ip = 127.0.0.1
port = 27017
dbpath = /Users/olifer/middle/mongo/master-slave/master/data/
master = true
从的配置文件
bind_ip = 127.0.0.1
port = 27018
dbpath = /Users/olifer/middle/mongo/master-slave/slave/data/
slave = true
source = 127.0.0.1:27017
启动主服务器,并且看到下面的日志,说明配置成功。
mongod -f ~/middle/mongo/master-slave/master/mongod.conf
2017-12-05T15:35:22.479+0800 I JOURNAL [initandlisten] journal dir=/Users/olifer/middle/mongo/master-slave/master/data/journal
2017-12-05T15:35:22.479+0800 I JOURNAL [initandlisten] recover : no journal files present, no recovery needed
2017-12-05T15:35:22.500+0800 I JOURNAL [durability] Durability thread started
2017-12-05T15:35:22.500+0800 I JOURNAL [journal writer] Journal writer thread started
2017-12-05T15:35:22.501+0800 I CONTROL [initandlisten] MongoDB starting : pid=6368 port=27017 dbpath=/Users/olifer/middle/mongo/master-slave/master/data/ master=1 64-bit host=oliferdeMacBook-Pro.local
2017-12-05T15:35:22.501+0800 I CONTROL [initandlisten] db version v3.0.7
2017-12-05T15:35:22.501+0800 I CONTROL [initandlisten] git version: nogitversion
2017-12-05T15:35:22.501+0800 I CONTROL [initandlisten] build info: Darwin yosemitevm.local 14.5.0 Darwin Kernel Version 14.5.0: Wed Jul 29 02:26:53 PDT 2015; root:xnu-2782.40.9~1/RELEASE_X86_64 x86_64 BOOST_LIB_VERSION=1_49
2017-12-05T15:35:22.501+0800 I CONTROL [initandlisten] allocator: system
2017-12-05T15:35:22.501+0800 I CONTROL [initandlisten] options: { config: "/Users/olifer/middle/mongo/master-slave/master/mongod.conf", master: true, net: { bindIp: "127.0.0.1", port: 27017 }, storage: { dbPath: "/Users/olifer/middle/mongo/master-slave/master/data/" } }
2017-12-05T15:35:22.505+0800 I INDEX [initandlisten] allocating new ns file /Users/olifer/middle/mongo/master-slave/master/data/local.ns, filling with zeroes...
2017-12-05T15:35:22.570+0800 I STORAGE [FileAllocator] allocating new datafile /Users/olifer/middle/mongo/master-slave/master/data/local.0, filling with zeroes...
2017-12-05T15:35:22.570+0800 I STORAGE [FileAllocator] creating directory /Users/olifer/middle/mongo/master-slave/master/data/_tmp
2017-12-05T15:35:22.786+0800 I STORAGE [FileAllocator] done allocating datafile /Users/olifer/middle/mongo/master-slave/master/data/local.0, size: 64MB, took 0.215 secs
2017-12-05T15:35:22.972+0800 I REPL [initandlisten] ******
2017-12-05T15:35:22.972+0800 I REPL [initandlisten] creating replication oplog of size: 192MB...
2017-12-05T15:35:22.972+0800 I STORAGE [FileAllocator] allocating new datafile /Users/olifer/middle/mongo/master-slave/master/data/local.1, filling with zeroes...
2017-12-05T15:35:24.061+0800 I STORAGE [FileAllocator] done allocating datafile /Users/olifer/middle/mongo/master-slave/master/data/local.1, size: 256MB, took 1.088 secs
2017-12-05T15:35:24.117+0800 I REPL [initandlisten] ******
2017-12-05T15:35:24.119+0800 I NETWORK [initandlisten] waiting for connections on port 27017
启动从服务器,并且看到下面的日志,说明配置成功。
mongod -f ~/middle/mongo/master-slave/slave/mongod.conf
2017-12-05T15:37:11.518+0800 I JOURNAL [initandlisten] journal dir=/Users/olifer/middle/mongo/master-slave/slave/data/journal
2017-12-05T15:37:11.518+0800 I JOURNAL [initandlisten] recover : no journal files present, no recovery needed
2017-12-05T15:37:11.535+0800 I JOURNAL [durability] Durability thread started
2017-12-05T15:37:11.535+0800 I JOURNAL [journal writer] Journal writer thread started
2017-12-05T15:37:11.535+0800 I CONTROL [initandlisten] MongoDB starting : pid=7304 port=27018 dbpath=/Users/olifer/middle/mongo/master-slave/slave/data/ slave=1 64-bit host=oliferdeMacBook-Pro.local
2017-12-05T15:37:11.535+0800 I CONTROL [initandlisten] db version v3.0.7
2017-12-05T15:37:11.535+0800 I CONTROL [initandlisten] git version: nogitversion
2017-12-05T15:37:11.535+0800 I CONTROL [initandlisten] build info: Darwin yosemitevm.local 14.5.0 Darwin Kernel Version 14.5.0: Wed Jul 29 02:26:53 PDT 2015; root:xnu-2782.40.9~1/RELEASE_X86_64 x86_64 BOOST_LIB_VERSION=1_49
2017-12-05T15:37:11.535+0800 I CONTROL [initandlisten] allocator: system
2017-12-05T15:37:11.535+0800 I CONTROL [initandlisten] options: { config: "/Users/olifer/middle/mongo/master-slave/slave/mongod.conf", net: { bindIp: "127.0.0.1", port: 27018 }, slave: true, source: "127.0.0.1:27017", storage: { dbPath: "/Users/olifer/middle/mongo/master-slave/slave/data/" } }
2017-12-05T15:37:11.536+0800 I INDEX [initandlisten] allocating new ns file /Users/olifer/middle/mongo/master-slave/slave/data/local.ns, filling with zeroes...
2017-12-05T15:37:11.605+0800 I STORAGE [FileAllocator] allocating new datafile /Users/olifer/middle/mongo/master-slave/slave/data/local.0, filling with zeroes...
2017-12-05T15:37:11.605+0800 I STORAGE [FileAllocator] creating directory /Users/olifer/middle/mongo/master-slave/slave/data/_tmp
2017-12-05T15:37:11.824+0800 I STORAGE [FileAllocator] done allocating datafile /Users/olifer/middle/mongo/master-slave/slave/data/local.0, size: 64MB, took 0.218 secs
2017-12-05T15:37:11.890+0800 I NETWORK [initandlisten] waiting for connections on port 27018
2017-12-05T15:37:12.894+0800 I REPL [replslave] repl: syncing from host:127.0.0.1:27017
2017-12-05T15:37:17.930+0800 I REPL [replslave] repl: sleep 2 sec before next pass
2017-12-05T15:37:19.935+0800 I REPL [replslave] repl: syncing from host:127.0.0.1:27017
2017-12-05T15:37:24.958+0800 I REPL [replslave] repl: syncing from host:127.0.0.1:27017
2017-12-05T15:37:33.188+0800 I REPL [replslave] repl: syncing from host:127.0.0.1:27017
2017-12-05T15:37:43.193+0800 I REPL [replslave] repl: syncing from host:127.0.0.1:27017
2017-12-05T15:37:53.198+0800 I REPL [replslave] repl: syncing from host:127.0.0.1:27017
我们可以通过slave的日志中可以看到,slave一直同master保持同步数据的联系。
我们通过客户端链接上服务端
mongo 127.0.0.1:27017 #主
mongo 127.0.0.1:27018 #从
在主上操作
> db.test.insert({"name":"linyang"})
WriteResult({ "nInserted" : 1 })
> db.test.find({});
{ "_id" : ObjectId("5a264e0ecb6f7d3713c516a7"), "name" : "linyang" }
往里面插入了一条记录,可以再从的日志中发现同步的记录
2017-12-05T15:43:10.387+0800 I INDEX [replslave] allocating new ns file /Users/olifer/middle/mongo/master-slave/slave/data/test.ns, filling with zeroes...
2017-12-05T15:43:10.450+0800 I STORAGE [FileAllocator] allocating new datafile /Users/olifer/middle/mongo/master-slave/slave/data/test.0, filling with zeroes...
2017-12-05T15:43:10.646+0800 I STORAGE [FileAllocator] done allocating datafile /Users/olifer/middle/mongo/master-slave/slave/data/test.0, size: 64MB, took 0.196 secs
2017-12-05T15:43:10.702+0800 I REPL [replslave] resync: dropping database test
2017-12-05T15:43:10.709+0800 I JOURNAL [replslave] journalCleanup...
2017-12-05T15:43:10.710+0800 I JOURNAL [replslave] removeJournalFiles
2017-12-05T15:43:10.713+0800 I JOURNAL [replslave] journalCleanup...
2017-12-05T15:43:10.713+0800 I JOURNAL [replslave] removeJournalFiles
2017-12-05T15:43:10.715+0800 I REPL [replslave] resync: cloning database test to get an initial copy
2017-12-05T15:43:10.718+0800 I INDEX [replslave] allocating new ns file /Users/olifer/middle/mongo/master-slave/slave/data/test.ns, filling with zeroes...
2017-12-05T15:43:10.803+0800 I STORAGE [FileAllocator] allocating new datafile /Users/olifer/middle/mongo/master-slave/slave/data/test.0, filling with zeroes...
2017-12-05T15:43:11.056+0800 I STORAGE [FileAllocator] done allocating datafile /Users/olifer/middle/mongo/master-slave/slave/data/test.0, size: 64MB, took 0.252 secs
2017-12-05T15:43:11.107+0800 I INDEX [replslave] build index on: test.test properties: { v: 1, key: { _id: 1 }, name: "_id_", ns: "test.test" }
2017-12-05T15:43:11.107+0800 I INDEX [replslave] building index using bulk method
2017-12-05T15:43:11.107+0800 I INDEX [replslave] build index done. scanned 1 total records. 0 secs
2017-12-05T15:43:11.107+0800 I STORAGE [replslave] copying indexes for: { name: "test", options: {} }
2017-12-05T15:43:11.108+0800 I REPL [replslave] resync: done with initial clone for db: test
我们再从的client中查看有没有信息。
mongo 127.0.0.1:27018
MongoDB shell version: 3.0.7
connecting to: 127.0.0.1:27018/test
> db.test.find({});
{ "_id" : ObjectId("5a264e0ecb6f7d3713c516a7"), "name" : "linyang" }
>
也确认同步过来了。然后再从上插入记录
> db.test.insert({"age":34});
WriteResult({ "writeError" : { "code" : undefined, "errmsg" : "not master" } })
>
从上不能插入记录。如果主挂掉了,从又不能写数据,那么是否mongo集群就挂掉了。答案当然是否定的,因为,mongo还有副本集。
副本集
我们使用三台mongo来模拟.
第一台的配置文件 /Users/olifer/middle/mongo/replica/a/mongod.conf
dbpath = /Users/olifer/middle/mongo/replica/a/data
port = 8001
bind_ip = 127.0.0.1
replSet = child/127.0.0.1:8002
第二台的配置文件 /Users/olifer/middle/mongo/replica/b/mongod.conf
dbpath = /Users/olifer/middle/mongo/replica/a/data
port = 8002
bind_ip = 127.0.0.1
replSet = child/127.0.0.1:8003
第三台的配置文件 /Users/olifer/middle/mongo/replica/c/mongod.conf
dbpath = /Users/olifer/middle/mongo/replica/a/data
port = 8003
bind_ip = 127.0.0.1
replSet = child/127.0.0.1:8001
启动服务器
mongod -f ~/middle/mongo/replica/a/mongod.conf
mongod -f ~/middle/mongo/replica/b/mongod.conf
mongod -f ~/middle/mongo/replica/c/mongod.conf
启动成功后,进入三个服务的任何一个客户端
mongo 127.0.0.1:8002
> config = {_id: 'child', members: [
{
"_id":1,
"host":"127.0.0.1:8001"
},
{
"_id":2,
"host":"127.0.0.1:8002"
},
{
"_id":3,
"host":"127.0.0.1:8003"
}
]
}
{
"_id" : "child",
"members" : [
{
"_id" : 1,
"host" : "127.0.0.1:8001"
},
{
"_id" : 2,
"host" : "127.0.0.1:8002"
},
{
"_id" : 3,
"host" : "127.0.0.1:8003"
}
]
}
> rs.initiate(config);
{ "ok" : 1 }
child:SECONDARY>
配置完后发现前缀发生了改变,我们进入另外两个渠道的客户端
mongo 127.0.0.1:8001
MongoDB shell version: 3.0.7
connecting to: 127.0.0.1:8001/test
child:PRIMARY>
mongo 127.0.0.1:8003
MongoDB shell version: 3.0.7
connecting to: 127.0.0.1:8003/test
child:SECONDARY>
其中child:PRIMARY>表示活跃节点。其余为备份节点。注意:只有活跃节点才能进行查询数据库的信息操作,备份节点不能进行会报错,在任意客户端执行rs.status() 来查看所有状态
child:PRIMARY> rs.status()
{
"set" : "child",
"date" : ISODate("2017-12-05T08:56:44.523Z"),
"myState" : 1,
"members" : [
{
"_id" : 1,
"name" : "127.0.0.1:8001",
"health" : 1,
"state" : 1,
"stateStr" : "PRIMARY",
"uptime" : 620,
"optime" : Timestamp(1512463818, 1),
"optimeDate" : ISODate("2017-12-05T08:50:18Z"),
"electionTime" : Timestamp(1512463821, 1),
"electionDate" : ISODate("2017-12-05T08:50:21Z"),
"configVersion" : 1,
"self" : true
},
{
"_id" : 2,
"name" : "127.0.0.1:8002",
"health" : 1,
"state" : 2,
"stateStr" : "SECONDARY",
"uptime" : 387,
"optime" : Timestamp(1512463818, 1),
"optimeDate" : ISODate("2017-12-05T08:50:18Z"),
"lastHeartbeat" : ISODate("2017-12-05T08:56:43.756Z"),
"lastHeartbeatRecv" : ISODate("2017-12-05T08:56:43.756Z"),
"pingMs" : 0,
"configVersion" : 1
},
{
"_id" : 3,
"name" : "127.0.0.1:8003",
"health" : 1,
"state" : 2,
"stateStr" : "SECONDARY",
"uptime" : 387,
"optime" : Timestamp(1512463818, 1),
"optimeDate" : ISODate("2017-12-05T08:50:18Z"),
"lastHeartbeat" : ISODate("2017-12-05T08:56:43.756Z"),
"lastHeartbeatRecv" : ISODate("2017-12-05T08:56:43.756Z"),
"pingMs" : 0,
"configVersion" : 1
}
],
"ok" : 1
}
搭建成功后,我们来验证一下。
在活跃的节点创建数据,在备份库拉取数据
child:PRIMARY> db.repl.insert({"name":"123"});
WriteResult({ "nInserted" : 1 })
child:SECONDARY> db.repl.find();
{ "_id" : ObjectId("5a26602f5cefb1fdb377843b"), "name" : "123" }
功能正常。
从服务器不能写
child:SECONDARY> db.repl.insert({"age":444});
WriteResult({ "writeError" : { "code" : undefined, "errmsg" : "not master" } })
关闭活跃的节点,从服务器会通过选举得到最新的活跃的节点
当我关闭原来活跃的服务器8001后,通过内部选举,8002成了最新的活跃节点没从前缀也可以看出
child:SECONDARY>
child:PRIMARY>
现在看一下最新的状态
child:PRIMARY> rs.status();
{
"set" : "child",
"date" : ISODate("2017-12-05T09:09:31.666Z"),
"myState" : 1,
"members" : [
{
"_id" : 1,
"name" : "127.0.0.1:8001",
"health" : 0,
"state" : 8,
"stateStr" : "(not reachable/healthy)",
"uptime" : 0,
"optime" : Timestamp(0, 0),
"optimeDate" : ISODate("1970-01-01T00:00:00Z"),
"lastHeartbeat" : ISODate("2017-12-05T09:09:31.193Z"),
"lastHeartbeatRecv" : ISODate("2017-12-05T09:07:12.773Z"),
"pingMs" : 0,
"lastHeartbeatMessage" : "Failed attempt to connect to 127.0.0.1:8001; couldn't connect to server 127.0.0.1:8001 (127.0.0.1), connection attempt failed",
"configVersion" : -1
},
{
"_id" : 2,
"name" : "127.0.0.1:8002",
"health" : 1,
"state" : 1,
"stateStr" : "PRIMARY",
"uptime" : 1308,
"optime" : Timestamp(1512464432, 2),
"optimeDate" : ISODate("2017-12-05T09:00:32Z"),
"electionTime" : Timestamp(1512464835, 1),
"electionDate" : ISODate("2017-12-05T09:07:15Z"),
"configVersion" : 1,
"self" : true
},
{
"_id" : 3,
"name" : "127.0.0.1:8003",
"health" : 1,
"state" : 2,
"stateStr" : "SECONDARY",
"uptime" : 1152,
"optime" : Timestamp(1512464432, 2),
"optimeDate" : ISODate("2017-12-05T09:00:32Z"),
"lastHeartbeat" : ISODate("2017-12-05T09:09:30.963Z"),
"lastHeartbeatRecv" : ISODate("2017-12-05T09:09:30.963Z"),
"pingMs" : 0,
"configVersion" : 1
}
],
"ok" : 1
}
说明内部的切换确实成功了。还剩下一个分片,我们下次再讲。
参考资料
MongoDB 3.0 常见集群的搭建(主从复制,副本集,分片....)