mongo 集群配置

master-slave

主的配置文件 /Users/olifer/middle/mongo/master-slave/master/mongod.conf

bind_ip = 127.0.0.1
port = 27017
dbpath = /Users/olifer/middle/mongo/master-slave/master/data/
master = true

从的配置文件

bind_ip = 127.0.0.1
port = 27018
dbpath = /Users/olifer/middle/mongo/master-slave/slave/data/
slave = true
source = 127.0.0.1:27017

启动主服务器,并且看到下面的日志,说明配置成功。

mongod -f ~/middle/mongo/master-slave/master/mongod.conf

2017-12-05T15:35:22.479+0800 I JOURNAL  [initandlisten] journal dir=/Users/olifer/middle/mongo/master-slave/master/data/journal
2017-12-05T15:35:22.479+0800 I JOURNAL  [initandlisten] recover : no journal files present, no recovery needed
2017-12-05T15:35:22.500+0800 I JOURNAL  [durability] Durability thread started
2017-12-05T15:35:22.500+0800 I JOURNAL  [journal writer] Journal writer thread started
2017-12-05T15:35:22.501+0800 I CONTROL  [initandlisten] MongoDB starting : pid=6368 port=27017 dbpath=/Users/olifer/middle/mongo/master-slave/master/data/ master=1 64-bit host=oliferdeMacBook-Pro.local
2017-12-05T15:35:22.501+0800 I CONTROL  [initandlisten] db version v3.0.7
2017-12-05T15:35:22.501+0800 I CONTROL  [initandlisten] git version: nogitversion
2017-12-05T15:35:22.501+0800 I CONTROL  [initandlisten] build info: Darwin yosemitevm.local 14.5.0 Darwin Kernel Version 14.5.0: Wed Jul 29 02:26:53 PDT 2015; root:xnu-2782.40.9~1/RELEASE_X86_64 x86_64 BOOST_LIB_VERSION=1_49
2017-12-05T15:35:22.501+0800 I CONTROL  [initandlisten] allocator: system
2017-12-05T15:35:22.501+0800 I CONTROL  [initandlisten] options: { config: "/Users/olifer/middle/mongo/master-slave/master/mongod.conf", master: true, net: { bindIp: "127.0.0.1", port: 27017 }, storage: { dbPath: "/Users/olifer/middle/mongo/master-slave/master/data/" } }
2017-12-05T15:35:22.505+0800 I INDEX    [initandlisten] allocating new ns file /Users/olifer/middle/mongo/master-slave/master/data/local.ns, filling with zeroes...
2017-12-05T15:35:22.570+0800 I STORAGE  [FileAllocator] allocating new datafile /Users/olifer/middle/mongo/master-slave/master/data/local.0, filling with zeroes...
2017-12-05T15:35:22.570+0800 I STORAGE  [FileAllocator] creating directory /Users/olifer/middle/mongo/master-slave/master/data/_tmp
2017-12-05T15:35:22.786+0800 I STORAGE  [FileAllocator] done allocating datafile /Users/olifer/middle/mongo/master-slave/master/data/local.0, size: 64MB,  took 0.215 secs
2017-12-05T15:35:22.972+0800 I REPL     [initandlisten] ******
2017-12-05T15:35:22.972+0800 I REPL     [initandlisten] creating replication oplog of size: 192MB...
2017-12-05T15:35:22.972+0800 I STORAGE  [FileAllocator] allocating new datafile /Users/olifer/middle/mongo/master-slave/master/data/local.1, filling with zeroes...
2017-12-05T15:35:24.061+0800 I STORAGE  [FileAllocator] done allocating datafile /Users/olifer/middle/mongo/master-slave/master/data/local.1, size: 256MB,  took 1.088 secs
2017-12-05T15:35:24.117+0800 I REPL     [initandlisten] ******
2017-12-05T15:35:24.119+0800 I NETWORK  [initandlisten] waiting for connections on port 27017

启动从服务器,并且看到下面的日志,说明配置成功。

mongod -f ~/middle/mongo/master-slave/slave/mongod.conf
2017-12-05T15:37:11.518+0800 I JOURNAL  [initandlisten] journal dir=/Users/olifer/middle/mongo/master-slave/slave/data/journal
2017-12-05T15:37:11.518+0800 I JOURNAL  [initandlisten] recover : no journal files present, no recovery needed
2017-12-05T15:37:11.535+0800 I JOURNAL  [durability] Durability thread started
2017-12-05T15:37:11.535+0800 I JOURNAL  [journal writer] Journal writer thread started
2017-12-05T15:37:11.535+0800 I CONTROL  [initandlisten] MongoDB starting : pid=7304 port=27018 dbpath=/Users/olifer/middle/mongo/master-slave/slave/data/ slave=1 64-bit host=oliferdeMacBook-Pro.local
2017-12-05T15:37:11.535+0800 I CONTROL  [initandlisten] db version v3.0.7
2017-12-05T15:37:11.535+0800 I CONTROL  [initandlisten] git version: nogitversion
2017-12-05T15:37:11.535+0800 I CONTROL  [initandlisten] build info: Darwin yosemitevm.local 14.5.0 Darwin Kernel Version 14.5.0: Wed Jul 29 02:26:53 PDT 2015; root:xnu-2782.40.9~1/RELEASE_X86_64 x86_64 BOOST_LIB_VERSION=1_49
2017-12-05T15:37:11.535+0800 I CONTROL  [initandlisten] allocator: system
2017-12-05T15:37:11.535+0800 I CONTROL  [initandlisten] options: { config: "/Users/olifer/middle/mongo/master-slave/slave/mongod.conf", net: { bindIp: "127.0.0.1", port: 27018 }, slave: true, source: "127.0.0.1:27017", storage: { dbPath: "/Users/olifer/middle/mongo/master-slave/slave/data/" } }
2017-12-05T15:37:11.536+0800 I INDEX    [initandlisten] allocating new ns file /Users/olifer/middle/mongo/master-slave/slave/data/local.ns, filling with zeroes...
2017-12-05T15:37:11.605+0800 I STORAGE  [FileAllocator] allocating new datafile /Users/olifer/middle/mongo/master-slave/slave/data/local.0, filling with zeroes...
2017-12-05T15:37:11.605+0800 I STORAGE  [FileAllocator] creating directory /Users/olifer/middle/mongo/master-slave/slave/data/_tmp
2017-12-05T15:37:11.824+0800 I STORAGE  [FileAllocator] done allocating datafile /Users/olifer/middle/mongo/master-slave/slave/data/local.0, size: 64MB,  took 0.218 secs
2017-12-05T15:37:11.890+0800 I NETWORK  [initandlisten] waiting for connections on port 27018
2017-12-05T15:37:12.894+0800 I REPL     [replslave] repl: syncing from host:127.0.0.1:27017
2017-12-05T15:37:17.930+0800 I REPL     [replslave] repl: sleep 2 sec before next pass
2017-12-05T15:37:19.935+0800 I REPL     [replslave] repl: syncing from host:127.0.0.1:27017
2017-12-05T15:37:24.958+0800 I REPL     [replslave] repl: syncing from host:127.0.0.1:27017
2017-12-05T15:37:33.188+0800 I REPL     [replslave] repl: syncing from host:127.0.0.1:27017
2017-12-05T15:37:43.193+0800 I REPL     [replslave] repl: syncing from host:127.0.0.1:27017
2017-12-05T15:37:53.198+0800 I REPL     [replslave] repl: syncing from host:127.0.0.1:27017

我们可以通过slave的日志中可以看到,slave一直同master保持同步数据的联系。

我们通过客户端链接上服务端

mongo 127.0.0.1:27017  #主
mongo 127.0.0.1:27018  #从

在主上操作

> db.test.insert({"name":"linyang"})
WriteResult({ "nInserted" : 1 })
> db.test.find({});
{ "_id" : ObjectId("5a264e0ecb6f7d3713c516a7"), "name" : "linyang" }

往里面插入了一条记录,可以再从的日志中发现同步的记录

2017-12-05T15:43:10.387+0800 I INDEX    [replslave] allocating new ns file /Users/olifer/middle/mongo/master-slave/slave/data/test.ns, filling with zeroes...
2017-12-05T15:43:10.450+0800 I STORAGE  [FileAllocator] allocating new datafile /Users/olifer/middle/mongo/master-slave/slave/data/test.0, filling with zeroes...
2017-12-05T15:43:10.646+0800 I STORAGE  [FileAllocator] done allocating datafile /Users/olifer/middle/mongo/master-slave/slave/data/test.0, size: 64MB,  took 0.196 secs
2017-12-05T15:43:10.702+0800 I REPL     [replslave] resync: dropping database test
2017-12-05T15:43:10.709+0800 I JOURNAL  [replslave] journalCleanup...
2017-12-05T15:43:10.710+0800 I JOURNAL  [replslave] removeJournalFiles
2017-12-05T15:43:10.713+0800 I JOURNAL  [replslave] journalCleanup...
2017-12-05T15:43:10.713+0800 I JOURNAL  [replslave] removeJournalFiles
2017-12-05T15:43:10.715+0800 I REPL     [replslave] resync: cloning database test to get an initial copy
2017-12-05T15:43:10.718+0800 I INDEX    [replslave] allocating new ns file /Users/olifer/middle/mongo/master-slave/slave/data/test.ns, filling with zeroes...
2017-12-05T15:43:10.803+0800 I STORAGE  [FileAllocator] allocating new datafile /Users/olifer/middle/mongo/master-slave/slave/data/test.0, filling with zeroes...
2017-12-05T15:43:11.056+0800 I STORAGE  [FileAllocator] done allocating datafile /Users/olifer/middle/mongo/master-slave/slave/data/test.0, size: 64MB,  took 0.252 secs
2017-12-05T15:43:11.107+0800 I INDEX    [replslave] build index on: test.test properties: { v: 1, key: { _id: 1 }, name: "_id_", ns: "test.test" }
2017-12-05T15:43:11.107+0800 I INDEX    [replslave]      building index using bulk method
2017-12-05T15:43:11.107+0800 I INDEX    [replslave] build index done.  scanned 1 total records. 0 secs
2017-12-05T15:43:11.107+0800 I STORAGE  [replslave] copying indexes for: { name: "test", options: {} }
2017-12-05T15:43:11.108+0800 I REPL     [replslave] resync: done with initial clone for db: test

我们再从的client中查看有没有信息。

mongo 127.0.0.1:27018
MongoDB shell version: 3.0.7
connecting to: 127.0.0.1:27018/test
> db.test.find({});
{ "_id" : ObjectId("5a264e0ecb6f7d3713c516a7"), "name" : "linyang" }
>

也确认同步过来了。然后再从上插入记录

> db.test.insert({"age":34});
WriteResult({ "writeError" : { "code" : undefined, "errmsg" : "not master" } })
>

从上不能插入记录。如果主挂掉了,从又不能写数据,那么是否mongo集群就挂掉了。答案当然是否定的,因为,mongo还有副本集。

副本集

我们使用三台mongo来模拟.
第一台的配置文件 /Users/olifer/middle/mongo/replica/a/mongod.conf

dbpath = /Users/olifer/middle/mongo/replica/a/data
port = 8001
bind_ip = 127.0.0.1
replSet = child/127.0.0.1:8002

第二台的配置文件 /Users/olifer/middle/mongo/replica/b/mongod.conf

dbpath = /Users/olifer/middle/mongo/replica/a/data
port = 8002
bind_ip = 127.0.0.1
replSet = child/127.0.0.1:8003

第三台的配置文件 /Users/olifer/middle/mongo/replica/c/mongod.conf

dbpath = /Users/olifer/middle/mongo/replica/a/data
port = 8003
bind_ip = 127.0.0.1
replSet = child/127.0.0.1:8001

启动服务器

mongod -f ~/middle/mongo/replica/a/mongod.conf
mongod -f ~/middle/mongo/replica/b/mongod.conf
mongod -f ~/middle/mongo/replica/c/mongod.conf

启动成功后,进入三个服务的任何一个客户端

mongo 127.0.0.1:8002

> config = {_id: 'child', members: [
  {
    "_id":1,
    "host":"127.0.0.1:8001"
  },
  {
     "_id":2,
     "host":"127.0.0.1:8002"
  },
  {
     "_id":3,
     "host":"127.0.0.1:8003"
   }
   ]
 }
{
    "_id" : "child",
    "members" : [
        {
            "_id" : 1,
            "host" : "127.0.0.1:8001"
        },
        {
            "_id" : 2,
            "host" : "127.0.0.1:8002"
        },
        {
            "_id" : 3,
            "host" : "127.0.0.1:8003"
        }
    ]
}
> rs.initiate(config);
{ "ok" : 1 }
child:SECONDARY>

配置完后发现前缀发生了改变,我们进入另外两个渠道的客户端

mongo 127.0.0.1:8001
MongoDB shell version: 3.0.7
connecting to: 127.0.0.1:8001/test
child:PRIMARY> 

mongo 127.0.0.1:8003
MongoDB shell version: 3.0.7
connecting to: 127.0.0.1:8003/test
child:SECONDARY> 

其中child:PRIMARY>表示活跃节点。其余为备份节点。注意:只有活跃节点才能进行查询数据库的信息操作,备份节点不能进行会报错,在任意客户端执行rs.status() 来查看所有状态

child:PRIMARY> rs.status()
{
    "set" : "child",
    "date" : ISODate("2017-12-05T08:56:44.523Z"),
    "myState" : 1,
    "members" : [
        {
            "_id" : 1,
            "name" : "127.0.0.1:8001",
            "health" : 1,
            "state" : 1,
            "stateStr" : "PRIMARY",
            "uptime" : 620,
            "optime" : Timestamp(1512463818, 1),
            "optimeDate" : ISODate("2017-12-05T08:50:18Z"),
            "electionTime" : Timestamp(1512463821, 1),
            "electionDate" : ISODate("2017-12-05T08:50:21Z"),
            "configVersion" : 1,
            "self" : true
        },
        {
            "_id" : 2,
            "name" : "127.0.0.1:8002",
            "health" : 1,
            "state" : 2,
            "stateStr" : "SECONDARY",
            "uptime" : 387,
            "optime" : Timestamp(1512463818, 1),
            "optimeDate" : ISODate("2017-12-05T08:50:18Z"),
            "lastHeartbeat" : ISODate("2017-12-05T08:56:43.756Z"),
            "lastHeartbeatRecv" : ISODate("2017-12-05T08:56:43.756Z"),
            "pingMs" : 0,
            "configVersion" : 1
        },
        {
            "_id" : 3,
            "name" : "127.0.0.1:8003",
            "health" : 1,
            "state" : 2,
            "stateStr" : "SECONDARY",
            "uptime" : 387,
            "optime" : Timestamp(1512463818, 1),
            "optimeDate" : ISODate("2017-12-05T08:50:18Z"),
            "lastHeartbeat" : ISODate("2017-12-05T08:56:43.756Z"),
            "lastHeartbeatRecv" : ISODate("2017-12-05T08:56:43.756Z"),
            "pingMs" : 0,
            "configVersion" : 1
        }
    ],
    "ok" : 1
}

搭建成功后,我们来验证一下。

在活跃的节点创建数据,在备份库拉取数据
child:PRIMARY> db.repl.insert({"name":"123"});
WriteResult({ "nInserted" : 1 })

child:SECONDARY> db.repl.find();
{ "_id" : ObjectId("5a26602f5cefb1fdb377843b"), "name" : "123" }

功能正常。

从服务器不能写
child:SECONDARY> db.repl.insert({"age":444});
WriteResult({ "writeError" : { "code" : undefined, "errmsg" : "not master" } })
关闭活跃的节点,从服务器会通过选举得到最新的活跃的节点

当我关闭原来活跃的服务器8001后,通过内部选举,8002成了最新的活跃节点没从前缀也可以看出

child:SECONDARY>
child:PRIMARY>

现在看一下最新的状态

child:PRIMARY> rs.status();
{
    "set" : "child",
    "date" : ISODate("2017-12-05T09:09:31.666Z"),
    "myState" : 1,
    "members" : [
        {
            "_id" : 1,
            "name" : "127.0.0.1:8001",
            "health" : 0,
            "state" : 8,
            "stateStr" : "(not reachable/healthy)",
            "uptime" : 0,
            "optime" : Timestamp(0, 0),
            "optimeDate" : ISODate("1970-01-01T00:00:00Z"),
            "lastHeartbeat" : ISODate("2017-12-05T09:09:31.193Z"),
            "lastHeartbeatRecv" : ISODate("2017-12-05T09:07:12.773Z"),
            "pingMs" : 0,
            "lastHeartbeatMessage" : "Failed attempt to connect to 127.0.0.1:8001; couldn't connect to server 127.0.0.1:8001 (127.0.0.1), connection attempt failed",
            "configVersion" : -1
        },
        {
            "_id" : 2,
            "name" : "127.0.0.1:8002",
            "health" : 1,
            "state" : 1,
            "stateStr" : "PRIMARY",
            "uptime" : 1308,
            "optime" : Timestamp(1512464432, 2),
            "optimeDate" : ISODate("2017-12-05T09:00:32Z"),
            "electionTime" : Timestamp(1512464835, 1),
            "electionDate" : ISODate("2017-12-05T09:07:15Z"),
            "configVersion" : 1,
            "self" : true
        },
        {
            "_id" : 3,
            "name" : "127.0.0.1:8003",
            "health" : 1,
            "state" : 2,
            "stateStr" : "SECONDARY",
            "uptime" : 1152,
            "optime" : Timestamp(1512464432, 2),
            "optimeDate" : ISODate("2017-12-05T09:00:32Z"),
            "lastHeartbeat" : ISODate("2017-12-05T09:09:30.963Z"),
            "lastHeartbeatRecv" : ISODate("2017-12-05T09:09:30.963Z"),
            "pingMs" : 0,
            "configVersion" : 1
        }
    ],
    "ok" : 1
}

说明内部的切换确实成功了。还剩下一个分片,我们下次再讲。

参考资料

MongoDB 3.0 常见集群的搭建(主从复制,副本集,分片....)

©著作权归作者所有,转载或内容合作请联系作者
  • 序言:七十年代末,一起剥皮案震惊了整个滨河市,随后出现的几起案子,更是在滨河造成了极大的恐慌,老刑警刘岩,带你破解...
    沈念sama阅读 159,015评论 4 362
  • 序言:滨河连续发生了三起死亡事件,死亡现场离奇诡异,居然都是意外死亡,警方通过查阅死者的电脑和手机,发现死者居然都...
    沈念sama阅读 67,262评论 1 292
  • 文/潘晓璐 我一进店门,熙熙楼的掌柜王于贵愁眉苦脸地迎上来,“玉大人,你说我怎么就摊上这事。” “怎么了?”我有些...
    开封第一讲书人阅读 108,727评论 0 243
  • 文/不坏的土叔 我叫张陵,是天一观的道长。 经常有香客问我,道长,这世上最难降的妖魔是什么? 我笑而不...
    开封第一讲书人阅读 43,986评论 0 205
  • 正文 为了忘掉前任,我火速办了婚礼,结果婚礼上,老公的妹妹穿的比我还像新娘。我一直安慰自己,他们只是感情好,可当我...
    茶点故事阅读 52,363评论 3 287
  • 文/花漫 我一把揭开白布。 她就那样静静地躺着,像睡着了一般。 火红的嫁衣衬着肌肤如雪。 梳的纹丝不乱的头发上,一...
    开封第一讲书人阅读 40,610评论 1 219
  • 那天,我揣着相机与录音,去河边找鬼。 笑死,一个胖子当着我的面吹牛,可吹牛的内容都是我干的。 我是一名探鬼主播,决...
    沈念sama阅读 31,871评论 2 312
  • 文/苍兰香墨 我猛地睁开眼,长吁一口气:“原来是场噩梦啊……” “哼!你这毒妇竟也来了?” 一声冷哼从身侧响起,我...
    开封第一讲书人阅读 30,582评论 0 198
  • 序言:老挝万荣一对情侣失踪,失踪者是张志新(化名)和其女友刘颖,没想到半个月后,有当地人在树林里发现了一具尸体,经...
    沈念sama阅读 34,297评论 1 242
  • 正文 独居荒郊野岭守林人离奇死亡,尸身上长有42处带血的脓包…… 初始之章·张勋 以下内容为张勋视角 年9月15日...
    茶点故事阅读 30,551评论 2 246
  • 正文 我和宋清朗相恋三年,在试婚纱的时候发现自己被绿了。 大学时的朋友给我发了我未婚夫和他白月光在一起吃饭的照片。...
    茶点故事阅读 32,053评论 1 260
  • 序言:一个原本活蹦乱跳的男人离奇死亡,死状恐怖,灵堂内的尸体忽然破棺而出,到底是诈尸还是另有隐情,我是刑警宁泽,带...
    沈念sama阅读 28,385评论 2 253
  • 正文 年R本政府宣布,位于F岛的核电站,受9级特大地震影响,放射性物质发生泄漏。R本人自食恶果不足惜,却给世界环境...
    茶点故事阅读 33,035评论 3 236
  • 文/蒙蒙 一、第九天 我趴在偏房一处隐蔽的房顶上张望。 院中可真热闹,春花似锦、人声如沸。这庄子的主人今日做“春日...
    开封第一讲书人阅读 26,079评论 0 8
  • 文/苍兰香墨 我抬头看了看天上的太阳。三九已至,却和暖如春,着一层夹袄步出监牢的瞬间,已是汗流浃背。 一阵脚步声响...
    开封第一讲书人阅读 26,841评论 0 195
  • 我被黑心中介骗来泰国打工, 没想到刚下飞机就差点儿被人妖公主榨干…… 1. 我叫王不留,地道东北人。 一个月前我还...
    沈念sama阅读 35,648评论 2 274
  • 正文 我出身青楼,却偏偏与公主长得像,于是被迫代替她去往敌国和亲。 传闻我的和亲对象是个残疾皇子,可洞房花烛夜当晚...
    茶点故事阅读 35,550评论 2 270

推荐阅读更多精彩内容

  • 一. 准备工作 安装docker 安装docker-compose 注意 docker 通过 unix socke...
    谁又失忆阅读 7,426评论 3 26
  • Spring Cloud为开发人员提供了快速构建分布式系统中一些常见模式的工具(例如配置管理,服务发现,断路器,智...
    卡卡罗2017阅读 134,100评论 18 139
  • Redis 配置文件示例 注意:想要读取配置文件,Redis的第一个参数必须是文件的路径 ./redis-serv...
    起个名忒难阅读 1,149评论 0 1
  • NoSql数据库优缺点 在优势方面主要体现在下面几点: 简单的扩展 快速的读写 低廉的成本 灵活的数据模型 在不足...
    dreamer_lk阅读 2,609评论 0 6
  • 我错过一个好时候, 头连着耳朵下垂, 上牙挤着下牙, 舌头叫史茅戈 被矮人刺死在牙缝, 滚滚的烫珠子 裹着薄皮不愿...
    葡萄美酒阅读 294评论 1 3