zookeeper 安装和配置

zookeeper 安装和配置

1.什么是zookeeper
  引用官方的说法:“Zookeeper是一个高性能,分布式的,开源分布式应用协调服务。它提供了简单原始的功能,分布式应用可以基于它实现更高级的服务,比如同步,配置管理,集群管理,名空间。它被设计为易于编程,使用文件系统目录树作为数据模型。服务端跑在java上,提供java和C的客户端API”。

2.设计目标
Zookeeper服务自身组成一个集群(2n+1个服务允许n个失效)。Zookeeper服务有两个角色,一个是leader,负责写服务和数据同步,剩下的是follower,提供读服务,leader失效后会在follower中重新选举新的leader
Zookeeper逻辑图如下,
zookeeper service

客户端可以连接到每个server,每个server的数据完全相同。
每个follower都和leader有连接,接受leader的数据更新操作。
Server记录事务日志和快照到持久存储。
大多数server可用,整体服务就可用

3.zookeeper能做什么
Zookeeper的主流应用场景实现思路(除去官方示例)
(1)配置管理
集中式的配置管理在应用集群中是非常常见的,一般商业公司内部都会实现一套集中的配置管理中心,应对不同的应用集群对于共享各自配置的需求,并且在配置变更时能够通知到集群中的每一个机器。

(2)集群管理
应用集群中,我们常常需要让每一个机器知道集群中(或依赖的其他某一个集群)哪些机器是活着的,并且在集群机器因为宕机,网络断链等原因能够不在人工介入的情况下迅速通知到每一个机器。

(3)数据发布与订阅 (我的业务用到这个特性,后面会有详细介绍)
应用配置集中到节点上,应用启动时主动获取,并在节点上注册一个watcher,每次配置更新都会通知到应用。

(4)名空间服务
分布式命名服务,创建一个节点后,节点的路径就是全局唯一的,可以作为全局名称使用。

(5)分布式通知/协调
不同的系统都监听同一个节点,一旦有了更新,另一个系统能够收到通知。

(6)分布式锁
Zookeeper能保证数据的强一致性,用户任何时候都可以相信集群中每个节点的数据都是相同的。一个用户创建一个节点作为锁,另一个用户检测该节点,如果存在,代表别的用户已经锁住,如果不存在,则可以创建一个节点,代表拥有一个锁。

zookeeper 安装
1、首选安装java运行环境。如果不知如何按照可以参加我的JAVA环境配置这篇文章,也可百度搜索。
2、下载zookeeper,目前最新的稳定版本为3.4.10;zookeeper官网下载
linux下可以使用wget命令下载

1
2
3
4
5
##下载
wget https://mirrors.tuna.tsinghua.edu.cn/apache/zookeeper/stable/zookeeper-3.4.10.tar.gz

##然后解压:
tar -xzvf zookeeper-3.4.10.tar.gz

##运行

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
### 1启动zookeeper
# bin/zkServer.sh start
ZooKeeper JMX enabled by default
Using config: /gzdata/zookeeper/zookeeper-3.4.10/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED

### 2查看状态
# bin/zkServer.sh status
ZooKeeper JMX enabled by default
Using config: /gzdata/zookeeper/zookeeper-3.4.10/bin/../conf/zoo.cfg
Mode: standalone

### 3停止zookeeper
# bin/zkServer.sh stop
ZooKeeper JMX enabled by default
Using config: /gzdata/zookeeper/zookeeper-3.4.10/bin/../conf/zoo.cfg
Stopping zookeeper ... STOPPED

### 4 重启
# bin/zkServer.sh restart
ZooKeeper JMX enabled by default
Using config: /gzdata/zookeeper/zookeeper-3.4.10/bin/../conf/zoo.cfg
ZooKeeper JMX enabled by default
Using config: /gzdata/zookeeper/zookeeper-3.4.10/bin/../conf/zoo.cfg
Stopping zookeeper ... STOPPED
ZooKeeper JMX enabled by default
Using config: /gzdata/zookeeper/zookeeper-3.4.10/bin/../conf/zoo.cfg
Starting zookeeper ... STARTED

管理zookeeper存储

对于长时间运行的生产系统,必须从外部(dataDir和日志)管理ZooKeeper的存储。有关更多细节,请参阅维护部分

连接Zookeeper

1
# ./zkCli.sh -server 127.0.0.1:2181

这样可以执行简单的文件操作。
连接后,你将看到类似下面的输出结果:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
Connecting to 127.0.0.1:2181
2017-05-11 20:21:23,599 [myid:] - INFO [main:Environment@100] - Client environment:zookeeper.version=3.4.10-39d3a4f269333c922ed3db283be479f9deacaa0f, built on 03/23/2017 10:13 GMT
2017-05-11 20:21:23,605 [myid:] - INFO [main:Environment@100] - Client environment:host.name=localhost
2017-05-11 20:21:23,605 [myid:] - INFO [main:Environment@100] - Client environment:java.version=1.8.0_91
2017-05-11 20:21:23,608 [myid:] - INFO [main:Environment@100] - Client environment:java.vendor=Oracle Corporation
2017-05-11 20:21:23,608 [myid:] - INFO [main:Environment@100] - Client environment:java.home=/usr/local/jdk/jdk1.8.0_91/jre
2017-05-11 20:21:23,608 [myid:] - INFO [main:Environment@100] - Client environment:java.class.path=/gzdata/zookeeper/zookeeper-3.4.10/bin/../build/classes:/gzdata/zookeeper/zookeeper-3.4.10/bin/../build/lib/*.jar:/gzdata/zookeeper/zookeeper-3.4.10/bin/../lib/slf4j-log4j12-1.6.1.jar:/gzdata/zookeeper/zookeeper-3.4.10/bin/../lib/slf4j-api-1.6.1.jar:/gzdata/zookeeper/zookeeper-3.4.10/bin/../lib/netty-3.10.5.Final.jar:/gzdata/zookeeper/zookeeper-3.4.10/bin/../lib/log4j-1.2.16.jar:/gzdata/zookeeper/zookeeper-3.4.10/bin/../lib/jline-0.9.94.jar:/gzdata/zookeeper/zookeeper-3.4.10/bin/../zookeeper-3.4.10.jar:/gzdata/zookeeper/zookeeper-3.4.10/bin/../src/java/lib/*.jar:/gzdata/zookeeper/zookeeper-3.4.10/bin/../conf:.:/usr/local/jdk/jdk1.8.0_91/lib/dt.jar:/usr/local/jdk/jdk1.8.0_91/lib/tools.jar
2017-05-11 20:21:23,608 [myid:] - INFO [main:Environment@100] - Client environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib
2017-05-11 20:21:23,608 [myid:] - INFO [main:Environment@100] - Client environment:java.io.tmpdir=/tmp
2017-05-11 20:21:23,608 [myid:] - INFO [main:Environment@100] - Client environment:java.compiler=<NA>
2017-05-11 20:21:23,608 [myid:] - INFO [main:Environment@100] - Client environment:os.name=Linux
2017-05-11 20:21:23,609 [myid:] - INFO [main:Environment@100] - Client environment:os.arch=amd64
2017-05-11 20:21:23,609 [myid:] - INFO [main:Environment@100] - Client environment:os.version=3.10.0-514.2.2.el7.x86_64
2017-05-11 20:21:23,609 [myid:] - INFO [main:Environment@100] - Client environment:user.name=root
2017-05-11 20:21:23,609 [myid:] - INFO [main:Environment@100] - Client environment:user.home=/root
2017-05-11 20:21:23,609 [myid:] - INFO [main:Environment@100] - Client environment:user.dir=/gzdata/zookeeper/zookeeper-3.4.10/bin
2017-05-11 20:21:23,610 [myid:] - INFO [main:ZooKeeper@438] - Initiating client connection, connectString=127.0.0.1:2181 sessionTimeout=30000 watcher=org.apache.zookeeper.ZooKeeperMain$MyWatcher@42110406
2017-05-11 20:21:23,641 [myid:] - INFO [main-SendThread(127.0.0.1:2181):ClientCnxn$SendThread@1032] - Opening socket connection to server 127.0.0.1/127.0.0.1:2181. Will not attempt to authenticate using SASL (unknown error)
Welcome to ZooKeeper!
JLine support is enabled
2017-05-11 20:21:23,809 [myid:] - INFO [main-SendThread(127.0.0.1:2181):ClientCnxn$SendThread@876] - Socket connection established to 127.0.0.1/127.0.0.1:2181, initiating session
[zk: 127.0.0.1:2181(CONNECTING) 0] 2017-05-11 20:21:23,998 [myid:] - INFO [main-SendThread(127.0.0.1:2181):ClientCnxn$SendThread@1299] - Session establishment complete on server 127.0.0.1/127.0.0.1:2181, sessionid = 0x15bf77496660000, negotiated timeout = 30000

WATCHER::

WatchedEvent state:SyncConnected type:None path:null

shell中输入help命令,可以获取一个命令列表,如:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
help
ZooKeeper -server host:port cmd args
stat path [watch]
set path data [version]
ls path [watch]
delquota [-n|-b] path
ls2 path [watch]
setAcl path acl
setquota -n|-b val path
history
redo cmdno
printwatches on|off
delete path [version]
sync path
listquota path
rmr path
get path [watch]
create [-s] [-e] path data acl
addauth scheme auth
quit
getAcl path
close
connect host:port
[zk: 127.0.0.1:2181(CONNECTED) 1]

在这,你可以尝试着输入一些简单的命令来体验下。首先,我们输入ls 命令,来看下结果:

1
2
[zk: 127.0.0.1:2181(CONNECTED) 3] ls /
[zookeeper]

接下来,我们通过 create /zk_test mydata 命令创建一个新的znode。新创建的znode将于字符串“mydata”关联起来,你应看到:

1
2
[zk: 127.0.0.1:2181(CONNECTED) 4] create /zk_test mydata
Created /zk_test

我们用ls /命令来查看刚创建好的znode。

1
2
[zk: 127.0.0.1:2181(CONNECTED) 5] ls /
[zookeeper, zk_test]

注意:上面的zk_test目录是我们刚刚创建的。

接下来,我们通过get命令验证数据是否与znode关联,如:

1
2
3
4
5
6
7
8
9
10
11
12
13
[zk: 127.0.0.1:2181(CONNECTED) 6] get /zk_test
mydata
cZxid = 0x2
ctime = Thu May 11 20:23:24 CST 2017
mZxid = 0x2
mtime = Thu May 11 20:23:24 CST 2017
pZxid = 0x2
cversion = 0
dataVersion = 0
aclVersion = 0
ephemeralOwner = 0x0
dataLength = 7
numChildren = 0

我们通过set命令改变zk_test节点上关联的数据

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
[zk: 127.0.0.1:2181(CONNECTED) 8] set /zk_test hello,world
cZxid = 0x2
ctime = Thu May 11 20:23:24 CST 2017
mZxid = 0x3
mtime = Thu May 11 20:55:08 CST 2017
pZxid = 0x2
cversion = 0
dataVersion = 1
aclVersion = 0
ephemeralOwner = 0x0
dataLength = 11
numChildren = 0
[zk: 127.0.0.1:2181(CONNECTED) 9] get /zk_test
hello,world
cZxid = 0x2
ctime = Thu May 11 20:23:24 CST 2017
mZxid = 0x3
mtime = Thu May 11 20:55:08 CST 2017
pZxid = 0x2
cversion = 0
dataVersion = 1
aclVersion = 0
ephemeralOwner = 0x0
dataLength = 11
numChildren = 0

注意:修改时一定要指明相应路径 ,如: /zk_test .我们通过get命令查看到数据确实发生了变化。

最后,我们来用delete删除节点。

1
2
3
4
[zk: 127.0.0.1:2181(CONNECTED) 10] delete /zk_test
[zk: 127.0.0.1:2181(CONNECTED) 11] ls /
[zookeeper]
[zk: 127.0.0.1:2181(CONNECTED) 12]

###复制模式

在zookeeper默认是运行在单机模式下以满足评估、开发、测试的需要。但生产环境中,你必须运行在复制模式下,在同一应用内的一个复制组成为quorum,在复制模式中,quorum中的所有服务器都有相同配置文件的副本。

注意:
For replicated mode, a minimum of three servers are required, and it is strongly recommended that you have an odd number of servers. If you only have two servers, then you are in a situation where if one of them fails, there are not enough machines to form a majority quorum. Two servers is inherently less stable than a single server, because there are two single points of failure.

The required conf/zoo.cfg file for replicated mode is similar to the one used in standalone mode, but with a few differences. Here is an example:

tickTime=2000
dataDir=/var/lib/zookeeper
clientPort=2181
initLimit=5
syncLimit=2
server.1=zoo1:2888:3888
server.2=zoo2:2888:3888
server.3=zoo3:2888:3888

The new entry, initLimit is timeouts ZooKeeper uses to limit the length of time the ZooKeeper servers in quorum have to connect to a leader. The entry syncLimit limits how far out of date a server can be from a leader.

With both of these timeouts, you specify the unit of time using tickTime. In this example, the timeout for initLimit is 5 ticks at 2000 milleseconds a tick, or 10 seconds.

The entries of the form server.X list the servers that make up the ZooKeeper service. When the server starts up, it knows which server it is by looking for the file myid in the data directory. That file has the contains the server number, in ASCII.

Finally, note the two port numbers after each server name: “ 2888” and “3888”. Peers use the former port to connect to other peers. Such a connection is necessary so that peers can communicate, for example, to agree upon the order of updates. More specifically, a ZooKeeper server uses this port to connect followers to the leader. When a new leader arises, a follower opens a TCP connection to the leader using this port. Because the default leader election also uses TCP, we currently require another port for leader election. This is the second port in the server entry.

Note
If you want to test multiple servers on a single machine, specify the servername as localhost with unique quorum & leader election ports (i.e. 2888:3888, 2889:3889, 2890:3890 in the example above) for each server.X in that server’s config file. Of course separate dataDirs and distinct clientPorts are also necessary (in the above replicated example, running on a single localhost, you would still have three config files).

Please be aware that setting up multiple servers on a single machine will not create any redundancy. If something were to happen which caused the machine to die, all of the zookeeper servers would be offline. Full redundancy requires that each server have its own machine. It must be a completely separate physical server. Multiple virtual machines on the same physical host are still vulnerable to the complete failure of that host.

参考:
http://coolxing.iteye.com/blog/1871009