通过MCollective更加安全地实现puppet的推送更新功能

来源:互联网 发布:埋点数据 编辑:程序博客网 时间:2024/06/08 17:47

1 介绍

1.1 Mcollective介绍

MCollective 是一个构建服务器编排(Server Orchestration)和并行工作执行系统的框架。
首先,MCollective 是一种针对服务器集群进行可编程控制的系统管理解决方案。在这一点上,它的功能类似:Func,Fabric 和 Capistrano。

其次,MCollective 的设计打破基于中心存储式系统和像 SSH 这样的工具,不再仅仅痴迷于 SSH 的 For 循环。它使用发布订阅中间件(Publish Subscribe Middleware)这样的现代化 工具和通过目标数据(meta data)而不是主机名(hostnames)来实时发现网络资源这样的现代化理念。提供了一个可扩展的而且迅速的并行执行环境。

MCollective 工具为命令行界面,但它可与数千个应用实例进行通信,而且传输速度惊人。无论部署的实例位于什么位置,通信都能以线速进行传输,使用的是一个类似多路传送的推送信息系统。MCollective 工具没有可视化用户界面,用户只能通过检索来获取需要应用的实例。Puppet Dashboard 提供有这部分功能。

MCollective特点:

  • 能够与小到大型服务器集群交互

  • 使用广播范式(broadcast paradigm)来进行请求分发,所有服务器会同时收到请求,而只有与请求所附带的过滤器匹配的服务器才会去执行这些请求。没有中心数据库来进行同步,网络是唯一的真理

  • 打破了以往用主机名作为身份验证手段的复杂命名规则。使用每台机器自身提供的丰富的目标数据来定位它们。目标数据来自于:Puppet, Chef, Facter, Ohai 或者自身提供的插件

  • 使用命令行调用远程代理

  • 能够写自定义的设备报告

  • 大量的代理来管理包,服务和其他来自于社区的通用组件

  • 允许写 SimpleRPC 风格的代理、客户端和使用 Ruby 实现 Web UIs

  • 外部可插件化(pluggable)实现本地需求

  • 中间件系统已有丰富的身份验证和授权模型,利用这些作为控制的第一道防线。

  • 重用中间件来做集群、路由和网络隔离以实现安全和可扩展安装。

MCollective 就是一个框架,一个空壳。它除了 MCO 命令之外都可以被替换被自定义。

备注:更多信息请参考http://docs.puppetlabs.com/

1.2 Middleware(RabbitMQ、ActiveMQ)介绍

RabbitMQ是一个实现了高级消息排队协议(AMQP)的消息队列服务。RabbitMQ基于OTP(Open Telecom Platform,开发电信平台)进行构建,并使用Erlang语言和运行时环境来实现。
ActiveMQ 是Apache出品,最流行的,能力强劲的开源消息总线。ActiveMQ 是一个完全支持JMS1.1和J2EE 1.4规范的 JMS Provider实现

备注:MCollective是基于Apache ActiveMQ中间件来进行开发和测试的,然而其对java和XML格式的配置文件的依赖使我们将更多的注意力和兴趣转移到RabbitMQ中间件服务上。如果考虑到性能和扩展性,部署ActivemMQ是一个更好的选择。

1.3 工作原理图

mcollective触发更新图mcollective触发更新图

备注:更多详细信息请参考
http://docs.puppetlabs.com/mcollective/reference/basic/messageflow.html

2 安装和配置Middleware

备注:ActiveMQ和RabbitMQ选其一进行部署即可

2.1 安装和配置ActiveMQ

2.1.1 安装ActiveMQ

1
[root@puppetserver rpms]# yum install tanukiwrapper activemq activemq-info-provider

2.1.2 配置ActiveMQ

主要配置MCollective连接的端口以及账号、密码及权限

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
[root@puppetserver rpms]# vim /etc/activemq/activemq.xml
<simpleAuthenticationPlugin>
            <users>
<!--              <authenticationUser username="${activemq.username}"password="${activemq.password}"groups="admins,everyone"/> --> #禁用
              <authenticationUser username="mcollective"password="secret" groups="mcollective,admins,everyone"/>  #配置通信的账号及密码
            </users>
          </simpleAuthenticationPlugin>
<authorizationPlugin>  #配置权限,默认即可
            <map>
              <authorizationMap>
                <authorizationEntries>
                  <authorizationEntry queue=">"write="admins" read="admins" admin="admins"/>
                  <authorizationEntry topic=">"write="admins" read="admins" admin="admins"/>
                  <authorizationEntry topic="mcollective.>"write="mcollective"read="mcollective"admin="mcollective"/>
                  <authorizationEntry topic="mcollective.>"write="mcollective"read="mcollective"admin="mcollective"/>
                  <authorizationEntry topic="ActiveMQ.Advisory.>"read="everyone" write="everyone" admin="everyone"/>
                </authorizationEntries>
              </authorizationMap>
            </map>
          </authorizationPlugin>
<transportConnectors>
            <transportConnector name="openwire"uri="tcp://0.0.0.0:61616"/>
            <transportConnector name="stomp+nio"uri="stomp://0.0.0.0:61613"/> #配置通信协议为stomp,监听61613端口
        </transportConnectors>

2.1.3 启动ActiveMQ

1
2
3
4
5
[root@puppetserver rpms]# /etc/rc.d/init.d/activemq start
Starting ActiveMQ Broker...
[root@puppetserver rpms]# chkconfig activemq on
[root@puppetserver rpms]# netstat -nlatp | grep61613  #查看监听端口
tcp        0     0 :::61613                   :::*                        LISTEN      33805/java

备注:更多详细配置信息请参考
http://docs.puppetlabs.com/mcollective/reference/plugins/connector_activemq.html

2.2 安装和配置RabbitMQ

2.2.1 安装RabbitMQ

1
2
3
4
5
6
[root@puppetserver rpms]# yum install erlang #RabbitMQ依赖erlang语言,需要安装大概65个左右的erlang依赖包
[root@puppetserver rpms]# yum install rabbitmq-server
[root@puppetserver rpms]# ll /usr/lib/rabbitmq/lib/rabbitmq_server-3.1.5/plugins/ #默认已经安装了stomp插件,老版本需要下载安装
-rw-r--r-- 1root root 242999 Aug 24 17:42amqp_client-3.1.5.ez
-rw-r--r-- 1root root  85847 Aug 24 17:42rabbitmq_stomp-3.1.5.ez

2.2.2 启动rabbitmq-server

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
[root@puppetserver rpms]# /etc/rc.d/init.d/rabbitmq-server start #启动rabbitmq服务
Starting rabbitmq-server: SUCCESS
rabbitmq-server.
[root@puppetserver rpms]# /etc/rc.d/init.d/rabbitmq-server status #查看rabbitmq状态
Status of node rabbit@puppetserver ...
[{pid,43198},
 {running_applications,[{rabbit,"RabbitMQ","3.1.5"},
                        {mnesia,"MNESIA  CXC 138 12","4.5"},
                        {os_mon,"CPO  CXC 138 46","2.2.7"},
                        {xmerl,"XML parser","1.2.10"},
                        {sasl,"SASL  CXC 138 11","2.1.10"},
                        {stdlib,"ERTS  CXC 138 10","1.17.5"},
                        {kernel,"ERTS  CXC 138 10","2.14.5"}]},
 {os,{unix,linux}},
 {erlang_version,"Erlang R14B04 (erts-5.8.5) [source] [64-bit] [rq:1] [async-threads:30] [kernel-poll:true]\n"},
 {memory,[{total,27101856},
          {connection_procs,2648},
          {queue_procs,5296},
          {plugins,0},
          {other_proc,9182320},
          {mnesia,57456},
          {mgmt_db,0},
          {msg_index,21848},
          {other_ets,765504},
          {binary,3296},
          {code,14419185},
          {atom,1354457},
          {other_system,1289846}]},
 {vm_memory_high_watermark,0.4},
 {vm_memory_limit,838362726},
 {disk_free_limit,1000000000},
 {disk_free,15992676352},
 {file_descriptors,[{total_limit,924},
                    {total_used,3},
                    {sockets_limit,829},
                    {sockets_used,1}]},
 {processes,[{limit,1048576},{used,122}]},
 {run_queue,0},
 {uptime,4}]
...done.
[root@puppetserver rpms]# netstat -nlp | grep beam #默认监听端口为5672
tcp        0      0 0.0.0.0:44422              0.0.0.0:*                   LISTEN     43198/beam        
tcp        0      0 :::5672                     :::*                        LISTEN      43198/beam

2.2.3 配置RabbitMQ

2.2.3.1 加载amqp_client和rabbit_stomp插件

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
[root@puppetserver sbin]#ln -s /usr/lib/rabbitmq/lib/rabbitmq_server-3.1.5/sbin/rabbitmq-plugins /usr/sbin/rabbitmq-plugins #创建命令rabbitmq-plugins的软连接
[root@puppetserver sbin]# ln -s /usr/lib/rabbitmq/lib/rabbitmq_server-3.1.5/sbin/rabbitmq-env /usr/sbin/rabbitmq-env #创建命令rabbitmq-env的软连接
[root@puppetserver sbin]# rabbitmq-plugins enable rabbitmq_stomp #开启rabbitmq_stomp插件
The following plugins have been enabled:
  amqp_client
  rabbitmq_stomp
Plugin configuration has changed. Restart RabbitMQfor changes to take effect.
[root@puppetserver sbin]# /etc/rc.d/init.d/rabbitmq-server restart
Restarting rabbitmq-server: SUCCESS
rabbitmq-server.
[root@puppetserver rabbitmq]# tailf /var/log/rabbitmq/rabbit\@puppetserver.log #可以从日志看到stomp插件加载成功
=INFO REPORT==== 3-Oct-2013::20:25:18===
started STOMP TCP Listener on [::]:61613
=INFO REPORT==== 3-Oct-2013::20:25:18===
Server startup complete;2 plugins started.
 * amqp_client
 * rabbitmq_stomp
 *

2.2.3.2 创建rabbitmq.config配置文件,修改监听端口为61613

1
2
3
4
5
6
7
8
9
10
11
[root@puppetserver rpms]# vim /etc/rabbitmq/rabbitmq.config
[
        {stomp,[ {tcp_listeners, [61613]} ]} #设置connector为stomp,监听端口为61613
].
[root@puppetserver rpms]# /etc/rc.d/init.d/rabbitmq-server restart
Restarting rabbitmq-server: SUCCESS
rabbitmq-server.
[root@puppetserver rpms]# netstat -nlp | grep beam  #默认监听端口为61613
tcp        0     0 0.0.0.0:56532              0.0.0.0:*                   LISTEN     1906/beam.smp     
tcp        0      0 :::61613                    :::*                        LISTEN      1906/beam.smp     
tcp        0      0 :::5672                     :::*                        LISTEN      1906/beam.smp

2.2.3.3 删除默认账户guest,为MCollective创建账户“mcollective”并设置密码为“secret”,然后设置权限。

1
2
3
4
5
6
7
8
9
10
11
12
13
[root@puppetserver rpms]# rabbitmqctl delete_user guest
Deleting user "guest" ...
...done.
[root@puppetserver rpms]# rabbitmqctl add_user mcollective secret
Creating user "mcollective" ...
...done.
[root@puppetserver rpms]# rabbitmqctl set_permissions -p"/" mcollective  ".*" ".*" ".*"
Setting permissionsfor user "mcollective"in vhost "/"...
...done.
[root@puppetserver sbin]# rabbitmqctl list_users #查看监听用户
Listing users ...
mcollective    []
...done.

备注:RabbitMQ拥有一个默认的guest账户,它默认对消息队列拥有全部权限。出于安全方面的考虑,建议删除这个账户。

更多详细配置信息请参考
http://www.rabbitmq.com/admin-guide.html

更多详细配置信息请参考:
http://docs.puppetlabs.com/mcollective/reference/plugins/connector_rabbitmq.html

3 安装和配置MCollective

3.1 安装MCollective

3.1.1 测试端安装MCollective客户端

1
[root@puppetserver rpms]# yum install mcollective-common  mcollective-client #依赖包rubygem-stomp

3.1.2 节点安装MCollective服务端

1
[root@agent1 ~]# yum install mcollective  mcollective-common #依赖rubygem-stomp、rubygems和ruby相关包

3.2 配置MCollective

3.2.1 测试端配置MCollective客户端

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
[root@puppetserver rpms]# vim /etc/mcollective/client.cfg
topicprefix = /topic/
main_collective = mcollective
collectives = mcollective
libdir = /usr/libexec/mcollective
logger_type = console
loglevel = warn
# Plugins
securityprovider = psk
plugin.psk = a36cd839414370e10fd281b8a38a4f48 #MCollective通信共享密钥,和MCollective服务端保持一致
connector = stomp  #通信协议
plugin.stomp.host =192.168.100.110 #Middleware地址
plugin.stomp.port =61613  #Middleware监听端口
plugin.stomp.user = mcollective  #Middleware通信账号
plugin.stomp.password = secret #Middleware通信密码
# Facts
factsource = yaml
plugin.yaml = /etc/mcollective/facts.yaml

3.2.2 节点配置MCollective服务端

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
[root@agent1 rpms]# vim /etc/mcollective/server.cfg
topicprefix = /topic/
main_collective = mcollective
collectives = mcollective
libdir = /usr/libexec/mcollective  #存放plugins的位置
logfile = /var/log/mcollective.log
loglevel = info
daemonize = 1
# Plugins
securityprovider = psk
plugin.psk = a36cd839414370e10fd281b8a38a4f48 #MCollective通信共享密钥,和MCollective客户端保持一致
connector = stomp #通信协议
plugin.stomp.host = 192.168.100.110 #Middleware地址
plugin.stomp.port = 61613  #Middleware监听端口
plugin.stomp.user = mcollective  #Middleware通信账号
plugin.stomp.password = secret #Middleware通信密码
# Facts
factsource = yaml
plugin.yaml = /etc/mcollective/facts.yaml
[root@agent1 ~]# /etc/rc.d/init.d/mcollective start
Starting mcollective:                                      [  OK  ]
[root@agent1 ~]# chkconfig mcollective on
[root@agent1 ~]#

3.3 测试Mcollective与Middleware通信

1
2
3
4
5
6
7
8
[root@puppetserver rpms]# mco ping  #检查所有存活的节点
agent2.kisspuppet.com                       time=119.98ms
agent1.kisspuppet.com                       time=159.31ms
---- ping statistics ----
2 replies max: 159.31 min: 119.98 avg: 139.64
[root@puppetserver rpms]# mco find
agent1.kisspuppet.com
agent2.kisspuppet.com

4 Mcollective插件的安装及测试

MCollective可以使用多种方式进行扩展。最普遍的一种扩展MCollective的方式就是重用已经写好的agent插件。这些小的Ruby库可以让MCollective在整个集群中执行自定义的命令。

一个agent插件通常包含一个Ruby库,它必须被分发到所有运行MCollective agent的节点上。另外,一个数据定义文件(DDL)提供了插件接受的传入参数的具体描述,整个DDL文件需要放在MCollective客户端系统上。最后,一个使用指定的agent插件运行MCollective的脚步也需要被安装到所有的MCollective客户端系统上。

备注:更多插件可以在https://github.com/puppetlabs/mcollective-plugins找到。

4.1 安装puppet agent插件

MCollective本身并不包含一个可以立即使用的Puppet agent插件,需要安装使用。这一插件可以让操作员在需要时运行Puppet agent。他不需要等待Puppet agent的默认运行间隔,也不需要使用其他工具来开始这些任务

4.1.1 安装MCollective的Agent插件

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
[root@agent1 rpms]# yum install mcollective-puppet-agent mcollective-puppet-common
[root@agent1 rpms]# ll /usr/libexec/mcollective/mcollective/agent/
total 36
-rw-r--r-- 1root root 1033 May 21 01:34discovery.rb
-rw-r--r-- 1root root 8346 May 14 07:28puppet.ddl
-rw-r--r-- 1root root 7975 May 14 07:25puppet.rb
-rw-r--r-- 1root root 5999 May 21 01:34rpcutil.ddl
-rw-r--r-- 1root root 3120 May 21 01:34rpcutil.rb
[root@puppetserver rpms]# yum install mcollective-puppet-client mcollective-puppet-common
[root@puppetserver rpms]# ll /usr/libexec/mcollective/mcollective/agent/
total 28
-rw-r--r-- 1 root root 1033May 21 01:34 discovery.rb
-rw-r--r-- 1 root root 8346May 14 07:28 puppet.ddl
-rw-r--r-- 1 root root 5999May 21 01:34 rpcutil.ddl
-rw-r--r-- 1 root root 3120May 21 01:34 rpcutil.rb

4.1.2 载入Agent插件

1
2
3
4
5
6
7
8
9
10
11
12
[root@puppetserver rpms]# mco  #客户端默认在自动载入
The Marionette Collective version2.2.4
usage: /usr/bin/mco command <options>
Known commands:
   completion           facts                find              
   help                 inventory            ping              
   plugin               puppet               rpc               
Type '/usr/bin/mco help'for a detailed listof commands and'/usr/bin/mco help command'
toget detailed helpfor a command
[root@agent1 ~]# /etc/rc.d/init.d/mcollective restart
Shutting down mcollective:                                 [  OK  ]
Starting mcollective:                                      [  OK  ]

4.1.3 验证Agent插件是否被载入

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
[root@puppetserver rpms]# mco inventory agent1.kisspuppet.com #查看节点agent1是否已经载入puppet插件
Inventory foragent1.kisspuppet.com:
   Server Statistics:
                      Version:2.2.4
                   Start Time: Thu Oct03 16:09:03+0800 2013
                  Config File: /etc/mcollective/server.cfg
                  Collectives: mcollective
              Main Collective: mcollective
                   Process ID:8902
               Total Messages:3
      Messages Passed Filters:3
            Messages Filtered:0
             Expired Messages:0
                 Replies Sent:2
         Total Processor Time:0.46 seconds
                  System Time:0.12 seconds
   Agents:
      discovery       puppet          rpcutil      
   Data Plugins:
      agent           fstat           puppet   #已经载入puppet插件    
      resource                                     
   Configuration Management Classes:
      No classes applied
   Facts:
      mcollective =>1

4.1.4 从MCollective中运行Puppet

在运行命令之前,可以在节点查看puppet日志和puppetd服务的启停来判断命令是否调用了puppetd进程。

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
[root@puppetserver ~]# mco puppet  --noop --verbose status #查看节点agent守护进程状态
Discovering hosts using the mc methodfor 2 second(s) .... 2
 * [ ============================================================> ]2 / 2
   agent2.kisspuppet.com: Currently stopped; last completed run9 hours 35minutes 36 seconds ago
   agent1.kisspuppet.com: Currently stopped; last completed run9 hours 35minutes 34 seconds ago
Summary of Applying:
   false= 2
Summary of Daemon Running:
   stopped =2
Summary of Enabled:
   enabled =2
[root@puppetserver rpms]# mco puppet -v runonce
Discovering hosts using the mc methodfor 2 second(s) .... 2
 * [ ============================================================> ]2 / 2
agent1.kisspuppet.com                      : OK
    {:summary=>     "Started a background Puppet run using the 'puppet agent --onetime --daemonize --color=false --splay --splaylimit 30' command"}
agent2.kisspuppet.com                      : OK
    {:summary=>     "Started a background Puppet run using the 'puppet agent --onetime --daemonize --color=false --splay --splaylimit 30' command"}
---- rpc stats ----
           Nodes:2 / 2
     Pass / Fail:2 / 0
      Start Time: Thu Oct03 16:12:03+0800 2013
  Discovery Time:2007.23ms
      Agent Time:3591.72ms
      Total Time:5598.94ms

备注:当使用MCollective运行Puppet时,要求在所有被管理的节点上Puppet agent守护进程都需要被关闭。在每次使用mco puppet -v runonce命令调用puppetd agent时,MCollective都会产生一个新的Puppet进程。这个进程会和任何已经运行的Puppet agent守护进程产生功能性的重复。

当Puppet使用—runonce参数运行时,agent会在后台运行。所以虽然MCollective成功运行了Puppet,但实际上的Puppet agent运行可能并不成功。需要查看Puppet报告来确定每一个Puppet agent运行的结果。MCollective返回的OK值表示MCollective服务器成功地启动了puppetd进程并且没有得到任何输出。

4.2 安装facter插件(测试多次发现存在不稳定性)

注意:通过facter插件获取节点facter变量信息不是很稳定,因此可将节点facts信息通过inline_template写入/etc/mcollective/facts.yaml中,并在/etc/mcollective/server.cfg中设置factsource = yaml,这样MCollective客户端只需要每次读取这个文件中的facter变量即可。而且在本地目录/var/lib/puppet/yaml/facts/也会生成一份节点的facter信息,模块部分信息如下:

1
2
3
4
5
6
7
8
9
10
class mcollective::facter {
    file{"/etc/mcollective/facts.yaml":
          owner    => root,
          group    => root,
          mode     =>0440,
          loglevel => debug, # reduce noisein Puppet reports
#          content  => inline_template("<%= scope.to_hash.reject { |k,v| k.to_s =~ /(uptime_seconds|timestamp|free)/ }.to_yaml %>"), # exclude rapidly changing fact
          content => inline_template('<%= scope.to_hash.reject { |k,v| k.to_s =~ /(uptime.*|path|timestamp|free|.*password.*|.*psk.*|.*key)/ }.to_yaml %>'),
    }
}
1
2
3
4
5
6
[root@agent1 ~]# yum install mcollective-facter-facts
[root@agent1 rpms]# ll /usr/libexec/mcollective/mcollective/facts/
total 12
-rw-r--r-- 1root root  422 Feb 21  2013facter_facts.ddl
-rw-r--r-- 1root root  945 Feb 21  2013facter_facts.rb
-rw-r--r-- 1root root 1530 May 21 01:34yaml_facts.rb
1
2
3
4
5
6
7
8
9
[root@agent1 ~]# vim /etc/mcollective/server.cfg
# Facts
#factsource = yaml #注释掉
factsource = facter
plugin.yaml = /etc/mcollective/facts.yaml
[root@agent1 rpms]# /etc/rc.d/init.d/mcollective restart
Shutting down mcollective:                                 [  OK  ]
Starting mcollective:                                      [  OK  ]
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
[root@puppetserver rpms]# mco inventory agent1.kisspuppet.com #查看节点agent1是否加载了facts插件
Inventory foragent1.kisspuppet.com:
   Server Statistics:
                      Version:2.2.4
                   Start Time: Thu Oct03 16:31:47+0800 2013
                  Config File: /etc/mcollective/server.cfg
                  Collectives: mcollective
              Main Collective: mcollective
                   Process ID:9485
               Total Messages:37
      Messages Passed Filters:33
            Messages Filtered:4
             Expired Messages:0
                 Replies Sent:32
         Total Processor Time:0.74 seconds
                  System Time:0.21 seconds
   Agents:
      discovery       puppet          rpcutil      
   Data Plugins:
      agent           fstat           puppet       
      resource                                     
   Configuration Management Classes:
      No classes applied
   Facts:  #可以看到获取的节点facter信息(获取信息需要一些等待时间)
      architecture => x86_64
      augeasversion =>0.10.0
      bios_release_date =>07/02/2012
      bios_vendor => Phoenix Technologies LTD
      bios_version =>6.00
      blockdevice_fd0_size =>4096
     
      uptime_days =>0
      uptime_hours =>20
      uptime_seconds =>74506
      uuid => 564DFBAB-CADC-FC69-36CA-955BFDB30F43
      virtual => vmware
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
[root@puppetserver rpms]# mco facts lsbdistdescription -v  #使用mco facts命令对操作系统类型进行显示
Discovering hosts using the mc methodfor 2 second(s) .... 2
Report forfact: lsbdistdescription
        Red Hat Enterprise Linux Server release5.7 (Tikanga)found 1 times
            agent2.kisspuppet.com
        Red Hat Enterprise Linux Server release5.8 (Tikanga)found 1 times
            agent1.kisspuppet.com
---- rpc stats ----
           Nodes:2 / 2
     Pass / Fail:2 / 0
      Start Time: Thu Oct03 16:59:04+0800 2013
  Discovery Time:2004.83ms
      Agent Time:67.32ms
      Total Time:2072.15ms
[root@puppetserver rpms]# mco facts lsbdistdescription  #使用mco facts命令对操作系统类型进行统计
Report forfact: lsbdistdescription
        Red Hat Enterprise Linux Server release5.7 (Tikanga)found 1 times
        Red Hat Enterprise Linux Server release5.8 (Tikanga)found 1 times
Finished processing2 / 2hosts in 79.15 ms
[root@puppetserver rpms]# mco facts -v --with-fact hostname='agent1'memoryfree #查看主机agent1的剩余内存
Discovering hosts using the mc methodfor 2 second(s) .... 1
Report forfact: memoryfree
        795.13MB                               found 1times
            agent1.kisspuppet.com
---- rpc stats ----
           Nodes:1 / 1
     Pass / Fail:1 / 0
      Start Time: Thu Oct03 17:02:13+0800 2013
  Discovery Time:2005.65ms
      Agent Time:49.37ms
      Total Time:2055.03ms

4.3 使用元数据定位主机

4.3.1 使用默认facter元数据定位主机

4.3.1.1 触发所有系统为RedHat,版本为5.7的所有节点puppetd守护进程

1
2
3
4
5
6
7
8
9
10
11
[root@puppetserver rpms]# mco puppet -v runonce   rpc --np -F  operatingsystemrelease='5.7'-F operatingsystem='RedHat' 
Discovering hosts using the mc methodfor 2 second(s) .... 1
agent2.kisspuppet.com                      : OK
    {:summary=>     "Started a background Puppet run using the 'puppet agent --onetime --daemonize --color=false --splay --splaylimit 30' command"}
---- rpc stats ----
           Nodes:1 / 1
     Pass / Fail:1 / 0
      Start Time: Thu Oct03 17:03:56+0800 2013
  Discovery Time:2008.09ms
      Agent Time:1187.69ms
      Total Time:3195.78ms

4.3.1.2 触发所有系统为RedHat,kernel版本为2.6.18的所有节点puppetd守护进程

1
2
3
4
5
6
7
8
9
10
11
12
13
[root@puppetserver rpms]# mco puppet -v runonce   rpc --np -F  kernelversion='2.6.18' -F operatingsystem='RedHat'
Discovering hosts using the mc methodfor 2 second(s) .... 2
agent2.kisspuppet.com                      : OK
    {:summary=>     "Started a background Puppet run using the 'puppet agent --onetime --daemonize --color=false --splay --splaylimit 30' command"}
agent1.kisspuppet.com                      : OK
    {:summary=>     "Started a background Puppet run using the 'puppet agent --onetime --daemonize --color=false --splay --splaylimit 30' command"}
---- rpc stats ----
           Nodes:2 / 2
     Pass / Fail:2 / 0
      Start Time: Thu Oct03 17:06:15+0800 2013
  Discovery Time:2004.32ms
      Agent Time:1308.34ms
      Total Time:3312.66ms

4.3.2 使用自定义facter元数据定位主机

备注:使用自定义facter元数据可以更加灵活的定位主机,如何定义fact可参考博文《通过自定义fact增强MCollective推送更新元数据的灵活性》

4.3.2.1 在agent1上定义facter my_apply1和my_apply2

1
2
3
[root@agent1 mcollective]# facter -p | grep my_apply
my_apply1 => apache
my_apply2 => mysql

4.3.2.2 在agent2上定义facter my_apply2和my_apply3

1
2
3
[root@agent2 mcollective]# facter -p | grep my_apply
my_apply2 => mysql
my_apply3 => php

4.3.2.3 在MCollective客户端测试节点自定义facter是否正确

1
2
3
4
5
6
[root@puppetserver facter]# mco inventory agent1.kisspuppet.com  | grep my_apply
      my_apply1 => apache
      my_apply2 => mysql
[root@puppetserver facter]# mco inventory agent2.kisspuppet.com  | grep my_apply
      my_apply2 => mysql
      my_apply3 => php

4.3.2.4 通过自定义facter定位主机触发更新

1
2
3
4
5
6
7
8
9
10
11
12
[root@puppetserver facter]# mco puppet -v runonce  mco facts -v --with-fact  my_apply3='php'#筛选节点facter变量my_apply3=php的主机进行触发puppetd守护进程
Discovering hosts using the mc methodfor 2 second(s) .... 1
 * [ ============================================================> ]1 / 1
agent2.kisspuppet.com                      : OK
    {:summary=>     "Started a background Puppet run using the 'puppet agent --onetime --daemonize --color=false --splay --splaylimit 30' command"}
---- rpc stats ----
           Nodes:1 / 1
     Pass / Fail:1 / 0
      Start Time: Thu Oct03 23:33:54+0800 2013
  Discovery Time:2005.35ms
      Agent Time:1078.86ms
      Total Time:3084.21ms
0 0