分类: 高性能计算
2013-08-23 01:46:54
This document describes the steps necessary to quickly set up a cluster with IBM system x, rack-mounted servers. Although the examples given in this document are specific to iDataplex hardware (because that's the most common server type used for clusters), the basic instructions apply to any x86_64, IPMI-controlled, rack-mounted servers.
Contents[] |
This document is meant to get you going as quickly as possible and therefore only goes through the most common scenario. For additional scenarios and setup tasks, see .
This configuration will have a single dx360 Management Node with 167 other dx360 servers as nodes. The OS deployed will be RH Enterprise Linux 6.2, x86_64 edition. Here is a diagram of the racks:
In our example, the management node is known as 'mgt', the node namess are n1-n167, and the domain will be 'cluster'. We will use the BMCs in shared mode so they will share the NIC on each node that the node's operating system communicates to the xCAT management node over. This is call the management LAN. We will use subnet 172.16.0.0 with a netmask of 255.240.0.0 (/12) for it. (This provides an IP address range of 172.16.0.1 - 172.31.255.254 .) We will use the following subsets of this range for:
The network is physically laid out such that port number on a switch is
equal to the U position number within a column, like this:
Here is a summary of the steps required to set up the cluster and what this document will take you through:
Install one of the supported distros on the Management Node (MN). It is recommended to ensure that dhcp, bind (not bind-chroot), httpd, nfs-utils, and perl-XML-Parser are installed. (But if not, the process of installing the xCAT software later will pull them in, assuming you follow the steps to make the distro RPMs available.)
Hardware requirements for your xCAT management node are dependent on your cluster size and configuration. A minimum requirement for an xCAT Management Node or Service Node that is dedicated to running xCAT to install a small cluster ( < 16 nodes) should have 4-6 Gigabytes of memory. A medium size cluster, 6-8 Gigabytes of memory; and a large cluster, 16 Gigabytes or more. Keeping swapping to a minimum should be a goal.
For a list of supported OS and Hardware, refer to .
Note: you can skip this step in xCAT 2.8.1 and above, because xCAT does it automatically when it is installed.
To disable SELinux manually:
echo 0 > /selinux/enforce sed -i 's/^SELINUX=.*$/SELINUX=disabled/' /etc/selinux/config
Note: you can skip this step in xCAT 2.8 and above, because xCAT does it automatically when it is installed.
The management node provides many services to the cluster nodes, but the firewall on the management node can interfere with this. If your cluster is on a secure network, the easiest thing to do is to disable the firewall on the Management Mode:
For RH:
service iptables stop chkconfig iptables off
If disabling the firewall completely isn't an option, configure iptables to allow the following services on the NIC that faces the cluster: DHCP, TFTP, NFS, HTTP, DNS.
For SLES:
SuSEfirewall2 stop
The xCAT installation process will scan and populate certain settings from the running configuration. Having the networks configured ahead of time will aid in correct configuration. (After installation of xCAT, all the networks in the cluster must be defined in the xCAT networks table before starting to install cluster nodes.) When xCAT is installed on the Management Node, it will automatically run makenetworks to create an entry in the networks table for each of the networks the management node is on. Additional network configurations can be added to the xCAT networks table manually later if needed.
The networks that are typically used in a cluster are:
In our example, we only deal with the management network because:
For more information, see .
Configure the cluster facing NIC(s) on the management node. For example edit the following files:
[RH]: /etc/sysconfig/network-scripts/ifcfg-eth1
[SLES]: /etc/sysconfig/network/ifcfg-eth1
DEVICE=eth1 ONBOOT=yes BOOTPROTO=static IPADDR=172.20.0.1 NETMASK=255.240.0.0
If the public facing NIC on your management node is configured by DHCP, you may want to set PEERDNS=no in the NIC's config file to prevent the dhclient from rewriting /etc/resolv.conf. This would be important if you will be configuring DNS on the management node (via makedns - covered later in this doc) and want the management node itself to use that DNS. In this case, set PEERDNS=no in each /etc/sysconfig/network-scripts/ifcfg-* file that has BOOTPROTO=dhcp.
On the other hand, if you want dhclient to configure /etc/resolv.conf on your management node, then don't set PEERDNS=no in the NIC config files.
The xCAT management node hostname should be configured before installing xCAT on the management node. The hostname or its resolvable ip address will be used as the default master name in the xCAT site table, when installed. This name needs to be the one that will resolve to the cluster-facing NIC. Short hostnames (no domain) are the norm for the management node and all cluster nodes. Node names should never end in "-enx" for any x.
To set the hostname, edit /etc/sysconfig/network to contain, for example:
HOSTNAME=mgt
If you run hostname command, if should return the same:
# hostname mgt
Ensure that at least the management node is in /etc/hosts:
127.0.0.1 localhost.localdomain localhost ::1 localhost6.localdomain6 localhost6 ### 172.20.0.1 mgt mgt.cluster
When using the management node to install compute nodes, the timezone configuration on the management node will be inherited by the compute nodes. So it is recommended to setup the correct timezone on the management node. To do this on RHEL, see . The process is similar, but not identical, for SLES. (Just google it.)
You can also optionally set up the MN as an NTP for the cluster. See .
It is not required, but recommended, that you create a separate file system for the /install directory on the Management Node. The size should be at least 30 meg to hold to allow space for several install images.
Note: in xCAT 2.8 and above, you do not need to restart the management node. Simply restart the cluster-facing NIC, for example: ifdown eth1; ifup eth1
For xCAT 2.7 and below, though it is possible to restart the correct services for all settings, the simplest step would be to reboot the Management Node at this point.
It is recommended that spanning tree be set in the switches to portfast or edge-port for faster boot performance. Please see the relevant switch documentation as to how to configure this item.
It is recommended that lldp protocol in the switches is enabled to collect the switch and port information for compute node during discovery process.
Note: this step is necessary if you want to use xCAT's automatic switch-based discovery (described later on in this document) for IPMI-controlled rack-mounted servers (including iDataPlex) and Flex chassis. If you have a small cluster and prefer to use the sequential discover method (described later) or manually enter the MACs for the hardware, you can skip this section. Although you may want to still set up your switches for management so you can use xCAT tools to manage them, as described in .
xCAT will use the ethernet switches during node discovery to find out which switch port a particular MAC address is communicating over. This allows xCAT to match a random booting node with the proper node name in the database. To set up a switch, give it an IP address on its management port and enable basic SNMP functionality. (Typically, the SNMP agent in the switches is disabled by default.) The easiest method is to configure the switches to give the SNMP version 1 community string called "public" read access. This will allow xCAT to communicate to the switches without further customization. (xCAT will get the list of switches from the table.) If you want to use SNMP version 3 (e.g. for better security), see the example below. With SNMP V3 you also have to set the user/password and AuthProto (default is 'md5') in the table.
If for some reason you can't configure SNMP on your switches, you can use sequential discovery or the more manual method of entering the nodes' MACs into the database. See for a description of your choices.
xCAT supports many switch types, such as BNT and Cisco. Here is an example of configuring SNMP V3 on the Cisco switch 3750/3650:
1. First, user should switch to the configure mode by the following commands:
[root@x346n01 ~]# telnet xcat3750 Trying 192.168.0.234... Connected to xcat3750. Escape character is '^]'. User Access Verification Password:
xcat3750-1>enable Password:
xcat3750-1#configure terminal Enter configuration commands, one per line. End with CNTL/Z. xcat3750-1(config)#
2. Configure the snmp-server on the switch:
Switch(config)# access-list 10 permit 192.168.0.20 # 192.168.0.20 is the IP of MN Switch(config)# snmp-server group xcatadmin v3 auth write v1default Switch(config)# snmp-server community public RO 10 Switch(config)# snmp-server community private RW 10 Switch(config)# snmp-server enable traps license?
3. Configure the snmp user id (assuming a user/pw of xcat/passw0rd):
Switch(config)# snmp-server user xcat xcatadmin v3 auth SHA passw0rd access 10
4. Check the snmp communication to the switch :
yum install net-snmp net-snmp-utils
snmpwalk -v 3 -u xcat -a SHA -A passw0rd -X cluster -l authnoPriv 192.168.0.234 .1.3.6.1.2.1.2.2.1.2
Later on in this document, it will explain how to make sure the switch and switches tables are setup correctly.
There are two options for installation of xCAT:
Pick either one, but not both.
If not able to, or not wishing to, use the live internet repository, choose this option.
Go to the site and download the level of xCAT tarball you desire. Go to the page and download the latest snap of the xCAT dependency tarball. (The latest snap of the xCAT dependency tarball will work with any version of xCAT.)
Copy the files to the Management Node (MN) and untar them:
mkdir /root/xcat2 cd /root/xcat2 tar jxvf xcat-core-2.*.tar.bz2 # or core-rpms-snap.tar.bz2 tar jxvf xcat-dep-*.tar.bz2
Point YUM to the local repositories for xCAT and its dependencies:
cd /root/xcat2/xcat-dep// ./mklocalrepo.sh cd /root/xcat2/xcat-core ./mklocalrepo.sh
[SLES 11]:
zypper ar file:///root/xcat2/xcat-dep/sles11/xCAT-dep zypper ar file:///root/xcat2/xcat-core xcat-core
You can check a zypper repository using "zypper lr -d", or remove a zypper repository using "zypper rr".
[SLES 10.2+]:
zypper sa file:///root/xcat2/xcat-dep/sles10/xCAT-dep zypper sa file:///root/xcat2/xcat-core xcat-core
You can check a zypper repository using "zypper sl -d", or remove a zypper repository using "zypper sd".
When using the live internet repository, you need to first make sure that name resolution on your management node is at least set up enough to resolve sourceforge.net. Then make sure the correct repo files are in /etc/yum.repos.d:
To get the current official release:
cd /etc/yum.repos.d wget
To get the deps package:
wget
for example:
wget
To setup to use SLES with zypper:
[SLES11]:
zypper ar -t rpm-md xCAT-core
zypper ar -t rpm-md/ xCAT-dep
[SLES10.2+]:
zypper sa xCAT-core
zypper sa/ xCAT-dep
xCAT uses on several packages that come from the Linux distro. Follow this section to create the repository of the OS on the Management Node.
See the following documentation:
[RH]: Use yum to install xCAT and all the dependencies:
yum clean metadata yum install xCAT
[SLES]: Use zypper to install xCAT and all the dependencies:
zypper install xCAT
Note: in xCAT 2.8.2 and above, xCAT supports cloning new nodes from a pre-installed/pre-configured node, we call this provisioning method as sysclone. It leverages the opensource tool systemimager. If you will be installing stateful(diskful) nodes using the sysclone provmethod, you need to install systemimager and all the dependencies (using sysclone is optional):
[RH]: Use yum to install systemimager and all the dependencies:
yum install systemimager-server
[SLES]: Use zypper to install systemimager and all the dependencies:
zypper install systemimager-server
Add xCAT commands to the path by running the following:
source /etc/profile.d/xcat.sh
Check to see the database is initialized:
tabdump site
The output should similar to the following:
key,value,comments,disable "xcatdport","3001",, "xcatiport","3002",, "tftpdir","/tftpboot",, "installdir","/install",, . . .
If the tabdump command does not work, see .
If you need to update the xCAT RPMs later:
To update xCAT:
[RH]:
yum clean metadata yum update '*xCAT*'
[SLES]:
zypper refresh zypper update -t package '*xCAT*'
Note: this will not apply updates that may have been made to some of the xCAT deps packages. (If there are brand new deps packages, they will get installed.) In most cases, this is ok, but if you want to make all updates for xCAT rpms and deps, run the following command. This command will also pick up additional OS updates.
[RH]:
yum update
[SLES]:
zypper refresh zypper update
Note: If you are updating from xCAT 2.7.x (or earlier) to xCAT 2.8 or later, there are some additional migration steps that need to be considered:
All networks in the cluster must be defined in the networks table. When xCAT was installed, it ran makenetworks, which created an entry in this table for each of the networks the management node is connected to. Now is the time to add to the networks table any other networks in the cluster, or update existing networks in the table.
For a sample Networks Setup, see the following example:
The password should be set in the passwd table that will be assigned to root when the node is installed. You can modify this table using tabedit. To change the default password for root on the nodes, change the system line. To change the password to be used for the BMCs, change the ipmi line.
tabedit passwd #key,username,password,cryptmethod,comments,disable "system","root","cluster",,, "ipmi","USERID","PASSW0RD",,,
To get the hostname/IP pairs copied from /etc/hosts to the DNS on the MN:
chdef -t site forwarders=1.2.3.4,1.2.5.6
search cluster nameserver 172.20.0.1
makedns -n
For more information about name resolution in an xCAT Cluster, see .
You usually don't want your DHCP server listening on your public (site) network, so set site.dhcpinterfaces to your MN's cluster facing NICs. For example:
chdef -t site dhcpinterfaces=eth1
Then this will get the network stanza part of the DHCP configuration (including the dynamic range) set:
makedhcp -n
The IP/MAC mappings for the nodes will be added to DHCP automatically as the nodes are discovered.
Nothing to do here - the TFTP server is done by xCAT during the Management Node install.
makeconservercf
If you want to run a discovery process, a dynamic range must be defined in the networks table. It's used for the nodes to get an IP address before xCAT knows their MAC addresses.
In this case, we'll designate 172.20.255.1-172.20.255.254 as a dynamic range:
chdef -t network 172_16_0_0-255_240_0_0 dynamicrange=172.20.255.1-172.20.255.254
Several xCAT database tables must be filled in while setting up an iDataPlex cluster. To make this process easier, xCAT provides several template files in /opt/xcat/share/xcat/templates/e1350/. These files contain regular expressions that describe the naming patterns in the cluster. With xCAT's regular expression support, one line in a table can define one or more attribute values for all the nodes in a node group. (For more information on xCAT's database regular expressions, see .) To load the default templates into your database:
cd /opt/xcat/share/xcat/templates/e1350/ for i in *csv; do tabrestore $i; done
These templates contain entries for a lot of different node groups, but we will be using the following node groups:
In our example, ipmi, idataplex, 42perswitch, and compute will all have the exact same membership because all of our iDataPlex nodes have those characteristics.
The templates automatically define the following attributes and naming conventions:
For a description of the attribute names in bold above, see the .
If these conventions don't work for your situation, you can either:
Now you can use the power of the templates to define the nodes quickly. By simply adding the nodes to the correct groups, they will pick up all of the attributes of that group:
nodeadd n[001-167] groups=ipmi,idataplex,42perswitch,compute,all nodeadd n[001-167]-bmc groups=84bmcperrack nodeadd switch1-switch4 groups=switch
To see the list of nodes you just defined:
nodels
To see all of the attributes that the combination of the templates and your nodelist have defined for a few sample nodes:
lsdef n100,n100-bmc,switch2
This is the easiest way to verify that the regular expressions in the templates are giving you attribute values you are happy with. (Or, if you modified the regular expressions, that you did it correctly.)
If not using a terminal server, SOL is recommended, but not required to be configured. To instruct xCAT to configure SOL in installed operating systems on dx340 systems:
chdef -t group -o compute serialport=1 serialspeed=19200 serialflow=hard
For dx360-m2 and newer use:
chdef -t group -o compute serialport=0 serialspeed=115200 serialflow=hard
Since the map between the xCAT node names and IP addresses have been added in the hosts table by the 31350 template, you can run the makehosts xCAT command to create the /etc/hosts file from the xCAT hosts table. (You can skip this step if creating /etc/hosts manually.)
makehosts switch,idataplex,ipmi
Verify the entries have been created in the file /etc/hosts. For example your /etc/hosts should look like this:
127.0.0.1 localhost.localdomain localhost ::1 localhost6.localdomain6 localhost6 ### 172.20.0.1 mgt mgt.cluster 172.20.101.1 n1 n1.cluster 172.20.101.2 n2 n2.cluster 172.20.101.3 n3 n3.cluster 172.20.101.4 n4 n4.cluster 172.20.101.5 n5 n5.cluster 172.20.101.6 n6 n6.cluster 172.20.101.7 n7 n7.cluster . . .
Add the node ip mapping to the DNS.
makedns
xCAT supports 3 approaches to discover the new physical nodes and define them to xCAT database:
This is a simple approach in which you give xCAT a range of node names to be given to the discovered nodes, and then you power the nodes on sequentially (usually in physical order), and each node is given the next node name in the noderange.
With this approach, xCAT assumes the nodes are plugged into your ethernet switches in an orderly fashion. So it uses each node's switch port number to determine where it is physically located in the racks and therefore what node name it should be given. This method requires a little more setup (configuring the switches and defining the switch table). But the advantage of this method is that you can power all of the nodes on at the same time and xCAT will sort out which node is which. This can save you a lot of time in a large cluster.
If you don't want to use either of the automatically discovery processes, just follow the manual discovery process.
Choose just one of these options and follow the corresponding section below (and skip the other two).
Note: This feature is only supported in xCAT 2.8.1 and higher.
Sequential Discovery means the new nodes will be discovered one by one. The nodes will be given names from a 'node name pool' in the order they are powered on.
Specify the node name pool by giving a noderange to the nodediscoverstart command:
nodediscoverstart noderange=n[001-010]
The value of noderange should be in the xCAT format.
Note: other node attributes can be given to nodediscoverstart so that xCAT will assign those attributes to the nodes as they are discovered. We aren't showing that in this document, because we already predefined the nodes, the groups they are in, and several attributes (provided by the e1350 templates). If you don't want to predefine nodes, you can give more attributes to nodediscoverstart and have it define the nodes. See the for details.
At this point you can physically power on the nodes one at a time, in the order you want them to receive their node names.
There are additional nodediscover commands you can run during the discovery process. See their for more details.
nodediscoverstatus
nodediscoverls -t seq -l
nodediscoverstop
Note: The sequential discovery process will be stopped automatically when all of the node names in the node name pool are used up.
This method of discovery assumes that you have the nodes plugged into your ethernet switches in an orderly fashion. So we use each nodes switch port number to determine where it is physically located in the racks and therefore what node name it should be given.
To use this discovery method, you must have already configured the switches as described in
The table templates already put group-oriented regular expression entries in the switch table. Use lsdef for a sample node to see if the switch and switchport attributes are correct. If not, use chdef or tabedit to change the values.
If you configured your switches to use SNMP V3, then you need to define several attributes in the switches table. Assuming all of your switches use the same values, you can set these attributes at the group level:
tabch switch=switch switches.snmpversion=3 switches.username=xcat switches.password=passw0rd switches.auth=sha
Prerequisite: The dynamic dhcp range has been configured before your power on the nodes.
If you have a few nodes which were not discovered by Sequential Discovery or Switch Discovery, you could find them in discoverydata table. The undiscovered nodes are identified as 'undef' method in discoverydata table.
Display the undefined nodes with nodediscoverls command:
nodediscoverls -t undef UUID NODE METHOD MTM SERIAL 61E5F2D7-0D59-11E2-A7BC-3440B5BEDBB1 undef undef 786310X 1052EF1 FC5F8852-CB97-11E1-8D59-E41F13EEB1BA undef undef 7914B2A 06DVAC9 96656F17-6482-E011-9954-5CF3FC317F68 undef undef 7377D2C 99A2007
If you want to manually define the 'undefined' node to a specific free node, using the 'nodediscoverdef' command ( available 2.8.2 or higher).
e.g. You have a free node n10 and you want to define the undefined node which uuid is '61E5F2D7-0D59-11E2-A7BC-3440B5BEDBB1' to the n10, run following command:
nodediscoverdef -u 61E5F2D7-0D59-11E2-A7BC-3440B5BEDBB1 -n n10
After the manually defining, the 'node name' and 'discovery method' attributes of undefined node will be changed. You could display the change by nodediscoverls command:
# nodediscoverls UUID NODE METHOD MTM SERIAL 61E5F2D7-0D59-11E2-A7BC-3440B5BEDBB1 n10 manual 786310X 1052EF1 FC5F8852-CB97-11E1-8D59-E41F13EEB1BA undef undef 7914B2A 06DVAC9 96656F17-6482-E011-9954-5CF3FC317F68 undef undef 7377D2C 99A2007
And you can run 'lsdef n10' to see the 'mac address' and 'mtm' has been updated to the node definition. If the next task like bmcsetup has been set in the chain table, the task chain will be continued to run after the running of nodediscoverdef command.
If you want to update node firmware when you discover the nodes, follow the steps in before continuing.
If you want to automatically deploy the nodes after they are discovered, follow the steps in before continuing. (But if you are new to xCAT we don't recommend this.)
To initiate any of the 3 discover methods, walk over to systems and hit the power buttons. For the sequential discovery method power the nodes on in the order that you want them to be given the node names. Wait a short time (about 30 seconds) between each node to ensure they will contact xcatd in the correct order. For the switch and manual discovery processes, you can power on all of the nodes at the same time.
On the MN watch nodes being discovered by:
tail -f /var/log/messages
Look for the dhcp requests, the xCAT discovery requests, and the "
A quick summary of what is happening during the discovery process is:
After a successful discovery process, the following attributes will
be added to the database for each node. (You can verify this by running
lsdef
If you cannot discover the nodes successfully, see the next section .
If at some later time you want to force a re-discover of a node, run:
makedhcp -d
and then reboot the node(s).
When the bmcsetup process completes on each node (about 5-10 minutes), xCAT genesis will drop into a shell and wait indefinitely (and change the node's currstate attribute to "shell"). You can monitor the progress of the nodes using:
watch -d 'nodels ipmi chain.currstate|xcoll'
Before all nodes complete, you will see output like:
==================================== n1,n10,n11,n75,n76,n77,n78,n79,n8,n80,n81,n82,n83,n84,n85,n86,n87,n88,n89,n9,n90,n91 ==================================== shell ==================================== n31,n32,n33,n34,n35,n36,n37,n38,n39,n4,n40,n41,n42,n43,n44,n45,n46,n47,n48,n49,n5,n50,n51,n52, n53,n54,n55,n56,n57,n58,n59,n6,n60,n61,n62,n63,n64,n65,n66,n67,n68,n69,n7,n70,n71,n72,n73,n74 ==================================== runcmd=bmcsetup
When all nodes have made it to the shell, xcoll will just show that the whole nodegroup "ipmi" has the output "shell":
==================================== ipmi ==================================== shell
When the nodes are in the xCAT genesis shell, you can ssh or psh to any of the nodes to check anything you want.
At this point, the BMCs should all be configured and ready for hardware management. To verify this:
# rpower ipmi stat | xcoll ==================================== ipmi ==================================== on
To get the remote console working for each node, some uEFI hardware settings must have specific values. First check the settings, and if they aren't correct, then set them properly. This can be done via the ASU utility. The settings are slightly different, depending on the hardware type:
show uEFI.Com1ActiveAfterBoot show uEFI.SerialPortSharing show uEFI.SerialPortAccessMode show uEFI.RemoteConsoleRedirectionAnd create a file called asu-set with contents:
set uEFI.Com1ActiveAfterBoot Enable set uEFI.SerialPortSharing Enable set uEFI.SerialPortAccessMode Dedicated set uEFI.RemoteConsoleRedirection Enable
show DevicesandIOPorts.Com1ActiveAfterBoot show DevicesandIOPorts.SerialPortSharing show DevicesandIOPorts.SerialPortAccessMode show DevicesandIOPorts.RemoteConsoleAnd create a file called asu-set with contents:
set DevicesandIOPorts.Com1ActiveAfterBoot Enable set DevicesandIOPorts.SerialPortSharing Enable set DevicesandIOPorts.SerialPortAccessMode Dedicated set DevicesandIOPorts.RemoteConsole Enable
Then for both types of machines, use the tool to check these settings:
pasu -b asu-show ipmi | xcoll # Or you can check just one node and assume the rest are the same
If the settings are not correct, then set them:
pasu -b asu-set ipmi | xcoll
For alternate ways to set the ASU settings, see .
Now the remote console should work. Verify it on one node by running:
rcons
To verify that you can see the genesis shell prompt (after hitting enter). To exit rcons type: ctrl-shift-E (all together), then "c", the ".".
You are now ready to choose an operating system and deployment method for the nodes....
There are two options to install your nodes as stateful (diskful) nodes:
This section describes the process for setting up xCAT to install nodes; that is how to install an OS on the disk of each node.
The command copies the contents of the linux distro media to
/install/
copycds/RHEL6.2-*-Server-x86_64-DVD1.iso
copycds /dev/dvd # or whatever the device name of your dvd drive is
Tip: if this is the same distro version as your management node, create a .repo file in /etc/yum.repos.d with content similar to:
[local-rhels6.2-x86_64] name=xCAT local rhels 6.2 baseurl=file:/install/rhels6.2/x86_64 enabled=1 gpgcheck=0
This way, if you need some additional RPMs on your MN at a later, you can simply install them using yum. Or if you are installing other software on your MN that requires some additional RPMs from the disto, they will automatically be found and installed.
The copycds command also automatically creates several osimage defintions in the database that can be used for node deployment. To see them:
lsdef -t osimage # see the list of osimages lsdef -t osimage# see the attributes of a particular osimage
From the list above, select the osimage for your distro, architecture, provisioning method (in this case install), and profile (compute, service, etc.). Although it is optional, we recommend you make a copy of the osimage, changing its name to a simpler name. For example:
lsdef -t osimage -z rhels6.2-x86_64-install-compute | sed 's/^[^ ]\+:/mycomputeimage:/' | mkdef -z
This displays the osimage "rhels6.2-x86_64-install-compute" in a format that can be used as input to mkdef, but on the way there it uses sed to modify the name of the object to "mycomputeimage".
Initially, this osimage object points to templates, pkglists, etc. that are shipped by default with xCAT. And some attributes, for example otherpkglist and synclists, won't have any value at all because xCAT doesn't ship a default file for that. You can now change/fill in any osimage attributes that you want. A general convention is that if you are modifying one of the default files that an osimage attribute points to, copy it into /install/custom and have your osimage point to it there. (If you modify the copy under /opt/xcat directly, it will be over-written the next time you upgrade xCAT.)
But for now, we will use the default values in the osimage definition and continue on. (If you really want to see examples of modifying/creating the pkglist, template, otherpkgs pkglist, and sync file list, see the section . Most of the examples there can be used for stateful nodes too.)
Using a postinstall script ( you could also use the updatenode method):
mkdir /install/postscripts/data cp/install/postscripts/data
Create the postscript updatekernel:
vi /install/postscripts/updatekernel
Add the following lines to the file
#!/bin/bash rpm -Uivh data/kernel-*rpm
Change the permission on the file
chmod 755 /install/postscripts/updatekernel
Add the script to the postscripts table and run the install:
chdef -p -t group -o compute postscripts=updatekernel rnetboot compute
After the initial install of the distro onto nodes, if you want to update the distro on the nodes (either with a few updates or a new SP) without reinstalling the nodes:
copycdsOr, for just a few updated rpms, you can copy the updated rpms from the distributor into a directory under /install and run createrepo in that directory./RHEL6.3-*-Server-x86_64-DVD1.iso
chdef -t osimage rhels6.2-x86_64-install-compute -p pkgdir=/install/rhels6.3/x86_64Note: the above command will add a 2nd repo to the pkgdir attribute. This is only supported for xCAT 2.8.2 and above. For earlier versions of xCAT, omit the -p flag to replace the existing repo directory with the new one.
updatenode compute -P ospkgs
This section describes how to install or configure a diskful node (we call it as golden-client), capture an osimage from this golden-client, the osimage can be used to clone other nodes later.
Note: this support is available in xCAT 2.8.2 and above.
If you want to use the sysclone provisioning method, you need a golden-client. In this way, you can customize and tweak the golden-client’s configuration according to your needs, verify it’s proper operation, so once the image is captured and deployed, the new nodes will behave in the same way as the golden-client.
To install a golden-client, follow the section .
To install the systemimager rpms onto the golden-client:
[RH]:
mkdir -p /install/post/otherpkgs/rhels6.3/x86_64/xcat cd /install/post/otherpkgs/rhels6.3/x86_64/xcat tar jxvf xcat-dep-*.tar.bz2
[RH]:
chdef -t osimage -ootherpkglist=/opt/xcat/share/xcat/install/rh/sysclone.rhels6.x86_64.otherpkgs.pkglist chdef -t osimage -o -p otherpkgdir=/install/post/otherpkgs/rhels6.3/x86_64 rpower reset # you could also use the updatenode method
[Fedora/CentOS]
Using to capture an osimage from the golden-client.
imgcapture-t sysclone -o
Tip: when imgcapture is run, it pulls an osimage from the
golden-client, and creates an osimage definition on xcat management
node. Use lsdef -t osimage
The command tells xCAT what you want to do next with this node, tells the node hardware to boot from the network for the next boot, and powering on the node using starts the installation process:
nodeset compute osimage=mycomputeimage rsetboot compute net rpower compute boot
Tip: when nodeset is run, it processes the kickstart or autoyast template associated with the osimage, plugging in node-specific attributes, and creates a specific kickstart/autoyast file for each node in /install/autoinst. If you need to customize the template, make a copy of the template file that is pointed to by the osimage.template attribute and edit that file (or the files it includes).
It is possible to use the wcons command to watch the installation process for a sampling of the nodes:
wcons n1,n20,n80,n100
or rcons to watch one node
rcons n1
Additionally, nodestat may be used to check the status of a node as it installs:
nodestat n20,n21 n20: installing man-pages - 2.39-10.el5 (0%) n21: installing prep
Note: the percentage complete reported by nodestat is not necessarily reliable.
You can also watch nodelist.status until it changes to "booted" for each node:
nodels compute nodelist.status | xcoll
Once all of the nodes are installed and booted, you should be able ssh to all of them from the MN (w/o a password), because xCAT should have automatically set up the ssh keys (if the postscripts ran successfully):
xdsh compute date
If there are problems, see .
Note: this section describes how to create a stateless image using the genimage command to install a list of rpms into the image. As an alternative, you can also capture an image from a running node and create a stateless image out of it. See for details.
The command copies the contents of the linux distro media to
/install/
copycds/RHEL6.2-Server-20080430.0-x86_64-DVD.iso
copycds /dev/dvd # or whatever the device name of your dvd drive is
Tip: if this is the same distro version as your management node, create a .repo file in /etc/yum.repos.d with content similar to:
[local-rhels6.2-x86_64] name=xCAT local rhels 6.2 baseurl=file:/install/rhels6.2/x86_64 enabled=1 gpgcheck=0
This way, if you need some additional RPMs on your MN at a later, you can simply install them using yum. Or if you are installing other software on your MN that requires some additional RPMs from the disto, they will automatically be found and installed.
Note: To use an osimage as your provisioning method, you need to be running xCAT 2.6.6 or later.
The provmethod attribute of your nodes should contain the name of the osimage object definition that is being used for those nodes. The osimage object contains paths for pkgs, templates, kernels, etc. If you haven't already, run to copy the distro rpms to /install. Default osimage objects are also defined when copycds is run. To view the osimages:
lsdef -t osimage # see the list of osimages lsdef -t osimage# see the attributes of a particular osimage
From the list found above, select the osimage for your distro, architecture, provisioning method (install, netboot, statelite), and profile (compute, service, etc.). Although it is optional, we recommend you make a copy of the osimage, changing its name to a simpler name. For example:
lsdef -t osimage -z rhels6.3-x86_64-netboot-compute | sed 's/^[^ ]\+:/mycomputeimage:/' | mkdef -z
This displays the osimage "rhels6.3-x86_64-netboot-compute" in a format that can be used as input to mkdef, but on the way there it uses sed to modify the name of the object to "mycomputeimage".
Initially, this osimage object points to templates, pkglists, etc. that are shipped by default with xCAT. And some attributes, for example otherpkglist and synclists, won't have any value at all because xCAT doesn't ship a default file for that. You can now change/fill in any osimage attributes that you want. A general convention is that if you are modifying one of the default files that an osimage attribute points to, copy it into /install/custom and have your osimage point to it there. (If you modify the copy under /opt/xcat directly, it will be over-written the next time you upgrade xCAT.)
You likely want to customize the main pkglist for the image. This is the list of rpms or groups that will be installed from the distro. (Other rpms that they depend on will be installed automatically.) For example:
mkdir -p /install/custom/netboot/rh cp -p /opt/xcat/share/xcat/netboot/rh/compute.rhels6.x86_64.pkglist /install/custom/netboot/rh vi /install/custom/netboot/rh/compute.rhels6.x86_64.pkglist chdef -t osimage mycomputeimage pkglist=/install/custom/netboot/rh/compute.rhels6.x86_64.pkglist
The goal is to install the fewest number of rpms that still provides the function and applications that you need, because the resulting ramdisk will use real memory in your nodes.
Also, check to see if the default exclude list excludes all files and directories you do not want in the image. The exclude list enables you to trim the image after the rpms are installed into the image, so that you can make the image as small as possible.
cp /opt/xcat/share/xcat/netboot/rh/compute.exlist /install/custom/netboot/rh vi /install/custom/netboot/rh/compute.exlist chdef -t osimage mycomputeimage exlist=/install/custom/netboot/rh/compute.exlist
Make sure nothing is excluded in the exclude list that you need on the node. For example, if you require perl on your nodes, remove the line "./usr/lib/perl5*".
The linuximage.pkgdir is the name of the directory where the distro packages are stored. It can be set multiple paths. The multiple paths must be separated by ",". The first path is the value of osimage.pkgdir and must be the OS base pkg directory path, such as pkgdir=/install/rhels6.2/x86_64,/install/updates/rhels6.2/x86_64 . In the os base pkg path, there is default repository data. In the other pkg path(s), the users should make sure there is repository data. If not, use "createrepo" command to create them.
If you have additional os updates rpms (rpms may be from the os website, or the additional os distro) that you also want installed, make a directory to hold them, create a list of the rpms you want installed, and add that information to the osimage definition:
mkdir -p /install/updates/rhels6.2/x86_64 cd /install/updates/rhels6.2/x86_64 cp /myrpms/* .
If there is no repository data in the directory, you can run "createrepo" to create it:
createrepo .
The createrepo command is in the createrepo rpm, which for RHEL is in the 1st DVD, but for SLES is in the SDK DVD.
NOTE: when the management node is rhels6.x, and the otherpkgs repository data is for rhels5.x, we should run createrepo with "-s md5". Such as:
createrepo -s md5 .
... myrpm1 myrpm2 myrpm3
chdef -t osimage mycomputeimage pkgdir=/install/rhels6.2/x86_64,/install/updates/rhels6.2/x86_64 pkglist=/install/custom/install/rh/compute.rhels6.x86_64.pkglist
If you add more rpms at a later time, you must run createrepo again.
Note: After the above setting,
If you have additional rpms (rpms not in the distro) that you also want installed, make a directory to hold them, create a list of the rpms you want installed, and add that information to the osimage definition:
mkdir -p /install/post/otherpkgs/rh/x86_64 cd /install/post/otherpkgs/rh/x86_64 cp /myrpms/* . createrepo .
NOTE: when the management node is rhels6.x, and the otherpkgs repository data is for rhels5.x, we should run createrepo with "-s md5". Such as:
createrepo -s md5 .
myrpm1 myrpm2 myrpm3
chdef -t osimage mycomputeimage otherpkgdir=/install/post/otherpkgs/rh/x86_64 otherpkglist=/install/custom/netboot/rh/compute.otherpkgs.pkglist
If you add more rpms at a later time, you must run createrepo again. The createrepo command is in the createrepo rpm, which for RHEL is in the 1st DVD, but for SLES is in the SDK DVD.
If you have multiple sets of rpms that you want to keep separate to keep them organized, you can put them in separate sub-directories in the otherpkgdir. If you do this, you need to do the following extra things, in addition to the steps above:
xcat/xcat-core/xCATsn xcat/xcat-dep/rh6/x86_64/conserver-xcat
There are some examples of otherpkgs.pkglist in
/opt/xcat/share/xcat/netboot/
Note: the otherpkgs postbootscript should by default be associated with every node. Use lsdef to check:
lsdef node1 -i postbootscripts
If it is not, you need to add it. For example, add it for all of the nodes in the "compute" group:
chdef -p -t group compute postbootscripts=otherpkgs
Postinstall scripts for diskless images are analogous to postscripts
for diskfull installation. The postinstall script is run by genimage
near the end of its processing. You can use it to do anything to your
image that you want done every time you generate this kind of image. In
the script you can install rpms that need special flags, or tweak the
image in some way. There are some examples shipped in
/opt/xcat/share/xcat/netboot/
chdef -t osimage mycomputeimage postinstall=/install/custom/netboot/rh/compute.postinstall
Note: This is only supported for stateless nodes in xCAT 2.7 and above.
Sync lists contain a list of files that should be sync'd from the management node to the image and to the running nodes. This allows you to have 1 copy of config files for a particular type of node and make sure that all those nodes are running with those config files. The sync list should contain a line for each file you want sync'd, specifying the path it has on the MN and the path it should be given on the node. For example:
/install/custom/syncfiles/compute/etc/motd -> /etc/motd /etc/hosts -> /etc/hosts
If you put the above contents in /install/custom/netboot/rh/compute.synclist, then:
chdef -t osimage mycomputeimage synclists=/install/custom/netboot/rh/compute.synclist
For more details, see .
You can configure any noderange to use this osimage. In this example, we define that the whole compute group should use the image:
chdef -t group compute provmethod=mycomputeimage
Now that you have associated an osimage with nodes, if you want to list a node's attributes, including the osimage attributes all in one command:
lsdef node1 --osimage
There are other attributes that can be set in your osimage definition. See the osimage man page for details.
If you are building an image for a different OS/architecture than is on the Management node, you need to follow this process: . Note: different OS in this case means, for example, RHEL 5 vs. RHEL 6. If the difference is just an update level/service pack (e.g. RHEL 6.0 vs. RHEL 6.3), then you can build it on the MN.
If the image you are building is for nodes that are the same OS and architecture as the management node (the most common case), then you can follow the instructions here to run genimage on the management node.
Run genimage to generate the image based on the mycomputeimage definition:
genimage mycomputeimage
Before you pack the image, you have the opportunity to change any files in the image that you want to, by cd'ing to the rootimgdir (e.g. /install/netboot/rhels6/x86_64/compute/rootimg). Although, instead, we recommend that you make all changes to the image via your postinstall script, so that it is repeatable.
The genimage command creates /etc/fstab in the image. If you want to, for example, limit the amount of space that can be used in /tmp and /var/tmp, you can add lines like the following to it (either by editing it by hand or via the postinstall script):
tmpfs /tmp tmpfs defaults,size=50m 0 2 tmpfs /var/tmp tmpfs defaults,size=50m 0 2
But probably an easier way to accomplish this is to create a postscript to be run when the node boots up with the following lines:
logger -t xcat "$0: BEGIN" mount -o remount,size=50m /tmp/ mount -o remount,size=50m /var/tmp/ logger -t xcat "$0: END"
Assuming you call this postscript settmpsize, you can add this to the list of postscripts that should be run for your compute nodes by:
chdef -t group compute -p postbootscripts=settmpsize
Now pack the image to create the ramdisk:
packimage mycomputeimage
The kerneldir attribute in linuximage table is used to
assign one directory to hold the new kernel to be installed into the
stateless/statelite image. Its default value is /install/kernels, you need to create the directory named
Assuming you have the kernel in RPM format in /tmp, the value of kerneldir is not set (which will take the default value: /install/kernels).
This procedure assumes you are using xCAT 2.6.1 or later. The rpm names are an example and you can substitute your level and architecture. The kernel will be installed directly from the rpm package.
The kernel RPM package is usually named kernel-
cp /tmp/kernel-2.6.32.10-0.5.x86_64.rpm /install/kernels/ createrepo /install/kernels/
Usually, the kernel files for SLES are separated into two parts: kernel-
kernel-ppc64-base-2.6.27.19-5.1.x86_64.rpm kernel-ppc64-2.6.27.19-5.1.x86_64.rpm
2.6.27.19-5.1.x86_64 is NOT the kernel version. 2.6.27.19-5-x86_64 is the kernel version . Follow this naming convention to determine the kernel version.
After the kernel version is determined for SLES, then:
cp /tmp/kernel-ppc64-base-2.6.27.19-5.1.x86_64.rpm /install/kernels/ cp /tmp/kernel-ppc64-2.6.27.19-5.1.x86_64.rpm /install/kernels/
Run genimage/packimage to update the image with the new kernel: (Use sles as example)
Since the kernel version is different from the rpm package version, the -g flag needs to be specified on the genimage command for the rpm version of kernel packages.
genimage -i eth0 -n ibmveth -o sles11.1 -p compute -k 2.6.27.19-5-x86_64 -g 2.6.27.19-5.1
The kernel drivers in the stateless initrd are used for the devices during the netboot. If you are missing one or more kernel drivers for specific devices (especially for the network device), the netboot process will fail. xCAT offers two approaches to add additional drivers to the stateless initrd during the running of genimage.
genimage-n
Generally, the genimage command has a default driver list which will
be added to the initrd. But if you specify the '-n' flag, the default
driver list will be replaced with your
The default driver list:
rh-x86: tg3 bnx2 bnx2x e1000 e1000e igb mlx_en virtio_net be2net rh-ppc: e1000 e1000e igb ibmveth ehea sles-x86: tg3 bnx2 bnx2x e1000 e1000e igb mlx_en be2net sels-ppc: tg3 e1000 e1000e igb ibmveth ehea be2net
Note: With this approach, xCAT will search for the drivers in the rootimage. You need to make sure the drivers have been included in the rootimage before generating the initrd. You can install the drivers manually in an existing rootimage (using chroot) and run genimage again, or you can use a postinstall script to install drivers to the rootimage during your initial genimage run.
Refer to the doc .
nodeset compute osimage=mycomputeimage
(If you need to update your diskless image sometime later, change your osimage attributes and the files they point to accordingly, and then rerun genimage, packimage, nodeset, and boot the nodes.)
Now boot your nodes...
rsetboot compute net rpower compute boot
This section gives some examples of using key commands and command combinations in useful ways. For any xCAT command, typing 'man
In this configuration, a handy convenience group would be the lower systems in the chassis, the ones able to read temperature and fanspeed. In this case, the odd systems would be on the bottom, so to do this with a regular expression:
# nodech '/n.*[13579]$' groups,=bottom
or explicitly
chdef -p n1-n9,n11-n19,n21-n29,n31-n39,n41-n49,n51-n59,n61-n69,n71-79,n81-n89, n91-n99,n101-n109,n111-119,n121-n129,n131-139,n141-n149,n151-n159,n161-n167 groups="bottom"
We can list discovered and expanded versions of attributes (Actual vpd should appear instead of *) :
# nodels n97 nodepos.rack nodepos.u vpd.serial vpd.mtm n97: nodepos.u: A-13 n97: nodepos.rack: 2 n97: vpd.serial: ******** n97: vpd.mtm: *******
You can also list all the attributes:
#lsdef n97 Object name: n97 arch=x86_64 . groups=bottom,ipmi,idataplex,42perswitch,compute,all . . . rack=1 unit=A1
xCAT provides parallel commands and the sinv (inventory) command, to analyze the consistency of the cluster. See
Combining the use of in-band and out-of-band utilities with the xcoll utility, it is possible to quickly analyze the level and consistency of firmware across the servers:
mgt# rinv n1-n3 mprom|xcoll ==================================== n1,n2,n3 ==================================== BMC Firmware: 1.18
The BMC does not have the BIOS version, so to do the same for that, use psh:
mgt# psh n1-n3 dmidecode|grep "BIOS Information" -A4|grep Version|xcoll ==================================== n1,n2,n3 ==================================== Version: I1E123A
To update the firmware on your nodes, see .
To do this, see .
xCAT has several utilities to help manage and monitor the Mellanox IB network. See .
If the configuration is louder than expected (iDataplex chassis should nominally have a fairly modest noise impact), find the nodes with elevated fanspeed:
# rvitals bottom fanspeed|sort -k 4|tail -n 3 n3: PSU FAN3: 2160 RPM n3: PSU FAN4: 2240 RPM n3: PSU FAN1: 2320 RPM
In this example, the fanspeeds are pretty typical. If fan speeds are
elevated, there may be a thermal issue. In a dx340 system, if near
10,000 RPM, there is probably either a defective sensor or misprogrammed
power supply.
To find the warmest detected temperatures in a configuration:
# rvitals bottom temp|grep Domain|sort -t: -k 3|tail -n 3 n3: Domain B Therm 1: 46 C (115 F) n7: Domain A Therm 1: 47 C (117 F) n3: Domain A Therm 1: 49 C (120 F)
Change tail to head in the above examples to seek the slowest fans/lowest temperatures. Currently, an iDataplex chassis without a planar tray in the top position will report '0 C' for Domain B temperatures.
For more options, see rvitals manpage: