FF32B00 Administration Operation V1-0
FF32B00 Administration Operation V1-0
User Guide
User Guide
Bestell-Nr./ Order No.:
U41608-J-Z145-1-76
*U41608-J-Z145-1-76*
U41608-J-Z145-1-76
FlexFrame™ for mySAP™ Business Suite
Version 3.2
All rights, including rights of translation, reproduction by printing, copying or similar methods,
in part or in whole, are reserved.
Offenders will be liable for damages.
All rights, including rights created by patent grant or registration of a utility model or design,
are reserved.
Delivery subject to availability. Right of technical modification reserved.
Contents
1 Introduction ..................................................................................................... 1
1.1 Requirements .................................................................................................... 1
1.2 Notational Conventions ..................................................................................... 1
1.3 Document History.............................................................................................. 2
1.4 Related Documents........................................................................................... 2
1.5 Special Hints for FlexFrame 3.2 ........................................................................ 2
17 Glossary....................................................................................................... 297
18 Index............................................................................................................. 303
1.1 Requirements
This document addresses administrators on FlexFrame environments. We assume that
the reader of this document has technical background knowledge in the areas of
operating systems (Linux®, Solaris™), IP networking and SAP® basis.
2.2 Hardware
A typical FlexFrame environment consists of the following hardware types:
Add-on FlexFrame
1 4
Control
2
Network
3
5
1. Control Nodes: PRIMERGY™ RX300 S2 with SuSE Linux Enterprise Server 8 (on
local disks).
2. Two or more identical network switches Cisco Catalyst 3750G (per switch group)
3. Network Attached Storage with one or more Network Appliance Filer heads, each
with a disk storage of at least 2*14 disks with 144 GB each hosting shared OS file
systems and application data.
4. Intel® based PRIMERGY servers (standard rack servers or blade servers) with SuSE
Linux Enterprise Server 8 (shared OS).
®
5. SPARC64 V based PRIMEPOWER servers with Solaris 8 and 9 (Shared OS).
2.3 Software
The FlexFrame solution consists of both, hardware and software. To grant proper
function of the whole landscape, the entire software set is strictly defined. Anything other
than the software components listed below is not part of FlexFrame. This applies
unchanged if software from the list below is missing, is installed in other versions than
below, or if software other that the actual SAP components is added.
PR IMER G Y FSC2
RX800
The network boot concept used for Solaris-based PRIMEPOWER servers is derived from
the Solaris Diskless Client Concept. A shared /usr area is mounted read-only. Each
server has its own root file system which is mounted read-write.
In Step (1) the PRIMEPOWER™ server sends an RARP (Reverse Address Resolution
Protocol) request into the Storage LAN network segment. The Control Node with the
active in.rarpd process is receiving this request. The request asks for an IP address
based on the MAC address of the initiating NIC (Network Interface Card).
The in.rarpd process on the Control Node will search the file /etc/ethers for the
MAC address provided. If it finds a line with the MAC address, it will return the associated
IP address as an RARP response (2).
Now the PRIMEPOWER server can setup the NIC with the IP address. Next step (3) is to
get a file via the TFTP (Trivial File Transfer Protocol) from the Control Node which
responded to the RARP request. The file name is build by the hex notation of the IP
address.
The in.tftpd of the Control Node will send this file to the PRIMERPOWER server. This
file contains the first step of the Solaris kernel for booting via the network.
Now additional parameters are required. The Solaris kernel sends a bootparam request
to the Control Node (5). The rpc.bootparamd process on the Control Node will read
the file /etc/bootparams for the given Application Node’s host name and replies with
the detail information about its NFS server and path to the root file system along with
mount options and other parameters.
Synopsis
Command Options
--op add-sw
Adds a new member to the switch group.
--group <switch_group_id>
Defines the switch group to be used.
--type <switch_type>
Defines the type of the new member to be added to the switch group. Currently the
type has to be cat3750g-24t or cat3750g-24ts.
Command Output
The command displays some information about processing steps. The output may look
like this:
cn1:/opt/FlexFrame/bin # ff_swgroup_adm.pl --op add-sw --group 1
--type cat3750g-24ts
To add the switch to switch stack (switch group 1) write down the
current switch ids as they may change inserting the new switch to
stack. To connect the switch to the stack use the provided
stacking cable and stacking ports at rear side. See Cisco
installation manual for details.
Synopsis
Command Options
--op rem-sw
Removes the last member from a switch group.
--group <switch_group_id>
Defines the switch group to be used.
--switch <switch_id>
Defines the stack ID of the switch to be removed from a switch group.
Command Output
The command displays some information about processing steps. The output may look
like this:
cn1:/opt/FlexFrame/bin # ff_swgroup_adm.pl --op rem-sw --group 1
--switch 3
Switch was successfully removed from LDAP data.
Synopsis
Command Options
--op list
Displays switch group configuration.
--group <switch_group_id>
Defines the switch group to be used.
Command Output
The output may look like this:
cn1:/opt/FlexFrame/bin # ff_swgroup_adm.pl --op list --group 1
Switch Group 1
Switch port usage: (switch id, used, free tx, free fx ports)
1 21 used 3 free tx 4 free fx
2 21 used 3 free tx 4 free fx
Switch port list: (switch id, port id, connected device, vlans)
1 1 pw250-10 t10,t12,u11
1 2 pw250-11 t10,t12,u11
1 3 swb-1-1/11 t1,t10,t11,t12,t13
1 4 swb-1-2/11 t1,t10,t11,t12,t13
1 5 pw650-20 t10,t12,u11
1 6 pw900-211 t10,t12,u11
1 7 rx300-13 t10,t12,u11
1 8 rx300-14 t10,t12,u11
1 9 pw250-12 t10,t12,u11
1 10 pw250-13 t10,t12,u11
1 11 cn1 t10,t11,t12,u13
1 12 cn2 t10,t11,t12,u13
1 13 filer-ip1 t11
1 14 extern. Connect t10,t11,t12
1 15 --- unused ---
1 16 --- unused ---
1 17 --- unused ---
1 18 cn1 mgmt u13
1 19 pw250-12 mgmt u13
1 20 rx300-13 mgmt u13
1 21 - u13
1 22 pw650-20 mgmt u13
1 23 pw250-10 mgmt u13
1 24 Corporate LAN u10
2 1 pw250-10 t10,t12,u11
2 2 pw250-11 t10,t12,u11
2 3 swb-1-1/12 t1,t10,t11,t12,t13
2 4 swb-1-2/12 t1,t10,t11,t12,t13
2 5 pw650-20 t10,t12,u11
2 6 pw900-211 t10,t12,u11
2 7 rx300-13 t10,t12,u11
2 8 rx300-14 t10,t12,u11
2 9 pw250-12 t10,t12,u11
2 10 pw250-13 t10,t12,u11
2 11 cn1 t10,t11,t12,u13
2 12 cn2 t10,t11,t12,u13
2 13 filer-ip1 t11
2 14 extern. Connect t10,t11,t12
2 15 --- unused ---
2 16 --- unused ---
2 17 --- unused ---
2 18 cn2 mgmt u13
2 19 pw250-13 mgmt u13
2 20 rx300-14 mgmt u13
2 21 - u13
2 22 - u13
2 23 pw250-11 mgmt u13
2 24 Corporate LAN u10
Synopsis
Command Options
--op pass
Changes the switch group access password.
--group <switch_group_id>
Defines the switch group to be used.
--passwd <password>
Defines the new password as clear text.
Command Output
The output may look like this:
cn1:/opt/FlexFrame/bin # ff_swgroup_adm.pl --op pass --group 1
--passwd passwort
update switch 1/1 configuration
Notice: Update will take about 1 minute.
............+
Synopsis
Command Options
--op name
Changes the switch group host name.
--group <switch_group_id>
Defines the switch group to be used.
--name <name>
Defines the new host name to be used.
Command Output
The output may look like this:
cn1:/opt/FlexFrame/bin # ff_swgroup_adm.pl --op name --group 1
--name swg1
update switch 1/1 configuration
Notice: Update will take about 1 minute.
...+
Synopsis
Command Options
--op add
Adds a switch port configuration.
--port <swgroup:switch:port>
Defines the switch group, switch and port ID of the port to be used.
--lan <pool:lan[:lan][,pool:lan[:lan]]>
Defines the accessible VLANs. For better readability, a VLAN is specified with its
pool and LAN name. Use only client, server or storage as LAN names. For more than
one LAN per pool the LAN names may be added to the same pool statement. The
VLANs are not restricted to belong to the same pool. To directly add VLAN IDs not
used within any pool, use '#' as pool name and the VLAN ID(s) as LAN(s).
The accessible VLANs. For better readability, a VLAN is specified with its pool and
LAN name. Use only client, server or storage as LAN names. For more than one LAN
per pool the LAN names may be added to the same pool statement. The VLANs are
not restricted to belong to the same pool. To directly add VLAN IDs not used within
any pool, use '#' as pool name and the VLAN ID(s) as LAN(s).
If only a single VLAN is configured this is accessible as native VLAN. This means the
data packet contains no VLAN tag. This is the behavior used by a standard server
network interface. For more than one LAN they are configured as tagged. To define
which of them should be used as native VLAN use the option --native.
Examples:
--lan poolA:client:server,poolB:client:server
--lan poolA:client,poolA:server,poolB:client:server
--lan poolA:storage,poolB:storage
--lan poolA:server
--lan '#:417:891'
--lan poolA:server,'#:417:891'
--lan 'poolA:server,#:417:891'
--native <pool:lan>
Use this option to define the native VLAN of the accessible VLANs defined with
option --lan. To directly add VLAN ID not used within any pool, use '#' as pool
name and the VLAN ID as LAN.
Examples:
--native poolA:server
--native '#:417'
--desc <description>
The description is added to configuration of switch port and the LDAP data of the
switch port configuration.
Command Output
The command displays some information about processing steps. The output may look
like this:
cn1:/opt/FlexFrame/bin # ff_swport_adm.pl --op add --port 1:1:15
--lan ip1:storage:server,’#:4000’ --native ip1:storage
Synopsis
Command Options
--op rem
Removes the configuration of a switch port.
--port <swgroup:switch:port>
Defines the switch group, switch and port ID of the port to be used.
Command Output
The command displays some information about processing steps. The output may look
like this:
cn1:/opt/FlexFrame/bin # ff_swport_adm.pl --op rem --port 1:1:15
Synopsis
Command Options
--op list
Displays configuration of the switch port.
--port <swgroup:switch:port>
Defines the switch group, switch and port ID of the port to be used.
Command Output
The output may look like this:
cn1:/opt/FlexFrame/bin # ff_swport_adm.pl --op list --port 1:1:4
native VLAN: 26
Port Peer Type: AN
Peer Node: rx300-1
Synopsis
Command Options
--op add
Adds a Filer.
--name <node_name>
Defines the node name of Filer.
--type <filer_type>
Defines the type of the Filer. See usage for a list of known Filer types.
--swgroup <switch_group_id>
Defines the switch group the Filer should be added to. See usage for a list of
configured switch group IDs.
--host <ip_host_part>
Defines the host part to be used to build IP addresses for the Control or Storage LAN
networks. If this option is omitted the script uses a free host number to calculate the
IP address.
--ports <port_count>
Defines the count of ports to be used with pool Storage LAN networks. If this option
is omitted the script uses the default of two ports.
Command Output
The command displays some information about processing steps. The output may look
like this:
cn1:/opt/FlexFrame/bin # ff_filer_adm.pl --op add --name filer2
--type FAS270 --swgroup 1
update LDAP
.....
update switch 1/1 configuration
Notice: Update will take about 1 minute.
................................+
Some manual interventions are nescessary to integrate the filer
into FlexFrame environment.
The following list of actions have to be performed in order to
integrate the filer into your FlexFrame landscape. Since your
exact configuration may vary, these steps have to be performed
manually. However, the VIF must be named 'storage'.
hostname filer2
vif create multi storage <-- add your NICs here, eg. e0a e0b
vlan create storage 13
ifconfig storage-13 192.168.20.4 netmask 255.255.255.0 broadcast
192.168.20.255 mtusize 1500 -wins up
options dns.enable off
options nis.enable off
savecore
/vol/vol0 -sec=sys,rw=192.168.20.2:192.168.20.1,anon=0
192.168.20.2 root
192.168.20.1 root
Synopsis
Command Option
--op list-all
Displays all configured Filers.
Command Output
The command displays some information about the Filer configuration. The output may
look like this:
cn1:/opt/FlexFrame/bin # ff_filer_adm.pl --op list-all
Filer configurations
filer
Control Lan
192.168.20.3 filer-co
Type: FAS940
Switch Link Aggregation
Port Count: 2
Link Aggr.ID: 5
Storage LAN switch ports
1 / 1 / 13 SwGroup / Switch / Port
1 / 2 / 13 SwGroup / Switch / Port
Control LAN switch ports
1 / 2 / 15 SwGroup / Switch / Port
Pools
pool1 192.168.2.131 filer-pool1-st master
pool2 10.3.1.3 filer-pool2-st master
pool5 192.168.40.3 filer-pool5 master
usr 192.168.5.3 filer-usr-st master
Synopsis
Command Options
--op list
Displays the configuration of a switch port.
--name <node_name>
Defines the node name of a Filer.
Command Output
The output may look like this:
cn1:/opt/FlexFrame/bin # ff_filer_adm.pl --op list --name filer
Filer configurations
filer
Control Lan
192.168.20.3 filer-co
Type: FAS940
Switch Link Aggregation
Port Count: 2
Link Aggr.ID: 5
Synopsis
Command Options
--op add-pool
Adds the given pool to named Filer.
--name <node_name>
Defines the node name of Filer.
--pool <pool_name>
Defines the pool name the Filer has to be support. See usage for a list of configured
pools.
--role {master|slave}
Defines the role of Filer within given pool. This information is used by the
ff_setup_sid_folder program. If this option is omitted the script uses the default
role master.
--host <ip_host_part>
Defines the host part to be used to build IP addresses for the Control or Storage LAN
networks. If this option is omitted, the script uses a free host number to calculate the
IP address.
Command Output
The output may look like this:
cn1:/opt/FlexFrame/bin # ff_filer_adm.pl --op add-pool
--name filer --pool pool4
update LDAP
....
update switch 1/1 configuration
Notice: Update will take about 1 minute.
...........+
vlan: storage-25 has been created
Pool pool4 successfully added to filer, LDAP and network.
Synopsis
Command Options
--op rem-pool
Removes the given pool from the named Filer.
--name <node_name>
Defines the node name of the Filer.
--pool <pool_name>
Defines the pool name the Filer has to be support. See usage for a list of configured
pools.
Command Output
The output may look like this:
cn1:/opt/FlexFrame/bin # ff_filer_adm.pl --op rem-pool
--name filer --pool pool4
update switch 1/1 configuration
Notice: Update will take about 1 minute.
...........+
update LDAP
....
Pool pool4 successfully removed from filer, LDAP and network.
Synopsis
Command Options
--op rem
Removes a Filer from FlexFrame landscape.
--name <node_name>
Defines the node name of the Filer.
Command Output
The command displays some information about processing steps. The output may look
like this:
cn1:/opt/FlexFrame/bin # ff_filer_adm.pl --op rem --name filer2
update switch 1/1 configuration
Notice: Update will take about 1 minute.
.............................+
update LDAP
.....
Filer successfully removed from network and LDAP.
2.6 LDAP
A directory service is typically a specialized database which is optimized for reading and
searching hierarchical structures. Unlike RDBMS, these databases do not support
transaction based processing. LDAP (Lightweight Directory Access Protocol, see RFC
2251) defines means of setting and retrieving information in such directories. The
structure of the LDAP data is called Domain Information Tree (DIT).
In case (only) the replica fails, the remaining master LDAP DB continues to work.
In case the master DB fails, the remaining replica LDAP DB will provide the functionality.
However, if the LDAP DB needs to be changed, the replica needs to act as master (i.e.
become read-writable).
This requires manual interaction. The procedure is described in the document "Disaster
Procedures for FlexFrame Landscapes", slide “FlexFrame Control Node (LDAP master)".
2.7 PRIMECLUSTER
This section briefly describes how PRIMECLUSTER is implemented in FlexFrame.
PRIMECLUSTER is used for high availability of Control Nodes only, not for Application
Nodes.
All these services are monitored by PRIMECLUSTER and automatically restarted, if they
fail. If a recovery is not possible, the application will be set into faulted state and is tried to
start on the other Control Node. This process is called “failover”.
Individual userApplications may be switched to the other Control Node, i.e. it is possible
to run ff_manage on the first and netboot_srv on the second Control Node.
PRIMECLUSTER uses the scripts in /etc/PRIMECLUSTER/bin for starting, stopping
and monitoring services. These scripts are delivered with the ffPCL RPM.
If you try to stop running services manually, RMS will restart them. To prevent
this, you should set the corresponding userApplication into maintenance mode.
See the command overview table below and section “PRIMECLUSTER
Administration” on page 66.
You may damage your configuration if you ignore this!
Control Control
Node 1 Node 2
optional
dhcpd mysql
start
stop
inetd ServerView
myAMC Ctrl
apache
Action Command
display state of all resources hvdisp -a
display and continuously update hvdisp -u
state of all resources
display state of all resources of type hvdisp [-u] -T <resource_type>
<resource type>
(-u = update)
start or switch over hvswitch [-f] <userApplication>
<userApplication> [<node>]
(-f = forced switch)
Action Command
stop <userApplication> hvutil -f <userApplication>
clear faulted <userApplication> hvutil -c <userApplication>
set <userApplication> into hvutil -m on <userApplication>
maintenance mode
set all userApplications into hvutil -M on
maintenance mode
exit <userApplication> hvutil -m [force]off
maintenance mode <userApplication>
(forced exit)
exit cluster maintenance mode hvutil -M [force]off
(forced exit)
stop RMS on all nodes hvshut -a
(do offline processing)
stop RMS on all nodes hvshut -A
(don’t do offline processing)
forced/emergency stop of RMS on hvshut -f
local node
(don’t do offline processing)
stop RMS on the local node hvshut -l
(do offline processing)
forces stop of RMS on local node hvshut -L
(don’t do offline processing)
start RMS on local node hvcm
verify shutdown facility status sdtool -s
reinitialize shutdown facility sdtool -r
verify CF status cftool -n
stop PRIMECLUSTER CF including /opt/SMAW/SMAWcf/dep/master unload
all dependent components
/etc/init.d/cf stop
Action Command
start PRIMECLUSTER CF including /etc/init.d/cf start
all dependent components
/opt/SMAW/SMAWcf/dep/master load
Test the IPMI Login (the ipmipower ipmipower -s <IP_address> -u OEM
tool can be found on the ServerView -p <password>
CD)
Display the Cluster Foundation cfconfig -n
Nodes and their current status
Display all LAN interfaces that were cfconfig -d
recognized by the OS
(Solaris/Linux)
Some commands (hvswitch, hvutil, etc.) return without showing any result.
Verify the result using hvdisp -a or hvdisp -u and/or check the logfiles
(especially switchlog, see below).
The suffix RMS of the target node is not necessary for most commands.
The target node is optional. If not specified, the application will be switched to the
preferred node, therefore in this case you could have also used:
control1:~ # hvswitch netboot_srv
As you can see after a few seconds, netboot_srv has been switched over to the first
Control Node.
The command hvswitch will also be used to start an application that is offline on both
nodes. If nothing happens after issuing the hvswitch command, see switchlog and
netboot_srv.log for the cause (see below).
Remember that some applications (especially ff_manage) can take minutes to stop and
start.
For information how to access the administration interface of PRIMECLUSTER,
please refer to section “PRIMECLUSTER Administration” on page 66.
The program will stop if a service is in maintenance mode. Since this indicates an
operator intervention, it must be fixed manually, e.g. by calling:
hvutil -m off <service>
2.8 Network
The network is the backbone of the FlexFrame solution. Communication between the
various nodes and storage devices is done exclusively via the IP network infrastructure. It
serves both, communication between servers and clients as well as delivering IO data
blocks between the NAS (Network Attached Storage) and the servers.
The IP network infrastructure is essential for every FlexFrame configuration. FlexFrame is
designed with a dedicated network for connections between servers and storage that is
reserved for FlexFrame traffic only. One network segment, the Client LAN (see below)
can be routed outside the FlexFrame network to connect to the existing network.
2.8.2 Segments
FlexFrame 3.2 introduces a new network concept, providing higher availability as well as
increased flexibility in virtualizing the whole FlexFrame landscape.
This concept relies on VLAN technology that allows running multiple virtual networks
across a single physical network. Additionally, in order to ensure high network availability,
LAN bonding is used on every node. This includes a double switch and wiring
infrastructure, to keep the whole environment working, even when a network switch or
cable should fail. Additionally, the networks segements are prioritized, to ensure that
important connections are prefered.
Similar to former versions, there are four virtual networks within FlexFrame. The
difference of FlexFrame 3.2 is that these networks are run through one logical redundant
NIC, using bonding on Linux and IPMP on Solaris.
The following figure outlines the basic network segments of a typical FlexFrame
landscape with Linux-based Application Nodes.
e0 e4a e4b
vif storage
VLANs:
Control-LAN
Storage-LAN
Server-LAN
Client-LAN
bond0 bond0 bond0
Mgmt.
RSB * eth0 eth1 IPMI * eth1 eth2 eth0 eth1 Blade
Clients IPMP
Application Nodes
● Server LAN
The Server LAN segment is used for the communication between SAP instances
among each other and the databases.
● Storage LAN
The Storage LAN segment is dedicated to NFS communication for accessing the
executables of SAP and the RDBMS as well as the IO of the database content and
SAP instances.
The network bandwidth has to be one Gigabit for all components, since intense storage
traffic has to be handled.
dn: ou=auto_FlexFrame,ou=Automount,ou=pool2,ou=Pools,ou=FlexFrame,
dc=flexframe,dc=wdf,dc=fujitsu-siemens,dc=com
objectClass: top
objectClass: automountMap
ou: auto_FlexFrame
...
dn: cn=/FlexFrame,ou=auto_master,ou=Automount,ou=pool1,ou=Pools,
ou=FlexFrame,dc=flexframe,dc=wdf,dc=fujitsu-siemens,dc=com
objectClass: top
objectClass: automount
cn: /FlexFrame
automountInformation: auto_FlexFrame
dn: cn=myAMC,ou=auto_FlexFrame,ou=Automount,ou=pool1,ou=Pools,
ou=FlexFrame,dc=flexframe,dc=wdf,dc=fujitsu-siemens,dc=com
objectClass: top
objectClass: automount
cn: myAMC
automountInformation:
-rw,nointr,hard,rsize=32768,wsize=32768,proto=tcp,nolock,
vers=3 filpool1-st:/vol/volFF/pool-pool1/pooldata/&
an_linux:~ #
mount filpool1-st:/vol/volFF/pool-pool1/pooldata/myAMC/mnt
If you get an error message like Permission denied, check the exports on the Filer
and the existence of the directory myAMC/ itself.
Other entries in LDAP make use of variables like ${OSNAME} which is either Linux or
SunOS:
dn: cn=/,ou=auto.oracle,ou=Automount,ou=pool2,ou=Pools,
ou=FlexFrame,dc=flexframe,dc=wdf,dc=fujitsu-siemens,dc=com
objectClass: top
objectClass: automount
cn: /
description: catch-all for Linux automount
automountInformation:
-rw,nointr,hard,rsize=32768,wsize=32768,proto=tcp,nolock,
vers=3 filpool2-st:/vol/volFF/pool-pool2/oracle/${OSNAME}/&
On Linux, the automount mount points can be read using the following command:
an_linux:~ # mount
rootfs on / type rootfs (rw)
/dev/root on / type nfs (ro,v3,rsize=32768,wsize=32768,reserved,
hard,intr,tcp,nolock,addr=192.168.11.206)
192.168.11.206:/vol/volFF/os/Linux/FSC_3.2B00-000.SLES-
9.X86_64/var_img/var-c0a80b36
on /var type nfs (rw,v3,rsize=32768,wsize=32768,reserved,
hard,intr,tcp,nolock,addr=192.168.11.206)
192.168.11.206:/vol/volFF/os/Linux/FSC_3.2B00-000.SLES-
9.X86_64/var_img/var-c0a80b36
/dev on /dev type nfs (rw,v3,rsize=32768,wsize=32768,reserved,
hard,intr,tcp,nolock,addr=192.168.11.206)
192.168.11.206:/vol/volFF/os/Linux/pool_img/pool-c0a80bff on
/pool_img type nfs (rw,v3,rsize=32768,wsize=32768,reserved,
hard,intr,tcp,nolock,addr=192.168.11.206)
proc on /proc type proc (rw)
devpts on /dev/pts type devpts (rw)
shmfs on /dev/shm type shm (rw)
/dev/ram on /var/agentx type ext2 (rw)
automount(pid1750) on /FlexFrame type autofs (rw)
automount(pid1772) on /saplog/mirrlogA type autofs (rw)
automount(pid1752) on /home_sap type autofs (rw)
dn: cn=/home_sap,ou=auto_master,ou=Automount,ou=pool1,ou=Pools,
ou=FlexFrame,dc=flexframe,dc=wdf,dc=fujitsu-siemens,dc=com
objectClass: top
objectClass: automount
cn: /home_sap
automountInformation: auto_home_sap
dn: cn=/myAMC,ou=auto_master,ou=Automount,ou=pool1,ou=Pools,
ou=FlexFrame,dc=flexframe,dc=wdf,dc=fujitsu-siemens,dc=com
objectClass: top
objectClass: automount
cn: /myAMC
automountInformation: auto_myAMC
dn: cn=/sapdata/sapdata1,ou=auto_master,ou=Automount,ou=pool1,
ou=Pools,ou=FlexFrame,dc=flexframe,dc=wdf,dc=fujitsu-siemens,
dc=com
objectClass: top
objectClass: automount
cn: /sapdata/sapdata1
automountInformation: auto_sapdata1
...
2.9.3 Snapshots
When a snapshot is taken, no data blocks are being copied. Just the information where
the data blocks are located is saved. If a data block is modified, it is written to a new
location, while the content of the original data block is preserved (also known as “copy on
write”). Therefore, the creation of a snapshot is done very quickly, since only few data
have to be copied. Besides that, the snapshot functionality provided by NetApp is unique,
because the usage of snapshots does not decrease the throughput and performance of
the storage system.
Snapshot functionality will allow the administrator to create up to 250 backup-views of a
volume. The functionality “SnapRestore” provided by NetApp significantly reduces the
time to restore any of the copies if required. Snapshots will be named and can be re-
named and deleted. Nested snapshots can be used to create e.g. hourly and daily
backups of all databases. In a FlexFrame landscape, a single backup server is sufficient
to create tape backups of all volumes. Even a server-less backup can be implemented.
Off-Line backups require a minimal down-time to the database because the backup to
tape can be done reading form a quickly taken snapshot.
Before an Application Nodes can be booted, the NTP server must be set up correctly.
This can be verified by runing the following command.
control1:~ # ntpq -p
remote refid st t when poll reach delay offset jitter
=================================================================
*LOCAL(0) LOCAL(0) 5 l 45 64 377 0.000 0.000 0.004
control2-se control1-se 7 u 248 1024 377 2.287 2.587 0.815
This command needs to be repeated until an asterisk (*) is displayed at the begining of
one of the data lines. This character indicates that the NTP server is now ready. If you do
not wait and continue with booting, the Application Nodes may work with a different time
than the Control Nodes and may (among other possible side effects) create files which
may have wrong time stamp information.
If this sequence is not used and all servers are powered on at the same time,
the Application Nodes will try to boot while the Control Nodes are not ready to
receive the Application Node’s boot request. If this is the case, manual
intervention is required to re-initiate the boot process of the Application Nodes.
Before shutting down all SAP and DB services, check if users, batch jobs, print
jobs and RFC connections to other SAP systems have finished working.
To shutdown the SAP and DB services, use the following command for each pool:
control1:~ # stop_all_sapservices <pool_name>
Before you can stop the Filer, you need to stop all processes on the Control Nodes. Since
those processes are under control of PRIMECLUSTER you can use a single command:
control1:~ # hvshut -a
To send the halt command to the Filer, use the following command:
control1:~ # rsh <filer_name> halt
There’s no explicit power-off sequence for the switches. Assuming there are no
other devices connected to the switches, they may simply be plugged-off after
all components were powered down.
If you do not send an explicit “halt” command to the Filer, the backup-battery of
the Filer maybe drained since the Filer assumes a power-loss and tries to
preserve the contents of its NVRAM. If the Filer is powered-off for too long the
result can be a loss of NVRAM data and a long waiting period during next
startup of the Filer.
4.1 Networks
The network of FlexFrame is its backbone. Here are some tips to get an overview of the
current situation on the various networks:
To double-check the network addresses, their names and pool assignment you can use
the getent command:
The loopback network is local for each host and always has the IP address 127.0.0.0.
The control network is the Control LAN network segment for the complete FlexFrame
landscape.
In the example we have configured two pools called pool1 and pool2. For each pool
there are the three dedicated and distinct segments storage, server and client. The
building rule of the network name is <segment>_<pool name>.
On the Control Nodes you can see the relation of each pool specific segment to its
interface using the netstat -r command like this:
control1:~ # netstat -r
Kernel IP routing table
Destination Gateway Genmask Flags MSS Window
irtt Iface
192.168.100.0 * 255.255.255.252 U 0 0
0 cip0
server_pool2 * 255.255.255.0 U 0 0
0 vlan42
control * 255.255.255.0 U 0 0
0 bond0
control * 255.255.255.0 U 0 0
0 eth1
control * 255.255.255.0 U 0 0
0 eth2
server_pool1 * 255.255.255.0 U 0 0
0 vlan32
storage_pool2 * 255.255.255.0 U 0 0
0 vlan41
client_pool1 * 255.255.255.0 U 0 0
0 vlan30
storage_pool1 * 255.255.255.0 U 0 0
0 vlan31
client_pool2 * 255.255.255.0 U 0 0
0 vlan40
default gw216p1 0.0.0.0 UG 0 0
0 vlan30
The cip0 interface is used for PRIMECLUSTER communication between the two Control
Nodes, this network does not have a name associated with it.
Here you can quickly see that the Server LAN segment of pool2 (server_pool2) is
using the VLAN ID 42 on interface vlan42.
Note that the control Control LAN segment is shown on the three interfaces bond0,
eth1 and eth2) which is because bond0 is the combination of eth1 and eth2 grouped
together for redundancy purposes.
Each line represents a request to a certain IP address. Since Application Nodes have
multiple IP addresses they are queried multiple times, which is required to detect wrong
configurations.
The Application Node an_0900 was “Not reached“ means that this IP address (read
from LDAP database) did not respond to an ICMP request. A possible cause may be that
this server is powered off.
The message No SNMP info. is printed if the IP address is available but did not respond
to the SNMP request. A possible cause is that the server may not be part of the
FlexFrame landscape. In our example it is a gateway which does not have an SNMP-
daemon running.
There are three IP addresses listed which do not have a host name associated with,
therefore the names in brackets are missing. In our case this is caused by the test
addresses for Solaris Application Nodes (IPMP).
After the last server was probed with SNMP, we scan for available switches.
Now each Control Node or Application Node is listed. There are different types of
components which produce slightly different output.
Here is a sample of a Control Node:
In the header of the block for this Control Node you see its name and the OS as cn which
denotes a Control Node. The OS Version indicates that this Control Node is running
SuSE SLES-8. The Model FSC (Fujitsu Siemens Computers) is displayed if no detailed
information on the servers exact model name can be found.
Now, each network device is listed along with information on the associated IP address,
the MTU size (maximum transfer unit), and MAC address and to which switch and which
port the device is connected.
Abbreviation Explanation
n.a. not applicable. Since the device lo (loopback) is a logical
interface with no connection to the outside world, no information on
MAC, switch or port can be displayed.
!undef! Marks an undefined status. The device is not configured or in an
undefined status. Please check these interfaces. As for a Control
Node, the eth0 is used for IPMI and hence not used for the OS. The
devices sit0 and cip1 through cip7 are not used.
no ip There is no IP address associated whit this device.
not up This device or interface is not up.
Gi<s>/0/<p> This denotes a port on a switch. Gi is short for “GigabitEthernet” and
<s> is the switch-number of the stack of switches. <p> is the port
number where this particular port is connected to.
The example Gi1/0/3 tells us that the eth1 interface is connected
to port 3 of switch #1 of this stack of switches.
The bond0 interface is built of eth1 and eth2. Unfortunately, Linux does not
deliver the correct information in such a case. The bond0 interface is actually up
and using the IP address shown for eth1.
The name of the switch and the short notation of the port are shown if possible. The
information is derived from ARP caches which are very volatile. There maybe cases
where !undef! is listed, but a port is associated. Furthermore, only the active interface
of eth1 and eth2 is seen.
The blade servers are connected to switch blades. The related port is shown as
a PortChannel. This represents the “view” from the Cisco switch.
Here you can see if there is a certain blade server in a given slot. If so, its serial number
is listed.
Only one interface (eth0) of the two available management blades is active at a
given time.
PRIMEPOWER 250 and 450 models have a built-in XSCF which is listed after
the Application Node itself. Here, the XSCF is connected to port Gi1/0/20.
The following example shows some error messages. This error messages indicates that
one NIC of this Application Node is not connected to the port where it should be
(according to the LDAP database). If, e.g. two ports of two Application Nodes of the same
pool are mixed up, the functionality is just the same. However, if one Application Node
gets moved into a different pool, the ports are changed. This would lead to a malfunction
of both ports.
If this error message is displayed check the cabling and run ff_netscan.sh again.
This error message indicates that the IP address shown does not belong to the interface
where it is configured on. It has to be in a different VLAN (network segment).
Not all combinations can be verified this way.
This indicates that the two MAC addresses (MAC1 and MAC2) which are known to the
LDAP database are not found on this Application Node (<node name>). One possible
cause is that a NIC was replaced but the LDAP database was not updated using the
appropriate tool.
If the known NIC fails and the Application Node tries to boot again it will not get an IP
address and cannot boot.
The IP address ldap_an_ip could not be found on this Application Node. Maybe there
was a problem during configuration of the network interfaces.
At the end of the ff_netscan.sh output you should find the line:
Netscan finished.
For further details on options refer to the online manual page of ff_netscan.sh using
the following command:
control1:~ # man ff_netscan.sh
Description
This command sets the default router for a FlexFrame pool. LDAP entries for DHCP
configuration and pool information are adjusted. Each Solaris Application Node (if
installed) will get an /etc/defaultrouter file with the default router you provided in
the command line.
The default router must be a valid router (resolvable name or IP address) of any
of the available networks for all Application Nodes in the given pool (by pool
name). This tool will not check if the default router is reachable from the
Application Nodes.
Synopsis
Command Options
-d Debugging information will be logged to /tmp/ff_pool_defrt.DEBUGLOG.
-p <pool_name>
Name of the pool. Note: The pool name is case sensitive.
-r <defrouter>
Name or IP address for the default router for the given pool.
Debugging
/tmp/ff_pool_defrt.DEBUGLOG will hold debugging information. In case of problems
please provide this file.
Description
This command adds or removes a DNS server for a pool. LDAP entries for DHCP
configuration and pool information are adjusted. Each Solaris Application Node will get an
/etc/resolv.conf file with the DNS server and DNS domain name given in the
command line. Each Linux node will get an entry in
/FlexFrame/vollFF/pool-<pool name>/pooldata/config/etc/resolv.conf,
which is a symbolic link from the /etc/resolv.conf file of the Application Node.
Synopsis
Command Options
-d Debugging information will be logged to /tmp/ff_pool_dnssrv.DEBUGLOG.
-p <pool_name>
Name of the pool. Note: The pool name is case sensitive.
-r <dnssrvr>
IP address for the DNS serverof the given pool. If the IP address do not match any of
the local pool networks (Client VLAN, Storage VLAN or Server VLAN) and no default
router is installed, processing is aborted.
-u <action>
Possible actions are add or remove.
-n <domain_name>
DNS domain name of the given pool. If the domain name do not match the one given
in LDAP and the -f option is not used, processing is aborted.
-f Use force option to change the domain name when adding a DNS server.
Debugging
/tmp/ff_pool_dnssrv.DEBUGLOG will hold debugging information. In case of
problems please provide this file.
You will see an overview of all Web interfaces installed on the Control Nodes.
If you cannot connect to any Control Node, the http service might be running on
the other Control node – if not, check the PRIMECLUSTER configuration.
You can directly access the PRIMECLUSTER Administration interface from any
Control Node by accessing the following URL:
http://localhost:8081/Plugin.cgi
This even works if the Apache web server is not running and therefore the
FlexFrame web portal is down.
After successful login you will see the Web-Based Admin View tool:
To run the Cluster Admin JAVA applet, you have to accept the security certificate by
Fujitsu Siemens Computers. The following warning is displayed:
Select Yes or Always. Selecting Always will automatically accept the certificate the next
time you call this applet.
Next, you will be asked to which management server Cluster Admin should connect:
Select a server and click on Ok. For the PRIMECLUSTER management of FlexFrame it
does not matter which one you select.
This part of PRIMECLUSTER is the base for Cluster interconnectivity and node
monitoring. Both Control Nodes should be displayed in green colour.
In some cases, PRIMECLUSTER cannot detect the correct state of the partner node and
will display LEFTCLUSTER and/or DOWN. For example this will happen if you remove all
network cables to the remote node including the IPMI interface.
In case of a fault:
● Check the connectivity and restart CF by selecting Tools - Stop CF and
Tools - Start CF. If this does not work, reboot at least one Control Node.
● If a node is really down and cannot be rebooted, select
Tools - Mark Node Down from the pull down menu.
Do not try to change CF configuration and do not unconfigure CF unless
instructed to do so by Fujitsu Siemens Computers support!
Select the tab labeled rms at the bottom of the left panel to proceed to PRIMECLUSTER
RMS administration.
For a description of how PRIMECLUSTER is implemented in FlexFrame, refer to section
PRIMECLUSTER on page 29.
To clear this state, right click on ff_manage on the second Control Node and select
Offline or Clear fault.
You could also select Online in this case; PRIMECLUSTER would then shutdown
ff_manage on the first Control Node and start it on the second one (this may take some
time). The application will be coloured yellow (Wait state) until the switch-over was
finished.
Faulted applications (see examples 2 and 3) may occur when a sub application crashes
and cannot be restarted or when a sub application cannot be started during
PRIMECLUSTER startup.
Before clearing a fault, you should look into the logfile to find out the reason for the fault.
To view the logfile, right click on the faulted application and select View logfile.
Common causes for faults are misconfigured services, a broken network connection to
the Filer or a damaged mySQL database which is used by myAMC Messenger, to name
just a few examples.
To clear the fault state, right click on the application and select Clear fault to tell
PRIMECLUSTER that it should do offline processing on the application.
You have to switch the application online on one node to make sure that
FlexFrame will work correctly. This can be established by using the Online
(hvswitch) or Switch (hvswitch) entry of the application context menu.
5.4 ServerView S2
ServerView S2 allows you to query information of PRIMERGY servers in the FlexFrame
environment.
This tool is not needed for administration of FlexFrame but is required by Fujitsu Siemens
Computers support.
If you would like to monitor the hardware status of PRIMERGY servers in your
FlexFrame™ environment, click on ServerView and on the next page, click on the
button labeled start:
An overview of all configured servers (initially only the local host) will be shown on the
main screen:
To add more servers, run the Server Browser by either clicking on Administration /
Server Browser on the top navigation bar or right clicking anywhere in the list and
selecting New Server from the context menu.
The easiest way to add servers is to scan a subnet for each pool, for example the
“Server” subnet for Application Nodes and the “Control” subnet for management blades.
To scan a subnet, enter the first three bytes of the network address in the Subnet input
field on the bottom left of the screen and click on Start Browsing:
The browsing process is finished when the button Stop Browsing changes its caption to
Start Browsing.
Please do not interrupt the browsing process as it may hang the JAVA applet or
the browsing processes. If this happens, the Server Browser or even ServerView
S2 must be restarted.
To restart ServerView S2, run the following commands on the console:
cn1:~ # /etc/PRIMECLUSTER/bin/srvst.sh stop
PRIMECLUSTER will detect the “faulted” application and try to restart it after
about 10 seconds. Verify this by calling:
cn1:~ # /etc/PRIMECLUSTER/bin/srvst.sh status
To add all manageable servers, right click anywhere on the list and click on
Select Manageables.
After clicking on Select Manageables, all servers with known type will be selected:
To finally add those selected servers, click on Apply in the upper right corner of the
Server Browser window.
Note that the Control Nodes are accessible from all subnets, so a message about already
existing servers will be shown which should be confirmed by clicking on OK:
If ServerView S2 detects blade servers, it asks to add the whole Blade Center instead of
each individual blade. This makes management easier and is highly recommended.
So in the following dialogue you should reply with Yes:
If ServerView can not detect the host name of the Management Blade, it will ask for it:
After closing the Server Browser, the main window containing the server list will show the
added servers. Further servers can be added by repeating the recent steps using a
different subnet address.
To view the state of a single server, click on the blue server name. To view the state of a
Server Blade, click on the Management Blade / Blade Center, then select the right blade
and click on ServerView on the bottom navigation bar.
BX300 / 600
● While booting, press F2
● Menu:
Advanced -> PCI Configuration -> OnBoard LAN1 + LAN2 [Enable]
● Reboot
● Menu: Boot 2 x MBA …. on top
● Menu: Exit -> Saving Changes
BX600 S2
● While booting, press F2
● Menu:
Advanced -> PCI Configuration -> OnBoard LAN1 + LAN2 [Enable]
● Reboot
● While booting, press F2
● Menu: Boot -> Boot Device Priority -> 1st Boot Device
● Select IBA GE Slot 0420 …
● Menu: Boot -> Boot Device Priority -> 2nd Boot Device
● Select IBA GE Slot 0421 …
● Menu: Exit -> Saving Changes
RX300
● While booting, press F2
● Menu: Advanced -> LAN Remote Boot A+B [PXE]
● Reboot
● While booting, press F2
RX300 S2
First, you have to make the PCI-X NICs net bootable.
Download Proboot.exe from the INTEL website:
http://downloadfinder.intel.com/scripts-df-external/Product_Filter.aspx?ProductID=412
Extract proboot.exe and move it to a bootable CD-Rom.
Power on the RX300 S2.
After booting start IBAutil from the DOS Prompt:
<LW>:\ cd <IBAutil-DIR>
<LW>:\ IBAutil.exe -all -flashenable
Reboot.
After booting start IBAutil from the DOS Prompt:
<LW>:\ cd <IBAutil-DIR>
<LW>:\ IBAutil.exe -all -install pxe
RX600
● While booting, press F2
● Menu: Advanced -> Ethernet on Board -> Rom Scan [Enable]
Enable Master [Enable]
● Menu: Advanced -> Embedded SCSI Bios Scan Order [LAST]
● Menu: Main -> Boot Options 2 x MBA …. on top
● Menu: Exit -> Saving Changes
RX600 S2
First, configure the PCI-E NICs as network-bootable.
Download proboot.exe from the INTEL website:
http://developer.intel.com/design/network/products/ethernet/linecard_ec.htm
Extract proboot.exe and move it to a bootable CD-Rom. Insert it into the drive to boot
the node from CD.
● Power on the RX600 S2 .
● While booting you will be requested to Press any key …
● Select Boot Manager
● Select Primary Master CD-ROM
<LW>:\ cd <IBAutil-DIR>
<LW>:\ IBAutil.exe -all -install pxe
RX800
● While booting, press F1.
● Menu: Configuration/Setup Utility
● Select Menu Devices and I/O Ports
● Set Planar Ethernet to [Enabled]
● Press ESC to go back
● Select Menu Start Options
● Select Menu Startup Sequence Options
● Set First Startup Device to [Network]
● Press ESC to go back
● Menu: Start Options
● Set Planar Ethernet PXE/DHCP to [Planar Ethernet 2]
● Press ESC to go back
● Select Menu Advanced Setup
● Select Menu CPU Options
● Set Hyper-Threading Technology to [Enabled]
● Press 2 x ESC to go back
● Select Save settings
● Press ENTER to save
● Select Exit Setup
● Select Yes, exit the Setup Utility
RX800 S2
● While booting, press F1.
● Menu Configuration/Setup Utility
● Select Menu Start Options
● Select Menu Startup Sequence Options
The program changed the MAC addresses of the node within DHCP or RARP
configuration and LDAP. No further intervention is necessary.
Some software licenses are derived from information based on server specific
numbers, such as MAC addresses of network cards. On Linux, double-check
your SAP licenses after you have replaced a network card.
In this case the Control Node hardware must be replaced by the same model with the
same disk controller and the same network interface controllers. The approach is very
simple, plug in your hard disks in the new Control Node in the same order as the old one,
power on the Control Node, enter the RAID controller BIOS and check the parameters;
hardware raid should be enabled as before.
See also the manual “Installation of a FlexFrame Environment“, chapter “Control Node
Host RAID Configuration”.
For further information on third party software within FlexFrame, please refer to section
“Third Party Software” on page 105.
Nevertheless we will describe here the general approach to update/install software on the
Control Nodes, on a few examples. In your special case this approach can differ to the
ones described here.
New FlexFrame software should be updated/installed as RPMs. Only the RPMs will be
documented in the RPM database. If there is no RPM package available (because the
vendor does not deliver this piece of software not as an RPM and SuSE does not provide
this version as RPM as well) you may also install this software from a tar archive. In this
case you should document this in the directory
/FlexFrame/volFF/FlexFrame/documentation/ as a plain text file.
Normally, an already installed software package must be updated with the rpm command
rpm -U <package_name>.rpm.
Don’t forget to install the same software package on the other Control Node, too.
..done
control1:/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8 #
control1:/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8 # rpm -U
srvmagt-eecd-3.10-14.suse.rpm
Running pre (1) for srvmagt-eecd-3.10-14
-rwxr-xr-x 1 7484 Sep 26 2004 /sbin/halt
Running post (1) for srvmagt-eecd-3.10-14
insserv: script ff_mysql: service mysql already provided!
insserv: script mysql.org: service mysql already provided!
insserv: script ff_myAMC.MessengerSrv: service myAMC.Messenger
already provided!
Starting eecd..done
You have new mail in /var/mail/root
control1:/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8 #
control1:/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8 # rpm -U
srvmagt-agents-3.10-14.suse.rpm
Running pre (1) for srvmagt-agents-3.10-14
ONUCDSNMP=true
Shutting down snmpd:..done
Running post (1) for srvmagt-agents-3.10-14
Linking 32-bit/64-bit binaries for i686
insserv: script ff_mysql: service mysql already provided!
insserv: script mysql.org: service mysql already provided!
insserv: script ff_myAMC.MessengerSrv: service myAMC.Messenger
already provided!
-rwxr-xr-x 1 47507 Oct 7 2004 /usr/sbin/snmpd
Running triggerin (1, 1) for srvmagt-3.10-14
lrwxrwxrwx 1 16 Oct 27 18:09 /usr/sbin/snmpd -> ucdsnmpd-
srvmagt
Starting snmpd..done
Starting agents: sc bus hd unix ether bios secur status inv
vv..done
control1:/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8 #
Kernel:
control1: # cp -p <somewhere>/k_smp-2.4.21-286.i586.rpm
/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8
Kernel source:
control1: # cp -p <somewhere>/kernel-source-2.4.21-286.i586.rpm
/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8
control1: # cp -p <somewhere>/initrd-2.4.21-286-smp
/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8
This is not the original initrd as created while installing the kernel-RPM. This initrd
already contains all necessary drivers for the Control Node while booting.
Special Fujitsu Siemens Computers modules:
control1: # cp -p <somewhere>/FSC-2.4.21-286-smp.tar
/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8
In this case only the net/bcm5700.o module is necessary. Normally all special Fujitsu
Siemens Computers drivers will be delivered in a zip archive. In our case we have
extracted only the necessary bcm5700 module and repacked it in a tar archive to simplify
the next installation steps.
control1: # cd /boot
control1:/boot # cp -p initrd-2.4.21-286-smp
initrd-2.4.21-286-smp.ORIG
control1:/boot # cp -p
/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8/initrd-2.4.21-286-smp.
Because various PRIMERGYs (i.e. RX300 and RX300 S2) with different SCSI
controllers are supported, we load more than one kernel module in our initrd.
The modules that could not find the appropriate hardware will display
corresponding (non-critical) error messages.
control1:/boot/mnt/lib/modules/2.4.21-286-smp/kernel/drivers/scsi
# insmod megaraid.o
megaraid.o: init_module: No such device
Hint: insmod errors can be caused by incorrect module parameters,
including invalid IO or IRQ parameters.
You may find more information in syslog or the output from
dmesg
control1:/boot/mnt/lib/modules/2.4.21-286-smp/kernel/drivers/scsi
Replace the original kernel modules with the Fujitsu Siemens Computers modules.
Rename the original driver:
control1:/lib/modules/2.4.21-286-smp/kernel/drivers/net/bcm #
mv bcm5700.o bcm5700.o.ORIG
Install the special Fujitsu Siemens Computers driver bc5700.o; unpack this tar in “/”:
control1:/ # cd /
control1:/ # tar -xvf
/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8/FSC-2.4.21-286-smp.tar
lib/modules/2.4.21-286-smp/net/bcm5700.o
Configure the old kernel as a fall-back in the boot loader. Create a new entry for the
previous kernel:
control1: # cd /boot/grub
control1:/boot/grub # cp -p menu.lst menu.lst.ORIG
control1:/boot/grub # vi menu.lst
gfxmenu (hd0,0)/message
color white/blue black/light-gray
default 0
timeout 8
title linux
kernel (hd0,0)/vmlinuz root=/dev/sda5 vga=788 acpi=off
initrd (hd0,0)/initrd
title floppy
root (fd0)
chainloader +1
title failsafe
kernel (hd0,0)/vmlinuz.shipped root=/dev/sda5 ide=nodma apm=off
acpi=off vga=normal nosmp disableapic maxcpus=0 3
initrd (hd0,0)/initrd.shipped
title oldkernel-251
kernel (hd0,0)/vmlinuz-2.4.21-251-smp root=/dev/sda3 vga=788
acpi=off
initrd (hd0,0)/initrd-2.4.21-251-smp
control1: # cd /FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8
control1:/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8 # rpm -i --
force kernel-source-2.4.21-286.i586.rpm
control1: # cd /usr/src/linux
control1:/usr/src/linux # make menuconfig
SAVE + EXIT
● Conversion and saving of old parameter files with support of the CLI
● Option of reverting to the previous FA Agent version in the event of an unsatisfactory
update
Versioning of the FlexFrame Autonomous Agents is linked to the FlexFrame versions via
a matrix. In other words, it is possible that multiple agent versions can exist for operation
with one FlexFrame version. What is decisive then is the scope of functions and rules that
is required. In concrete terms this means that the FlexFrame Autonomous Agent Version
2.0 can, for example, also be used in compatibility mode for FlexFrame 3.0. In this way
you can profit from enhanced detector and rule characteristics of a new FlexFrame
Autonomous Agent without at the same time having to update the entire FlexFrame.
As soon as new functions of a release are used, functional compatibility only applies to a
limited degree. A matrix will show you which functions are compatible with the release in
the FlexFrame Agents or which are upward- or downward-compatible to a FlexFrame
version.
# Version V20K16
pool.release.current=V20K16
pool.release.base=V20K16
To switch to another version or revision level you have to adjust the value of the
pool.release.current entry. The syntax is
V<version_number>K<revision_number>
This syntax is mandatory. Ensure that the installed version is entered .
For further information on migrating an FA Agent version please refer to section 7.3.3 ff.
For further information on the migration of FA Agent versions on pool level, please see
documentation "FA Agents - Installation and Administration", section 4.8.
Parameter Description
-V/--verbose Detailed output during migration.
-lf/--logfile <file> Writes log messages to log file <file>.
-lp/--logpath <path> Generates the log file in the specified directory.
-h/--help Prints usage.
rd
If you have installed additional 3 party software to your FlexFrame landscape,
please make sure to back up this software components and their configuration
data for later restore. Please consider: If you install a new FlexFrame image, the
whole file system and directories are overwritten by a complete new root file
rd
system image. Afterwards, any 3 party software and configurations have to be
re-installed or restored.
Installing third party software on Control Nodes or Application Nodes may cause
functional restrictions and other malfunctions of the FlexFrame system software.
Fujitsu Siemens Computers shall have no liability whatsoever whether of a
direct, indirect or consequential nature with regard to damages caused by the
third party software or its erroneous installation.
Synopsis
Command Options
--op list
Lists the configuration details of an Application Node.
--name <node_name>
The name of the Application Node to be listed. Use operation mode list-all to get
a list of all configured Application Nodes with their node names (see 8.1.2).
The output is structured in sections: hardware, software, network, assigned pool and
group, switch ports.
● Hardware
This section contains information about system type, rack ID, device rack name,
shutdown facility with IP address and host name, mac addresses and on blade
servers or partition servers the chassis and slot/partition ID.
● Software
This section contains information on OS type, vendor and version and the root
image path.
● Network
This section lists the VLAN ID, IP address and host name of all configured networks,
sorted by LAN segments.
● Pool and Group
This section lists the names of the assigned pool and group.
● Switch ports
In case link aggregates are configured, this section identifies the aggregate and its
ports. Each used switch port is shown with the switch group ID, switch ID and port ID
(cabinet ID, switch blade ID and port on blade servers) for the common LANs
(Storage, Server and Client LAN) and the Control LAN (if used).
Definition of Switch Group:
A number of Cisco Catalyst 3750G switches within one system cabinet. The
switches are connected as a loop with Cisco StackWise cables at the rear of
each switch. With connected cables, the switches form a stack that behaves like
a virtual switch including all ports of connected switches. To identify a switch
port within the entire FlexFrame environment, three items are used:
• the number of the switch group (it is like a numbering of the virtual switches)
starting with 1.
• the number of the switch within the switch group starting with 1.
• the number of the switch port starting with 1.
Definition of Switch Port:
A switch has a number of ports where other network devices and the host net-
work interfaces are connected. The port is identified by a number starting at 1.
Within a switch group, the port number is prefixed by the switch number (the
identification number of the switch within the switch group).
Command Output
The command displays detailed information on the selected Application Node. The output
differs between blade servers and all others.
A PRIMERGY blade server output may look like this:
cn1:/opt/FlexFrame/bin # ff_an_adm.pl --op list --name bx2-6
Hardware
System: BX620S2 RackID: 2 AN Name: AN6
Shut.Facil.: Mgmt Blade bx2-co (195.40.224.78)
MAC Addr.: 00:C0:9F:99:E9:F4 00:C0:9F:99:E9:F5
IDs: 2 / 6 (System Cabinet / Slot|Partition)
Software
OS: SuSE / Linux / SLES-9.X86_64 (Vendor / Type /
Version)
OS Path: f1-pool2-st:/vol/volFF/os/Linux/FSC_3.2B00-
000.SLES-9.X86_64/root_img
Network
VlanID Host IP Hostname
Storage LAN: 890 195.40.231.98 bx2-6-st
Server LAN: 741 195.40.231.2 bx2-6-se
Client LAN: 740 172.28.26.2 bx2-6
Switch ports
Cabinet SwBlade Port
Common LANs: 2 1 6
Common LANs: 2 2 6
Hardware
System: PW650 RackID: 1 AN Name: AN5
Shut.Facil.: RPS pw651-co (192.168.10.171)
MAC Addr.: 00:e0:00:a6:0a:10 00:e0:00:a6:da:0c
Software
OS: Sun / SunOS / 5.8 (Vendor / Type / Version)
OS Path: filer-p1-
st:/vol/volFF/os/Solaris/FSC_5.8_202_20050530/bi_FJSV,GPUZC-M_PW-
CMZ/root/pw651-st
Network
VlanID Host IP Hostname
Storage LAN: 200 192.168.21.171 pw651-st
200 192.168.21.181 pw651-stt1
200 192.168.21.191 pw651-stt2
Server LAN: 300 192.168.31.171 pw651-se
300 192.168.31.181 pw651-set1
300 192.168.31.191 pw651-set2
Client LAN: 400 192.168.41.171 pw651
400 192.168.41.181 pw651-clt1
400 192.168.41.191 pw651-clt2
Switch ports
SW Grp Switch Port
Common LANs: 1 1 9
Common LANs: 1 2 9
Control LAN: 1 1 21
Synopsis
Command Options
--op list-all
Lists all configured Application Nodes.
--pool <pool_name>
The name of the pool of which the Application Nodes have to be listed.
Command Output
The output may look like this:
cn1:/opt/FlexFrame/bin # ff_an_adm.pl --op list-all
bx1-2
Node Type: BX600 Rack/Cabinet/Slot|Partition ID: 1/1/2
OS: SuSE / Linux / SLES-9.X86_64 (Vendor / Type /
Version)
OS Path: f1-pool1-st:/vol/volFF/os/Linux/FSC_3.2B00-
000.SLES-9.X86_64/root_img
Host IP Hostname
Storage LAN: 195.40.224.223 bx1-2-st
Server LAN: 195.40.224.31 bx1-2-se
Client LAN: 143.161.72.31 bx1.2
MAC Addr.: 00:c0:9f:95:5f:8a 00:c0:9f:95:5f:8b
bx2-1
Node Type: BX600 Rack/Cabinet/Slot|Partition ID: 2/2/1
OS: SuSE / Linux / SLES-9.X86_64 (Vendor / Type /
Version)
OS Path: f1-pool1-st:/vol/volFF/os/Linux/FSC_3.2B00-
000.SLES-9.X86_64/root_img
Host IP Hostname
Storage LAN: 195.40.224.227 bx2-1-st
Server LAN: 195.40.224.35 bx2-1-se
Client LAN: 143.161.72.35 vies1pyx
MAC Addr.: 00:c0:9f:95:60:60 00:c0:9f:95:60:61
bx2-2
Node Type: BX600 Rack/Cabinet/Slot|Partition ID: 2/2/2
OS: SuSE / Linux / SLES-9.X86_64 (Vendor / Type /
Version)
OS Path: f1-pool1-st:/vol/volFF/os/Linux/FSC_3.2B00-
000.SLES-9.X86_64/root_img
Host IP Hostname
Storage LAN: 195.40.224.228 bx2-2-st
Server LAN: 195.40.224.36 bx2-2-se
Client LAN: 143.161.72.36 bx2-2
MAC Addr.: 00:c0:9f:93:7f:cc 00:c0:9f:93:7f:cd
Pool pool2
Pool Group bx600_o
bx1-6
Node Type: BX620S2 Rack/Cabinet/Slot|Partition ID: 1/1/6
OS: SuSE / Linux / SLES-9.X86_64 (Vendor / Type /
Version)
OS Path: f1-pool2-st:/vol/volFF/os/Linux/FSC_3.2B00-
000.SLES-9.X86_64/root_img
Host IP Hostname
Storage LAN: 195.40.231.97 bx1-6-st
Server LAN: 195.40.231.1 bx1-6-se
Client LAN: 172.28.26.1 bx1-6
MAC Addr.: 00:C0:9F:99:E6:CC 00:C0:9F:99:E6:CD
bx2-6
The output of list-all is less detailed than the list output. It is used to get an
overview. It shows the Application Nodes sorted by pool and group in alphabetical order.
For each node the system type, the cabinet and slot/partition ID (if node is a blade or
partition server), the OS type, vendor, version, the root image path, the main IP
addresses, host names and the MAC addresses are listed.
Synopsis
Command Options
--op add
Adds an Application Node.
--type <system_type>
Specifies the product name and type like “PW250” or “RX300S2”. These are the
common system type (family) terms. More detailed product identifiers are not
necessary. See usage (call ff_an_adm.pl without any parameter) to get a list of
supported system types.
--name <node_name>
The name of the Application Node. This name has to be unique for the entire
FlexFrame system. All interface names are based on this node name. We
recommend using lower case names if possible.
--pool <pool_name>
The name of the pool this node should belong to. See usage (call ff_an_adm.pl
without any parameter) to get a list of currently configured pools.
--group <group_name>
The name of the pool group this node is a member of. A group must consist of
Application Nodes of the same OS image version and should be of the same
capacity (CPU, Memory etc.). There should be at least one spare Node in a group.
Otherwise, take-over of failing services will not be possible. Use command
ff_pool_adm.pl with op mode list or list-all to get the pool groups.
--swgroup <switch_group_id>
Defines the switch group the Application Node is connected to. This information is
necessary to assign and configure switch ports. The switch group was numbered
during installation with the FlexFrame PlanningTool. Use this number here too. See
usage (call ff_an_adm.pl without any parameter) to get a list of currently
configured switch group IDs.
--mac <mac_addresses>
Add here the both MAC addresses of the data NICs used with booting. Use the
double colon separated hex notation for each MAC address. Concatenate them with
a comma. The MAC address syntax is six colon separated hex values, eg.
00:e0:00:c5:19:41.
--ospath <path_to_os_image>
Defines the OS image to be used. Add the relative path to
/FlexFrame/volFF/os/ as seen from the Control Node. See usage (call
ff_an_adm.pl without any parameter) to get a list of available OS pathes.
--host <ip_host_number>[,<test1_host_number>,<test2_host_number>]
Host part to be used to build IP addresses for the three networks. On Solaris systems
(PRIMEPOWER) you have to give three different host numbers, separated with a
comma. If this option is omitted, the script uses free host numbers to calculate the IP
addresses.
With some server types you have to give additional information, i.e. PRIMERGY server
blades or PRIMEPOWER server partitions:
--slot <BXxxx_cabinet/slot>
With PRIMERGY server blades use this option to define the cabinet and slot ID of the
server blade. New cabinets have to be defined with the
/opt/FlexFrame/bin/ff_bx_cabinet_adm.pl command.
--part <PW_cabinet/partition>
With PRIMEPOWER server partitions (models PW900, PW1500 and PW2500) use
this option to define the cabinet and partition ID of the server partition. With new
cabinets, a separate SMC Control LAN IP address and switch port are assigned
additionally.
Command Output
The command displays some information about processing steps. The output for a blade
server may look like this:
cn1:/opt/FlexFrame/bin # ff_an_adm.pl --op add --type BX620S2
--name bx1-6 --pool pool1 --group bx600_o --ospath
Linux/FSC_3.2B00-000.SLES-9.X86_64 --host 1 --slot 1/6 --mac
00:C0:9F:99:E6:CC,00:C0:9F:99:E6:CD
update swblade 1/1 configuration
Notice: Update will take about 1 minute.
update swblade 1/2 configuration
Notice: Update will take about 1 minute.
If not reported any error all precautions are done to create
application nodes os image. To do this call:
ff_new_an.sh -n bx1-6
Creating and customizing an image may take some minutes.
Don't get anxious.
1 / 1 / 14 data NIC-2
1 / 2 / 15 mgmt NIC-1
The script first checks all arguments and aborts with error messages in case of errors.
Then it fetches free IP addresses and switch ports. The switch ports are reconfigured to
match requirements, the LDAP data is created and a netboot file is written. The netboot
file is used by ff_new_an.sh to create Application Node images and extend the Filer’s
exports list.
At the end you get a cabling advice and instructions how to call ff_new_an.sh script to
finish the Application Node creation.
Synopsis
Command Options
--op rem
Removes an Application Node.
--name <node_name>
The name of the Application Node to be removed. Use operation mode list-all to
get all configured Application Nodes and their names (see 8.1.2).
Command Output
The command displays only errors and warnings. An output may look like this:
cn1:/opt/FlexFrame/bin # ff_an_adm.pl --op rem --name pw250-3
Synopsis
Command Options
--op list
Lists the configuration details of a blade server cabinet.
--name <cabinet_name>
The name of the blade server cabinet to be listed.
The output is structured in sections: hardware, software, network, assigned pool and
group, switch ports.
Command Output
The command displays detailed information on the selected blade server cabinet. The
output may look like this:
Switch Blade
SwitchID Type Switch name Hostname IP Address
1 Quanta bx600-2-swb1 bx600-2-swb1 195.40.224.76
2 Quanta bx600-2-swb2 bx600-2-swb2 195.40.224.77
As seen from the sample above, the cabinet ID and name, the cabinet system type, the
management blade and the switch blades are listed.
For the management blade the host name, the IP address and both core switch ports are
displayed.
The switch blade information shows the switch and host name, the IP address and the
switch blade port to core switch port connections, structured by switch blade ID.
Synopsis
Command Option
--op list-all
Lists all configured blade server cabinets.
Command Output
The command displays the configured blade server cabinets. An output may look like
this:
Primergy Cabinets
1 (cab1) BX600
Management Blade: cab1-co / 195.40.224.75
Switch Group ID: 1
Server Blades (by slot id)
1 (blade1) BX600
Pool / Group: pool1 / PROD
2 (blade2) BX600
Pool / Group: pool1 / PROD
3 (blade3) BX600
Pool / Group: pool1 / PROD
4 (blade4) BX600S2D
Pool / Group: pool2 / DEV
5 (blade5) BX600
Pool / Group: pool2 / DEV
For each cabinet the ID, the cabinet name, the management host name and IP address
and the server blades are displayed.
Each server blade is shown with its slot ID and name, the system type and the pool and
group it belongs to.
Synopsis
Command Options
--op add
Adds a blade server cabinet.
--type <system_type>
PRIMERGY blade system type e.g. BX300 or BX600.
Call ff_bx_cabinet_adm.pl without any parameter to get a list of supported
system types.
--name <cabinet_name>
Name of the subsystem (cabinet). It is used to generate a new name for the
management blade (has to be unique within entire FlexFrame).
--swgroup <switch_group_id>
Switch group number (starts with 1) the cabinet has to be connected to (physically).
See usage (call ff_bx_cabinet_adm.pl without any parameter) to get a list of
currently configured switch group IDs.
--swblades <type_of_switch_blades>
The type of switch blades. Default is set to the newer type of switch blades for this
cabinet type (Quanta). Valid types are: Quanta, Accton, Pass-Through. For the
default switch blades this option may be omitted. Else give the type for the switch
blade like "Accton,Accton", or any combination of valid types for all used switch
blades.
Command Output
At the end of the output, the command displays further instructions. The output may look
like this:
SwitchID/Port Mgmt/SwitchBlade
1 / 8 slave ManagementBlade
1 / 11 SwitchBlade 1 Port 12
1 / 12 SwitchBlade 2 Port 12
2 / 8 master ManagementBlade
2 / 11 SwitchBlade 1 Port 11
2 / 12 SwitchBlade 2 Port 11
Console> enable
Console # configure
Console(config)# interface vlan 200
Console(config-if)# ip address 195.40.224.77 255.255.255.0
Console(config-if)# end
Console # copy tftp startup-config
TFTP server ip address: 195.40.224.3
Source configuration file name: swblade-2-2.config
Startup configuration file name [startup-config]:
\Write to FLASH Programming.
-Write to FLASH finish.
Success.
Console # reload
(Vty-0) #configure
(Vty-0) (Config)#vlan database
(Vty-0) (Vlan)#vlan 200
(Vty-0) (Vlan)#exit
(Vty-0) (Config)#interface vlan 200
(Vty-0) (if-vlan 200)#ip address 195.40.224.76 255.255.255.0
(Vty-0) (if-vlan 200)#exit
(Vty-0) (Config)#end
(Vty-0) #copy tftp://195.40.224.3/swblade-2-1.config script
config.scr
Mode....................................... TFTP
Set TFTP Server IP......................... 195.40.224.3
TFTP Path..................................
TFTP Filename.............................. swblade-2-1.config
Data Type.................................. Config Script
Destination Filename....................... config.scr
(swblade-2-1) #
Plug all other uplink ports to core switches. Now the switch blade
should be fully operational.
Set up the management blade initially with name and IP address listed by the output as
seen above. Use console redirection of management blade to connect to the console of
the switch blades, and upload configuration as described by FlexFrame Installation
Guide.
Finally, plug in the network cables according to the wiring plan given by the command
output.
Synopsis
Command Options
--op rem
Removes a blade server cabinet.
--id <cabinet_id>
Specifies the subsystem (cabinet) ID of the cabinet to be removed. Use the
list-all option to get the ID (see section 8.5.1.2).
Command Output
If there are any server blades configured for this cabinet, an error message is displayed
like in the sample below:
Use the list operation mode to list the configured server blades. You have to remove
them before you can remove the cabinet they are in.
If no server blades are configured for this cabinet, the command displays a summary at
the end. The output may look like this:
The cabinet has been removed successfully from LDAP and the core switch ports used
by the cabinet have been reconfigured to default.
Synopsis
Command Options
--op swb-change
Selects the operation mode. Change the type of a switch blade.
--id <cabinet_id>
Specifiesd the subsystem (cabinet) ID of the cabinet. Use the list-all option to
get the ID.
--swbid <switch_blade_id>
Specifies the ID of the switch blade. The ID is the slot number of the selected switch
blade.
--swbtype <switch_blade_type>
Defines the new type of the switch blade. See usage for the currently supported
types.
Command Output
The output may look like this:
The switch blade type was changed at LDAP database. To get the initial configuration
use operation mode swb-config of this program. It will display instructions how to
upload the configurations too.
Synopsis
Command Options
--op swb-name
Selects the operation mode. Change the name of a switch blade.
--id <cabinet_id>
Specifies the subsystem (cabinet) ID of the cabinet. Use the list-all option to get
the ID.
--swbid <switch_blade_id>
Specifies the ID of the switch blade. The ID is the slot number of the selected switch
blade.
--swbname <switch_blade_name>
Defines the new name of the switch blade.
Command Output
The output may look like this:
As noted by the program the name of the switch will be changed at /etc/hosts of both
control nodes, the LDAP database and at least the hostname and, if possible, the SNMP
sysname at the selected switch blade itself.
Synopsis
Command Options
--op swb-passwd
Selects the operation mode. Change the login password of a switch blade.
--id <cabinet_id>
Specifies the subsystem (cabinet) ID of the cabinet. Use the list-all option to get
the ID.
--swbid <switch_blade_id>
Specifies the ID of the switch blade. The ID is the slot number of the selected switch
blade.
--swbpwd <password>
Defines the new login and enable password of the switch blade.
Command Output
The output may look like this:
As noted by the program the password will be changed at the LDAP database and the
selected switch blade. At the switch blade the login password and theenable password
are changed and have to be the same.
Synopsis
Command Options
--op swb-config
Selects the operation mode. Create the initial switch blade configuration.
--id <cabinet_id>
Specifies the subsystem (cabinet) ID of the cabinet. Use the list-all option to get
the ID.
--swbid <switch_blade_id>
Specifies the ID of the switch blade. The ID is the slot number of the selected switch
blade.
Command Output
The output may look like this:
Console> enable
Console # configure
Console(config)# interface vlan 200
Console(config-if)# ip address 195.40.224.77 255.255.255.0
Console(config-if)# end
Console # copy tftp startup-config
TFTP server ip address: 195.40.224.3
Source configuration file name: swblade2-2.config
Startup configuration file name [startup-config]:
\Write to FLASH Programming.
-Write to FLASH finish.
Success.
Console # reload
System will be restarted, continue <y/n>? y
control1:/FlexFrame/volFF/os/Linux # ls -l
total 20
drwxr-xr-x 5 root root 4096 Apr 29 16:33 .
drwxr-xr-x 4 root root 4096 Mar 15 15:10 ..
drwxr-xr-x 4 root root 4096 Apr 22 13:41
FSC_3.2B00-000.SLES-9.X86_64
drwxr-xr-x 4 root root 4096 Apr 22 14:15 pool_img
Synopsis
Command Option
-p <path_to_images>
Each image version has its own path to images i.e. FSC_3.2.x. If this path already
exist you have to specify a new image path with the -p option.
Example for destination path FSC_3.2.x:
control1:/ # mount /media/dvd
control1:/ # /media/dvd/ff_install_an_linux_images.sh
-p /FlexFrame/volFF/os/Linux/FSC_3.2.x
..
..
control1:/FlexFrame/volFF/os/Linux # ls -l
total 20
drwxr-xr-x 5 root root 4096 Apr 29 16:33 .
drwxr-xr-x 4 root root 4096 Mar 15 15:10 ..
drwxr-xr-x 4 root root 4096 Apr 22 13:41
FSC_3.2B00-000.SLES-9.X86_64
drwxr-xr-x 4 root root 4096 Apr 29 17:10 FSC_3.2.x
drwxr-xr-x 4 root root 4096 Apr 22 14:15 pool_img
The location for the Linux Application Node images always has to be
/FlexFrame/volFF/os/Linux/.
Synopsis
Example:
control1:/ # ff_create_an_cfg.pl --name an_0504
--lin_os_path /FlexFrame/volFF/os/Linux/FSC_3.2.x
Synopsis
ff_new_an.sh -n <app_node_name>
Example:
control1:/ # ff_new_an.sh -n an_0504
You now have a new Application Node Image tree with one newly configured Application
Node. The newly configured Application Node will use this image tree upon the next
reboot and can be used for maintenance purposes.
Boot
Boot or
or Reboot
Reboot
Create
Create next maintenance cycle
Remaining
Remaining
Custom
Custom Image
Image Tree
Tree Application
Application Nodes
Nodes
Create
Create and
and Configure
Configure
Disable
Disable FA
FA Agents
Agents Remaining
Remaining
in
in Custom
Custom root
root Image
Image Application
Application Node
Node
Images
Images
Create
Create
Isolate
Isolate one
one Netboot
Netboot Configuration
Configuration
Application
Application Node
Node for
for Remaining
Remaining
for
for Maintenance
Maintenance Application
Application Nodes
Nodes
Create
Create Change
Change exports
exports
Netboot
Netboot Configuration
Configuration on
on Filer
Filer to
to
for
for Maintenance
Maintenance Read-Only
Read-Only
Application
Application Node
Node for
for root
root Image
Image
Create
Create and
and Configure
Configure
Maintenance
Maintenance Enable
Enable FA
FA Agents
Agents
Application
Application Node
Node in
in Custom
Custom root
root Image
Image
Image
Image
Change
Change exports
exports Create
Create var_template
var_template
on
on Filer
Filer to
to from
from Customized
Customized
Read-Write
Read-Write var
var Image
Image
for
for root
root Image
Image
Boot
Boot or
or Reboot
Reboot test failed?
Reboot
Reboot and
and Test
Test
Maintenance
Maintenance Maintenance
Maintenance
Application
Application Node
Node Application
Application Node
Node
Remount
Remount Remount
Remount
root
root Filesystem
Filesystem root
root Filesystem
Filesystem
as
as Read-Write
Read-Write as
as Read-Only
Read-Only
Start
Start Software
Software Update,
Update, Finalize
Finalize
Maintenance
Maintenance Actions
Actions Kernel
Kernel Update,
Update, ...
... Maintenance
Maintenance Actions
Actions
control1:/FlexFrame/volFF/os/Linux/FSC_3.2B00-000.SLES-9.X86_64 #
find root_img var_img/var_template | cpio -pdum ../CustomImage_3.2
control1:/FlexFrame/volFF/os/Linux/FSC_3.2B00-000.SLES-9.X86_64 #
cd ..
control1:/FlexFrame/volFF/os/Linux # ls -l
total 20
drwxr-xr-x 5 root root 4096 Apr 29 16:33 .
drwxr-xr-x 4 root root 4096 Mar 15 15:10 ..
drwxr-xr-x 4 root root 4096 Apr 22 13:41
FSC_3.2B00-000.SLES-9.X86_64
drwxr-xr-x 4 root root 4096 Apr 29 17:10
CustomImage_3.2
drwxr-xr-x 4 root root 4096 Apr 22 14:15 pool_img
This image directory tree contains all necessary images, the root image and all var
images:
control1:/FlexFrame/volFF/os/Linux # ls -l CustomImage_3.2
total 16
drwxr-xr-x 4 root root 4096 Apr 29 17:10 .
drwxr-xr-x 5 root root 4096 Apr 29 16:33 ..
drwxr-xr-x 37 root root 4096 May 3 11:07 root_img
drwxr-xr-x 9 root root 4096 Apr 22 14:15 var_img
control1:/FlexFrame/volFF/os/Linux/CustomImage_3.2/root_img/etc/
init.d # insserv -r ./myAMC*
You now have a new Application Node image tree with one newly configured Application
Node.
The newly configured Application Node will use this image tree upon the next reboot and
can be used for maintenance purposes.
control1:/FlexFrame/<filer>/vol0/etc # vi exports
....
/vol/volFF/os/Linux/CustomImage_3.2/root_img
-sec=sys,rw=192.168.11.53,anon=0
control1:/vol/volFF/os/Linux/CustomImage_3.2/var_img #
mv var_template var_template.OLD
control1:/vol/volFF/os/Linux/CustomImage_3.2/var_img #
cp -pdR var-coa80b35 var_template
Cleanup:
control1:/vol/volFF/os/Linux/CustomImage_3.2/var_img/var_template/
tmp # rm -rf *
control1:/vol/volFF/os/Linux/CustomImage_3.2/var_img/var_template/
FlexFrame/etc/sysconfig/network # rm ifcfg-vlan*
control1:/FlexFrame/volFF/os/Linux/CustomImage_3.2/root_img/etc/
init.d # insserv ./myAMC*
Default kernel:
control1: # cp -p <somewhere>/k_default-2.4.21-286.i586.rpm
/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8
Kernel source:
control1: # cp -p <somewhere>/kernel-source-2.4.21-286.i586.rpm
/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8
control1: # cp -p <somewhere>/initrd_2.4.21.gz
/FlexFrame/volFF/FlexFrame/stage/SuSE/SLES8
This is not the original initrd as created while installing the kernel RPM. This initrd
already contains all necessary drivers for the Application Nodes while net-booting.
To install any software on the Application Nodes it is necessary to get write permissions
to the mounted root file system. Mount the root file system as read-write (example).
Export the root image as read-write only for the maintenance Application Node
192.168.12.19.
Edit vol0/etc/exports on the Filer:
control1:/FlexFrame/<filer>/vol0/etc # vi exports
...
/vol/volFF/os/Linux/CustomImage_3.2/root_img -
ro,anon=0,rw=192.168.12.19
Export the temporary permissions via rsh on the Filer. Telnet to the Filer is also possible:
control1:/ # rsh <filer> exportfs -a
Do not forget to export the root image as read-only after the maintenance!
On the Application Node:
Remount the Application Nodes root filesystem with write permissions:
an_0504:/ # mount -o remount rw /
Install the new Linux kernel 2.4.21-286, -default for Mono Blades BX300 with
Pentium-M processor and -smp for the rest.
For SMP kernels:
an_0504:/mnt/stage/SuSE/SLES8 # rpm -i k_smp-2.4.21-286.i586.rpm
Kernel source:
Save the old symbolic links:
an_0504:/usr/share/doc/packages # mv kernel-source
kernel-source-2.4.21-251
Because we support different PRIMERGYs with different network, SCSI and IDE
controllers, the appropriate modules are loaded depending from the netboot
configuration parameter DRV in the special Application Node initrd.
control1:/FlexFrame/volFF/os/Linux/CustomImage_3.2/root_img/boot #
cp -p vmlinuz-2.4.21-286-smp /tftpboot
control1:/FlexFrame/volFF/os/Linux/CustomImage_3.2/root_img/boot #
cp -p vmlinuz-2.4.21-286-default /tftpboot
control1:/tftpboot/pxelinux.cfg # vi rx300_pool1_CustomImage_3.2
default SLES8-286
..
..
LABEL SLES8-286
KERNEL SLES8_2.4.21-286
APPEND initrd=initrd_2.4.21.gz
nfsroot=192.168.12.204:/vol/volFF/os/Linux/CustomImage_3.2/
root_img,intr,v3,nolock,rsize=32768,wsize=32768,tcp BI=1:2
DRV=BCM5700,E1000,AIC79 apm=off
ServerView will generate (compile) new modules for the new Linux Kernel during the next
reboot. The root image has to be writable for this, too.
While rebooting the new kernel, this message should appear on the console:
Because of the mount option “ro” on the Application Node, automatic module compilation
will fail. In this case this has to be done manually with:
an_0504:/ # mount -o remount,rw /
an_0504:/ # /etc/init.d/eecd_mods_src start
Compiling modules for 2.4.21-286-smp:
copa(Ok) cop(Ok) ihpci(Ok) ipmi(Ok) smbus(Ok)
done
For SLES8 Application Node images, it is required to do this for both, the SMP kernel and
the default kernel. With SLES9 there is only an smp kernel and you are finished now.
Therefore, for SLES8, the netboot configuration should be changed like this:
control1:/tftpboot/pxelinux.cfg # vi rx300_pool1_CustomImage_3.2
default SLES8-286-def
..
LABEL SLES8-286-def
KERNEL SLES8_2.4.21-286-def
APPEND initrd=initrd_2.4.21.gz
nfsroot=192.168.12.204:/vol/volFF/os/Linux/CustomImage_3.2/
root_img,intr,v3,nolock,rsize=32768,wsize=32768,tcp BI=1:2
DRV=BCM5700,E1000,AIC79 apm=off
Reboot the Application Node and compile the ServerView modules again:
an_0504:/ # mount -o remount,rw /
an_0504:/ # /etc/init.d/eecd_mods_src start
Compiling modules for 2.4.21-286-default:
copa(Ok) cop(Ok) ihpci(Ok) ipmi(Ok) smbus(Ok)
done
control1:/FlexFrame/<filer>/vol0/etc # vi exports
..
/vol/volFF/os/Linux/CustomImage_3.2/root_img -ro,anon=0
Export the regular permissions via rsh on the Filer. Telnet to the Filer is also possible:
8.6.4.7 Changing All Netboot Configuration Templates to the New Linux Kernel
In case the kernel tests were run successfully, the netboot configuration templates for
creating new Application Nodes should be switched to the new configuration.
In this example the old Linux kernel descriptor was “-251” and the new one is “-286”:
control1:/tftpboot/pxelinux.cfg/templates #
perl -p -i.$(date +'%Y%m%d%H%M%S') -e 's/-251/-286/' *_template
an_0504:/mnt/stage/ServerView #
rpm -U srvmagt-mods_src-3.10-14.suse.rpm
insserv: can not stat(_sap_acc_agents)
Loading modules: ipmi smbus msr cpuid
..done
an_0504:/mnt/stage/ServerView #
rpm -U srvmagt-eecd-3.10-14.suse.rpm
Running pre (2) for srvmagt-eecd-3.10-14
Shutting down eecd: TERM..done
lrwxrwxrwx 1 12 Sep 24 18:01 /sbin/halt -> halt-srvmagt
Running post (2) for srvmagt-eecd-3.10-14
insserv: can not stat(_sap_acc_agents)
Starting eecd..done
an_0504:/mnt/stage/ServerView #
rpm -U srvmagt-agents-3.10-14.suse.rpm
Running pre (2) for srvmagt-agents-3.10-14
ONUCDSNMP=true
Stopping agents: sc bus hd unix ether bios secur status inv
vv..done
Shutting down snmpd:..done
Running post (2) for srvmagt-agents-3.10-14
Linking 32-bit/64-bit binaries for i686
insserv: can not stat(_sap_acc_agents)
lrwxrwxrwx 1 16 Sep 24 17:47 /usr/sbin/snmpd -> ucdsnmpd-
srvmagt
Running triggerin (2, 1) for srvmagt-3.10-14
lrwxrwxrwx 1 16 Sep 24 17:47 /usr/sbin/snmpd -> ucdsnmpd-
srvmagt
Starting snmpd..done
Starting agents: sc bus hd unix ether bios secur status inv
vv..done
an_0504:/mnt/stage/ServerView #
The RPM database is linked from the var image to the root image.
Synopsis
To modify the netboot configuration files for a single Application Node, i.e. Application
Node an_0506, invoke:
control1:/ #
ff_an_adm.pl --name an_0506 --ospath Linux/CustomImage_3.2
Synopsis
To create a new var image for one Application Node, i.e. an_0506 invoke:
To create new var images for all Application Nodes belonging the new Application Node
image CustomImage_3.2, invoke:
control1:/ # ff_new_an.sh
-s 'PATH_TO_IMAGES=/FlexFrame/volFF/os/Linux/CustomImage_3.2'
This will create all var-images for the new Application Node image at once.
The newly configured Application Nodes will use their new images upon the next reboot.
8.7.1 Introduction
<identifier> denotes the version and origin of the images below. For images created
at Fujitsu Siemens Computers the identifier is built as follows:
<identifier> = FSC_<SunOS_version>_<version_details>
Names for customer specific Solaris images must not start with “FSC”. It is
recommended to use a similar naming convention.
In this directory you will find one or more boot images (bi*) or master boot image
directories (mbi*) in the form
bi_<server_class> or mbi_<server_class>
<server class> denotes groups of PRIMEPOWER servers within the same server
class which share the same basic technology.
As of the time this document was written, the following server classes are supported:
FJSV,GPUZC-M_PW-P (PW 250 and PW 450)
FJSV,GPUZC-M_PW-CMZ (PW 650 and PW 850)
FJSV,GPUZC-L_PW-CLZ (PW 900, PW 1500 and PW 2500)
The classification in different hardware platforms is necessary because of the ESF
(Enhanced Shutdown Facility) software, which has to be installed on each
PRIMEPOWER system.
The next directory level holds a structure similar to the Solaris diskless client concept:
Solaris_<version>
exec
root
The root folder holds the clone and directories named like the Application Nodes using
this image (including -st suffix for Storage LAN). The clone is the basis for creating an
Application Node specific image.
Now you have direct access to the Application Node’s console. We recommend keeping
the consoles of the Application Nodes open during boot and maintenance work.
If the console is not accessible (locked by another administrator or the Application Node
is not responding, etc.) you can connect directly to the XSCF itself using port 8010:
control1:~ # telnet an_0800-co 8010
Trying 192.168.20.30...
Connected to an_0800-co.
Escape character is '^]'.
SCF Shell
login:root
Password:
SCF Version 03190001
ALL RIGHTS RESERVED, COPYRIGHT (C) FUJITSU LIMITED 2003
ff.ff.ff[192.168.20.30]
SCF>
If the Application Node got stuck, check the console logs using the following command:
SCF> show-console-logs
If this setting is not done, strange network behaviour may be the result. Some
network switches cannot handle the same MAC address at different
ports/VLANs.
The address reported by banner may differ! banner typically reports hme0's
mac address, which is not used in FlexFrame.
properly configured in /etc/sysidcfg. The host name must match the network name
of the NIC in the Storage LAN you are booting from.
On the very first boot /etc/sysidcfg is read by a couple of tools, which in turn edit e.g.
/etc/hosts, /etc/default/init etc. After the next reboot /etc/sysidcfg is no
longer used. After having mounted its root file system, the client mounts /usr as defined
in /etc/vfstab. The system should now go to multiuser level as usual.
We start the boot process from the console of the PRIMEPOWER (from the OBP).
More information is displayed with the options -dv.
control1:~ #
touch /FlexFrame/volFF/os/Solaris/FSC_5.8_202_20050211/
bi_ FJSV,GPUZC-M_PW-P/root/an_0800-st/etc/.trace.sd
...
##### rc2: /etc/rc2.d/S01SMAWswapmirror start
##### rc2: /etc/rc2.d/S05RMTMPFILES start
##### rc2: /etc/rc2.d/S06log3 start
Starting the logging daemon
Started!
##### rc2: /etc/rc2.d/S10lu start
##### rc2: /etc/rc2.d/S19FJSVdmpsnap start
##### rc2: /etc/rc2.d/S20sysetup start
...
If you want turn of the messages again, remove the file .trace.sd, e.g. like:
control1:~ # rm /FlexFrame/volFF/os/Solaris/FSC_5.8_202_20050211/
bi_ FJSV,GPUZC-M_PW-P/root/an_0800-st/etc/.trace.sd
You need sufficient free space on the file system of the filer, where the Solaris
Boot Images will be stored. Before creating Solaris Boot Images it has to be
assured that the minimum available space of this file system should not exceed
an usage level of 80 percent *after* creating the image. Be in mind, one Solaris
Boot Image will need about 5GB of space in the file system!
Otherwise the commands called by nb_unpack_bi will fail and for that reason
the Solaris Boot Image will be unusable! If this happens, you will get error
messages like:
Patch cluster install script for Solaris 9 Recommended
Determining if sufficient save space exists...
expr: syntax error
expr: syntax error
expr: syntax error
.....
expr: syntax error
./Rinstall_cluster: test: argument expected
For each hardware class of the Application Nodes you have to create the corresponding
Boot Image on your Control Node respectively Filer.
The following graphic and explanations will give you a short overview to the necessary
preparations for creating a Boot Image.
and follow the instructions. For further details, please refer to the section
“Preparation for Solaris Application Nodes” in the chapter “Installation Scripts” of the
“Installation of a FlexFrame Environment” manual.
5. Install the DVD on the second Control Node as well. Here, only some of the tools will
be installed, not the entire Boot Image. The Boot Image has already been copied to
the Filer by installing the DVD on the first Control Node.
mount /media/dvd
cd /media/dvd
./nb_unpack_bi
You will be asked for creating the connection to the Control Node.
The authenticity of host <CN_name> (IP.IP.IP.IP) can't be established.
RSA key fingerprint is 12:f2:7f:c4:ea:98:8b:32:dd:49:02:f5:ae:3b:95:f7.
root@<CN_name>-st's password:
Then you are asked for the password of user root on the Control Node:
root@<CN_name>-st's password:
PermitRootLogin=yes
8.2 Check if ssh connection from the Control Node to the helper system is working
without any interaction (yes or password).
On Control Node:
ssh <AN_name>-st
After all these preparations please go back to section 8.7.3.1 (see page 154) to start the
Solaris Boot Image creation using nb_unpack_bi command.
8.8.1 Introduction
Running a Solaris Image Maintenance Cycle means maintaining an existing Solaris
Image: You might want to apply patches, install or upgrade software, modify several
configurations or change other settings on a Solaris Image where your Solaris Application
Nodes run on.
In order to run a Solaris Image Maintenance Cycle you will need a
● Solaris Image
● Solaris Application Node running on that Solaris Image
● time slot of min. 2 hours
● the software to install and/or a plan what you want to maintain
The Solaris Application Node selected for running the Solaris Image
Maintenance Cycle won't be visible to FlexFrame i.e. in particular won't serve as
a spare node during that time.
We would like to point out that changes due to 3rd party software or patches or
customer modifications may have heavy (negative) side effects to the
FlexFrame functionality.
Before installing any 3rd party software, please see section “Third Party
Software” on page 105.
We would also like to point out that there is software on the market which can
not be installed on NFS or is not ready for Adaptive Computing (e.g. moving
SAP services from one to another node). If this is the case, contact your 3rd
party software vendor. It is usually not an issue of FlexFrame.
8.8.2 Overview
This section lists an example in order to give an idea how the Solaris Image Maintenance
Cycle works. Basically the Solaris Image Maintenance Cycle comprises of just 3 steps:
1. Transfer a Boot Image into a Maintenance Image.
2. Boot a Solaris Application Node from it and maintain it: install or upgrade software,
modify, apply.
3. Transfer the Maintenance Image into a Boot Image and boot all of the Solaris
Application Nodes of that group from the maintained Boot Image.
Example:
Assuming you run a couple of Solaris Application Nodes on a certain Image. To make it
short let's call it GOODIMAGE.
Now you decide that it is needed to have a certain piece of software SOFTWARE on these
Solaris Application Nodes.
You know that you can't install software directly on the Solaris Application Nodes:
● On the one hand the Solaris Application Nodes of a group share the same /usr file
system. This is only read-only for all of them.
● On the other hand you don't like to install software to all of them (via the shared
filesystem) without a careful test procedure.
● Beside that you must prevent FlexFrame from taking over this Solaris Application
Node in case of failures of another Solaris Application Node during the installation of
your software SOFTWARE.
Therefore you have to select one of your Solaris Application Nodes and bring it into a so
called maintenance state. This comprises of:
1. Choose one Solaris Application Node which is running on the Image you like to
maintain.
2. Save the configuration data of this Solaris Application Node - you might want to bring
it up with the same Internet Address, group, pool, etc.
3. Remove this Solaris Application Node.
4. Create a so called Maintenance Image out of the Solaris Image GOODIMAGE the
chosen Solaris Application Node run on.
5. Create configuration information for this Solaris Application Node.
6. Configure this Solaris Application Node.
7. Boot this Solaris Application Node.
Now this Solaris Application Node runs on a Maintenance Image which is separated from
the original GOODIMAGE. All the other Solaris Application Nodes on the GOODIMAGE run
without getting touched. The Maintenance Image allows software installation since it has
in particular it's own writable /usr file system and the FlexFrame (FA Agents) don't control
it. That's why you can continue now:
8. Install the software SOFTWARE on the Solaris Application Node and test it. Configure,
reboot, change settings of SOFTWARE until you are satified.
The result is a modified GOODIMAGE which contains the software SOFTWARE. To make it
short let's call it BETTERIMAGE. In order to have the software SOFTWARE available on all
the Solaris Application Nodes in the group you have to:
9. Shut down the Solaris Application Node.
10. Transfer the Maintenance Image BETTERIMAGE back to a Boot Image (now
recognizable by ff_an_adm.pl).
11. Remove the Solaris Application Node you used in Maintenance.
12. Create configuration information for this Solaris Application Node: just like #5 but
now on the BETTERIMAGE instead of on the GOODIMAGE.
13. Configure this Solaris Application Node.
14. Boot this Solaris Application Node.
Now the Maintenance cycle is actually done but we have to switch the Solaris Application
Nodes onto the new maintained Image.
15. Set by help of ff_an_adm.pl and ff_new_an.sh all of the Solaris Application
Nodes of the group to the BETTERIMAGE. Just reboot them and all of them contain
the modifications you made during maintenance.
2. Save the configuration data of this Solaris Application Node. This is helpful since we
need most of the configuration data in the next steps. In addition we might want to
bring up this Solaris Application Node after the Maintenance Cycle with the same
Internet Address, group, pool and so on.
control1: ff_an_adm.pl --op list --name kid20_4
--cmdline > kid20_4.config
For our task the last line of the generated file kid20_4.config is the most
important: it contains the command line used to configure the node.
3. Remove this Solaris Application Node. This allows the use of the configuration data
of kid20_4 during maintenance cycle.
4. Create a so called Maintenance Image out of the Boot Image. The command to do
that is nb_boot2maintenance (for details and example output refer to the
respective chapter in the User Guide). The nb_boot2maintenance needs two
parameters to work: The Boot Image we like to maintain and a target directory. The
latter will get the Maintenance Image. Please note that it must not exist and should
get a meaningful name: It must not start with FSC but should start with
MAINTENANCE and reflect the date or version or other tags we need to identify it. We
choose /FlexFrame/volFF/os/Solaris/MAINTENANCE_5.9_904_20060704.
control1: ./nb_boot2maintenance
-s /FlexFrame/volFF/os/Solaris/FSC_5.9_904_20060607/
bi_FJSV,GPUZC-M_PW-P
-d /FlexFrame/volFF/os/Solaris/MAINTENANCE_5.9_904_20060704
5. Create configuration information for the Solaris Application Node during its life in the
Maintenance cycle. The command to do this is again ff_an_adm.pl. The command
line to do this looks complicated but most of it is contained in the saved file
kid20_4.config. Just pick the last line but change the following:
At first the ospath parameter to our Maintenance Image path and at second the
name parameter to maintenance:
8. Since the Solaris Application Node runs on a separate Image with writable
filesystems the real maintenance can start. We can apply patches, install or upgrade
software, modify several configurations or change other settings on the system. We
should carefully check the system after applying the changes and respect the
warning we read in the “Introduction” section.
Any change we do will remain in the image i.e. any Solaris Application Node
which will run on that image later on will "inherit" everything. This is good for
sure in case of intended changes. But keep an eye on unwanted log files,
notes, and so on. Clean them up.
9. After the test run of the Solaris Application Node on the Maintenance Image we have
to shut it down again. Please do this using the console of the Solaris Application
Node:
maintenance-st console # init 0
10. Now we have to transfer the Maintenance Image into a Boot Image. The command
to do that is nb_maintenance2boot (for details and example output refer to the
respective chapter in the User Guide). This is the counterpart to
nb_boot2maintenance and transfers the Maintenance Image into a Boot Image
by moving appropriate contents and cleaning up the Image. The command reads the
well known configuration file:
control1: ./nb_maintenance2boot
-r /tftpboot/config/netboot_pool1_maintenance.cfg
11. Now we will bring up our former Solaris Application Node up again and therefore
have to release its configuration at first: Remove the Solaris Application Node - the
Node in Maintenance state is no longer needed.
control1: ff_an_adm.pl --op rem --name maintenance
12. Create configuration information for the Solaris Application Node kid20_4 in order
to bring it back to life. The command to do this is again ff_an_adm.pl. The
command line to do this looks complicated but most of it is contained in the saved file
kid20_4.config. Just pick the last line but change the ospath parameter to our
Maintenance Image path:
control1: ff_an_adm.pl --op add --name kid20_4 --type PW250
--pool pool1 --group test1 --swgroup 1
--mac 00:E0:00:C5:46:A0,00:E0:00:A6:F4:7B
--ospath Solaris/MAINTENANCE_5.9_904_20060704
--host 129,130,131
14. Now we are able boot the Solaris Application Node but we have to consider: The test
run we did was a test run without the specific FlexFrame software. The
ff_new_an.sh -n kid20_4 call enabled the FlexFrame functions. Therefore after
coming up the FlexFrame (FA Agents) will recognize the Application Node. In order
to prevent this new Application Node from being treated as a spare node it is wise to
hide the Node. Remove the respective link:
control1: cd /FlexFrame/volFF/os/Solaris/MAINTENANCE_5.9_904
_20060704/bi_FJSV,GPUZC-M_PW-P/root/kid20_4-st/etc/rc3.d
control1: mv S20myAMC.FA_AppAgent _S20myAMC.FA_AppAgent
The kid20_4 system will come up and will contain all the changes we made during
maintenance. It runs in the FlexFrame environment without serving as a spare node. If
this works now without any problems as well we are done.
The Maintenance cycle is actually finished but we have to switch the Solaris Application
Nodes onto the new maintained Image.
Up to this point the Solaris Application Nodes can still operate on the previous
version of the Solaris Image. Finally we need only to reboot the Solaris
Application Nodes to change their Image from the previous to the new
maintained Image.
And now we can activate the kid20_4 as a fully functional FlexFrame Solaris Application
Node by moving back the link and rebooting the system:
control1: cd /FlexFrame/volFF/os/Solaris/MAINTENANCE_5.9_904
_20060704/bi_FJSV,GPUZC-M_PW-P/root/kid20_4-st/etc/rc3.d
control1: mv _S20myAMC.FA_AppAgent S20myAMC.FA_AppAgent
control1: ssh kid20_4 init 6
Since all the Solaris Application Nodes of a group must have the very same
software it is recommended to bring up the Solaris Application Nodes on the
new maintained Image in a short period of time..
8.10 Troubleshooting
The messages can be turned off by removing the file .trace.sd, e.g. like:
control1:~ # rm /FlexFrame/volFF/os/Solaris/
FSC_5.8_202_20050211/bi_FJSV,GPUZC-M_PW-P/root/
an_0800-st/etc/.trace.sd
an_Solaris# rm /usr/testfile
Check /usr in the vfstab and the Filer’s exports. Check that it is actually exported via
exportfs. Check /etc/hosts.
Synopsis
ff_pool_adm.pl
--op add --name <pool_name>
--storage <vlan_id>,<network_ip>,<netmask>
--server <vlan_id>,<network_ip>,<netmask>
--client <vlan_id>,<network_ip>,<netmask>
--dns <domain_name>,[<dns_server_ ip>]
[--sapdata <nas_name>[,<volume_path>]]
[--saplog <nas_name>[,<volume_path>]]
[--volff <nas_name>[,<volume_path>]]
[--volff_common <nas_name>[,<volume_path>]]
[--defrouter <default_router_ip>] [--switchgrp <id>[,<id>]]
Command Options
--op add
Adds a pool.
--name <pool_name>
Name of new pool (has to be unique within entire FlexFrame). We recommend using
short lowercase names for <pool_name>.
--storage <vlan_id>,<network_ip>,<netmask>
Pool specific storage network segment. The option is followed by a comma
separated list of the VLAN ID, the network IP and the netmask of the network IP
address.
--server <vlan_id>,<network_ip>,<netmask>
Pool specific server network segment. The option is followed by a comma separated
list of the VLAN ID, the network IP and the netmask of the network IP address.
--client <vlan_id>,<network_ip>,<netmask>
Pool specific client network segment. The option is followed by a comma separated
list of the VLAN ID, the network IP and the netmask of network IP.
--dns <domain_name>,[<dns_server_ip>]
DNS domain name and servers to be used for this pool. More than one DNS server
IP address may be given. Keep in mind to use the default router option if the server
IP addresses do not match any of the pool networks. A DNS option may look like
this: my.domain.com,192.168.251.17,192.168.251.18
--sapdata <nas_name>[,<volume_path>]
Optional NAS name and volume path the pool should use for sapdata. A missing
volume path is auto-filled with the default (/vol/sapdata/<pool_name>). e.g.
filer1,/vol/sapdata/pool1. The entire option defaults to common NAS name
with default path. <nas_name> is the Filer's node name for this pool (without
-st suffix).
--saplog <nas_name>[,<volume_path>]
Optional NAS name and volume path the pool should use for saplog. A missing
volume path is auto filled with the default (/vol/saplog/<pool_name>). e.g.
filer1,/vol/saplog/pool1. The entire option defaults to common NAS name
with default path. <nas_name> is the Filer's node name for this pool (without
-st suffix).
--volff <nas_name>[,<volume_path>]
Optional NAS name and volume path the pool should use for volFF. A missing
volume path is auto filled with the default (/vol/volFF/pool-<pool_name>). e.g.
filer1,/vol/volFF/pool-pool1. The entire option defaults to common NAS
name with default path. <nas_name> is the Filer's node name for this pool (without
-st suffix).
--volff_common <nas_name>[,<volume_path>]
Optional NAS name and volume path the pool should use for common volFF data. A
missing volume path is auto filled with the default (/vol/volFF). e.g.
filer1,/vol/volFF. The entire option defaults to common NAS name with default
path. <nas_name> is the Filer's node name for this pool (without -st suffix).
Currently it has to be the first Filer of the FlexFrame landscape and
/vol/volFF/pool-<pool_name>.
--defrouter <default_router_ip>
The default router is a gateway to route IP data to other, non-pool local networks. All
IP data that can not be addressed to a local network will be sent to the default router
to be forwarded to the destination network. The option parameter is an IP address of
this default router. Keep in mind to use a default router IP address matching one of
the local pool networks, because otherwise it will not be accessible by Application
Nodes.
--switchgrp <id>[,<id>]
The switch group ID(s) the Client LAN to corporate LAN ports should be configured.
If not given, the client VLAN is assigned to the existing trunk ports or a new port pair
at first both switch groups. Not more than two switch group IDs are accepted.
Command Output
The command displays information about processing steps, errors and warnings. The
output may look like this:
cn1:~ # ff_pool_adm.pl --op add --name pool4
--storage 30,192.168.30.0,255.255.255.0
--server 31,192.168.31.0,255.255.255.0
--client 32,192.168.32.0,255.255.255.0 --sapdata filer
--saplog filer --volff filer --volff_common filer
--dns my.domain.com --defrouter 192.168.32.254
update LDAP
..........................
..........................
..........................
..........................
..........................
..........................
.....
update switch 1/1 configuration
Notice: Update will take about 1 minute.
vlan: storage-30 has been created
restart cluster service ldap_srv1
Notice: restart will take up to 1 minute.
stop and wait until service is offline
start and wait until service is online
restart cluster service ldap_srv2
Notice: restart will take up to 1 minute.
stop and wait until service is offline
start and wait until service is online
restart cluster service netboot_srv
Notice: restart will take up to 2 minutes.
stop and wait until service is offline
start and wait until service is online
Synopsis
Command Options
--op rem
Removes a pool.
--name <pool_name>
Name of pool to be removed. Use ff_pool_adm.pl --op list-all to get a list
of currently configured pools (see 8.11.4).
Command Output
The command displays only errors and warnings. The output may look like this:
cn1:~ # ff_pool_adm.pl --op rem --name pool4
update LDAP
..........................
..........................
..........................
..........................
..........................
..........................
Synopsis
Command Options
--op list
Lists pool details
--name <pool_name>
Name of pool to be listed. Use ff_pool_adm.pl -op list-all to get a list of
currently configured pools (see 8.11.4).
--list <part>[,<part>]
To reduce the output to the interesting parts, use this option. The parameters to this
option are the display sections. Add them as a comma separated list. The default
sections are: network,storage,dns,group,sid,an,cn,filer. You may also use a two
character abbreviation instead the full section name like ne for network.
Command Output
The command displays the pool configuration details. The output may look like this:
cn1:/opt/FlexFrame/bin # ff_pool_adm.pl --op list --name p1
Pool configuration details of pool p1
Networks
Client-LAN
Network: 10.10.10.0 Netmask: 255.255.255.0 VLAN ID: 100
Server-LAN
Network: 192.168.10.0 Netmask: 255.255.255.0 VLAN ID: 110
Storage-LAN
Network: 192.168.20.0 Netmask: 255.255.255.0 VLAN ID: 120
Def.Router: 192.168.10.254
Storage Volumes
sapdata fas01-p1-st:/vol/sapdata/p1
saplog fas01-p1-st:/vol/saplog/p1
volFF fas01-p1-st:/vol/volFF/pool-p1
volFF shared fas01-p1-st:/vol/volFF
DNS data
Domain Name: my.domain.com
Pool Groups
Linux
OS: SuSE Linux SLES-9.X86_64
Solaris
OS: Sun SunOS 5.8
Application Nodes
PW250-1
Type: PW250
OS: Sun SunOS 5.8
Group: Solaris
Client-LAN PW250-1 10.10.10.33
local test1 PW250-1-clt1 10.10.10.34
blade03
Type: BX600 Cabinet ID: 1 Slot/Partition ID: 3
OS: SuSE Linux SLES-9.X86_64
Group: Linux
Client-LAN blade03 10.10.10.25
Server-LAN blade03-se 192.168.10.25
Storage-LAN blade03-st 192.168.20.25
pw250-2
Type: PW250
OS: Sun SunOS 5.8
Group: Solaris
Client-LAN pw250-2 10.10.10.36
local test1 pw250-2-clt1 10.10.10.37
local test2 pw250-2-clt2 10.10.10.38
Server-LAN pw250-2-se 192.168.10.36
local test1 pw250-2-set1 192.168.10.37
local test2 pw250-2-set2 192.168.10.38
Storage-LAN pw250-2-st 192.168.20.36
local test1 pw250-2-stt1 192.168.20.37
local test2 pw250-2-stt2 192.168.20.38
rx801
Type: RX800
OS: SuSE Linux SLES-9.X86_64
Group: Linux
Client-LAN rx801 10.10.10.2
Server-LAN rx801-se 192.168.10.2
Storage-LAN rx801-st 192.168.20.2
Control Nodes
cn1
Client-LAN cn1-p1 10.10.10.10
Server-LAN cn1-p1-se 192.168.10.10
Storage-LAN cn1-p1-st 192.168.20.10
cn2
Client-LAN cn2-p1 10.10.10.11
Server-LAN cn2-p1-se 192.168.10.11
Storage-LAN cn2-p1-st 192.168.20.11
Filer Nodes
fas01-p1
Storage-LAN fas01-p1-st 192.168.20.14
Networks
Client-LAN
Network: 10.10.10.0 Netmask: 255.255.255.0 VLAN ID: 100
Server-LAN
Network: 192.168.10.0 Netmask: 255.255.255.0 VLAN ID: 110
Storage-LAN
Network: 192.168.20.0 Netmask: 255.255.255.0 VLAN ID: 120
Def.Router: 192.168.10.254
Pool Groups
Linux
OS: SuSE Linux SLES-9.X86_64
Solaris
OS: Sun SunOS 5.8
Synopsis
Command Options
--op list-all
Displays all configured pools.
--list <part>[,<part>]
To reduce output to interesting parts use this option. The parameters to this option
are the display sections. Add them as a comma separated list. The default sections
are: network,storage,group,sid,cn,filer. You may also use a two character
abbreviation instead the full section name like ne for network.
Command Output
The command displays the pool configuration details. The output may look like this:
cn1:/opt/FlexFrame/bin # ff_pool_adm.pl --op list-all
Pool configurations
p1
Pool Networks
Client-LAN
Network: 10.10.10.0 Netmask: 255.255.255.0 VLAN ID:
100
Server-LAN
Network: 192.168.10.0 Netmask: 255.255.255.0 VLAN ID:
110
Storage-LAN
Network: 192.168.20.0 Netmask: 255.255.255.0 VLAN ID:
120
Pool Groups
Linux
OS: SuSE Linux SLES-9.X86_64
Solaris
OS: Sun SunOS 5.8
Pool SIDs
D01
SAP Version: SAP-6.20 DB Version: Oracle-9
P01
SAP Version: SAP-6.20 DB Version: Oracle-9
Q01
SAP Version: SAP-6.20 DB Version: Oracle-9
cn2
Client-LAN cn2-p1 10.10.10.11
Server-LAN cn2-p1-se 192.168.10.11
Storage-LAN cn2-p1-st 192.168.20.11
p1
Pool Groups
Linux
OS: SuSE Linux SLES-9.X86_64
Solaris
OS: Sun SunOS 5.8
Pool SIDs
D01
SAP Version: SAP-6.20 DB Version: Oracle-9
P01
SAP Version: SAP-6.20 DB Version: Oracle-9
Q01
Synopsis
Command Options
--op add
Adds a group to a pool.
--pool <pool_name>
Name of pool the group should be added. We recommend using short lowercase
names for <pool_name>.
--group <group name>
The name of the group to be added.
--ostype {Linux|SunOS}
Type of operating system (OS) the systems of this group work with. Currently only
Linux and SunOS (Solaris) are valid choices.
--osversion <version_string>
The version of the OS. Can be omitted if only one image version (Linux or Solaris) is
installed.
--osvendor {Sun|SuSE}
The vendor of the OS. Currently only Sun (Solaris) and SuSE (Linux) are supported.
Can be omitted if only one image version (Linux or Solaris) is installed.
Synopsis
Command Options
--op rem
Removes a group from a pool.
--pool <pool_name>
Name of the pool the group should be removed from.
--group <group_name>
The name of the group to be removed.
Synopsis
Command Options
--op group
Changes pool group of Application Node.
--name <node_name>
Name of the Application Node the pool group has to be changed.
--group <group_name>
The name of the pool group the Application Node should be assigned to. Use
ff_pool_adm.pl --op list-all to get available pool groups (see 8.11.4).
You cannot remove names or addresses which are essential to the FlexFrame
landscape.
Each pool has its own hosts database. Therefore you have to maintain each
pool individually.
Synopsis
Command Options
-d debug. This option will log debug information which can be found in the file
/tmp/ff_hosts.DEBUGLOG
-l List all the hosts entries for the pool as provided by option -p.
-p <pool_name>
Pool where the host name should be added to.
-a <ip>
Add the IP addess <ip>. Has to be used togehter with option -n. If an entry with the
IP address <ip> already exists, the name provided will be added as an alias.
-n <name>
Host name <name> will be added to the list.
-r <name>
Deletes the host name or alias of name. The host name cannot be deleted if it has
additional aliases. Remove the aliases first.
Debugging
/tmp/ff_hosts.DEBUGLOG will hold debugging information. In case of problems,
please provide this file.
Examples:
The following example will list all additional hosts entries for pool poolname created
using this tool:
cn1:~ # ff_hosts.sh -l -p poolname
The following example will add a host newhost with the IP address 1.2.3.4 to the
pool poolname:
The following example will remove the hosts entry for newhost:.
Synopsis
ff_reboot_all_an.sh
Debugging
/tmp/ff_reboot_all_an.DEBUGLOG will hold debugging information. In case of
problems, please provide this file.
Please note that the passwords are displayed on the screen – the input will not
be hidden!
/opt/FlexFrame/bin/ff_ssh_tool.sh –i
You will be asked to enter the passwords of both Control Nodes.
cn1:~ # ff_ssh_tool.sh -i
Copyright (C) 2005 Fujitsu Siemens Computers. All rights reserved.
Continue? yes
In order to change the password for the LDAPadmins use -p LDAPadmins instead of
providing a pool name, and provide admins or replicate as the user option's value. This
will change passwords of admins or replicate LDAP users.
In cases where existing SAP systems are transfered into FlexFrame or recovered from an
earlier installation, it might be necessary to change user or group IDs to different values
than the ones provided by the installation tools. In order to do that, ff_change_id.pl can
be utilized:
cn1:~ # vi /etc/shadow
cn1:~ # passwd
Changing password for root.
New password:
Password will be truncated to 8 characters
Re-enter new password:
Password changed
The user passwords of the Control Nodes may (but should not) differ and can be
changed at any moment. A change has no other impact.
Modify the password on both, the root and the replica of the LDAP server.
The passwords can be changed at any moment. A change has no other impact.
If the LDAP server and its replica are not rebooted both the old password and the new
password will be valid. After rebooting both only the new password will be valid.
The LDAP servers should be rebooted from the PRIMECLUSTER Interface (ldap_srv1
and ldap_srv2) .
The password can be changed at any moment. A change has no other impact.
Modify this password with an editor. The user control must not be changed. This
password must be identical for user control in all SAPDBs/MaxDBs across the site.
The password can be changed at any moment. A change has no other impact.
Navigate to the switch group where you want to change the password:
telnet session:
enable
configure terminal
enable password <new_password>
line vty 0 15
password <new_password>
end
copy running config startup-config
3. On Control Node:
cd /tftpboot
touch sw<n>-<m>.config
chmod 666 sw<n>-<m>.config
4. On switch:
copy running config tftp <control_lan_ip_of_CN>
sw<n>-<m>.config
telnet session:
enable
configure
enable password level 0 <new passwd>
line vty
password <new password>
end
copy running config startup-config
3. On control node:
cd /tftpboot
touch bx[3,6]00-<n>-swb<m>.config
chmod 666 bx[3,6]00-<n>-swb<m>.config
4. On switch:
copy running config tftp <control lan ip fo CN>
bx[3,6]00-<n>-swb<m>.config
Synopsis
Description
● Setting up the root password in the root image
You need a password to log onto the Application Nodes as root. This password is
valid for Application Nodes with the same shared root Image. The script asks for the
root password twice. If both the entered strings are the same, the password will be
accepted.
Enter password: ********
Enter password again: ********
Password accepted.
If you plan to use more root images, for example a separate root image for
every pool with different root passwords, you must copy these root images
manually after this installation procedure.
● Setting up ssh in the root image
To log on from the Control Nodes to the Application Nodes without entering the root
password, set up ssh as follows:
– Generate host keys for the root image
– Deploy authorized keys from the Control Nodes into the root image
Synopsis
Description
Please enter the password for your Solaris Application Nodes. It applies for all Solaris
Application Nodes.
SWITCHTYPE=cat3750g-24ts
HOSTNAME=sw1-1
IP=100.100.13.18;255.255.255.0;14
SNMP=public;ro
SYSLOG=100.100.13.13;local0
PASSWORD=passwort;root
NTP=100.100.13.12;server
NTP=100.100.13.13;server
9.5 FA Agents
messengerdb.jdbc.password
messengerdb.jdbc.url=jdbc:mysql://localhost:3306/messenger
messengerdb.jdbc.username=myAMC
messengerdb.jdbc.password=FlexFrame
9.5.1.1 BX300/600
Default: <Username> = root, <Password> = root
Use the same user name, password and community string in config file
/opt/myAMC/vFF/vFF_<pool_name>/config/myAMC_FA_SD_Sec.xml, config
section Security_default or use a separate config section entry for this server. The
SNMP community string should be “read-write”.
9.5.1.2 RX300/RX300 S2
IPMI configuration:
● LAN A onboard should only be used for IPMI on the Control LAN
● LAN B onboard and the upper NIC of the separate LAN card should be used for
bonding.
Use the same user name and password in config file
/opt/myAMC/vFF/vFF_<pool_name>/config/myAMC_FA_SD_Sec.xml, config
section Security_default or use a separate config section entry for this server.
9.5.1.3 RX600
RSB configuration:
● RX600: RSB1 onboard
Connect the Control LAN cable with the LAN port for the remote management controller.
Use the same user name and password in config file
/opt/myAMC/vFF/vFF_<pool_name>/config/myAMC_FA_SD_Sec.xml, config
section "Security_default" or use a separate config section entry for this server.
Use the same community string in config file
/opt/myAMC/vFF/vFF_<pool_name>/config/myAMC_FA_SD_Sec.xml, config
section Security_default or use a separate config section entry for this server.
9.5.1.4 RX800
ASM, RSB and the connection between must be configured.
Set user name and password (same as in
/opt/myAMC/vFF/vFF_<pool_name>/config/myAMC_FA_SD_Sec.xml, config
section Security_default or separate entry for this server).
Set Permissions for Reset/SwitchOff to 1 and then press <F1>.
Use the same user name and password in config file
/opt/myAMC/vFF/vFF_<pool_name>/config/myAMC_FA_SD_Sec.xml, config
section Security_default or use a separate config section entry for this server.
Create new SNMP Community 2 for tcc and then press F1.
Use the same community string in config file
/opt/myAMC/vFF/vFF_<pool_name>/config/myAMC_FA_SD_Sec.xml, config
section Security_default or use a separate config section entry for this server.
9.5.2.1 General
The TrapTargets.xml file contains all the trap destinations, i.e. information which is
needed to send SNMP traps. Two parameters are required for each target:
● Host name or IP address
● SNMP community
9.6 PRIMECLUSTER
Read access with the PCL Web-based Admin View is possible with every Unix user
defined on the control nodes. Write access is only possible with root
See section “User Administration“on page 191.
Command Options
--op list
Determines the list operation.
--pool <pool_name>
Determines the FlexFrame pool to which the operation should be applied.
--sid <SAP_system_id>
Determines the SID being used.
Examples:
%> ff_sid_adm.pl –-op list –-pool Pan
OL1
OL4
OL2
SHT
Synopsis
ff_sid_adm.pl –-op add –-pool <pool_name> --sid <SAP_system_id>
--sapversion {4.6|6.20|6.40|7.0}
--db {ORACLE9|ORACLE10|SAPDB73|SAPDB74|MAXDB75|MAXDB76}
--group <groupname1>:<gidnumber1>,<groupname2>:<gidnumber2>,...
--user <username1>:<uidnumber1>,<username2>:<uidnumber2>,...
<db loghost>:{<db_loghost_ip>|*}
--sap {ci|app|jc|j|scs|ascs}:<SYSNR>:
<loghost-client>:{<loghost_client_ip>|*}
<loghost-server>:{<loghost_server_ip>|*}
Command Options
--op add
Determines the add operation.
--op del
Determines the del operation.
--pool <pool_name>
Determines the FlexFrame pool to which the operation should be applied.
--sid <SAP_system_id>
Determines the SID being used.
--sapversion {4.6|6.20|6.40|7.0}
Specifies the SAP basis version being used.
--db {ORACLE9|ORACLE10|SAPDB73|SAPDB74|MAXDB75|MAXDB76}
Specifies the database type as well as the respective version being used.
--group <groupname1>:<gidnumber1>,<groupname2>:<gidnumber2>,...
--user <username1>:<uidnumber1>,<username2>:<uidnumber2>,...
user and group enable specially selected user numbers and group numbers to be
assigned to SAP users and SAP groups respectively. In this case a check is made to
see whether the user or group has already been defined for the DB system involved.
A user or group is created only if they do not already exist. For example, a group dba
which already exists cannot be assigned a group number which deviates from the
default value.
<db loghost>:{<db_loghost_ip>|*}}
The logical host name is used for the database as well as the IP address for that host
name. Use an asterisk if you want it to be chosen automatically. All the entries need
to be specified in a colon separated format.
--sap {ci|app|jc|j|scs|ascs}:<SYSNR>:
<loghost>-client:{<loghost client ip>|*}
<loghost>-server:{<loghost server ip>|*}
Specifies an SAP instance (optionally multiple of those) through its type (ci, app, jc,
j, scs, ascs), its SAP system number, the logical host name in the client network,
the respective IP address, the logical host name in the server network and the
respective IP address. Again, the IP addresses can be replaced with asterisks in
order to have them chosen automatically. All the entries need to be specified in a
colon separated format.
--sysnr <SYSNR>
Removes a specific SAP instance instead of the entire system (SID).
Examples:
Adding an SID with one Central Instance:
control1:~ # ff_sid_adm.pl –-op add –-sid SHT –-pool Otto
--sapversion 6.40 –-db ORACLE9:dbsht:192.168.1.1
--sap ci:00:sht00-client:\*:sht00-server:\*
Synopsis
Synopsis
ff_change_id.pl --pool=<pool_name>
[{--uid=<id> <user_name>|--gid=<id> <group_name>}]
Example:
ff_change_id.pl --pool=pool1 --uid=501 orasht
The following directories have to be copied from the source pool to the target pool:
/FlexFrame/volFF/pool-<pool_name>/<dbtype>/<OS>/<SID>
/FlexFrame/volFF/pool-<pool_name>/sap/sapmnt/<SID>
/FlexFrame/volFF/pool-<pool_name>/sap/usr_sap/<SID>
/FlexFrame/volFF/pool-<pool_name>/sap/home_sap/<sid>adm
/FlexFrame/sapdata/<pool_name>/<SID>
/FlexFrame/saplog/<pool_name>/<SID>
/FlexFrame/volFF/pool-<pool_name>/oracle/<OS>/client/*
(Only necessary if the Oracle client software is not already installed in the target pool.)
control1:/ # cd /FlexFrame/volFF/pool-p1/oracle/SunOS
control1:/FlexFrame/volFF/pool-p1/oracle/SunOS # cp -r OSI
../../../pool-p2/oracle/SunOS
control1:/FlexFrame/volFF/pool-p1/oracle/SunOS # cd ../../sap
control1:/FlexFrame/volFF/pool-p1/sap # cp -r sapmnt/OSI
../../pool-p2/sap/sapmnt
control1:/FlexFrame/volFF/pool-p1/sap # cp -r usr_sap/OSI
../../pool-p2/sap/usr_sap
control1:/FlexFrame/volFF/pool-p1/sap # cp -r home_sap/osiadm
../../pool-p2/sap/home_sap
control1:/FlexFrame/volFF/pool-p1/sap # cd /FlexFrame/sapdata/p1
control1:/FlexFrame/sapdata/p1 # cp -r OSI ../p2
control1:/FlexFrame/sapdata/p1 # cd ../../saplog/p1
control1:/FlexFrame/saplog/p1 # cp -r OSI ../p2
control1:/ # cd /FlexFrame/volFF/pool-p1/sapdb/Linux
control1:/FlexFrame/volFF/pool-p1/sapdb/Linux # cp -r MLI
../../../pool-p2/sapdb/Linux
control1:/FlexFrame/volFF/pool-p1/sapdb/Linux # cd ../../sap
control1:/FlexFrame/volFF/pool-p1/sap # cp -r sapmnt/MLI
../../pool-p2/sap/sapmnt
control1:/FlexFrame/volFF/pool-p1/sap # cp -r usr_sap/MLI
../../pool-p2/sap/usr_sap
control1:/FlexFrame/volFF/pool-p1/sap # cp -r home_sap/mliadm
../../pool-p2/sap/home_sap
control1:/FlexFrame/volFF/pool-p1/sap # cp -r home_sap/sqdmli
../../pool-p2/sap/home_sap
control1:/FlexFrame/volFF/pool-p1/sap # cd /FlexFrame/sapdata/p1
control1:/FlexFrame/sapdata/p1 # cp -r MLI ../p2
control1:/FlexFrame/sapdata/p1 # cd ../../saplog/p1
control1:/FlexFrame/saplog/p1 # cp -r MLI ../p2
Synopsis
ff_change_id.pl --pool=<pool_name>
[{--uid=<id> <user_name>|--gid=<id> <group_name>}]
Example:
# ff_change_id.pl --pool=pool1 --uid=501 orasht
Here, dataC11 and logC11 are the new names of the volumes and 4 and 10 are the
numbers of disks to be used.
We recommend using the volume names sapdata and saplog (if on a different
Filer) or data<SID> and log<SID> if SID specific volumes on the same Filer.
Next, create qtrees for each FlexFrame pool which will store data and logs in those
volumes:
filer2> qtree create /vol/dataC11/pool1
filer2> qtree create /vol/logC11/pool1
If you use more than the first Filer, make sure it is reachable using, e.g.:
control1:~ # ping filer2-st
PING filer2-st (192.168.10.203) from 192.168.10.201 : 56(84) bytes
ofdata.
64 bytes from filer2-st (192.168.10.203): icmp_seq=1 ttl=255
time=0.117 ms
64 bytes from filer2-st (192.168.10.203): icmp_seq=2 ttl=255
time=0.107 ms
64 bytes from filer2-st (192.168.10.203): icmp_seq=3 ttl=255
time=0.103 ms
Now, the LDAP database has to be told that this SID is not using the default (first) Filer
and sapdata/saplog volumes:
control1:~ # ff_sid_mnt_adm.pl --op=add --pool=pool1 --sid=C11
--sapdata=filer2:/vol/dataC11/pool1/C11
--saplog=filer2:/vol/logC11/pool1/C11
Now, the volumes on the Control Nodes need to be mounted. To do so you should add
the following lines to each Control Node's /etc/fstab:
Repeat the sapdata and saplog-lines for each pool, if there's more than one pool
for those volumes.
Use the volume name for the last directory in the mount point.
/vol/dataC11/pool1 -sec=sys,rw=192.168.10.0/24,anon=0
/vol/logC11/pool1 -sec=sys,rw=192.168.10.0/24,anon=0
Now you need to make the file re-read its configuration file:
control1:~ # rsh filer2-st exporfs -a
Synopsis
Command Options
-d Writes debugging information to a log file (see below).
-l Lists all SAP shadow service ports of the pool provided with the -p option. If the
option -s is used, only the port of the specified SID is displayed.
-a Adds an entry.
-r Removes an entry.
-p <pool_name>
Specifies the name of the pool (e.g. pool1).
-s <sid>
Specifies the SAP System ID (SID) by a 3 character string (e.g. C11).
-o <port_no>
Specifies the service port number. The default is 3694 (optional).
Debugging
/tmp/ff_sap_shadowport.DEBUGLOG holds debugging information if option -d was
used. In case of problems, please provide this file.
10.5.2 FA Agents
Please make sure that the FA Application Agents are stopped on the hosts while you are
installing, updating or removing any SAP or database software:
Stop the FA Agent:
/etc/init.d/myAMC.FA_AppAgent stop
Node states
RUNNING
SWITCH INT
SWITCH EXT
PowerOff
Service states
SHUTDOWN
DOWN
WATCH
NOWATCH
UNKNOWN
NULL
RUNNING
STOPPING
STARTING
REBOOT
RESTART
RESTARTING
REBOOT
REBOOTING
RBGET
SWITCH
SWITCHOVER
SWGET
ERROR
If you don't want to configure the parameters in the plain xml files, you may use the FA
WebGUI to do this more conveniently. Details are provided in the “FA Agents –
Installation and Administration“ manual.
Synopsis
Command Options
-i Shows inactive services
-v Verbose mode
The application or service must not be started directly, e.g. for an SAP
instance as <sid>adm with startsap, since in this case the interfaces are
neither supplied with IP addresses, nor is the service control file maintained.
The started application will not work due to the lack of a network connection.
The call syntax for sapdb, sapci, sapscs, sapascs and sapjc is:
sapdb <sid> <action>
sapci <sid> <action>
sapscs <sid> <action>
sapascs <sid> <action>
sapjc <sid> <action>
<sid>
System ID (SID), 3-digit, in lower case.
<action>
The action to be performed with the application or service. Actions are start, stop,
restart, status, cleanup, watch and nowatch.
Call the following from the Control Node, using ssh [-t] with reference to the example
of sapapp:
ssh <application_node_name> sapapp <ID> <SID> <action>
Cleanup
This kills application processes that are still running and deletes occupied resources
such as shared memory, semaphores and the message queue.
Note that this action may only be performed after stopping the application has failed.
Nowatch
This removes the application from monitoring by the high-availability software (FA
Agent) without the application having to be restarted. The application itself retains its
current status.
Watch
This includes the application again into monitoring by the high-availability software
(FA Agent) without the application having to be restarted. The application itself
retains its current status.
Control1: # ln -s /FlexFrame/pooldata/config/scripts_config/
user_script /FlexFrame/scripts/user_script
If no pool dependent function is needed, the script user_script can be located directly
in /FlexFrame/scripts.
For actions like start, stop, nowatch and watch, the exit code should be no_error
normally. For the action status, the exit code is dependent from various factors:
● The installation check against the SAP service may have failed
(plausibility_error)
● the SAP service is possibly not running or not running on this host (any_error)
● the SAP service is running on this host (no_error).
In addition to these exit codes, the SAP service scripts print out messages as described
in the section “Start/Stop Script Errors” on page 283.
stop_all_sapservices_local
Shutdown script for all running sap services on one Application Node. This script will
be integrated via runlevels. The information base for these
stop_all_sapservices* scripts are the /FlexFrame/scripts/log/*_host
files.
Example:
The central instance of BW1 is to be included into monitoring by the high-availability
software while running:
blade1 # sapci bw1 watch
● Remove the application from monitoring by the high-availability software and run it.
● r3up can now shut down and start the application directly with stopsap and
startsap; the IP addresses remain assigned to the interfaces and the host name is
retained for <sid>adm without any changes on the correct virtual host name.
● Include the application again into monitoring by the high-availability software after
the upgrade.
The output should be means that the script view_hosts does not really check the
state of the SAP services. It only checks the FlexFrame/script/log/*_host files
which are created immediately after a successfully start of the corresponding SAP
service.
For a detailed status information you can use:
control1 # ssh klinge4 sapci ol4 status
The exit code from our SAP service script will be “0” in this case. For any other errors
while executing our SAP service script, it will return an exit code not equal “0”.
control1:/ # ssh klinge1 sapci ol4 status
Central-Instance OL4 should be running on this host.
Central-Instance OL4 has really running processes on this host.
client_lan interface vlan2002:1 10.1.7.151 is already up.
server_lan interface vlan2006:2 172.16.4.151 is already up.
ciol4: klinge1 /FlexFrame/scripts/sapci done!
ciol4: exit_code=0
And now start the stopped Central Instance OL4 on the new destination klinge5:
Assumption
Running ACC and SolMan services.
SolMan
Transaction smsy_setup started (see chapter “Activate and check Transfer from SLD to
Solution Manager”).
Making services adaptive (see chapter “Integration of SAP Services”).
ACC
Click on refresh button.
New services should be visible in the ACC WebGUI.
13.1 Groups
13.1.1 General
FlexFrame offers advanced functions for partitioning a FlexFrame environment into
service or customer specific server pools and groups. This may be interesting for large
installations or application service providers.
Server pools
On FlexFrame 3.2, a pool is a number of Application Nodes belonging to the same
department (or customer) with exclusive hardware requirements. FlexFrame systems can
be divided into pools. Each FlexFrame system consists of at least one pool. Within a
pool, all servers may communicate with each other, but not with the systems of other
pools.
Servers of different pools can use different copies of the OS and can be separated into
different network segments.
The autonomous reactions of FlexFrame Autonomy are always pool-related. Pools can
be configured in LDAP by using the FF administration tools. Details about the pooling
within FA Agents are provided in the “FA Agents – Installation and Administration“
manual. The FA configuration file myAMC_FA_Pools.xml is used as a cache file in the
case LDAP is not responding.
Server groups
Within a server pool, various types of hardware can be used with different characteristics,
such as operating system and architecture, number of CPUs and RAM size. The bulk of
servers can be divided into groups (pool groups) of servers with similar operating
systems and hardware performance. This may be very useful for groups of high-
performance database servers or groups of medium-performance application servers.
Each pool consists of at least one group.
Each SAP application running in a pool can use one or more servers in one or more
groups of servers in the same pool. Each instance of this application runs in a selected
group. In case of a failure, switchover to a spare server is possible in the same group.
For example, a pair of servers can be divided into a high-performance database server
and a smaller application server. The group configuration makes sure that database
instances run in the group of high-performance servers, while application instances stay
on groups of smaller servers without interfering with each other.
All servers in a pool share the same VLAN network segments, even if they belong to
different groups. They do not share VLAN segments with other pools, except the Control
VLAN.
13.2 Traps
The FA Agents are able to send SNMP traps to configured trap destinations. The traps
contain status information respectively changes of services in the FlexFrame
environment. For example there is an SNMP trap if a service is starting. SNMP traps can
be used to connect an external system monitoring application to FlexFrame.
13.2.1 General
The TrapTargets.xml file contains all the trap destinations, i.e. information which is
needed to send SNMP traps. Two parameters are required for each target:
● Host name or IP address
● SNMP community
The community roughly corresponds to a password.
Generally public is configured as the default value.
Details are provided in the ”FA Agents – Installation and Administration“ manual.
ControlAgentTime
This parameter specifies how often the Control Agent checks the Livelists of the
Application Agents. The parameter should thus be about the same as the
LiveListWriterTime.
MaxHeartbeatTime
This parameter specifies the maximum time which may elapse between two Livelist
entries of an Application Agent before the Control Agent intervenes. The intervention
is a check if the Application Node is alive. If not, there will be an external switchover
to another Application Node. The MaxHeartbeatTime must therefore always be
greater than the ControlAgentTime and the LivelistWriterTime. In practice
the factor of 3 between LivelistWriterTime and MaxHeartbeatTime has
proved practical.
MaxRebootTime
This parameter specifies the maximum time which may elapse between two Livelist
entries of an Application Agent before the Control Agent intervenes during a reboot of
this Application Node.
MaxFailedReachNumber
Number of tries by the FA Control Agent to reach the Application Node after the
application has exceeded the MaxHeartbeatTime. After this number of tries, an
external switchover is initiated. This parameter specifies how often the Control Agent
attempts to reach a node after the MaxHeartbeatTime has been exceeded before
an external switchover is initiated.
Node_PowerDownTime
This parameter specifies the maximum time a Control Agent will wait for a node to
shut down before the services are started on a spare node by means of a switchover.
Node_CheckAvailabilityTime
This parameter specifies the maximum time a Control Agent will wait for complete
execution of the Node_CheckAvailabilityCommand to be completed. If no
positive acknowledgment from the Application Agent is received in that time, the
node is regarded as unavailable.
Node_SendTrapsAllowed
This parameter releases or blocks the sending of node traps.
Node_RebootCommand
This parameter specifies which command is executed when the Application Agent
initiates a reboot. Normally this is a shutdown with a subsequent reboot.
Node_ShutdownCommand
This parameter specifies which command is executed when the Application Agent
initiates a switchover. Normally this is a shutdown without a subsequent reboot.
Node_PowerDownCommand
This parameter specifies which command is executed by the Control Agent before an
external switchover is initiated. In this way it is ensured that the services on the node
being switched over are really stopped and can be taken over without any problem
by other nodes. The Control Agent waits at most for the period specified with the
Node_PowerDownTime parameter before it continues with the switchover.
Node_CheckAvailabilityCommand
This parameter specifies which command is executed by the Control Agent to check
the availability of a node. A return value of 0 is interpreted as a positive result, every
other return value as negative. The Control Agent waits at most for the period
specified with Node_CheckAvailabilityTime. If the command has not been
executed completely by then, it is assumed that the test is negative, i.e. the node is
no longer available, resulting in an external switchover.
Node_RemoteExecutionCommand
This parameter specifies which command the Control Agent puts ahead of a
command to be executed on another node. This is used, for example, to start or stop
a service remotely on an Application Node. Normally ssh is used here.
Service_ReactionDelayTime
This parameter interacts directly with CheckCycleTime. It can be set individually for
each service type. This time defines how long the triggering of a reaction is delayed
after a failure has been detected.
Examples:
CheckCycleTime = 10 sec; ServiceReactionDelayTime = 30 sec
In this example a failed service is detected in a cycle. However, the reaction only
takes place after 30 seconds. The failure must therefore have been identified as a
failure over at least three detection cycles. This allows preventing a detection error
resulting in an incorrect reaction.
CheckCycleTime = 10 sec; ServiceReactionDelayTime = 0 sec
In this example the required reaction takes place immediately in the cycle in which
the problem was detected.
Service_MaxRestartTime
This parameter defines the maximum time which may be required for a service type
in the event of a restart. If this time is exceeded, a second or further attempt is made
in accordance with Service_MaxRestartNumber. Thus if the time is selected to
short for the service to be monitored and the hardware used, i.e. the service requires
longer to restart than permitted by Service_MaxRestartTime, a problem situation
is triggered incorrectly.
Service_MaxStartTime
This parameter defines how long a service may take to start up. If this time is
exceeded, the Agent interprets the service as not started and initiates further
reactions.
Service_MaxStopTime
This parameter defines how long a service may take to stop. If this time is exceeded,
the Agent interprets the service as not stopped and initiates appropriate reactions.
Service_PingVirtualServiceInterface
This parameter defines whether the associated virtual FlexFrame service interface is
pinged to determine the availability of a service. If it is set to 0, the virtual LAN
interfaces of the client and server network are not queried. Interface availability then
has no influence on the status change of a service.
SAPScriptFilePath
This parameter specifies the directory in which the start and stop scripts for the SAP
services (sapdb, sapci, sapapp etc.) can be found. The default path
(/opt/myAMC/scripts/sap) is normally a symbolic link to the actual script
directory.
ControlFilePath
This parameter specifies the directory of the control files
(<service type><service ID><service SID>_host) generated by the
start/stop scripts.
BlackboardFilePath
This parameter specifies the directory in which the BlackBoard file can be found.
Commands can be entered in it that are executed by the FA Application Agents.
GroupConfigFile
This parameter specifies the file in which the group affiliation is configured.
13.4.1 General
The power shutdown concept of FlexFrame Autonomy provides an easy-to-configure
method for implementing secure shutdown of various hardware platforms. Various blade,
PRIMERGY and PRIMEPOWER systems (Midrange, Enterprise) can be installed
simultaneously in a FlexFrame. Each of these systems has different requirements which
must be taken into consideration for the power shutdown.
The FlexFrame Autonomy Control Agents make direct use of the Shutdown Agents from
the PRIMECLUSTER shutdown facility. These are part of the high-availability solution
PRIMECLUSTER (PCL) which is used on all Control Nodes of a FlexFrame system.
Normally, these agents are provided with their configuration information in the course of
the PCL configuration. However, in a FlexFrame solution only the Control Node PCLs are
configured; no configuration information on any Application Nodes exists in
PRIMECLUSTER.
The FlexFrame Autonomous Agents ascertain this lack of configuration information
automatically at runtime and then generate the configuration information required for the
agents.
Different configuration information is generated in accordance with the type of system for
which the power shutdown is performed.
13.4.2 Architecture
The figure below provides an overview of the components involved and how these
interact in a FlexFrame environment.
The PRIMECLUSTER software runs on the two Control Nodes in a FlexFrame solution.
The FA Control Agent runs on the active Control Node defined by PCL. The FA
Application Agents provide the Control Agent with information on the computer type and
further information which is required for generic creation of configuration files, insofar as
this is technically possible and the information is unambiguous and can be ascertained
securely.
Information which cannot be ascertained generically must be entered manually in the
BrutForceShutdown config section of the myAMC_FA.xml file.
For further information on configuring the power shutdown manually, please see the "FA
Agents - Installation and Administration" manual.
CFG
PRIMEPOWER
SCON
Midrange Application Node PRIMERGY
Application Node
PRIMEPOWER Enterprise
if you like to measure the time needed for a crash dump. Remember to disable the
Shutdown Facility before testing. Otherwise it may be interrupted if it takes too much
time.
You should see increased activity of the local disks during the dump.
On the next reboot, the crash dump will be saved from the swap devices to
/var/log/dump.
To analyze this dump, change to this directory and call lcrash -n <number>, where
<number> has to be the suffix of the dump files, for example map.0, dump.0 and
kerntypes.0.
For the following volumes in a FlexFrame 3.2 environment a snapshot schedule should
be configured:
vol0
Root volume with Data ONTAP.
VolFF
FlexFrame specific volume with the OS images, FA Agents, scripts and pool specific
data (e.g. SAP and DB executables, profiles etc.).
Sapdata
FlexFrame specific volume with all data files of the SAP databases.
Saplog
FlexFrame specific volume with all log files of the SAP databases.
Hourly
Data ONTAP creates these snapshots on the hour or at specified hours, except at
midnight if a nightly or weekly snapshot is scheduled to occur at the same time.
Hourly snapshots are called hourly.n, where n is an integer. The most recent
hourly snapshot is hourly.0, and hourly.1 is the next most recent hourly
snapshot.
To display the snapshot schedule for one or all volumes on a Filer, enter the following
command:
filer> snap sched [volume_name]
Example:
filer> snap sched volFF
Volume volFF: 2 6 8@8,12,16,20
The result means that for volume volFF weekly (max. 2), nightly (max. 6) and hourly
(max. 8) snapshots will be created. The hourly snapshots will be at 8:00, 12:00, 16:00
and 20:00 h.
You can create different snapshot schedules for different volumes on a Filer. On a very
active volume like SAPLOG, schedule snapshots every hour and keep them for just a few
hours. For example, the following schedule creates a snapshot every hour and keeps the
last three:
filer> snap sched saplog 0 0 3
This schedule does not consume much disk space, and it lets users recover files in
recent snapshots as long as they notice their mistake within a couple of hours.
Example for an Oracle backup script using snapshot (only parts visible)
##################################################################
# Name: sapsnap
# Create a snapshot
# Turning tablespaces in hot backup mode or shut down DB
#
# Syntax sapsnap <online|offline> SID
#
# Parameter: online Turns all tablespaces in hotbackup mode
before
# doing the snapshot. Ends hotbackup mode
after
# creating the snapshot
# offline Shuts down the DB before doing the
snapshot
# Starts the DB after creating the
snapshot
#
##################################################################
# functions
##################################################################
# Sending Database tables in BEGIN BACKUP
hotbackup_start (){
echo "sqlplus \"/ as sysdba\" @/tmp/prebackup_${SID}.sql"
> /tmp/prebackup_${SID}.sh;
For more information about backup of SAP systems, see the NetApp Technical Library
report 3365 about “SAP Backup and Recovery with NetApp Filers”:
http://www.netapp.com/tech_library/ftp/3365.pdf
filer> WARNING! This will restore a file from a snapshot into the
active filesystem. If the file already exists in the active
filesystem, it will be overwritten with the contents from the
snapshot.
Example with copying the same file from the .snapshot directory using a Control Node
(no SnapRestore license required):
control1> df -k
Filesystem 1K-blocks Used Available Use% Mounted on
filer:/vol/vol0/ 50119928 620012 49499916 2%
/FlexFrame/vol0
control1> cd /FlexFrame/vol0/etc
Copying will allocate new space while snap restore will use the old data blocks.
The data on volume volFF is shared among multiple SAP systems. Restoring
the complete volume has an effect on all SAP systems of the complete
FlexFrame landscape.
14.4.1 Arcserve
Detailed information on a backup solution with Arcserve is available at
http://extranet.fujitsu-
siemens.com/com/ep/storage/management/brightstor/arcserve_backup/general/arcserve
_with_flexframe
as well as
http://extranet.fujitsu-siemens.com/flexframe
at the bottom.
14.4.2 NetWorker
This concept is based on snapshots and uses the NDMP protocol for transferring data
from the NetApp Filer directly to the tape library. As a backup tool NetWorker is used
including add-on products like NSR-ORA-NDMP. A dedicated backup server will be used
for maintaining NetWorker.
Configuration Example for Oracle:
Detailed information on
● FlexFrame Backup with NetWorker
● slidesets for different target groups, also with implementation and configuration
information
● order units - example
will be provided by Fujitsu Siemens Computers Storage Consulting and is available at
http://extranet.fujitsu-siemens.com/com/ep/storage/solutions/it-solutions/FlexFrame-
Backup
as well as
http://extranet.fujitsu-siemens.com/flexframe
at the bottom.
Synopsis
Command Options
-d Debug messages to /tmp/ff_cn_backup.DEBUGLOG
-r Restore
-f <file_name>
Use file name for restore
-l <list>
Blank separated list of files (quoted)
The default behavior (without any options) is to backup the configuration files in a zipped
tar file at /FlexFrame/volFF/FlexFrame/backup/<name_of_cn><date>.tgz.
Additional files can be backup up using the -l option.
The Control Nodes are not inteded to have additional software installed on.
Therefore backup in the regular sense is not requried. Installation from the DVD
is much faster than any restore.
3. Abort the installation at the first “<YES> <NO>” screen with the title "FlexFrame(TM)
Setup for Control Nodes ".
4. Enter s to get into a subshell.
5. Install the latest patch set.
6. Reboot the Control Node using reboot -f
7. Again the setup screen will appear.
8. If the passwords are requested, enter the current passwords.
9. Enter the correct Control Node (first or second) for this Control Node during
installation.
10. Once the system is booted execute the following command:
control1:~# ff_cn_folders.pl -notrans
If there are differences in files listed, copy them over from the other Control Node
using:
scp -p control2:/<path> /<path>
Synopsis
Command Options
--silent
The silent mode supresses any message, even error messages.
--running
In normal case the configuration, which the switch reads on booting, is backed up,
the so called startup-config. Using the --running option the current running switch
configuration is used instead of startup-config. For a Quanta BX600 switch blade
(has six external ports) the running switch configuration is backed up only.
--name <switch_name>
Backup only the switch with given name. Keep in mind to use the name known at
LDAP database. It may differ from current switch name if it was not synchronized.
--ip <control_node_ip>
Use control node with given control lan ip address to save configurations to. In
normal case the program is able to detect the control node with running netboot_srv
service itself. If it is not able to detect the active node or can not determine the
control lan ip of the control node control, the program requests to use the --ip
option.
control1:/ # cd /var
control1:/var # ls -ld log*
drwxr-xr-x 20 root root 1440 May 2 14:01 log
lrwxrwxrwx 1 root root 58 Apr 29 17:28
log_pool1_klinge1 -> /FlexFrame/volFF/os/Linux/FSC_3.2B00-
000.SLES-9.X86_64/var_img/var-ac10020e/log
lrwxrwxrwx 1 root root 58 Apr 29 17:29
log_pool1_klinge2 -> /FlexFrame/volFF/os/Linux/FSC_3.2B00-
000.SLES-9.X86_64/var_img/var-ac10020f/log
lrwxrwxrwx 1 root root 58 Apr 29 17:30
log_pool1_klinge3 -> /FlexFrame/volFF/os/Linux/FSC_3.2B00-
000.SLES-9.X86_64/var_img/var-ac100210/log
lrwxrwxrwx 1 root root 58 Apr 29 17:30
log_pool1_klinge4 -> /FlexFrame/volFF/os/Linux/FSC_3.2B00-
000.SLES-9.X86_64/var_img/var-ac100211/log
lrwxrwxrwx 1 root root 58 Apr 29 17:30
log_pool1_klinge5 -> /FlexFrame/volFF/os/Linux/FSC_3.2B00-
000.SLES-9.X86_64/var_img/var-ac100212/log
lrwxrwxrwx 1 root root 58 Apr 29 17:31
log_otto_RX300-01 -> /FlexFrame/volFF/os/Linux/FSC_3.2B00-
000.SLES-9.X86_64/var_img/var-ac100113/log
lrwxrwxrwx 1 root root 58 Apr 29 17:33
log_otto_RX300-02 -> /FlexFrame/volFF/os/Linux/FSC_3.2B00-
000.SLES-9.X86_64/var_img/var-ac100114/log
control1:/var #
Or enter the log directory from this Application node and look for logfiles:
control1:/ # cd /var/log_pool1_klinge5
control1:/var/log_pool1_klinge5 # ls -lrt|tail
-rw-r--r-- 1 root root 278 May 4 16:23 log.scagt
-rw-r--r-- 1 root root 174 May 4 16:24 log.vvagt
-rw-r--r-- 1 root root 257 May 4 16:24
log.statusagt
-rw-r--r-- 1 root root 612 May 4 16:28 ntp
-rw-r--r-- 1 root root 5542 May 4 16:35 warn
-rw-r--r-- 1 root root 1696 May 4 16:35 auth.log
-rw-r--r-- 1 root root 484 May 4 18:16 log.busagt
drwxr-xr-x 2 root root 4096 May 4 18:19 sa
-rw-r--r-- 1 root root 27622 May 4 18:19 messages
prw------- 1 root root 0 May 4 18:24 psadfifo
control1:/var/log_pool1_klinge5 #
Most LDAP problems are related to access control and connection management.
If you have changed the loglevel send a -HUP to the slapd to advice it to reread
configuration files (use pkill -HUP slapd).
Another reason may be the LDAP client configuration. Check the following issues:
● Does the client use the correct server address?
● Does it use the correct user and password (Solaris only)?
Look at the client messages file to get this information.
If ldapsearch -x on Linux and ldaplist on Solaris works properly the problem may
be the nsswitch configuration. These are the top most problems.
When reporting problems with the administration scripts, please provide the
following information:
● The error message of the admin tool
● A dump of the LDAP database (ldapsearch -x > <file name>)
The configuration files /etc/openldap/slapd.conf and
/FlexFrame/volFF/FlexFrame/ldap/common/slapd.acl.conf
Error: Agents do not have the authorization to write to the directories assigned
Diagnosis:
The FA Autonomy production and log files are not written.
Response:
Provide the directories/files with the required permissions.
15.6.1 General
The activities and dynamic states of the FA Agents are documented in various files.
These files may not be changed manually as this can impact error free operation of the
FA Agents or result in incorrect reactions.
These files are created dynamically during ongoing operation. Deleting these files leads
to a status in which the Autonomous Agents reorganize themselves, and from this point
they re-evaluate the situation from the current viewpoint without any previous knowledge.
Subdirectories Content
Subdirectories Content
./FA_AppAgent/config myAMC.FA_AppAgent-specific
configuration data
./FA_AppAgent/log empty
./FA_CtrlAgent/config myAMC.FA_CtrlAgent-specific
configuration data
./FA_CtrlAgent/log empty
Subdirectories Content
./vFF/Common/.vFF_template.V<v>K<r>/ Configuration
config
Subdirectories Content
./vFF/Common/.vFF_template.V<v>K<r>/data Livelist
/FA/livelist livelist.log
./vFF/Common/.vFF_template.V<v>K<r>/data BlackBoard
/FA/blackboard blackboard.txt
15.6.3.1 Livelist
Each FA Application Agent regularly enters itself in this list. Through these entries the
Control Agent recognizes whether the various Application Agents are available and
functioning without error.
15.6.3.4 Reboot
The contents of this file are identical to those of the Services-List file. The file serves
as information storage when a reboot takes place. It is written only for the autonomous
reaction reboot and is deleted again after the reboot has been completed and the
services have been started up.
15.6.3.5 Switchover
The contents of this file are identical to those of the Services-List file. The file serves
as information storage (testament) when a switchover takes place. It is written only for the
autonomous reaction switchover and is deleted again after the services were taken
over.
15.6.3.7 BlackBoard
The BlackBoard is an input interface for the FA Agents. Commands can be entered here
which are executed by the FA Application Agents. The commands have a specific validity
period and are secured against manipulation. The file is written manually using a tool
which guarantees, among other things, protection against manipulation.
The functions of the FA Agents are documented in various files. These files may not be
changed manually as this can impair error-free operation of the FA Agents or result in
incorrect reactions.
These files are created dynamically during ongoing operation. Deleting these files leads
to a status in which the Autonomous Agents reorganize themselves, and from this point
they re-evaluate the situation from the current viewpoint without any previous knowledge.
For further information on collecting diagnosis data see “FA Agents - Installation and
Administration“ manual, section 4.7.3.
Clean up is the ultimative method to kill still running processes, freeing occupied
shared memory and semaphores caused by this SAP service. You should never
invoke this if you are not really sure!
Interface errors
start_flag errors
announce_start_flag errors
● ACC: Press the Refresh button after a while and check if the server is visible in the
Physical Landscape.
15.9 PRIMECLUSTER
Further on, shell scripts can be traced using the option "-x".
For details, see the man page of the shell in use (man sh; man bash; etc.)
sh -x ff_netscan.sh
Functions:
h Help
x <expression>
Evaluate an expression (hash/array) and display the result
p <string>
Returns strings
s Execute next command; follow subroutines
n Execute next command; skip subroutines
b <line_number>
Set break point at line <line number>
q Quit.
For further information, see the man page of perldebug.
Control Station
An Application Node running SAP ACC.
Dynamic Host Configuration Protocol
DHCP is a protocol for assigning dynamic IP addresses to devices on a network.
Dynamic Host Configuration Protocol server
DHCP service program.
Enterprise Resource Planning
Enterprise Resource Planning systems are management information systems that
integrate and automate many of the business practices associated with the
operations or production aspects of a company.
Ethernet
A Local Area Network which supports data transfer rates of 10 megabit per second.
Fiber Channel
Fibre Channel is a serial computer bus intended for connecting high speed storage
devices to computers.
Filer
Network attached storage for file systems.
FlexFrame Autonomous Agent
Central system management and high availability software component of FlexFrame.
FlexFrame for mySAP Business Suite
FlexFrame for mySAP Business Suite is a radically new architecture for mySAP
environments. It exploits the latest business critical computing technology to deliver
major cost savings for SAP customers.
Gigabit Ethernet
A Local Area Network which supports data transfer rates of 1 gigabit (1,024
megabits) per second.
Host name
Name of a physical server as seen from outside the network. One physical server
may have multiple host names.
Image
In the FlexFrame documentation, “Image” is used as a term for “Hard Disk Image”.
Internet Protocol Address
A unique number used by computers to refer to each other when sending information
through networks using the Internet Protocol.
Lightweight Directory Access Protocol
Protocol for accessing on-line directory services.
D ff_netscan.sh 54
data protection 255 ff_pool_defrt.sh 61
desaster repair 37 ff_pool_dnssrv.sh 62
document history 2 Filer
Domain Information Tree (DIT) 28 add new 21
E display all configured 23
error handling 269 display configuration 24
F remove 27
FA Agent version Filer cluster 45
install new 99 Filer configuration 11
migration on pool level 101 Filer volumes
FA Agents backup with NetApp Snapshot 256
configuring 239 Filers
error diagnosis 272 multiple 218
FlexFrame Autonomy 243 FlexFrame
groups 240 architecture 3
operation and log files 276 backup with tape library 263
power management 249 basic administration 46
traps 207, 242 general notes 4
FA Autonomous Agents 66 network 52
FA Autonomy FlexFrame Autonomous Agents
diagnostic tool 282 migration at pool level 100
FA migration tool 103 FlexFrame Autonomy
file mode 103 command line interface 101
parameters 104 FlexFrame configuration state,
displaying 51
pool mode 103
FlexFrame landscape
ff_change_id.pl 217
accessing 46
ff_install_an_linux_images.sh 201
power-off 48
ff_list_services.sh 224
power-on 46
e-mail: [email protected]
http://manuals.fujitsu-siemens.com
Submitted by
e-mail: [email protected]
http://manuals.fujitsu-siemens.com
Submitted by