-
AIX Booting Process explained below.
I. Read Only Storage Kernel Init Phase
Motherboard is Checked - POST --> look for Bootlist if found (Detected by ROS)------> look for boot image---Bootstrap code is loaded from disk
- BLV contains AIX Kernel, rc.boot scripts, reduced ODM, and boot commands.- BLV get uncompressed in RAM and kernel gets released from it.- Then AIX kernel gets the control- AIX Kernel creates a RAMFS- Kernel Starts the init process from BLV- init executes rc.boot script from the BLV in the RAM
II. Base Device Configuration Phase
1. All devices are configured with cfgmgr command - init process from RAMFS executes rc.boot 1
- restbase command copies ODM from BLV to RAMFS- cfgmgr will run and configure all the base devices
III. System Boot Phase
1. Logical volumes are varied on - rc.boot 2 is executed.
- ipl_varyon will varyon the rootvg- run fsck on / and mount the same to RAMFS- run fsck on /usr and mount the same to RAMFS- run fsck on /var and mount the same to RAMFS
2. Paging is started - copycore command checks the occurrunce of the dump and copies the same from primary dump device to /var/ras/adm
- unmount /var and activate the paging space- mount /var- Now /, /usr and /var are mounted in rootvg in hdisk
3. /etc/inittab is processed - Kernal remove RAMFS
- init process is started from / in rootvg- in rc.boot3, init processes /etc/inittab and remaining devices are configured.- /etc/init starts /etc/inittab and runs /sbin/rc.boot 3- run fsck on /tmp and mount the same.- syncvg for rootvg and reports the stale PPs
4. Relevant Services according to run level starts. srcmstr daemon controls this - starts the deamons like errdaemon, syncd etc..
* AIX 6.1, November 9, 2007o Workload Partitions o Live Application Mobilityo Role Based Access Control RBAC o AIX Security Expert hardening toolo Trusted AIX o Encrypting JFS2 filesystemo Trusted Execution o Concurrent Kernel Maintenanceo Kernel exploitation of POWER6 storage keys o probevue dynamic tracingo Systems Director Console for AIX o Integrated filesystem snapshoto Partition mobility on POWER6
* AIX 5L 5.3, August 13, 2004o NFS Version 4 support o Advanced Accountingo Virtual SCSI o Virtual Etherneto Simultaneous multithreading (SMT) support o Micro-Partitioning supporto POWER5 support o JFS2 quota supporto JFS2 filesystem shrink support
* AIX 5L 5.2, October 18, 2002o Introduced support for the IBM BladeCenter JS20 with the PowerPC 970.o Minimum level required for POWER5 hardwareo Support for MPIO Fibre Channel diskso iSCSI Initiator softwareo Dynamic LPAR support* AIX 5L 5.1, May 4,2001o Introduced support for the IA-64 architecture, although this never went beyond beta,[1]o Minimum level required for POWER4 hardware and the last release that supported Micro Channel architectureo Introduction of 64-bit kernel, installed but not activated by defaulto JFS2o introduced Logical Partitioning on POWER4o The L stands for Linux affinityo Trusted Computing Base (TCB)
Startup ModeNormal ModeLogin PromptsAll processes runningMulti User ModeSystem Management ServicesNot AIXRuns from firmwaresets boot listMaintainance ModeMaintainance MenuRecover root passwordFix machines that wont bootDignosticAIX diagnostic
If device state is
undefined------>Supported Devices------->Predefined Databasedefined--------->Not Usable------------------>Customized DatabaseAvailable--- --->Ready for use-------------->Customized Database
Seven types supported:Journaled File System (JFS)Enhanced Journaled File System (JFS2)CD-ROM File System (CDRFS)DVD-ROM File System (UDFS)Network File System (NFS)Common Internet Filesystem (CIFS)Proc File System (PROCFS)
Defaults Lvs and Filesystems on it
AIX Installation methodNew and Complete Overwrite Installation:You have a new machine without a prior or useful system installation. You want to install onto a hard disk that contains an existing root volume group that you wish to completely overwrite. You want to reassign your hard disks, that is, to make your rootvg smaller and assign less disk space to it.Migration installationUse the migration installation method to upgrade AIX Version 4.2, 4.3, AIX 5L Version 5.1, or Version 5.2 to AIX 5L Version 5.3 while preserving the existing root volume group. The installation process determines which optional software products must be installed. With the exception of /tmp, this method preserves most file systems, including the root volume group, logical volumes, and system configuration files.Preservation installationUse the preservation installation method when a version of the BOS is installed on your system, and you want to preserve the user data in the root volume group. However, this method overwrites the /usr, /tmp, /var, and / (root) file systems by default, so any user data in these directories is lost. These file systems are removed and recreated, so any other LPPs or filesets that you installed on the system will also be lost. System configuration must be done after doing a preservation installation.The /etc/preserve.list file contains a list of system files to be copied and saved during a preservation BOS installation. The /etc/filesystems file is listed by default. You can add the full path names of any additional files that you want tosave during the Preservation installation to the preserve.list file.
Trusted Computing BaseThe Trusted Computing Base (TCB) is the part of the system that is responsible for enforcing the information security policies of the system. By installing a system with the TCB option, you enable the trusted path, trusted shell, trustedprocesses, and system-integrity checking. Because every device is part of the TCB, every file in the /dev directory is monitored by the TCB
Network Installation ManagementNetwork Installation Management (NIM) allows you to manage an automated installation of the Base Operating System (BOS) and optional software on one or more machines. You can install a group of machines with a common configuration or customize an installation for your specific needs. The number of machines you can install simultaneously depends on the throughput of your network, the disk access throughput of the installation servers, and the platform type of your servers.The types of machines that can be managed in the NIM environment are stand-alone, diskless, and dataless clients.NIM resource objects represent files and directories that are used to support some type of NIM operation a copy of the current running rootvg to another disk, be sure to have all the file systems mounted that you want cloned across.
Alternate Disk InstallationAlternate disk installation, available starting with AIX Version 4.3, allows for system installation on a system while it is still up and running. Install or upgrade downtime decreases considerably. It also allows large facilities to manage anupgrade, while the systems are still running at the existing version. The switch over to the new version can then happen with a simple reboot, with the possibility to roll back to the original situation in case of problems.Cloning the rootvg to an alternate disk has many advantages. One advantage is having an online backup available, as in the case of a disk crash. Keeping an online backup requires an extra disk or disks to be available on the system. where altinst_rootvg is the cloned non-active/varied off rootvg. The cloned rootvg has all its logical volumes prefixed with the name 'alt'. The boot list is also changed to boot off altinst_rootvg. AIX likes to do things like this; it assumes you will want to boot off the cloned and not the real rootvg.Alternate disk installation can be used in one of two ways:1) Cloning the current running rootvg to an alternate disk. smitty alt_clone2) Installing a mksysb image on another disk. smitty alt_mksysbWith a mksysb image, you can clone one system image onto multiple target systems. However, the target systems might not contain the same hardware.
By default, using the alt_disk_install command does the following:1. Creates an /image.data file based on the current rootvg's configuration. A customized image.data file can be used.2. Creates an alternate rootvg (altinst_rootvg).3. Creates logical volumes and file systems with the alt_inst prefix.4. Generates a backup file list from the rootvg, and if an exclude.list file is given, those files are excluded from the list.5. Copies the final list to the altinst_rootvg's file systems.6. If specified, the installp command installs updates, fixes, or new filesets into the alternate file system.7. The bosboot command creates a boot logical volume on the alternate boot disk.8. If a customization script is specified, it runs at this point.9. The file systems are then unmounted, and the logical volumes and file systems are renamed.10.The logical volume definitions are exported from the system to avoid confusion with identical ODM names, but the altinst_rootvg definition is left as an ODM placeholder.11.By default, the bootlist is set to the new cloned rootvg for the next reboot.
Maintenance LevelA Maintenance Level (ML) is the service updates that are necessary to upgrade the base operating system (BOS) or an optional software product to the current release level.
Technology LevelA Technology Level (TL) is the new term for the twice yearly AIX 5L releases, which contain new hardware and software features and service updates. The first TL will be restricted to hardware features and enablement, as well as software service. The second TL will include hardware features and enablement, software service, and new software features.
FilesetA fileset is the smallest installable base unit for the AIX 5L operating system.
PackageA package is a group of separately installable filesets that provide a set of related functions. For example, bos.net is a package
Licensed Program ProductA Licensed Program Product (LPP) is a complete software product including all packages associated with that licensed program. For example, the BOS is a licensed program.
BundleA bundle is a list of software that can contain filesets, packages, and LPPs that are suited for a particular use, such as providing personal productivity software or software for a client machine in a network environment
PTFPTF is an acronym for Program Temporary Fix. A PTF is an updated fileset or a new fileset that fixes a previous system problem. PTFs are installed in the same way as regular filesets by the use of the installp command
APARAPAR is an acronym for Authorized Program Analysis Report. An APAR is an emergency fix, or interim fix, to a unique problem on the system. APARs will eventually become PTFs after testing and verification. APARs are applied to thesystem through the use of the instfix command
Applied StateThe applied state places software on the system and retains the previous version of the software. When an update is in the applied state, the previous version is stored in the /usr/lpp/PackageName directory, where PackageName could be bos.net which is a package.
Commited StateThe committed state places software on the system and removes all previous levels of the software from the /usr/lpp/PackageName directory. If committed software needs to be removed, you cannot go back to the previous version without a complete reinstall of the previous version software.
SUMAautomatic download, scheduling, and notification capabilities through the new Service Update Management Assistant (SUMA) tool
Object Data ManagerThe ODM is a repository in which the operating system keeps information regarding your system, such as devices, software, or TCP/IP configuration. The ODM is an object-oriented database that contains vital data, keeps this dataconsistent with the actual state of the system, and prevents the administrator from altering it by mistake.System data managed by the ODM includes:Device configuration informationDisplay information for SMIT (menus, selectors, and dialogs)Vital product data for installation and update proceduresCommunication configuration informationSystem resource controller data informationError log and dump informationNetwork Installation Manager (NIM) informationODM data is stored in binary format. You cannot modify ODM files with a text editor. You must use special commands that are designed to interact with the ODM.
The basic components of the ODM areObject classes Each file of the database is an object class. Each object class consists of objects having similar definitions.Objects Each object is one record in an object class. It is a stand-alone entity and has one or more descriptors.Descriptors The descriptors describe the layout of the objects. They determine the name and data type of the fields that are part of the object class. The descriptors of an object and their associated values can be located and changed using ODM commands.
ODM information is divided in three parts in order to support diskless or dataless systems. The names of these three directories as follows:/usr/lib/objrepos Contains the predefined objects classes, SMIT menu object classes and the four object classes used by SWVPD for the /usr part of the installable software product. The object classes in this repository can be shared across the network by /usr clients, dataless and diskless workstations. Software installed in the /usr part can be shared among several machines with compatible hardware architecture.
/usr/share/lib/objrepos Contains the four object classes used by the SWVPD for the /usr/share part of the installable software product. The /usr/share part of a software product contains files that are not hardware dependent. They can be shared among several systems, even if these have a different hardware architecture. An example are terminfo files that describe terminal capabilities. Because terminfo is used on many UNIX systems, terminfo files are part of the /usr/share part of the software product.
/etc/objrepos Contains the customized devices object classes and the four object classes used by SWVPD for the / part of the installable software product. To access information in the other directories, this directory contains symbolic links to the predefined devices object classes. These links are needed because the ODMDIR variable points to only/etc/objrepos. It contains the part of the product that cannot be shared with other systems. Most of thissoftware requiring a separate copy for each machine is associated with the configuration of the machine orproduct.
The ODM commands are:odmadd Adds objects to an object class.odmchange Changes specific objects in a specified object class.odmcreate Creates empty object classes.odmdelete Removes objects from an object class.odmdrop Removes an entire object class.odmget Retrieves objects from object classes and puts the object information into odmadd command format.odmshow Displays the description of an object class.
System Management Interface ToolSystem Management Interface Tool (SMIT) provides an alternative to the typical method of using complex command syntax, valid parameter values, and custom shell path names for managing and maintaining your operatingsystem configuration.Software Installation and MaintenanceSoftware License ManagementDevicesSystem Storage Management (Physical & Logical Storage)Security & UsersCommunications Applications and ServicesPrint SpoolingAdvanced AccountingProblem DeterminationPerformance & Resource SchedulingSystem EnvironmentsProcesses & SubsystemsApplicationsInstallation AssistantCluster Systems ManagementUsing SMIT (information only)
SMIT runs in two modes: ASCII (non-graphical) and X-Window (graphical). ASCII SMIT can run on both terminals and graphical displays. The graphical mode, which supports a mouse and point-and-click operations, can be run only on agraphical display and with X-Window support. The ASCII mode is often the preferred way to run SMIT because it can be run from any machine.To start the ASCII mode, type at the command line:smitty or smit -aTo start the graphical mode, type:smit or smit -mSMIT is an interactive, menu-driven user interface that allows you to more easily perform routine system management tasks and to manage and maintain your operating system configuration.
SMIT uses three types of screens: menu, selector, and dialog screens. Menu screens display a list of items that you can select. Selector screens, often presented as a pop-up menu, display a list of items from which you specify or select a particular item Dialog screens are the interface to a command or task that you perform
Trusted Program,A trusted program, or trusted process, is a shell script, a daemon, or a program that meets a particular standard of security.Examples of trusted daemons are:_ ftpd_ rexecd_ telnetdExamples of non-trusted daemons are:_ rshd_ rlogind_ tftpd
Bootlist CommandThe bootlist command allows the user to display and alter the list of possible boot devices from which the system may be booted. This command supports the updating of the following:Normal boot list The normal list designates possible boot devices for when the system is booted in normal mode.Service boot list The service list designates possible boot devices for when the system is booted in service mode.Previous boot device This entry designates the last device from which the system booted. Some hardware platforms may attempt to boot from the previous boot device before looking for a boot device in one of the other lists.
Network Configuration Files$HOME/.netrcThe $HOME/.netrc file contains information used by the automatic login feature of the rexec and ftp commands. It is a hidden file in a user's home directory and must be owned either by the user executing the command or by the root user.$HOME/.forwardWhen mail is sent to a local user, the sendmail command checks for the $HOME/.forward file for that user. The $HOME/.forward file can contain one or more addresses or aliases. Any messages are then sent to the addresses oraliases in the $HOME/.forward file./etc/hosts.equiv and $HOME/.rhostsThe /etc/hosts.equiv file, along with any local $HOME/.rhosts files, defines the hosts (computers on a network) and user accounts that can invoke remote commands on a local host without supplying a password. A user or host that isnot required to supply a password is considered trusted, though the daemons that initiate the connections may be non-trusted in nature (for example, rlogind).
Intreduction to rc.* filesrc.boot fileThe /sbin/rc.boot file is a shell script that is called by the simple shell init and the standard init command to bring up a system. It controls the machine boot process. When the system is booting, the /sbin/rc.boot file is called on each bootphases, each time being passed a different parameter.Depending upon the type of boot device, the rc.boot file configures devices and also calls the appropriate applications.Appropriate applications include:_ Booting from disk (boot phase 1)_ Varying on a root volume group (boot phase 2)_ Enabling file systems (boot phase 2)_ Calling the BOS installation programs or diagnostics/etc/rc fileThe /etc/rc file performs normal startup initialization; its entry in the /etc/inittab file is located after the rc.boot entry. The init command reads the /etc/inittab file and creates a process for the /etc/rc file. The contents of the /etc/rc file are installation specific. If all of the necessary operations complete successfully, the file exits with a zero return code that allows the init command to start loggers to complete normal initialization and startup.Many bringup functions are done by the /etc/rc file, such as:_ Vary on all volume groups marked as auto-varyon._ Activate all paging spaces listed on /etc/swapspaces (using the swapon –a command)._ Configure all dump devices (using the sysdumpdev -q command)._ Perform file system checks (using the fsck -fp command)._ Perform mounting of file systems marked as mount=true on the /etc/filesystems file (using the mount all command).rc.net fileThe /etc/rc.net file is a shell script that contains network configuration information The stanzas allow you to enable the network interfaces and set the host name, the default gateway, and any static routes for the current host. This file can be used as a one-step configuration alternative to using individually the set of commands and files necessary to configure a host.The rc.net shell script is run by the configuration manager program during the second phase of configuration. If TCP/IP is installed, a second script, rc.tcpip, is run from the init command after the second phase of configuration has completed and after the init command has started the SRC master.rc.tcpip fileThe /etc/rc.tcpip file is a shell script that, when executed, uses SRC commands to initialize selected daemons. The rc.tcpip shell script is automatically executed with each system restart. It can also be executed at any time from the command line.Most of the daemons that can be initialized by the rc.tcpip file are specific to TCP/IP. These daemons are:_ inetd (started by default)_ gated_ routed_ named_ timed_ rwhod/etc/tcp.clean can be used to stop TCP/IP daemons.
System boot without starting rc.tcpipRemoving the rc.tcpip entry in /etc/inittab means that you are not starting any server applications during IPL.Without the server applications started, you will not be able to telnet or ftp to this machine from another host.However, as long as you have not brought down the network interface, you can still utilize the client network services. You can still ping other hosts, and you can still telnet or ftp to other hosts.
The /etc/services fileThe /etc/services file contains information about the known services used in the DARPA Internet network by inetd. Each service listed in /etc/services runs on a specific port number for communications, in a specific format, such as TCP or UDP.If you edit the /etc/services file, run the refresh -s inetd command, in order for your changes to be used.
File System Issues
Network File SystemThe Network File System (NFS) is a distributed file system that allows users to access files and directories of remote servers as though they were local. For example, you can use operating systems commands to create, remove, read,write, and set file attributes for remote files and directories. NFS is independent of machine types, operating systems, and network architectures because of its use of remote procedure calls (RPC) for these services.For the successful implementation of an NFS environment, you need the following things:1. The NFS daemons should be running on the server and the clients.2. The file systems that need to be remotely available will have to be exported.3. The exported file systems need to be mounted on the remote (client) systems.
The following are a list of terms that are used throughout this discussion:Server A computer that makes its file systems, directories, and other resources available for remote access.Clients The computers, or their processes, that use a server’s resources.Export The act of making file systems available to remote clients.Mount The act a client needs to do to access the file systems that a server exports.
The following describes the mounting process:When the server starts, the /etc/rc.nfs script runs the exportfs command, which reads the server /etc/exports file and then tells the kernel which directories are to be exported and which access restrictions they require.The rpc.mountd daemon and several nfsd daemons (eight, by default) are then started by the /etc/rc.nfs script.When the client starts, the /etc/rc.nfs script starts several biod daemons (eight, by default), which forward client mount requests to the appropriate server.Then the /etc/rc.nfs script executes the mount command, which reads the file systems listed in the /etc/filesystems file.The mount command locates one or more servers that export the information the client wants and sets up communication between itself and that server. This process is called binding.The mount command then requests that one or more servers allow the client to access the directories in the client /etc/filesystems file.The server rpc.mountd daemon receives the client mount requests and either grants or denies them. If the requested directory is available to that client, the rpc.mountd daemon sends the client's kernel an identifier called a file handle.The client kernel then ties the file handle to the mount point (a directory) by recording certain information in a mount record.
There are three types of NFS mounts: predefined, explicit, and automatic.Predefined mounts are specified in the /etc/filesystems file.Explicit mounts are usually made for short periods of time when there is a requirement for occasional unplanned mounts. There is no entry for shared file system in /etc/filesystemsAutomatic mounts are controlled by the automount command, which causes the AutoFS kernel extension to monitor specified directories for activity. If a program or user attempts to access a directory that is not currently mounted, then AutoFSintercepts the request, arranges for the mount of the file system, and then services the request.
The following describes the mounting process:1. When the server starts, the /etc/rc.nfs script runs the exportfs command, which reads the server /etc/exports file and then tells the kernel which directories are to be exported and which access restrictions they require.2. The rpc.mountd daemon and several nfsd daemons (eight, by default) are then started by the /etc/rc.nfs script.3. When the client starts, the /etc/rc.nfs script starts several biod daemons (eight, by default), which forward client mount requests to the appropriate server.4. Then the /etc/rc.nfs script executes the mount command, which reads the FS listed in the /etc/filesystems file.5. The mount command locates one or more servers that export the information the client wants and sets up communication between itself and that server. This process is called binding.6. The mount command then requests that one or more servers allow the client to access the directories in the client /etc/filesystems file.7. The server rpc.mountd daemon receives the client mount requests and either grants or denies them. If the requested directory is available to that client, the rpc.mountd daemon sends the client's kernel an identifier called a file handle.8. The client kernel then ties the file handle to the mount point (a directory) by recording certain information in a mount record.
The inetd daemonThe /usr/sbin/inetd daemon provides Internet service management for a network. This daemon reduces system load by invoking other daemons only when they are needed.If you change the /etc/inetd.conf using SMIT, then the inetd daemon will be refreshed automatically and will read the new /etc/inetd.conf file. If you change the file using an editor, run the# refresh -s inetdor# kill -1 InetdPIDExisting sessions are not affected when the inetd daemon is stopped, but no new telnet and ftp sessions can be established without first restarting the inetd daemon.
The portmap daemonThe portmap daemon converts remote procedure call (RPC) program numbers into Internet port numbers.
/etc/netsvc.conf Specifies default order for Host name resolutions/etc/hosts provides a list of server names or aliases and their IP addresses/etc/resolv.conf Name resolution file
Maximum transmission unit:MTU size is critical for proper network communications. Packets that are too small in length may be lost during transmission. Packets that are too long in length may collide with other packets that are being transmitted. These factors can lead to slower transmission rates and other network problems as packets must then be retransmitted.
To automate the NTP start after reboot:# vi /etc/inittabntp:2:wait:/usr/bin/startsrc -sxntpdORUncommented the commented the line from /etc/rc.tcpipstart /usr/sbin/xntpd "$src_running"
Virtual EthernetVirtual Ethernet enables inter-partition communication without the need for physical network adapters assigned to each partition. Virtual Ethernet allows the administrator to define in-memory connections between partitions handled at system level
Paging Space.To accommodate a large virtual memory space with a limited real memory space, the system uses real memory as a work space and keeps inactive data and programs on disk. The area of the disk that contains this data is called thesystem paging space.*Do not allocate more than one paging space logical volume on a physical volume.*Avoid putting a paging space logical volume on a heavily active logical volume.*Make each paging space logical volume equal in size.*Do not extend a paging space logical volume onto multiple physical volumes.
LVM ConceptsThe fundamental concepts used by LVM are physical volumes, volume groups, physical partitions, logical volumes, logical partitions, file systems, and raw devices. Some of their characteristics are presented as follows:Each individual disk drive is a named physical volume (PV) and has a name such as hdisk0 or hdisk1.One or more PVs can make up a volume group (VG). A physical volume can belong to a maximum of one VG.You cannot assign a fraction of a PV to one VG. A physical volume is assigned entirely to a volume group.Physical volumes can be assigned to the same volume group even though they are of different types, such as SCSI or SSA.Storage space from physical volumes is divided into physical partitions (Pps). The size of the physical partitions is identical on all disks belonging to the same VG.Within each volume group, one or more logical volumes (LVs) can be defined. Data stored on logical volumes appears to be contiguous from the user point of view, but can be spread on different physical volumes from the samevolume group.Logical volumes consist of one or more logical partitions (LPs). Each logical partition has at least one corresponding physical partition. A logical partition and a physical partition always have the same size. You can have up to threecopies of the data located on different physical partitions. Usually, physical partitions storing identical data are located on different physical disks for redundancy purposes.Data from a logical volume can be stored in an organized manner, having the form of files located in directories. This structured and hierarchical form of organization is named a file system.Data from a logical volume can also be seen as a sequential string of bytes. This type of logical volumes are named raw logical volumes. It is the responsibility of the application that uses this data to access and interpret it correctly.
Volume Group Descriptor AreaThe volume group descriptor area (VGDA) is an area on the disk that contains information pertinent to the volume group that the physical volume belongs to. It also includes information about the properties and status of all physical and logical volumes that are part of the volume group. The information from VGDA is used and updated by LVM commands. There is at least one VGDA per physical volume. Information from VGDAs of all disks that are part of thesame volume group must be identical. The VGDA internal architecture and location on the disk depends on the type of the volume group (original, big, or scalable).
Volume Group Status AreaThe volume group status area (VGSA) is used to describe the state of all physical partitions from all physical volumes within a volume group. The VGSA indicates if a physical partition contains accurate or stale information. VGSA is used for monitoring and maintained data copies synchronization. The VGSA is essentially a bitmap and its architecture and location on the disk depends on the type of the volume group.
Logical Volume Control BlockA logical volume control block (LVCB) contains important information about the logical volume, such as the number of the logical partitions or disk allocation policy. Its architecture and location on the disk depends on the type of the volume group it belongs to. For standard volume groups, the LVCB resides on the first block of user data within the LV. For big volume groups, there is additional LVCB information in VGDA on the disk. For scalable volume groups, all relevant logical volume control information is kept in the VGDA as part of the LVCB information area and the LV entry area.
Logical Track GroupLogical track group (LTG) size is the maximum allowed transfer size for an I/O disk operation.# lquerypv -M hdisk0
QuorumThe quorum is one of the mechanisms that the LVM uses to ensure that a volume group is ready to use and contains the most up-to-date data.A quorum is a vote of the number of Volume Group Descriptor Areas and Volume Group Status Areas (VGDA/VGSA) that are active. A quorum ensures data integrity of the VGDA/VGSA areas in the event of a disk failure. Each physical disk in a volume group has at least one VGDA/VGSA. When a volume group is created onto a single disk, it initially has two VGDA/VGSA areas residing on the disk. If a volume group consists of two disks, one disk still has two VGDA/VGSA areas, but the other disk has one VGDA/VGSA. When the volume group is made up of three or more disks, then each disk is allocated just one VGDA/VGSA.A quorum is lost when at least half of the disks (meaning their VGDA/VGSA areas) are unreadable by LVM. In a two-disk volume group, if the disk with only one VGDA/VGSA is lost, a quorum still exists because two of the three VGDA/VGSA areas still are reachable. If the disk with two VGDA/VGSA areas is lost, this statement is no longer true. The more disks that make up a volume group, the lower the chances of quorum being lost when one disk fails.When a quorum is lost, the volume group varies itself off so that the disks are no longer accessible by the LVM. This prevents further disk I/O to that volume group so that data is not lost or assumed to be written when physical problems occur. Additionally, as a result of the vary-off, the user is notified in the error log that a hardware error has occurred and service must be performed.There are cases when it is desirable to continue operating the volume group even though a quorum is lost. In these cases, quorum checking can be turned off for the volume group. This type of volume group is referred to as a nonquorum volume group. The most common case for a nonquorum volume group occurs when the logical volumes have been mirrored. When a disk is lost, the data is not lost if a copy of the logical volume resides on a disk that is not disabled and can be accessed. However, there can be instances in nonquorum volume groups, mirrored or nonmirrored, when the data (including copies) resides on the disk or disks that have become unavailable. In those instances, the data might not be accessible even though the volume group continues to be varied on.
Unlock VGA volume group can become locked after an abnormal termination of an LVM command. You can remove the lock using the chvg -u command.
Journaled file systemThis type of file system is named journaled because the system uses journaling techniques to maintain theintegrity of control structures. Each journaled file system must reside on a distinct jfs logical volume. Therefore, thefile system size will be a multiple of the size of a logical partition.
Enhanced journaled file systemThis is the enhanced version of the initial journalized file system. It uses extent based allocation to allow higherperformance, larger file systems, and a larger file size. Each enhanced journaled file system must reside on adistinct jfs2 logical volume. When the operating system is installed using the default options, it creates JFS2 filesystems.
Controll Growing file/var/adm/wtmp/etc/security/failedlogin/var/adm/sulog/var/spool/*/*$HOME/smit.log$HOME/smit.script$HOME/websm.log$HOME/websm.scriptJFS and JFS2 Difference
SuperblockThe superblock contains control information about a file system, such as the overall size of the file system in 512 byte blocks, the file system name, the file system log device, the version number, the number of inodes, the list of freeinodes, the list of free data blocks, date and time of creation, and file system state. All this data is stored in the first logical block of the file system. Corruption of this data may render the file system unusable. This is why the system keeps a second copy of the superblock on logical block 31.
Allocation groupAn allocation group consists of inodes and its corresponding data blocks. An allocation groups spans multiple adjacent disk blocks and improves the speed of I/O operations. Both JFS and JFS2 file systems use allocation groups. For a JFS file system, the allocation group size can be specified when the file system is created.
InodesThe inode contains control information about the file, such as type, size, owner, and the date and time when the file was created, modified, or last accessed. It also contains pointers to data blocks that store the actual data of the file. Every file has a corresponding inode.For JFS file systems, the maximum number of inodes, and hence the maximum number of files, is determined by the number of bytes per inode (nbpi) value, which is specified when the file system is created. For every nbpi bytes of your file system, there will be an inode created. The total number of inodes is fixed. The nbpi values needs to be correlated with allocation group size. The JFS restricts all file systems to 16 MB (224) inodes. JFS2 file systems manages the necessary space for inodes dynamically so there is not any nbpi parameter.
Data blocksData blocks store the actual data of the file or pointers to other data blocks. The default value for disk block size is 4 KB.
FragmentsFragments of logical blocks can be used to support files smaller than the standard size of the logical block (4 KB). This rule applies only to the last block of a file smaller than 32 KB.For JFS file systems only, you have the option to use compression to allow all logical blocks of a file to be stored as a sequence of contiguous fragments. Compression for a file system will increase the amount of CPU and I/O activitywhen using that file system. These features can be useful to support a large number of small files. Fragmentsize must be specified for a file system at installation time. Different file systems can have different fragment sizes.
Device logsThe journaled file system log stores transactional information about file system metadata changes.This data can be used to roll back incomplete operations if the machine crashes. JFS file systems are used for logging logical volumes of type jfslog, while JFS2 file systems are used for logging logical volumes of type jfs2log.Data from data blocks are not journaled. Log devices ensure file system integrity, not data integrity. After the operating system is installed, all file systems within the rootvg volume group use logical volume hd8 as a common log. You can create a JFS2 file system that can use inline logs. This means the log data is written into the same logical volume as the file system, and not into the log logical volume.Full backupAll the files are put on media during a full backup.
Differential BackupThe differential backup strategy first looks for the modification time of a file and compares it with the last full backup time. In other words, only modified files are backed up, but only if they changed after the latest full backup. Differential backups are cumulative; once a file has been modified, it will be included in every differential backup until the next full backup.
Incremental backupsIncremental backups are similar to differential backups in that both back up only modified files. However, incremental backup checks the difference between the modification time of a file and the last backup time (either being full orincremental backup). If the modification date is more recent than the last backup date, the file is backed up.
MKSYSBThe mksysb command creates a bootable image of all mounted file systems on the rootvg volume group. You can use this backup command to restore a system to its original state. The tape format includes a BOS boot image, a BOS install image, and a dummy table of contents (TOC) followed by the system backup (root volume group) image. The root volume group image is in backup-file format, starting with the data files and then any optional map files.User-defined paging spaces, unmounted file systems, and raw devices are not backed up.
The BOS boot image contains a copy of the system’s kernel and device drivers needed to boot from the mksysb tape. It is created by the bosboot command.There are three important files in the mkinsttape image (the second section of the mksysb tape): ./tapeblksz, ./bosinst.data, and ./image.data. The ./tapeblksz file contains the block size the tape drive was set to when the mksysb command was run.The ./bosinst.data file allows you to specify the requirements at the target system and how the user interacts with the target system. This file contains the customized BOS install procedures and dictates how the BOS install program willbehave. You can customize this file before issuing the mksysb command or use a procedure to customize this file after the image backup is done.The ./image.data file contains information describing the image installed during the BOS installation process. This information includes the sizes, names, maps, and mount points of logical volumes and file systems in the rootvg. You can customize this file before using the mksysb command, or run mksysb -i to generate a new ./image.data file on the tape during a backup. The mkszfile command generates the ./image.data file. The ./image.data file is arranged instanza format. Each stanza contains one or more fields.The dummy table of contents (TOC) is used so that the mksysb tape contains the same number of images as a BOS install tape.The rootvg data area contains all data in the rootvg volume group backed up by the mksysb command. The mksysb command uses the backup command to save the contents of mounted JFS data in rootvg, excluding raw data.
When you need to make a mksysb backup of a system, and you want to exclude some data file systems from the system, you need to edit the /etc/exclude.rootvg file. Make sure that there are no empty lines.^./tmp/
Restoring mksysbAn mksysb image enables you to restore the system image onto target systems that might not contain the same hardware devices or adapters, require the same kernel (uniprocessor or microprocessor), or be the same hardware platform (rs6k, rspc, or chrp) as the source system.You have several possibilities to restore the mksysb image:-If restoring on exactly the same machine, you can boot directly from the mksysb media and restore from there.-If restoring on a different type of machine, you use the cloning function-If you do not want to interfere with the production environment, you can use the alt_mksysb command.
MultibosThe multibos command allows the root level administrator to create multiple instances of AIX on the same rootvg. The multibos setup operation creates a standby Base Operating System (BOS) that boots from a distinct boot logical volume (BLV). This creates two bootable sets of BOS on a given rootvg. The administrator can boot from either instance of BOS by specifying the respective BLV as an argument to the bootlist command or using system firmware boot operations. Two bootable instances of BOS can be simultaneously maintained. The instance of BOS associated with the booted BLV is referred to as the active BOS. The instance of BOS associated with the BLV that has not been booted is referred to as the standby BOS. Currently, only two instances of BOS are supported per rootvg.
The multibos command allows the administrator to access, install maintenance and technology levels for, update, and customize the standby BOS either during setup or in subsequent customization operations. Installing maintenance and technology updates to the standby BOS does not change system files on the active BOS. This allows for concurrent update of the standby BOS, while the active BOS remains in production.In this case, the biggest difference is that multibos will create and copy only the following Logical Volumes (LVs):/;/usr;/var;/opt, and;hd5 (Boot logical volume).All others LVs will be shared with the original root volume group, although more logical volumes can be specified and copied.Multibos supports a new TL update, but AIX version upgrades are not supported by multibos.
Failed Disk Replacement
If the disk you are going to replace is mirrored, we recommend following these steps:1. Remove copies of all logical volumes that were residing on that disk using either the rmlvcopy command or unmirrorvg command.2. Remove the disk from the volume group using the reducevg command.3. Remove the disk definition using the rmdev command.4. Physically remove the disk. If the disk is not hot-swappable, you may be required to reboot the system.5. Make the replacement disk available. If the disk is hot-swappable, you can run cfgmgr; otherwise, you may need to reboot the system.6. Include the newly added disk into the volume group using the extendvg command.7. Recreate and synchronize the copies for all logical volumes using either mklvcopy or mirrorvg.
If the disk you are going to replace is not mirrored and is still functional, we recommend following these steps:1. Make the replacement disk available. If the disk is hot-swappable, you can run cfgmgr; otherwise, you may need to reboot the system.2. Include the newly added disk into the volume group using the extendvg command.3. Migrate all partitions from the failing disk to the new disk using either the migratepv command or the migratelp command. If the disks are part of the rootvg, you should consider the following:– If the disk to be replaced contains a copy of the BLV, you have to clear it using the chpv -c command.– A new BLV image must be created on the new disk using the bosboot command.– The bootlist must be updated to reflect these changes using the bootlist command.– If the disk to be replaced contains a paging space or a primary dump device, you should disable them. After the migratepv command completes, you should reactivate them.4. Remove the failing disk from the volume group using the reducevg command.5. Remove the disk definition using the rmdev command
If the disk is not mirrored, has failed completely, and there are other disks available in the volume group, we recommend following these steps:1. Identify all logical volumes that have at least one partition located on the failed disk.2. Close the logical volumes and unmount all the corresponding file systems using the umount command.3. Remove the file systems and logical volumes using the rmfs command.4. Remove the failing disk from the volume group using the reducevg command.5. Remove the disk definition using the rmdev command.6. Physically remove the disk. If the disk is not hot-swappable, you may be required to reboot the system.7. Make the replacement disk available. If the disk is hot-swappable, you can run cfgmgr; otherwise, you may need to reboot the system.8. Include the newly added disk into the volume group using the extendvg command.9. Recreate all the logical volumes and the corresponding file systems using the mklv command and the crfs command.10.If you have a backup of your data, restore your data from backup.
If the disk is not mirrored, has failed completely, there are no other disks available in the volume group (the volume group contained only one physical volume or all the physical volumes failed simultaneously), and the volume group is not rootvg, we recommend the following steps:1. Export the volume group definition from the system using the exportvg command.2. Ensure that /etc/filesystems does not contain any incorrect stanzas.3. Remove the disk definition using the rmdev command.4. Physically remove the disk. If the disk is not hot-swappable, you may be required to reboot the system.5. Make the replacement disk available. If the disk is hot-swappable, you can run cfgmgr; otherwise, you may need to reboot the system.6. If you have a volume group backup, restore it using the restvg command.7. If you do not have volume group backup, recreate the volume group, all the logical volumes, and the corresponding file systems using the mkvg command, the mklv command, and the crfs command.8. If you have a backup of your data, restore your data from backup.
If the disk is not mirrored, has failed completely, there are not other disks available in the volume group (the volume group contained only one physical volume or all physical volumes failed simultaneously), and the volume group isrootvg, we recommend following these steps:1. Replace the failing disk.2. Boot the system in maintenance mode.3. Restore the system from an mksysb image.
RAID LevelsRAID 0 Also known as striping. Data is split into blocks of equal size and stored on different disks. There is noredundancy. If one disk fails, your data is lost.RAID 1 Also known as mirroring. Data is split into blocks of equal size. Duplicate copies are kept on separatephysical disks. If one disk fails, the second copy is used and data is still available.RAID 5 Also known as striping with parity. Data is split into blocks of equal size. An additional data blockcontaining parity information is generated. Both data blocks and the parity block are written onto separatephysical disks. If one disk storing a block of data fails, the system will reconstruct the data from the faileddrive using the parity information and the remaining data from the other drives.RAID 10 Also known as RAID 0+1 or enhanced RAID. It is a combination of mirroring (RAID 1) and striping (RAID0).
SSH Securities:# vi /etc/ssh/sshd_configAnd make following changes for better securityPermitRootLogin no Disabled the Direct root loginProtocol 2 Good security conceptsBanner file Set the banner in a fileIgnoreRhosts yes If .rhosts is configured then ignoredPermitEmptyPassword no Disabled the acceptance of null passwordHostBasedAuthentication no Disabled the host based authentication
TCP Wrapper:TCP wrapper often called a s just a wrapper. It allows a system administrator to control access of TCP-based services or daemons that are wrapper aware. Download the “tcp_wrappers_7.6-ipv6.4.tar.gz” and install it.Replaced
To
and# refresh -s inetdWrapper first check host.allow, if match found access is granted. If match found in host.deny the access is resticted for perticular services.# vi /etc/host.allowtelnetd,sshd: .mydomain.com :allowtelnetd,sshd:192.168.4.10 , 192.168.6. : allow
# vi /etc/host.denytelnetd : 192.168.8., 192.168.9. : denytelnetd : 192.168.6. : allow# tcpdchk-v check the syntax in host.allow and host.deny
Audits in AIX:AIX auditing provides a framework within which to capture pertinent system and security related information, such as failed login attempts, cron usage etc. It is recommended that auditing is enabled as part of a group of measures designed to provide enhanced logging of system and security changes.Create a /audit filesystem, at least 100 MB in size. The script always creates a standard jfs filesystem. If the /audit filesystem exists, this step is skipped:# mklv -y <LV name> -t jfs -u 1 -c 1 rootvg 1 hdisk0# crfs -v jfs -d auditlv -m /audit -A yes -t no mount /audit# vi /etc/security/audit/configt /auditstart:binmode = onstreammode = off
bin:trail = /audit/trailbin1 = /audit/bin1bin2 = /audit/bin2binsize = 10240cmds = /etc/security/audit/bincmds
Add the auditing entries for root and all other users below the pre-defined audit classes users:root = general,SRC,mail,cron,tcpip,ipsec,lvm<user 1> = general,SRC,cron,tcpip<user 2> = general,SRC,cron,tcpipetc.
Update the /usr/lib/security/mkuser.default auditclasses entry to ensure that auditing is set up for any newly created users# chsec –f /usr/lib/security/mkuser.default –s user –a auditclasses=general,SRC,cron,tcpip
Add the audit startup command into /etc/inittab.# mkitab "audit:2:boot:audit start > /dev/console 2>&1 # Start audit"
Simultaneous Multi-ThreadingSimultaneous multithreading is the ability of a single physical processor to simultaneously dispatch instructions from more than one hardware thread context. Because there are two hardware threads per physical processor, additional instructions can run at the same time. The POWER5™ processor is a superscalar processor that is optimized to read and run instructions in parallel. Simultaneous multithreading allows you to take advantage of the superscalar nature of the POWER5™ processor by scheduling two applications at the same time on the same processor.Simultaneous multithreading is primarily beneficial in commercial environments where the speed of an individual transaction is not as important as the total number of transactions that are performed. Simultaneous multithreading is expected to increase the throughput of workloads with large or frequently changing working sets, such as database servers and Web servers.Multiple Hardware threads can run on one physical processor at the same time-A processor appears as two or four logical CPUs (lcpu) to AIX-Beneficial for most commercial environments-SMT is enabled by default with max no of logical CPUsCan change between SMT2 to SMT4 (POWER 7 only)Can Disabled and enabled
AIX 6.1 Terminology
What are WPARs?In contrast to LPARs, which are created and managed at the server's firmware level, AIX WPARs are software partitions that are created from, and share the resources of, a single instance of the AIX operating system. This means that you must have AIX 6.1 to create WPARs, but you can create WPARs on any System p hardware that supports AIX 6.1, including POWER4, POWER5, and POWER6 hardware. You don’t need an HMC or IVM to create or manage WPARs.There are two kinds of WPARs:System WPARsApplication WPARs
System WPARsSystem WPARs System WPARs are autonomous virtual system environments that have their own private file systems, users and groups, login, network space, and administrative domain. To users and applications, a system WPAR appears almost exactly like a full AIX system. Operating system services, such as telnet, are supported, so if network information has been configured, users can telnet into a system WPAR as root or any other defined user, issue commands, and run applications as they would on any other AIX system.-Autonomous virtual system environmentShared file system with global environment: /usr and /optPrivate file system for wpar’s own use:/, /var, /tmpUnique set of users, groups, and network addresses-Can be access via Network protocol (telnet or ssh)Login from the global environment using the clogin command-Can be stopped and restarted
Application WPARsApplication WPARs provide an environment for isolation of applications and their resources to enable checkpoint, restart, and relocation at the application level. An application WPAR is essentially a wrapper around a running application or process for the purposes of isolation and mobility. It lacks some of the system services provided by system WPARs—for example, it’s not possible to log in or telnet into an application WPAR. When the application running in an application WPAR terminates, the WPAR also ceases to exist. Application WPARs are most useful when you want to enable Live Application Mobility—that is, when you want to be able to move a running application from one AIX system to another. You might want to relocate applications to avoid downtime resulting from scheduled maintenance or to improve performance by moving an application to a more powerful server-Isolate an individual application-Light weight; quick to create and removeCreated with wparexec commandRemoved when stoppedStopped when the application finished-Shares file systems and devices with the global environment-No user login capabilities
WPAR ManagerIBM Workload Partitions Manager for AIX (WPAR Manager) is a platform management solution that provides a centralized point of control for managing workload partitions across a collection of managed systems running AIX 6.1. The managed systems might all be LPARs on a single physical server, or they might be located on multiple physical servers. Using WPAR Manager, you can monitor the health and status of multiple WPARs on multiple managed AIX systems. You can also perform all the basic WPAR life cycle operations—including create, view and manage properties, start, stop, and delete.WPAR Manager also supports relocation of WPARs between systems in a collection of managed servers. WPAR Managersupports two kinds of relocation:Manual relocation—This type of relocation is initiated by the user.Policy-based relocation—This type of relocation is initiated by WPAR Manager in response to workload conditions defined in a relocation policy.WPAR Manager is not part of AIX—it’s a separately purchased licensed program
Multi-system managementManaging WPARs on multiple AIX systems using the WPAR Manager requires two initial installation and configuration steps. First, you install and configure the management server software on an AIX system in your environment. Then, you install the WPAR Manager agent software on each AIX system that will be managed by WPAR Manager. The agent must then be configured to share WPAR data with a specific management server. After the WPAR Manager and agent components have been configured and started, the WPAR Manager automatically discovers all the managed systems, and begins to record data transmitted by the agents in an internal database.
Live Application MobilityLive Application Mobility is the capability to relocate a WPAR from one hosting system to another without having to restart any applications or processes running in the WPAR. Partition mobility refers to the ability to move an entire running AIX LPAR from one physical server to another.
Live Partition MobilityLive Partition Mobility allows you to migrate partitions that are running AIX and Linux operating systems and their hosted applications from one physical server to another without disrupting the infrastructure services. The migration operation, which takes just a few seconds, maintains complete system transactional integrity. The migration transfers the entire system environment, including processor state, memory, attached virtual devices, and connected users.Live Partition Mobility helps you meet increasingly stringent service-level agreements (SLAs) because it allows you to proactively move running partitions and applications from one server to another.Live Partition Mobility may also be used as a mechanism for server consolidation, because it provides an easy path to move applications from individual, stand-alone servers to consolidation servers. If you have partitions with workloads that have widely fluctuating resource requirements over time.Live Partition Mobility can be automated and incorporated into system management tools and scripts. Support for multiple concurrent migrations allows you to liberate system resources very quickly. For single-partition, point-in-timemigrations, the Hardware Management Console (HMC) and the Integrated Virtualization Manager (IVM) interfaces offer easy-to-use migration wizardsPartition Mobility is a hardware-based function, partition mobility is only supported on POWER6 hardware; application mobility is supported on any hardware that supports AIX 6.
Partition MigrationA partition migration operation can occur either when a partition is powered off (inactive), or when a partition is providing service (active).The migration process can be performed either with a powered-off or a live partition. The following two available migration types are discussed in more detail in the next sections:Inactive migration The logical partition is powered off and moved to the destination system.Active migration The migration of the partition is performed while service is provided, without disrupting user activities.
Policy-based relocationWPAR Manager also has the capability to monitor performance of workloads running in WPARs and relocate those workloads to different AIX systems to improve performance. For example, if CPU or memory usage in a WPAR or group of WPARs is, on average, higher than a value that you specify, then the WPAR Manager might attempt to relocate one or more WPARs to a more powerful, or less busy, server in your datacenter. The detailed mechanics of how the policy engine works are beyond the scope of this article, but will be covered in a later article.
Encrypted File SystemThe Encrypted File System (EFS) is a J2 filesystem-level encryption through individual key stores. This allows for file encryption in order to protect confidential data from attackers with physical access to the computer. User authentication and access control lists can protect files from unauthorized access while the operating system is running; however, it’s easy to circumvent the control lists if an attacker gains physical access to the computer.One solution is to store the encrypted files on the disks of the computer.In EFS, a key is associated to each user. These keys are stored in a cryptographically protected key store and upon successful login, and the user's keys are loaded into the kernel and associated with the process credentials. When the process needs to open an EFS-protected file, the system tests the credentials. If the system finds a key matching the file protection, the process is able to decrypt the file key and file content. The cryptographic information is kept in the extended attributes for each file. EFS uses extended attribute Version 2, and each file is encrypted before being written on the disk. The files are decrypted when they are read from the disk into memory so that the file data kept in memory is in clear format. The data is decrypted only once, which is a major advantage. When another user requires access to the file, their security credentials are verified before being granted access to the data even though the file data is already in memory and in clear format. If the user is not entitled to access the file, the access is refused. File encryption does not eliminate the role of traditional access permissions, but it does add more granularity and flexibility.
SnapshotThe snapshot functionality was provided so that administrators could create consistent, integrated snapshots of an online (or offline) JFS2 file system. A point-in-time copy of the source file system is created. The snapshot operation executes quickly and requires very little disk space. Multiple snapshots can be taken, which can be very handy if you need to perform multiple changes to a system with corresponding backups at each stage.The snapshot retains the same permissions as the original file system. Once the image has been created, you can use it to restore files or an entire file system to a known point in time. Or, you can mount the snapped file system and perform a backup of it to tape (via Tivoli Storage Manager for example). This could help reduce your backup time requirements.There are two types of JFS2 snapshots.The first is an external snapshot which uses a separate logical volume for the snapshot image.The second type allocates space out of the original file system, this is known as an internal snapshot.
An external snapshot can be mounted separately on its own mount point. A file system can use only one type of snapshot method at a time; that is, you cannot take an external snapshot of a file system and then take an internal snapshot of the same file system while the external exists (and vice versa).It is also important to note that you cannot use internal snapshots unless the file system was enabled to support them at the time of file system creation. So if you want to use snapshots on an existing file system and you did not enable it for internal snapshots, you will need either to recreate the file system with the snapshot option or use an external snapshot.One of the big advantages to using JFS2 snapshots is that we do not need to create a complete copy of all of the data. Nor do we need to recover the entire contents of the data. Thus reducing overhead and saving time.
Network Installation Management
Benefits of NIMNIM allows system administrators to install, upgrade, back up and maintain AIX 5L systems remotelyMultiple machines can be installed or maintained at the same timeMultiple AIX 5L versions, Maintenance Levels (MLs) or Technology Levels (TLs) can be used in the same NIM environmentDisaster recovery - local systems can be recovered quickly with NIMSystems can be installed with standard and consistent AIX 5L operating system imagesRecovering a system backup (mksysb) from tape can be difficult when booting from CD mediaCloning systems is another excellent feature of NIM which can assist you in moving from one server to another, or even from one site to another site
NIM OverviewA basic NIM environment consists of several IBM System p machines (at least two) connected via a TCP/IP network. A physical network can be shared by several NIM environments, therefore the machines attached to the samephysical network can be part of several NIM environmentsThe machines (or clients) can be standalone systems, LPARs, diskless, and dataless clients (also known as “thin clients”). The machines can be controlled by one (or more) Hardware Service Console
MasterMaster refers the machine where you set up and maintain your NIM environment. You can also initiate installations from here (push mode). It is the key piece of the NIM environment. Refer to 2.1.4, “NIM master selection” on page 26, for information about how to chose a machine to become the NIM master.ClientThe NIM client can be the target for NIM master-initiated operations such as installation, updates, and so forth (push mode). Also, a client can initiate its own installation or update (pull mode). A NIM client automatically becomes aresource server when it holds NIM resources.
Resource serverAny machine (the master or a standalone client) can be configured by the master as a server for a particular software resource. In most environments, the master is also resource server.If other machines are already reporting to the master and they are installed (AIX), you can chose one of them to act as a resource server, thus relieving the NIM master of the heavy I/O load (disk and network). In such cases, the NIM master is only used to run administrative tasks.
Push and pull modesThe push mode operation is initiated from the master. The pull mode operation is initiated from the client. The very first time a client is installed, only the pull mode can be used.
NIM databaseThe NIM database is stored in the AIX Object Data Management (ODM) repository on the NIM master and is divided into four classes: machines, networks, resources, groups.
Networks classThe network is what allows the machines in a NIM environment to communicate which each other. If the network is a simple local area network (LAN), the definition of the network object is simplified. The purpose of the network object is to depict the network topology used by the NIM environment. Each modification made on the physical network must be reflected to the NIM database.
Resources classA NIM resource is a pointer to a file or a directory located on a Resource Server, and the NIM Master is the default Resource Server. For example, the bosinst_data resource points to a file that contains information used to performthe installation process without requiring any manual intervention. The basic characteristics of a resource are the NIM name, the location (the location of the file or directory), and the resource server hosting the resource
Groups classA group is a collection of either machines or resources. Only two types are available:# res_group, for grouping a set of resources# mac_group, for grouping a set of machines
SPOTSPOT, or Shared Product Object Tree, is a directory of code (installed filesets) that is used during client booting procedure. It is equivalent in content to the code that resides in the /usr file system on a system running AIX 5L(binary objects - executables and libraries, header files and shell scripts). This resource (directory) replaces the content of the basic ramdisk image available on installation CDs.The SPOT also contains the code needed to generate the boot images (kernels, which will be stored in the /tftpboot directory) that the client uses until it can manage to NFS-mount the SPOT directory.
lpp_sourceAn lpp_source is a directory similar to AIX install CDs. It contains AIX Licensed Program Products (LLPs) in Backup File Format (BFF) format and RPM Package Manager (RPM) filesets that you can install
mksysbThis resource is a file containing the image of the root volume group (generated with the AIX mksysb command) of a machine. It is used to restore a machine, or to install it from scratch (also known as “cloning” a client). A mksysb resource can be defined by first creating the mksysb image with SMIT or the mksysb command, and then defining the resource that references the backup image file.
bosinst_dataThe bosinst_data resource is a flat ASCII file similar to the bosinst.data file used for restoring system backup images from tape or CD/DVD. When you start the installation of a IBM System p machine, you will be prompted toselect the console, language, installation method, the target disk(s) for rootvg, and so on. This resource comes in handy for both push or pull installation of multiple machines at the same time, because it will automate the installationprocess without requiring interactive answers for the installation parameters.
scriptThe script resource is a file. After the BOS installation is finished on your client, you can perform customization such as file system resizing, additional user creation, and so forth. The script resource points to a file (usually a shellscript) that contains all the commands needed to perform customization.
To rebuild and recover a NIM client /etc/niminfo file, use the niminit command on the client:niminit -a master=<MASTER_HOSTNAME> -a name=<CLIENT_NIM_NAME>
NIM improovement in AIX 5.3nimshAlternate NIM masterCreating an SPOT resource from a mksysbmksysb migration using NIM
Removing a NIM client definition/tftpboot This is used for storing NIM master boot images (kernels) and info files for the NIM clients./export/lpp_source This is used for storing versions of AIX base level filesets in specific directories; a NIM lpp_source resource represents a directory in which software installation images are stored/export/spot This is used for storing non-/usr Shared Product Object Three (SPOT) in individual directories (one per SPOT version). A SPOT is required to install or initialize all types of machine configurations, as well as to generate aspecific kernel image used for network booting a client system. Everything that a client machine requires in a /usrfile system, such as the AIX kernel, executable commands, libraries, and applications, should be included in theSPOT./export/images This is used for storing system backup images. These images can be created by using the NIM mksysb operation (“pulling” a mksysb image from a NIM client), or by copying a system backup as a file from a previously created backup file, tape, or DVD. To create subdirectories during NIM mksysb creation, you mustexport the NIM_MKSYSB_SUBDIRS=yes variables before using the command# nim -o define -t mksysb
Basic Naming Conventionslpp_source lpp<OS#><TL#><SP#>spot spot<OS#><TL#><SP#>mksysb mksysb<OS#><TL#><SP#>network net_<IP#NET>
Initializing NIM masternimconfig command can be used to perform a nim initializingConfiguring the NIM master using the nimconfig command creates the master, boot, the first network object, nim_script with the /export/nim directory; the master (nimesis) daemon is also started.root@master:/: lsnimmaster machines masterboot resources bootnim_script resources nim_scriptnet_10_1_1 networks ent
To convert an old NIM master into a client:Create a client definition on the NIM master.Remove the /etc/niminfo file (this will remove also the nim command from the system).Uninstall the bos.sysmgt.nim.master fileset.Check whether the bos.sysmgt.nim.client fileset is installed.Use the niminit command to register with the new NIM master (client registration must be enabled for this procedure to work).
Verify auxiliary services (inetd, bootpd and tftpd)root@msater:/: lssrc -ls inetdtftp /usr/sbin/tftpd tftpd -n activebootps /usr/sbin/bootpd bootpd /etc/bootptab active
The NIM master uses the bootpd (BOOT Protocol) and tftpd (Trivial File Transfer Protocol) daemons (services). Both services are normally started by the inetd daemon (InterNET master daemon).
SMS and Console flow during NIM client installationAt the Initial Program Load (IPL) menu, press the 1 key to access SMS Menu.select 2. Setup Remote IPL (Initial Program Load
Select the desired adapter from the NIC Adapters list,
select 1. IP Parameters to access the IP Parameters menu
1. From the IP Parameters menu select 1. Client IP Address, to set the Client IP Address.2. From the IP Parameters menu select 2. Server IP Address, to set the Server IP Address (NIM master).3. From the IP Parameters menu select 3. Gateway IP Address. Leave it if the NIM master and client are on the same IP subnet.4. From the IP Parameters menu select 4. Subnet Mask, to set the Subnet Mask for the IP network.5. Press ESC and return to the previous menu
Network Parameters menu, select 3. Ping Test
From the Network Parameters menu, select 1. Execute Ping Test
Next, press M from the Ping Test menu to return to the Main Menu
From the Main Menu, choose 5. Select Boot Options
From the Multiboot menu, choose 1. Select Install/Boot Device
From the Select Device Type menu, select 6. Network to choose a network
From the NIC Adapters list, in this case, select 1. to access the Select Task menu
From the Select Task menu, select 2. Normal Mode Boot
From the exit SMS menu select 1. Yes to let the network installation begin
First the network boot (IPL) is initiated for the LPAR.Next, the LPAR issues BOOTP requests to the server IP-address specified in the IP Parameters menu. It receives the BOOTP record information from /etc/bootptab on the NIM master. This record specifies the TFTP server for sending a bootable kernel. The LPAR issues TFTP requests to the server listed in the /etc/bootptab sa field from the NIM master.
Then control is passed to the transferred and loaded AIX 5L kernel.In this case the installation is made using a NIM mksysb resource.After installing the software, the boot image is created.Once the installation is finished and the LPAR is restarted and The newly installed kernel is loaded and the LPAR is restarted
LPAR and VIO Server Terminology
Logical partitioningLogical partitioning (LPAR) is the ability to logically slice up a single system's CPU, memory, and other resources to create multiple and separate servers. Each LPAR has its own allocation of CPU, memory, and I/O devices. This type of partitioning is done at the firmware level, not at the physical resource level. Thus, with the IBM System p5 servers, you are able to create LPARs using Micro-Partitioning™ -- assigning less than one physical CPU per LPAR.
Virtual I/O ServerIt is a special kind of partition called a Virtual I/O Server (VIO Server). The VIO Server provides the ability to share I/O resources among several LPARs. You define virtual Ethernet and disk devices on the VIO Server and then make them available to the other LPARs on the system. Without the ability to share I/O devices on the managed system, each LPAR would require its own dedicated devices. If you have two LPARs on a system, you would need at least two Ethernet cards and two disk controllers. However, if you have 15 LPARs, then you would need at least 15 Ethernet cards and 15 disk controllers. With the use of VIO LPARs, you could host 15 LPARs with a much smaller number of Ethernet and disk controllers.Of course, you are still able to dedicate I/O devices to an AIX partition if you need to for performance reasons. For example, if you have a large database LPAR that requires a large amount of throughput, you can dedicate a disk controller to that LPAR so that it won't have to compete for I/O resources with other LPARs using the VIO's disk controller.On the other hand, you might have several small LPARs that do not require a large amount of data throughput. They would be ideal candidates for using the VIO Server since each of them would not completely utilize the full capacity of an Ethernet card or disk controller (see Figure 3). Together they could take full advantage of the hardware's capacity
Dynamic logical partitioningNot only are you able to logically carve up your hardware into multiple LPARs, but IBM provides the capability of dynamically adding, removing, or moving resources between the partitions while they are up and running. You can add or remove CPU, memory, or I/O slots to a running partition without first shutting that LPAR down.Dynamic logical partitioning (DLPAR) provides a great deal of flexibility to your computing environment. As conditions change in your environment, you are able to react by moving hardware resources to where they are needed without interrupting services to the LPARs. The options for using this feature are endless. A word of caution, though, not all applications are able to handle the removal of CPUs or memory gracefully, so testing is needed in your particular situation before you begin to use this feature in a production environment.In the rest of this article, I'll discuss the steps you need to follow to configure an installed p5 system. It is impossible, in an article of this length, to discuss all the options available to you in configuring an LPAR system. Throughout the article, I'll cover the configuration basics and then point you to documents that provide more detailed information for that particular component. I'm also assuming the hardware is installed and connected to your network.
Micro-PartitioningMicro-Partitioning is the ability to distribute the processing capacity of one or more physical processors among one or more logical partitions. Thus, processors are shared among logical partitions.
System ProfileA system profile is a collection of one or more partition profiles. System profiles can be used to specify which partition profiles are activated at the same time.
Partition ProfileA partition profile represents a particular configuration for a logical partition. A partition profile contains information about the resources assigned to the partition. These resources include memory, processor processing capacity, and physical I/O slots. Partition profiles also include information about which devices provide important partition functions
Memory in LPARThere are three memory values for a partition profile:# Minimum memory: This amount of memory is required for the partition. The profile will fail to activate if the minimum memory is not met.# Desired memory: This is the requested amount of memory for the partition. On profile activation, the partition will receive an amount of memory between the minimum and desired amounts depending on what is available.# Maximum memory: The maximum represents the upper limit for memory. Memory allocations cannot go above the maximum. In order to go above the maximum, the maximum value needs to be changed.
Dedicated ProcessorDedicated processor means you are working with whole processors that are dedicated to a single partition. The dedicated processor handles the processing for one specific logical partition. If you choose to assign dedicated processors to a logical partition, you must assign at least one processor to that partition.
Shared ProcessorThe shared processing pool allows you to assign partial processors to a logical partition. The physical processors are held in the shared processing pool and are shared among logical partitions. At partition creation time, the shared processing units can be defined as 0.1.There are two typesCapped andUncappedCapped prohibits the LPAR from using more processing units than is currently allocated to it. If an LPAR is allocated 1.5 processing units but the workload requires 2.2 units, the LPAR would not be allowed access to the additional units, even if they were available in the shared processor pool.Uncapped allows the LPAR to use more processing units than is currently allocated. In this example, the LPAR would have access to the additional 0.7 processing units needed for the workload. If Uncapped is selected, you need to provide the Weight value for this LPAR. LPARs with higher Weight values receive higher priority when shared pool resources are fully used.
Virtual ProcessorVirtual processors tell the LPAR's OS how many physical processors it has. The OS cannot use parts of a processor, so you have to tell it in whole numbers how many processors it can use. Also, the number of virtual processors limits how many physical processors an LPAR can use. For example, if you give an LPAR two virtual processors, it cannot use more than two physical processors even though Maximum processing units is higher than two.■Minimum number of virtual processors: You must have at least one virtual processor for every part of a physical processor assigned to the LPAR. For example, if you have assigned 2.5 processing units to the LPAR, the minimum number of virtual processors is three.■Desired number of virtual processors: You can assign a virtual processor to every 0.1 processing unit allocated to the LPAR. There is no recommended value for this configuration setting. You need to use trial and error to determine what is the best setting for your application(s).■Maximum number of virtual processors: You can only have 10 virtual processors per processing unit. Therefore, you cannot assign a value greater than 10 times the maximum processing units value.
Virtual I/O AdapterVirtual IO adapters give a partition the flexibility to use certain types of resources with requiring the physical hardware to be present. If the partition will be using virtual IO, make sure that the Yes radio button is selected. Otherwise, select the No radio button.The following virtual IO adapters can be created:Virtual EthernetVirtual SerialVirtual SCSI
Virtual EthernetVirtual Ethernet technology is supported on AIX® 5.3 on POWER5™ hardware. This technology enables IP-based communication between logical partitions on the same system using a VLAN capable software switch in POWER5 systems. Shared Ethernet Adapter technology (part of the optional Virtual I/O Server feature on POWER5 hardware) enables the logical partitions to communicate with other systems without assigning physical Ethernet slots to the logical partitions.
The number of virtual Ethernet adapters for each LPAR varies by operating system. AIX 5.3 limits the number of virtual Ethernet adapters to 256 virtual Ethernet for each LPAR. Besides a PVID, the number of additional VID values that can be assigned for each virtual Ethernet adapter is 20, which indicates that each virtual Ethernet adapter can be used to access 21 networks. The HMC generates a locally administered Ethernet MAC address for the virtual Ethernet adapters so that these addresses do not conflict with physical Ethernet adapter MAC addresses. To ensure uniqueness among the virtual Ethernet adapters, the address generation is based on the system serial number, LPAR ID, and adapter ID.
When using the Integrated Virtualization Manager, only PVID is allowed (no additional VLANs), and only the PVID may only be 1-4.
Virtual Serial
Virtual serial adapters provide a point-to-point connection from one logical partition to another, or from the Hardware Management Console (HMC) to each logical partition on the managed system. Virtual serial adapters are used primarily to establish terminal or console connections to logical partitions. When you create a logical partition, the HMC automatically creates two virtual server serial adapters on the logical partition. These virtual server serial adapters allow you to establish a terminal or console connection to the logical partition through the HMC.
You can also create pairs of virtual serial adapters on logical partitions so that you can access and control one logical partition directly from another logical partition. For example, one logical partition uses the disk resources of another logical partition using virtual SCSI adapters. You can create a server serial adapter on the logical partition that uses the disk resources and a client serial adapter on the logical partition that owns the disk resources. This connection allows the logical partition that owns the disk resources to shut down the logical partition that uses the disk resources before you back up data on the logical partition that owns the disk resources.
Virtual SCSI
Virtual SCSI (Small Computer Systems Interface) adapters provide one logical partition with the ability to use storage I/O (disk, CD, and tape) that is owned by another logical partition.A virtual SCSI client adapter in one logical partition can communicate with a virtual SCSI server adapter in another logical partition. The virtual SCSI client adapter allows a logical partition to access a storage device being made available by the other logical partition. The logical partition owning the hardware is the server logical partition, and the logical partition that uses the virtualized hardware is the client logical partition. With this arrangement, the system can have many server logical partitions.
NPIV (Network Port ID Virtualization)N_Port ID Virtualization is a fibre channel industry standerd method for virtualizing a physical fibre channel port. NPIV allows one F_Port to be associated with multiple N_Port Ids, so a physical fibre channel HBA can be shared accross multiple guest operating systems. On POWER, NPIV allows logical partitions (LPARs) to have dedicated N_Port IDs, giving the OS a unique identity to the SAN, just as if it had a dedicated physical HBA(s).
"If I share an idea; you have one and I too have one but if you share one with me I have two and you too have two” - Dipak Warade
Saturday, July 5, 2014
Must know page about AIX and Related technologies
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment