Saturday, July 5, 2014

Daily Management - Level one



/etc/passwd Contains the basics of a user
/etc/security/.profile Providing the user profile
/etc/security/limits Contains all the ulimits, or users' system resource limitations
/etc/security/passwd Contains the AIX user's password information
/etc/security/user Contains the most important settings
/usr/lib/security/mkuser.default Contains values used when creating user through mkuser

# getconf LOGIN_NAME_MAX To verify the setting in AIX 5.3 and later
# lsattr -El sys0 To verify the setting in AIX 5.3 and later
# chdev -l sys0 -a max_logname=129 To change the value of sys0

# mkuser xander To create a user with default settings
# finger xander Gives the short information of user
# chuser core=1048576 xander To change the User attribute
# chsh xander To change the user login shell
# chfn xander To change GECOS information
# lsuser -f xander List the user in stanza format
# passwd xander To set or change password of user
# passwd -f xander To change user GECOS using passwd command
# passwd -s xander To change the user SHELL using passwd
# rmuser xander Remove the user
# rmuser –p xander To remove fully user information

/etc/group Contains the basics of a group
/etc/security/group Contains extended attributes to the specified group

# mkgroup atctest To create a group
# mkgroup -a atcadmin To create an admin group
# mkgroup adms=xander xangroup To create a group and add Xander as the administrator of the group
# chgroup id=204 users=xander,atc,amdc xangroup Change the group's GID and add users
# chgrpmem Another way to modify a group's members is with chgrpmem
# chgrpmem xangroup List the group information
# chgrpmem -m - atc xangroup Remove a user from a group
# lsgroup xangroup List the info of group
# lsgroup -f xangroup List info of group in stanza fromat
# rmgroup atctest Remove the group

/etc/security/environ Contains the environment attributes for users
/etc/security/lastlog Contains the last login attributes for users
/usr/lib/security/mkuser.sys Customizes new user accounts
/etc/security/login.cfg Contains system default login parameters
/etc/utmp Contains a record of users logged into the system
/var/adm/wtmp Contains connect-time accounting records
/etc/security/failedlogin Records all failed login attempts
/etc/motd Contains the message to be displayed every time user login
/etc/environment Specifies the basic environment for all processes.
/etc/profile Specifies additional environment settings for all users.
$HOME/.profile Specifies environment settings for a specific user.

# ps -ef To display all processes
# ps -f -l -ujim To list processes owned by specific users
# ps -M To list all the 64-bit processes

# kill 1095 To stop a given process
# kill -kill 2098 1569 To stop several processes that ignore the default signal
# kill -kill 0 To stop all of your processes and log yourself off
# kill -9 -1 To stop all processes that you own

# fuser -u /etc/filesystems List a process numbers and user login names of processes
# fuser -k -x -u -c /dev/hd1 To terminate all of the processes using a given file system
# fuser -kxuc /home To terminate all of the processes using a given file system
# fuser -d /usr To list all processes that are using a file that has been deleted from a given file system
# fuser -xc /tmp List open references within a specified file system
# find /home -type d -exec fuser -u {} \; Process is using a directory within the file system as its current working directory
# topas -i5 -n0 -p10 To view the top 10 processes

The svmon command captures and analyzes a snapshot of virtual memory
The svmon command creates nine types of reports
# svmon -G Global Report
# svmon -U User Report
# svmon -C Command Report
# svmon -W Workload Management Class Report
# svmon -T Workload Management Tier Report
# svmon -P Process Report
# svmon -S Segment Report
# svmon -D Detailed Report
# svmon -F Framed Report

# aclget status To display the access control information of status file
# aclput -i acldefs status To set the access control information of status from acldefs
# aclget plans | aclput status To set the access control information of status from plans
# acledit plans To edit the access control information of the plans

#chown -R john:build /tmp/src To change the owner & group of all files in the directory
# chmod 644 text To use the absolute mode
#chgrp -R staff proposals Change the group ownership of the directory named proposals of dir stuff
# at -f filename -t CCYYMMDDhhmmSS Increment Submit a job to be run at a later time
# at now -f appl/program > /dev/null 2>&1 To start at job now
# atq List the at jobs
# at -r root.1134169200.a Remove the at job
# ls /var/spool/cron/atjobs Location of at jobs

# ls -l /var/spool/cron/crontabs Location of cron jobs
# crontab -l List the all cron jobs
# crontab -e To edit crontab file
# crontab -v username Lists the status of the user's cron jobs
# crontab -r Remove the crontab file
# crontab ~deploy/deploy.schedule Runs the crontab file under user deploy

/var/adm/cron/log The cron daemon logs file
/var/adm/cron/cron.deny Deny cron access to user for cron schedular
/var/adm/cron/cron.allow Allow cron access to user for cron schedular
/var/adm/cron/at.allow Allow cron access to user for at schedular
/var/adm/cron/at.deny Deny cron access to user for at schedular

# telinit Directs the actions of the init process
# telinit M Goes into the maintainance mode
# telinit q Tell the init command to reprocess the /etc/inittab

Normally, you do not need to restart srcmstr. If the srcmstr daemon terminates abnormally, the respawn action specified in the /etc/inittab restarts the srcmstr daemon.

A command is a request to perform an operation or run a program. A program or command that is actually running on the computer is referred to as a process.

The common types of processes are:
Foreground and background processes
Processes that require a user to start them or to interact with them are called foreground processes. Processes that are run independently of a user are referred to as background processes. Programs and commands run as foreground processes by default.

Daemon processes
Daemons are processes that run unattended. They are constantly in the background and are available at all times. Daemons are usually started when the system starts, and they run until the system stops. A daemon process typically performs system services and is available at all times to more than one task or user. Daemon processes are started by the root user or root shell and can be stopped only by the root user. For example, the qdaemon process provides access to system resources such as printers. Another common daemon is the sendmail daemon.

Zombie processes
A zombie process is a dead process that is no longer executing but is still recognized in the process table (in other words, it has a PID number). It has no other system space allocated to it. Zombie processes have been killed or have exited and continue to exist in the process table until the parent process dies or the system is shut down and restarted. Zombie processes display as <defunct> when listed by the ps command.

Ctrl-C To intrrupt the process
Ctrl-Z To stop process

# fg 589934 To bring the process in to the foreground
# find / -type f > dir.paths & Run the find command in the background
# nohup find / -type f & To run the find command and leave it running after you log off in Bg

0 Represents standard input (stdin) <
1 Represents standard output (stdout) > or >> (append)
2 Represents standard error (stderr) 2> or 2>> (append)
2>&1 Redirects stderr to stdout.
1>&2 Redirects stdout to stderr


Performance Monitoring

# nice -10 foo Add 10 to current nice value (Lower Priority)
# nice -n 10 foo Add 10 to current nice value (Lower Priority)
# nice - -10 foo Subtract 10 from current value (Higher Priority)
# nice -n -10 foo Subtract 10 from current value (Higher Priority)

# renice -10 –p 563 Add 10 to default nice value (Lower Priority)
# renice –n 10 –p 563 Add 10 to current nice value (Lower Priority)
# renice - -10 –p 563 Subtract 10 from default nice value (Higher Priority)
# renice –n -10 –p 563 Subtract 10 from current nice value (Higher Priority)

# ps –ekl Long listing of kernel processes with priority
# ps –L 483445 –l It list the child the processes of parent process
# ps –kmo THREAD –p 16396 It list the threads of particular processes

# schedo To change the CPU usage priority decay rate

Context Switch : A context switch is when one thread is taken off a CPU and another thread is despatched onto the same CPU.

User Mode : User mode is when thread is executing its own application code or shared library code. Time spent in user mode is reflected as %user time in output of commands such as vmstat, topas, iostat, sar.

System Mode : System Mode is when the CPU is executing code in the kernel. CPU time spent in kernel mode is reflected as system time in output of vmstat, topas, iostat, sar commands. Context switch time, system calls, device interrupts, NFS I/O, and anything else in the kernel is considered as system time.

# time It show the elapsed time in system and user mode
# vmstat 5 3


# vmstat -f To display fork statistics
# vmstat -s To display count of various events

# sar –P ALL 5 1 All CPU related usage
# sar –q 5 3 Queue details
# ps aux Locating the dominant processes
# tprof –x sleep 60 Reports processor usages

# lparstat 2 3 System wide CPU report
# sar –P ALL 2 2 Viewing CPU statistic with SMT
# sar -d 1 2 Disk I/O statistic
# smtctl To check SMT is enabled or not
# smtctl –m on –w now Tern on SMT immediately
# smtctl –m on –w boot Tern on SMT at next reboot
# bindprocessor –q To check logical CPU

# iostat I/O device statistics


# iostat -d hdisk2 2 Continuous disk report
# netstat network statistic
# netstat -rn Network routing table
# netstat -rs Network routing statistic

CPU bound A system is said to be CPU-bound if the total system (sy) and user (us) CPU usage is approaching 100 percent. This would imply that idle time and wait time for CPU are approaching zero.

Memory bound A system is memory-bound if some virtual memory is forced out to disk, meaning the system is waiting on a relatively slow disk instead of relatively fast RAM. This is indicated by a non-zero value in the page-in (pi) and page-out (po) values.

Paging rate is the average number of page-ins and page-outs per CPU cycle.

Idle time calculation:
Total CPU Idle Time % = wait % + Idle Time %
Ex:
Average CPU Idle Time percentage = ((99+1) + (97+2) + (95+4) + (99+1))/4 = 99.5%




Must know page about AIX and Related technologies

  1. AIX Booting Process explained below.

    I. Read Only Storage Kernel Init Phase

    Motherboard is Checked - POST --> look for Bootlist if found (Detected by ROS)------> look for boot image---Bootstrap code is loaded from disk

    - BLV contains AIX Kernel, rc.boot scripts, reduced ODM, and boot commands.
    - BLV get uncompressed in RAM and kernel gets released from it.
    - Then AIX kernel gets the control
    - AIX Kernel creates a RAMFS
    - Kernel Starts the init process from BLV
    - init executes rc.boot script from the BLV in the RAM

    II. Base Device Configuration Phase

    1. All devices are configured with cfgmgr command - init process from RAMFS executes rc.boot 1

    - restbase command copies ODM from BLV to RAMFS
    - cfgmgr will run and configure all the base devices

    III. System Boot Phase

    1. Logical volumes are varied on - rc.boot 2 is executed.

    - ipl_varyon will varyon the rootvg
    - run fsck on / and mount the same to RAMFS
    - run fsck on /usr and mount the same to RAMFS
    - run fsck on /var and mount the same to RAMFS

    2. Paging is started - copycore command checks the occurrunce of the dump and copies the same from primary dump device to /var/ras/adm

    - unmount /var and activate the paging space
    - mount /var
    - Now /, /usr and /var are mounted in rootvg in hdisk

    3. /etc/inittab is processed - Kernal remove RAMFS

    - init process is started from / in rootvg
    - in rc.boot3, init processes /etc/inittab and remaining devices are configured.
    - /etc/init starts /etc/inittab and runs /sbin/rc.boot 3
    - run fsck on /tmp and mount the same.
    - syncvg for rootvg and reports the stale PPs

    4. Relevant Services according to run level starts. srcmstr daemon controls this - starts the deamons like errdaemon, syncd etc..


    * AIX 6.1, November 9, 2007
    o Workload Partitions o Live Application Mobility
    o Role Based Access Control RBAC o AIX Security Expert hardening tool
    o Trusted AIX o Encrypting JFS2 filesystem
    o Trusted Execution o Concurrent Kernel Maintenance
    o Kernel exploitation of POWER6 storage keys o probevue dynamic tracing
    o Systems Director Console for AIX o Integrated filesystem snapshot
    o Partition mobility on POWER6

    * AIX 5L 5.3, August 13, 2004
    o NFS Version 4 support o Advanced Accounting
    o Virtual SCSI o Virtual Ethernet
    o Simultaneous multithreading (SMT) support o Micro-Partitioning support
    o POWER5 support o JFS2 quota support
    o JFS2 filesystem shrink support

    * AIX 5L 5.2, October 18, 2002
    o Introduced support for the IBM BladeCenter JS20 with the PowerPC 970.
    o Minimum level required for POWER5 hardware
    o Support for MPIO Fibre Channel disks
    o iSCSI Initiator software
    o Dynamic LPAR support
    * AIX 5L 5.1, May 4,2001
    o Introduced support for the IA-64 architecture, although this never went beyond beta,[1]
    o Minimum level required for POWER4 hardware and the last release that supported Micro Channel architecture
    o Introduction of 64-bit kernel, installed but not activated by default
    o JFS2
    o introduced Logical Partitioning on POWER4
    o The L stands for Linux affinity
    o Trusted Computing Base (TCB)

    Startup Mode
    Normal Mode
    Login Prompts
    All processes running
    Multi User Mode
    System Management Services
    Not AIX
    Runs from firmware
    sets boot list
    Maintainance Mode
    Maintainance Menu
    Recover root password
    Fix machines that wont boot
    Dignostic
    AIX diagnostic

    If device state is

    undefined------>Supported Devices------->Predefined Database
    defined--------->Not Usable------------------>Customized Database
    Available--- --->Ready for use-------------->Customized Database

    Seven types supported:
    Journaled File System (JFS)
    Enhanced Journaled File System (JFS2)
    CD-ROM File System (CDRFS)
    DVD-ROM File System (UDFS)
    Network File System (NFS)
    Common Internet Filesystem (CIFS)
    Proc File System (PROCFS)

    Defaults Lvs and Filesystems on it

    AIX Installation method
    New and Complete Overwrite Installation:
    You have a new machine without a prior or useful system installation. You want to install onto a hard disk that contains an existing root volume group that you wish to completely overwrite. You want to reassign your hard disks, that is, to make your rootvg smaller and assign less disk space to it.
    Migration installation
    Use the migration installation method to upgrade AIX Version 4.2, 4.3, AIX 5L Version 5.1, or Version 5.2 to AIX 5L Version 5.3 while preserving the existing root volume group. The installation process determines which optional software products must be installed. With the exception of /tmp, this method preserves most file systems, including the root volume group, logical volumes, and system configuration files.
    Preservation installation
    Use the preservation installation method when a version of the BOS is installed on your system, and you want to preserve the user data in the root volume group. However, this method overwrites the /usr, /tmp, /var, and / (root) file systems by default, so any user data in these directories is lost. These file systems are removed and recreated, so any other LPPs or filesets that you installed on the system will also be lost. System configuration must be done after doing a preservation installation.
    The /etc/preserve.list file contains a list of system files to be copied and saved during a preservation BOS installation. The /etc/filesystems file is listed by default. You can add the full path names of any additional files that you want to
    save during the Preservation installation to the preserve.list file.

    Trusted Computing Base
    The Trusted Computing Base (TCB) is the part of the system that is responsible for enforcing the information security policies of the system. By installing a system with the TCB option, you enable the trusted path, trusted shell, trusted
    processes, and system-integrity checking. Because every device is part of the TCB, every file in the /dev directory is monitored by the TCB

    Network Installation Management
    Network Installation Management (NIM) allows you to manage an automated installation of the Base Operating System (BOS) and optional software on one or more machines. You can install a group of machines with a common configuration or customize an installation for your specific needs. The number of machines you can install simultaneously depends on the throughput of your network, the disk access throughput of the installation servers, and the platform type of your servers.
    The types of machines that can be managed in the NIM environment are stand-alone, diskless, and dataless clients.
    NIM resource objects represent files and directories that are used to support some type of NIM operation a copy of the current running rootvg to another disk, be sure to have all the file systems mounted that you want cloned across.

    Alternate Disk Installation
    Alternate disk installation, available starting with AIX Version 4.3, allows for system installation on a system while it is still up and running. Install or upgrade downtime decreases considerably. It also allows large facilities to manage an
    upgrade, while the systems are still running at the existing version. The switch over to the new version can then happen with a simple reboot, with the possibility to roll back to the original situation in case of problems.
    Cloning the rootvg to an alternate disk has many advantages. One advantage is having an online backup available, as in the case of a disk crash. Keeping an online backup requires an extra disk or disks to be available on the system. where altinst_rootvg is the cloned non-active/varied off rootvg. The cloned rootvg has all its logical volumes prefixed with the name 'alt'. The boot list is also changed to boot off altinst_rootvg. AIX likes to do things like this; it assumes you will want to boot off the cloned and not the real rootvg.
    Alternate disk installation can be used in one of two ways:
    1) Cloning the current running rootvg to an alternate disk. smitty alt_clone
    2) Installing a mksysb image on another disk. smitty alt_mksysb
    With a mksysb image, you can clone one system image onto multiple target systems. However, the target systems might not contain the same hardware.

    By default, using the alt_disk_install command does the following:
    1. Creates an /image.data file based on the current rootvg's configuration. A customized image.data file can be used.
    2. Creates an alternate rootvg (altinst_rootvg).
    3. Creates logical volumes and file systems with the alt_inst prefix.
    4. Generates a backup file list from the rootvg, and if an exclude.list file is given, those files are excluded from the list.
    5. Copies the final list to the altinst_rootvg's file systems.
    6. If specified, the installp command installs updates, fixes, or new filesets into the alternate file system.
    7. The bosboot command creates a boot logical volume on the alternate boot disk.
    8. If a customization script is specified, it runs at this point.
    9. The file systems are then unmounted, and the logical volumes and file systems are renamed.
    10.The logical volume definitions are exported from the system to avoid confusion with identical ODM names, but the altinst_rootvg definition is left as an ODM placeholder.
    11.By default, the bootlist is set to the new cloned rootvg for the next reboot.

    Maintenance Level
    A Maintenance Level (ML) is the service updates that are necessary to upgrade the base operating system (BOS) or an optional software product to the current release level.

    Technology Level
    A Technology Level (TL) is the new term for the twice yearly AIX 5L releases, which contain new hardware and software features and service updates. The first TL will be restricted to hardware features and enablement, as well as software service. The second TL will include hardware features and enablement, software service, and new software features.

    Fileset
    A fileset is the smallest installable base unit for the AIX 5L operating system.

    Package
    A package is a group of separately installable filesets that provide a set of related functions. For example, bos.net is a package

    Licensed Program Product
    A Licensed Program Product (LPP) is a complete software product including all packages associated with that licensed program. For example, the BOS is a licensed program.

    Bundle
    A bundle is a list of software that can contain filesets, packages, and LPPs that are suited for a particular use, such as providing personal productivity software or software for a client machine in a network environment

    PTF
    PTF is an acronym for Program Temporary Fix. A PTF is an updated fileset or a new fileset that fixes a previous system problem. PTFs are installed in the same way as regular filesets by the use of the installp command

    APAR
    APAR is an acronym for Authorized Program Analysis Report. An APAR is an emergency fix, or interim fix, to a unique problem on the system. APARs will eventually become PTFs after testing and verification. APARs are applied to the
    system through the use of the instfix command

    Applied State
    The applied state places software on the system and retains the previous version of the software. When an update is in the applied state, the previous version is stored in the /usr/lpp/PackageName directory, where PackageName could be bos.net which is a package.

    Commited State
    The committed state places software on the system and removes all previous levels of the software from the /usr/lpp/PackageName directory. If committed software needs to be removed, you cannot go back to the previous version without a complete reinstall of the previous version software.

    SUMA
    automatic download, scheduling, and notification capabilities through the new Service Update Management Assistant (SUMA) tool

    Object Data Manager
    The ODM is a repository in which the operating system keeps information regarding your system, such as devices, software, or TCP/IP configuration. The ODM is an object-oriented database that contains vital data, keeps this data
    consistent with the actual state of the system, and prevents the administrator from altering it by mistake.
    System data managed by the ODM includes:
    Device configuration information
    Display information for SMIT (menus, selectors, and dialogs)
    Vital product data for installation and update procedures
    Communication configuration information
    System resource controller data information
    Error log and dump information
    Network Installation Manager (NIM) information
    ODM data is stored in binary format. You cannot modify ODM files with a text editor. You must use special commands that are designed to interact with the ODM.

    The basic components of the ODM are
    Object classes Each file of the database is an object class. Each object class consists of objects having similar definitions.
    Objects Each object is one record in an object class. It is a stand-alone entity and has one or more descriptors.
    Descriptors The descriptors describe the layout of the objects. They determine the name and data type of the fields that are part of the object class. The descriptors of an object and their associated values can be located and changed using ODM commands.

    ODM information is divided in three parts in order to support diskless or dataless systems. The names of these three directories as follows:
    /usr/lib/objrepos Contains the predefined objects classes, SMIT menu object classes and the four object classes used by SWVPD for the /usr part of the installable software product. The object classes in this repository can be shared across the network by /usr clients, dataless and diskless workstations. Software installed in the /usr part can be shared among several machines with compatible hardware architecture.

    /usr/share/lib/objrepos Contains the four object classes used by the SWVPD for the /usr/share part of the installable software product. The /usr/share part of a software product contains files that are not hardware dependent. They can be shared among several systems, even if these have a different hardware architecture. An example are terminfo files that describe terminal capabilities. Because terminfo is used on many UNIX systems, terminfo files are part of the /usr/share part of the software product.

    /etc/objrepos Contains the customized devices object classes and the four object classes used by SWVPD for the / part of the installable software product. To access information in the other directories, this directory contains symbolic links to the predefined devices object classes. These links are needed because the ODMDIR variable points to only
    /etc/objrepos. It contains the part of the product that cannot be shared with other systems. Most of this
    software requiring a separate copy for each machine is associated with the configuration of the machine or
    product.

    The ODM commands are:
    odmadd Adds objects to an object class.
    odmchange Changes specific objects in a specified object class.
    odmcreate Creates empty object classes.
    odmdelete Removes objects from an object class.
    odmdrop Removes an entire object class.
    odmget Retrieves objects from object classes and puts the object information into odmadd command format.
    odmshow Displays the description of an object class.

    System Management Interface Tool
    System Management Interface Tool (SMIT) provides an alternative to the typical method of using complex command syntax, valid parameter values, and custom shell path names for managing and maintaining your operating
    system configuration.
    Software Installation and Maintenance
    Software License Management
    Devices
    System Storage Management (Physical & Logical Storage)
    Security & Users
    Communications Applications and Services
    Print Spooling
    Advanced Accounting
    Problem Determination
    Performance & Resource Scheduling
    System Environments
    Processes & Subsystems
    Applications
    Installation Assistant
    Cluster Systems Management
    Using SMIT (information only)

    SMIT runs in two modes: ASCII (non-graphical) and X-Window (graphical). ASCII SMIT can run on both terminals and graphical displays. The graphical mode, which supports a mouse and point-and-click operations, can be run only on a
    graphical display and with X-Window support. The ASCII mode is often the preferred way to run SMIT because it can be run from any machine.
    To start the ASCII mode, type at the command line:
    smitty or smit -a
    To start the graphical mode, type:
    smit or smit -m
    SMIT is an interactive, menu-driven user interface that allows you to more easily perform routine system management tasks and to manage and maintain your operating system configuration.

    SMIT uses three types of screens: menu, selector, and dialog screens. Menu screens display a list of items that you can select. Selector screens, often presented as a pop-up menu, display a list of items from which you specify or select a particular item Dialog screens are the interface to a command or task that you perform

    Trusted Program,
    A trusted program, or trusted process, is a shell script, a daemon, or a program that meets a particular standard of security.
    Examples of trusted daemons are:
    _ ftpd
    _ rexecd
    _ telnetd
    Examples of non-trusted daemons are:
    _ rshd
    _ rlogind
    _ tftpd

    Bootlist Command
    The bootlist command allows the user to display and alter the list of possible boot devices from which the system may be booted. This command supports the updating of the following:
    Normal boot list The normal list designates possible boot devices for when the system is booted in normal mode.
    Service boot list The service list designates possible boot devices for when the system is booted in service mode.
    Previous boot device This entry designates the last device from which the system booted. Some hardware platforms may attempt to boot from the previous boot device before looking for a boot device in one of the other lists.

    Network Configuration Files
    $HOME/.netrc
    The $HOME/.netrc file contains information used by the automatic login feature of the rexec and ftp commands. It is a hidden file in a user's home directory and must be owned either by the user executing the command or by the root user.
    $HOME/.forward
    When mail is sent to a local user, the sendmail command checks for the $HOME/.forward file for that user. The $HOME/.forward file can contain one or more addresses or aliases. Any messages are then sent to the addresses or
    aliases in the $HOME/.forward file.
    /etc/hosts.equiv and $HOME/.rhosts
    The /etc/hosts.equiv file, along with any local $HOME/.rhosts files, defines the hosts (computers on a network) and user accounts that can invoke remote commands on a local host without supplying a password. A user or host that is
    not required to supply a password is considered trusted, though the daemons that initiate the connections may be non-trusted in nature (for example, rlogind).

    Intreduction to rc.* files
    rc.boot file
    The /sbin/rc.boot file is a shell script that is called by the simple shell init and the standard init command to bring up a system. It controls the machine boot process. When the system is booting, the /sbin/rc.boot file is called on each boot
    phases, each time being passed a different parameter.
    Depending upon the type of boot device, the rc.boot file configures devices and also calls the appropriate applications.
    Appropriate applications include:
    _ Booting from disk (boot phase 1)
    _ Varying on a root volume group (boot phase 2)
    _ Enabling file systems (boot phase 2)
    _ Calling the BOS installation programs or diagnostics
    /etc/rc file
    The /etc/rc file performs normal startup initialization; its entry in the /etc/inittab file is located after the rc.boot entry. The init command reads the /etc/inittab file and creates a process for the /etc/rc file. The contents of the /etc/rc file are installation specific. If all of the necessary operations complete successfully, the file exits with a zero return code that allows the init command to start loggers to complete normal initialization and startup.
    Many bringup functions are done by the /etc/rc file, such as:
    _ Vary on all volume groups marked as auto-varyon.
    _ Activate all paging spaces listed on /etc/swapspaces (using the swapon –a command).
    _ Configure all dump devices (using the sysdumpdev -q command).
    _ Perform file system checks (using the fsck -fp command).
    _ Perform mounting of file systems marked as mount=true on the /etc/filesystems file (using the mount all command).
    rc.net file
    The /etc/rc.net file is a shell script that contains network configuration information The stanzas allow you to enable the network interfaces and set the host name, the default gateway, and any static routes for the current host. This file can be used as a one-step configuration alternative to using individually the set of commands and files necessary to configure a host.
    The rc.net shell script is run by the configuration manager program during the second phase of configuration. If TCP/IP is installed, a second script, rc.tcpip, is run from the init command after the second phase of configuration has completed and after the init command has started the SRC master.
    rc.tcpip file
    The /etc/rc.tcpip file is a shell script that, when executed, uses SRC commands to initialize selected daemons. The rc.tcpip shell script is automatically executed with each system restart. It can also be executed at any time from the command line.
    Most of the daemons that can be initialized by the rc.tcpip file are specific to TCP/IP. These daemons are:
    _ inetd (started by default)
    _ gated
    _ routed
    _ named
    _ timed
    _ rwhod
    /etc/tcp.clean can be used to stop TCP/IP daemons.

    System boot without starting rc.tcpip
    Removing the rc.tcpip entry in /etc/inittab means that you are not starting any server applications during IPL.
    Without the server applications started, you will not be able to telnet or ftp to this machine from another host.
    However, as long as you have not brought down the network interface, you can still utilize the client network services. You can still ping other hosts, and you can still telnet or ftp to other hosts.

    The /etc/services file
    The /etc/services file contains information about the known services used in the DARPA Internet network by inetd. Each service listed in /etc/services runs on a specific port number for communications, in a specific format, such as TCP or UDP.
    If you edit the /etc/services file, run the refresh -s inetd command, in order for your changes to be used.

    File System Issues




    Network File System
    The Network File System (NFS) is a distributed file system that allows users to access files and directories of remote servers as though they were local. For example, you can use operating systems commands to create, remove, read,
    write, and set file attributes for remote files and directories. NFS is independent of machine types, operating systems, and network architectures because of its use of remote procedure calls (RPC) for these services.
    For the successful implementation of an NFS environment, you need the following things:
    1. The NFS daemons should be running on the server and the clients.
    2. The file systems that need to be remotely available will have to be exported.
    3. The exported file systems need to be mounted on the remote (client) systems.

    The following are a list of terms that are used throughout this discussion:
    Server A computer that makes its file systems, directories, and other resources available for remote access.
    Clients The computers, or their processes, that use a server’s resources.
    Export The act of making file systems available to remote clients.
    Mount The act a client needs to do to access the file systems that a server exports.

    The following describes the mounting process:
    When the server starts, the /etc/rc.nfs script runs the exportfs command, which reads the server /etc/exports file and then tells the kernel which directories are to be exported and which access restrictions they require.
    The rpc.mountd daemon and several nfsd daemons (eight, by default) are then started by the /etc/rc.nfs script.
    When the client starts, the /etc/rc.nfs script starts several biod daemons (eight, by default), which forward client mount requests to the appropriate server.
    Then the /etc/rc.nfs script executes the mount command, which reads the file systems listed in the /etc/filesystems file.
    The mount command locates one or more servers that export the information the client wants and sets up communication between itself and that server. This process is called binding.
    The mount command then requests that one or more servers allow the client to access the directories in the client /etc/filesystems file.
    The server rpc.mountd daemon receives the client mount requests and either grants or denies them. If the requested directory is available to that client, the rpc.mountd daemon sends the client's kernel an identifier called a file handle.
    The client kernel then ties the file handle to the mount point (a directory) by recording certain information in a mount record.

    There are three types of NFS mounts: predefined, explicit, and automatic.
    Predefined mounts are specified in the /etc/filesystems file.
    Explicit mounts are usually made for short periods of time when there is a requirement for occasional unplanned mounts. There is no entry for shared file system in /etc/filesystems
    Automatic mounts are controlled by the automount command, which causes the AutoFS kernel extension to monitor specified directories for activity. If a program or user attempts to access a directory that is not currently mounted, then AutoFSintercepts the request, arranges for the mount of the file system, and then services the request.

    The following describes the mounting process:
    1. When the server starts, the /etc/rc.nfs script runs the exportfs command, which reads the server /etc/exports file and then tells the kernel which directories are to be exported and which access restrictions they require.
    2. The rpc.mountd daemon and several nfsd daemons (eight, by default) are then started by the /etc/rc.nfs script.
    3. When the client starts, the /etc/rc.nfs script starts several biod daemons (eight, by default), which forward client mount requests to the appropriate server.
    4. Then the /etc/rc.nfs script executes the mount command, which reads the FS listed in the /etc/filesystems file.
    5. The mount command locates one or more servers that export the information the client wants and sets up communication between itself and that server. This process is called binding.
    6. The mount command then requests that one or more servers allow the client to access the directories in the client /etc/filesystems file.
    7. The server rpc.mountd daemon receives the client mount requests and either grants or denies them. If the requested directory is available to that client, the rpc.mountd daemon sends the client's kernel an identifier called a file handle.
    8. The client kernel then ties the file handle to the mount point (a directory) by recording certain information in a mount record.

    The inetd daemon
    The /usr/sbin/inetd daemon provides Internet service management for a network. This daemon reduces system load by invoking other daemons only when they are needed.
    If you change the /etc/inetd.conf using SMIT, then the inetd daemon will be refreshed automatically and will read the new /etc/inetd.conf file. If you change the file using an editor, run the
    # refresh -s inetd
    or
    # kill -1 InetdPID
    Existing sessions are not affected when the inetd daemon is stopped, but no new telnet and ftp sessions can be established without first restarting the inetd daemon.

    The portmap daemon
    The portmap daemon converts remote procedure call (RPC) program numbers into Internet port numbers.

    /etc/netsvc.conf Specifies default order for Host name resolutions
    /etc/hosts provides a list of server names or aliases and their IP addresses
    /etc/resolv.conf Name resolution file

    Maximum transmission unit:
    MTU size is critical for proper network communications. Packets that are too small in length may be lost during transmission. Packets that are too long in length may collide with other packets that are being transmitted. These factors can lead to slower transmission rates and other network problems as packets must then be retransmitted.


    To automate the NTP start after reboot:
    # vi /etc/inittab
    ntp:2:wait:/usr/bin/startsrc -sxntpd
    OR
    Uncommented the commented the line from /etc/rc.tcpip
    start /usr/sbin/xntpd "$src_running"


    Virtual Ethernet
    Virtual Ethernet enables inter-partition communication without the need for physical network adapters assigned to each partition. Virtual Ethernet allows the administrator to define in-memory connections between partitions handled at system level

    Paging Space.
    To accommodate a large virtual memory space with a limited real memory space, the system uses real memory as a work space and keeps inactive data and programs on disk. The area of the disk that contains this data is called the
    system paging space.
    *Do not allocate more than one paging space logical volume on a physical volume.
    *Avoid putting a paging space logical volume on a heavily active logical volume.
    *Make each paging space logical volume equal in size.
    *Do not extend a paging space logical volume onto multiple physical volumes.


    LVM Concepts
    The fundamental concepts used by LVM are physical volumes, volume groups, physical partitions, logical volumes, logical partitions, file systems, and raw devices. Some of their characteristics are presented as follows:
    Each individual disk drive is a named physical volume (PV) and has a name such as hdisk0 or hdisk1.
    One or more PVs can make up a volume group (VG). A physical volume can belong to a maximum of one VG.
    You cannot assign a fraction of a PV to one VG. A physical volume is assigned entirely to a volume group.
    Physical volumes can be assigned to the same volume group even though they are of different types, such as SCSI or SSA.
    Storage space from physical volumes is divided into physical partitions (Pps). The size of the physical partitions is identical on all disks belonging to the same VG.
    Within each volume group, one or more logical volumes (LVs) can be defined. Data stored on logical volumes appears to be contiguous from the user point of view, but can be spread on different physical volumes from the same
    volume group.
    Logical volumes consist of one or more logical partitions (LPs). Each logical partition has at least one corresponding physical partition. A logical partition and a physical partition always have the same size. You can have up to three
    copies of the data located on different physical partitions. Usually, physical partitions storing identical data are located on different physical disks for redundancy purposes.
    Data from a logical volume can be stored in an organized manner, having the form of files located in directories. This structured and hierarchical form of organization is named a file system.
    Data from a logical volume can also be seen as a sequential string of bytes. This type of logical volumes are named raw logical volumes. It is the responsibility of the application that uses this data to access and interpret it correctly.

    Volume Group Descriptor Area
    The volume group descriptor area (VGDA) is an area on the disk that contains information pertinent to the volume group that the physical volume belongs to. It also includes information about the properties and status of all physical and logical volumes that are part of the volume group. The information from VGDA is used and updated by LVM commands. There is at least one VGDA per physical volume. Information from VGDAs of all disks that are part of the
    same volume group must be identical. The VGDA internal architecture and location on the disk depends on the type of the volume group (original, big, or scalable).

    Volume Group Status Area
    The volume group status area (VGSA) is used to describe the state of all physical partitions from all physical volumes within a volume group. The VGSA indicates if a physical partition contains accurate or stale information. VGSA is used for monitoring and maintained data copies synchronization. The VGSA is essentially a bitmap and its architecture and location on the disk depends on the type of the volume group.

    Logical Volume Control Block
    A logical volume control block (LVCB) contains important information about the logical volume, such as the number of the logical partitions or disk allocation policy. Its architecture and location on the disk depends on the type of the volume group it belongs to. For standard volume groups, the LVCB resides on the first block of user data within the LV. For big volume groups, there is additional LVCB information in VGDA on the disk. For scalable volume groups, all relevant logical volume control information is kept in the VGDA as part of the LVCB information area and the LV entry area.


    Logical Track Group
    Logical track group (LTG) size is the maximum allowed transfer size for an I/O disk operation.
    # lquerypv -M hdisk0

    Quorum
    The quorum is one of the mechanisms that the LVM uses to ensure that a volume group is ready to use and contains the most up-to-date data.
    A quorum is a vote of the number of Volume Group Descriptor Areas and Volume Group Status Areas (VGDA/VGSA) that are active. A quorum ensures data integrity of the VGDA/VGSA areas in the event of a disk failure. Each physical disk in a volume group has at least one VGDA/VGSA. When a volume group is created onto a single disk, it initially has two VGDA/VGSA areas residing on the disk. If a volume group consists of two disks, one disk still has two VGDA/VGSA areas, but the other disk has one VGDA/VGSA. When the volume group is made up of three or more disks, then each disk is allocated just one VGDA/VGSA.
    A quorum is lost when at least half of the disks (meaning their VGDA/VGSA areas) are unreadable by LVM. In a two-disk volume group, if the disk with only one VGDA/VGSA is lost, a quorum still exists because two of the three VGDA/VGSA areas still are reachable. If the disk with two VGDA/VGSA areas is lost, this statement is no longer true. The more disks that make up a volume group, the lower the chances of quorum being lost when one disk fails.
    When a quorum is lost, the volume group varies itself off so that the disks are no longer accessible by the LVM. This prevents further disk I/O to that volume group so that data is not lost or assumed to be written when physical problems occur. Additionally, as a result of the vary-off, the user is notified in the error log that a hardware error has occurred and service must be performed.
    There are cases when it is desirable to continue operating the volume group even though a quorum is lost. In these cases, quorum checking can be turned off for the volume group. This type of volume group is referred to as a nonquorum volume group. The most common case for a nonquorum volume group occurs when the logical volumes have been mirrored. When a disk is lost, the data is not lost if a copy of the logical volume resides on a disk that is not disabled and can be accessed. However, there can be instances in nonquorum volume groups, mirrored or nonmirrored, when the data (including copies) resides on the disk or disks that have become unavailable. In those instances, the data might not be accessible even though the volume group continues to be varied on.

    Unlock VG
    A volume group can become locked after an abnormal termination of an LVM command. You can remove the lock using the chvg -u command.

    Journaled file system
    This type of file system is named journaled because the system uses journaling techniques to maintain the
    integrity of control structures. Each journaled file system must reside on a distinct jfs logical volume. Therefore, the
    file system size will be a multiple of the size of a logical partition.

    Enhanced journaled file system
    This is the enhanced version of the initial journalized file system. It uses extent based allocation to allow higher
    performance, larger file systems, and a larger file size. Each enhanced journaled file system must reside on a
    distinct jfs2 logical volume. When the operating system is installed using the default options, it creates JFS2 file
    systems.

    Controll Growing file
    /var/adm/wtmp
    /etc/security/failedlogin
    /var/adm/sulog
    /var/spool/*/*
    $HOME/smit.log
    $HOME/smit.script
    $HOME/websm.log
    $HOME/websm.script
    JFS and JFS2 Difference

    Superblock
    The superblock contains control information about a file system, such as the overall size of the file system in 512 byte blocks, the file system name, the file system log device, the version number, the number of inodes, the list of free
    inodes, the list of free data blocks, date and time of creation, and file system state. All this data is stored in the first logical block of the file system. Corruption of this data may render the file system unusable. This is why the system keeps a second copy of the superblock on logical block 31.

    Allocation group
    An allocation group consists of inodes and its corresponding data blocks. An allocation groups spans multiple adjacent disk blocks and improves the speed of I/O operations. Both JFS and JFS2 file systems use allocation groups. For a JFS file system, the allocation group size can be specified when the file system is created.

    Inodes
    The inode contains control information about the file, such as type, size, owner, and the date and time when the file was created, modified, or last accessed. It also contains pointers to data blocks that store the actual data of the file. Every file has a corresponding inode.
    For JFS file systems, the maximum number of inodes, and hence the maximum number of files, is determined by the number of bytes per inode (nbpi) value, which is specified when the file system is created. For every nbpi bytes of your file system, there will be an inode created. The total number of inodes is fixed. The nbpi values needs to be correlated with allocation group size. The JFS restricts all file systems to 16 MB (224) inodes. JFS2 file systems manages the necessary space for inodes dynamically so there is not any nbpi parameter.

    Data blocks
    Data blocks store the actual data of the file or pointers to other data blocks. The default value for disk block size is 4 KB.

    Fragments
    Fragments of logical blocks can be used to support files smaller than the standard size of the logical block (4 KB). This rule applies only to the last block of a file smaller than 32 KB.
    For JFS file systems only, you have the option to use compression to allow all logical blocks of a file to be stored as a sequence of contiguous fragments. Compression for a file system will increase the amount of CPU and I/O activity
    when using that file system. These features can be useful to support a large number of small files. Fragment
    size must be specified for a file system at installation time. Different file systems can have different fragment sizes.

    Device logs
    The journaled file system log stores transactional information about file system metadata changes.This data can be used to roll back incomplete operations if the machine crashes. JFS file systems are used for logging logical volumes of type jfslog, while JFS2 file systems are used for logging logical volumes of type jfs2log.
    Data from data blocks are not journaled. Log devices ensure file system integrity, not data integrity. After the operating system is installed, all file systems within the rootvg volume group use logical volume hd8 as a common log. You can create a JFS2 file system that can use inline logs. This means the log data is written into the same logical volume as the file system, and not into the log logical volume.
    Full backup
    All the files are put on media during a full backup.

    Differential Backup
    The differential backup strategy first looks for the modification time of a file and compares it with the last full backup time. In other words, only modified files are backed up, but only if they changed after the latest full backup. Differential backups are cumulative; once a file has been modified, it will be included in every differential backup until the next full backup.

    Incremental backups
    Incremental backups are similar to differential backups in that both back up only modified files. However, incremental backup checks the difference between the modification time of a file and the last backup time (either being full or
    incremental backup). If the modification date is more recent than the last backup date, the file is backed up.

    MKSYSB
    The mksysb command creates a bootable image of all mounted file systems on the rootvg volume group. You can use this backup command to restore a system to its original state. The tape format includes a BOS boot image, a BOS install image, and a dummy table of contents (TOC) followed by the system backup (root volume group) image. The root volume group image is in backup-file format, starting with the data files and then any optional map files.
    User-defined paging spaces, unmounted file systems, and raw devices are not backed up.

    The BOS boot image contains a copy of the system’s kernel and device drivers needed to boot from the mksysb tape. It is created by the bosboot command.
    There are three important files in the mkinsttape image (the second section of the mksysb tape): ./tapeblksz, ./bosinst.data, and ./image.data. The ./tapeblksz file contains the block size the tape drive was set to when the mksysb command was run.
    The ./bosinst.data file allows you to specify the requirements at the target system and how the user interacts with the target system. This file contains the customized BOS install procedures and dictates how the BOS install program will
    behave. You can customize this file before issuing the mksysb command or use a procedure to customize this file after the image backup is done.
    The ./image.data file contains information describing the image installed during the BOS installation process. This information includes the sizes, names, maps, and mount points of logical volumes and file systems in the rootvg. You can customize this file before using the mksysb command, or run mksysb -i to generate a new ./image.data file on the tape during a backup. The mkszfile command generates the ./image.data file. The ./image.data file is arranged in
    stanza format. Each stanza contains one or more fields.
    The dummy table of contents (TOC) is used so that the mksysb tape contains the same number of images as a BOS install tape.
    The rootvg data area contains all data in the rootvg volume group backed up by the mksysb command. The mksysb command uses the backup command to save the contents of mounted JFS data in rootvg, excluding raw data.

    When you need to make a mksysb backup of a system, and you want to exclude some data file systems from the system, you need to edit the /etc/exclude.rootvg file. Make sure that there are no empty lines.
    ^./tmp/

    Restoring mksysb
    An mksysb image enables you to restore the system image onto target systems that might not contain the same hardware devices or adapters, require the same kernel (uniprocessor or microprocessor), or be the same hardware platform (rs6k, rspc, or chrp) as the source system.
    You have several possibilities to restore the mksysb image:
    -If restoring on exactly the same machine, you can boot directly from the mksysb media and restore from there.
    -If restoring on a different type of machine, you use the cloning function
    -If you do not want to interfere with the production environment, you can use the alt_mksysb command.

    Multibos
    The multibos command allows the root level administrator to create multiple instances of AIX on the same rootvg. The multibos setup operation creates a standby Base Operating System (BOS) that boots from a distinct boot logical volume (BLV). This creates two bootable sets of BOS on a given rootvg. The administrator can boot from either instance of BOS by specifying the respective BLV as an argument to the bootlist command or using system firmware boot operations. Two bootable instances of BOS can be simultaneously maintained. The instance of BOS associated with the booted BLV is referred to as the active BOS. The instance of BOS associated with the BLV that has not been booted is referred to as the standby BOS. Currently, only two instances of BOS are supported per rootvg.

    The multibos command allows the administrator to access, install maintenance and technology levels for, update, and customize the standby BOS either during setup or in subsequent customization operations. Installing maintenance and technology updates to the standby BOS does not change system files on the active BOS. This allows for concurrent update of the standby BOS, while the active BOS remains in production.
    In this case, the biggest difference is that multibos will create and copy only the following Logical Volumes (LVs):
    /;
    /usr;
    /var;
    /opt, and;
    hd5 (Boot logical volume).
    All others LVs will be shared with the original root volume group, although more logical volumes can be specified and copied.
    Multibos supports a new TL update, but AIX version upgrades are not supported by multibos.

    Failed Disk Replacement

    If the disk you are going to replace is mirrored, we recommend following these steps:
    1. Remove copies of all logical volumes that were residing on that disk using either the rmlvcopy command or unmirrorvg command.
    2. Remove the disk from the volume group using the reducevg command.
    3. Remove the disk definition using the rmdev command.
    4. Physically remove the disk. If the disk is not hot-swappable, you may be required to reboot the system.
    5. Make the replacement disk available. If the disk is hot-swappable, you can run cfgmgr; otherwise, you may need to reboot the system.
    6. Include the newly added disk into the volume group using the extendvg command.
    7. Recreate and synchronize the copies for all logical volumes using either mklvcopy or mirrorvg.

    If the disk you are going to replace is not mirrored and is still functional, we recommend following these steps:
    1. Make the replacement disk available. If the disk is hot-swappable, you can run cfgmgr; otherwise, you may need to reboot the system.
    2. Include the newly added disk into the volume group using the extendvg command.
    3. Migrate all partitions from the failing disk to the new disk using either the migratepv command or the migratelp command. If the disks are part of the rootvg, you should consider the following:
    If the disk to be replaced contains a copy of the BLV, you have to clear it using the chpv -c command.
    A new BLV image must be created on the new disk using the bosboot command.
    The bootlist must be updated to reflect these changes using the bootlist command.
    If the disk to be replaced contains a paging space or a primary dump device, you should disable them. After the migratepv command completes, you should reactivate them.
    4. Remove the failing disk from the volume group using the reducevg command.
    5. Remove the disk definition using the rmdev command

    If the disk is not mirrored, has failed completely, and there are other disks available in the volume group, we recommend following these steps:
    1. Identify all logical volumes that have at least one partition located on the failed disk.
    2. Close the logical volumes and unmount all the corresponding file systems using the umount command.
    3. Remove the file systems and logical volumes using the rmfs command.
    4. Remove the failing disk from the volume group using the reducevg command.
    5. Remove the disk definition using the rmdev command.
    6. Physically remove the disk. If the disk is not hot-swappable, you may be required to reboot the system.
    7. Make the replacement disk available. If the disk is hot-swappable, you can run cfgmgr; otherwise, you may need to reboot the system.
    8. Include the newly added disk into the volume group using the extendvg command.
    9. Recreate all the logical volumes and the corresponding file systems using the mklv command and the crfs command.
    10.If you have a backup of your data, restore your data from backup.

    If the disk is not mirrored, has failed completely, there are no other disks available in the volume group (the volume group contained only one physical volume or all the physical volumes failed simultaneously), and the volume group is not rootvg, we recommend the following steps:
    1. Export the volume group definition from the system using the exportvg command.
    2. Ensure that /etc/filesystems does not contain any incorrect stanzas.
    3. Remove the disk definition using the rmdev command.
    4. Physically remove the disk. If the disk is not hot-swappable, you may be required to reboot the system.
    5. Make the replacement disk available. If the disk is hot-swappable, you can run cfgmgr; otherwise, you may need to reboot the system.
    6. If you have a volume group backup, restore it using the restvg command.
    7. If you do not have volume group backup, recreate the volume group, all the logical volumes, and the corresponding file systems using the mkvg command, the mklv command, and the crfs command.
    8. If you have a backup of your data, restore your data from backup.

    If the disk is not mirrored, has failed completely, there are not other disks available in the volume group (the volume group contained only one physical volume or all physical volumes failed simultaneously), and the volume group is
    rootvg, we recommend following these steps:
    1. Replace the failing disk.
    2. Boot the system in maintenance mode.
    3. Restore the system from an mksysb image.

    RAID Levels
    RAID 0 Also known as striping. Data is split into blocks of equal size and stored on different disks. There is no
    redundancy. If one disk fails, your data is lost.
    RAID 1 Also known as mirroring. Data is split into blocks of equal size. Duplicate copies are kept on separate
    physical disks. If one disk fails, the second copy is used and data is still available.
    RAID 5 Also known as striping with parity. Data is split into blocks of equal size. An additional data block
    containing parity information is generated. Both data blocks and the parity block are written onto separate
    physical disks. If one disk storing a block of data fails, the system will reconstruct the data from the failed
    drive using the parity information and the remaining data from the other drives.
    RAID 10 Also known as RAID 0+1 or enhanced RAID. It is a combination of mirroring (RAID 1) and striping (RAID
    0).

    SSH Securities:
    # vi /etc/ssh/sshd_config
    And make following changes for better security
    PermitRootLogin no Disabled the Direct root login
    Protocol 2 Good security concepts
    Banner file Set the banner in a file
    IgnoreRhosts yes If .rhosts is configured then ignored
    PermitEmptyPassword no Disabled the acceptance of null password
    HostBasedAuthentication no Disabled the host based authentication

    TCP Wrapper:
    TCP wrapper often called a s just a wrapper. It allows a system administrator to control access of TCP-based services or daemons that are wrapper aware. Download the “tcp_wrappers_7.6-ipv6.4.tar.gz” and install it.
    Replaced

    To



    and
    # refresh -s inetd
    Wrapper first check host.allow, if match found access is granted. If match found in host.deny the access is resticted for perticular services.
    # vi /etc/host.allow
    telnetd,sshd: .mydomain.com :allow
    telnetd,sshd:192.168.4.10 , 192.168.6. : allow

    # vi /etc/host.deny
    telnetd : 192.168.8., 192.168.9. : deny
    telnetd : 192.168.6. : allow
    # tcpdchk-v check the syntax in host.allow and host.deny


    Audits in AIX:
    AIX auditing provides a framework within which to capture pertinent system and security related information, such as failed login attempts, cron usage etc. It is recommended that auditing is enabled as part of a group of measures designed to provide enhanced logging of system and security changes.
    Create a /audit filesystem, at least 100 MB in size. The script always creates a standard jfs filesystem. If the /audit filesystem exists, this step is skipped:
    # mklv -y <LV name> -t jfs -u 1 -c 1 rootvg 1 hdisk0
    # crfs -v jfs -d auditlv -m /audit -A yes -t no mount /audit
    # vi /etc/security/audit/configt /audit
    start:
    binmode = on
    streammode = off

    bin:
    trail = /audit/trail
    bin1 = /audit/bin1
    bin2 = /audit/bin2
    binsize = 10240
    cmds = /etc/security/audit/bincmds

    Add the auditing entries for root and all other users below the pre-defined audit classes users:
    root = general,SRC,mail,cron,tcpip,ipsec,lvm
    <user 1> = general,SRC,cron,tcpip
    <user 2> = general,SRC,cron,tcpip
    etc.

    Update the /usr/lib/security/mkuser.default auditclasses entry to ensure that auditing is set up for any newly created users
    # chsec –f /usr/lib/security/mkuser.default –s user –a auditclasses=general,SRC,cron,tcpip

    Add the audit startup command into /etc/inittab.
    # mkitab "audit:2:boot:audit start > /dev/console 2>&1 # Start audit"


    Simultaneous Multi-Threading
    Simultaneous multithreading is the ability of a single physical processor to simultaneously dispatch instructions from more than one hardware thread context. Because there are two hardware threads per physical processor, additional instructions can run at the same time. The POWER5™ processor is a superscalar processor that is optimized to read and run instructions in parallel. Simultaneous multithreading allows you to take advantage of the superscalar nature of the POWER5™ processor by scheduling two applications at the same time on the same processor.
    Simultaneous multithreading is primarily beneficial in commercial environments where the speed of an individual transaction is not as important as the total number of transactions that are performed. Simultaneous multithreading is expected to increase the throughput of workloads with large or frequently changing working sets, such as database servers and Web servers.
    Multiple Hardware threads can run on one physical processor at the same time
    -A processor appears as two or four logical CPUs (lcpu) to AIX
    -Beneficial for most commercial environments
    -SMT is enabled by default with max no of logical CPUs
    Can change between SMT2 to SMT4 (POWER 7 only)
    Can Disabled and enabled



    AIX 6.1 Terminology

    What are WPARs?
    In contrast to LPARs, which are created and managed at the server's firmware level, AIX WPARs are software partitions that are created from, and share the resources of, a single instance of the AIX operating system. This means that you must have AIX 6.1 to create WPARs, but you can create WPARs on any System p hardware that supports AIX 6.1, including POWER4, POWER5, and POWER6 hardware. You don’t need an HMC or IVM to create or manage WPARs.
    There are two kinds of WPARs:
    System WPARs
    Application WPARs

    System WPARs
    System WPARs System WPARs are autonomous virtual system environments that have their own private file systems, users and groups, login, network space, and administrative domain. To users and applications, a system WPAR appears almost exactly like a full AIX system. Operating system services, such as telnet, are supported, so if network information has been configured, users can telnet into a system WPAR as root or any other defined user, issue commands, and run applications as they would on any other AIX system.
    -Autonomous virtual system environment
    Shared file system with global environment: /usr and /opt
    Private file system for wpar’s own use:/, /var, /tmp
    Unique set of users, groups, and network addresses
    -Can be access via Network protocol (telnet or ssh)
    Login from the global environment using the clogin command
    -Can be stopped and restarted

    Application WPARs
    Application WPARs provide an environment for isolation of applications and their resources to enable checkpoint, restart, and relocation at the application level. An application WPAR is essentially a wrapper around a running application or process for the purposes of isolation and mobility. It lacks some of the system services provided by system WPARs—for example, it’s not possible to log in or telnet into an application WPAR. When the application running in an application WPAR terminates, the WPAR also ceases to exist. Application WPARs are most useful when you want to enable Live Application Mobility—that is, when you want to be able to move a running application from one AIX system to another. You might want to relocate applications to avoid downtime resulting from scheduled maintenance or to improve performance by moving an application to a more powerful server
    -Isolate an individual application
    -Light weight; quick to create and remove
    Created with wparexec command
    Removed when stopped
    Stopped when the application finished
    -Shares file systems and devices with the global environment
    -No user login capabilities

    WPAR Manager
    IBM Workload Partitions Manager for AIX (WPAR Manager) is a platform management solution that provides a centralized point of control for managing workload partitions across a collection of managed systems running AIX 6.1. The managed systems might all be LPARs on a single physical server, or they might be located on multiple physical servers. Using WPAR Manager, you can monitor the health and status of multiple WPARs on multiple managed AIX systems. You can also perform all the basic WPAR life cycle operations—including create, view and manage properties, start, stop, and delete.
    WPAR Manager also supports relocation of WPARs between systems in a collection of managed servers. WPAR Managersupports two kinds of relocation:
    Manual relocation—This type of relocation is initiated by the user.
    Policy-based relocation—This type of relocation is initiated by WPAR Manager in response to workload conditions defined in a relocation policy.
    WPAR Manager is not part of AIX—it’s a separately purchased licensed program

    Multi-system management
    Managing WPARs on multiple AIX systems using the WPAR Manager requires two initial installation and configuration steps. First, you install and configure the management server software on an AIX system in your environment. Then, you install the WPAR Manager agent software on each AIX system that will be managed by WPAR Manager. The agent must then be configured to share WPAR data with a specific management server. After the WPAR Manager and agent components have been configured and started, the WPAR Manager automatically discovers all the managed systems, and begins to record data transmitted by the agents in an internal database.

    Live Application Mobility
    Live Application Mobility is the capability to relocate a WPAR from one hosting system to another without having to restart any applications or processes running in the WPAR. Partition mobility refers to the ability to move an entire running AIX LPAR from one physical server to another.

    Live Partition Mobility
    Live Partition Mobility allows you to migrate partitions that are running AIX and Linux operating systems and their hosted applications from one physical server to another without disrupting the infrastructure services. The migration operation, which takes just a few seconds, maintains complete system transactional integrity. The migration transfers the entire system environment, including processor state, memory, attached virtual devices, and connected users.
    Live Partition Mobility helps you meet increasingly stringent service-level agreements (SLAs) because it allows you to proactively move running partitions and applications from one server to another.
    Live Partition Mobility may also be used as a mechanism for server consolidation, because it provides an easy path to move applications from individual, stand-alone servers to consolidation servers. If you have partitions with workloads that have widely fluctuating resource requirements over time.
    Live Partition Mobility can be automated and incorporated into system management tools and scripts. Support for multiple concurrent migrations allows you to liberate system resources very quickly. For single-partition, point-in-time
    migrations, the Hardware Management Console (HMC) and the Integrated Virtualization Manager (IVM) interfaces offer easy-to-use migration wizards
    Partition Mobility is a hardware-based function, partition mobility is only supported on POWER6 hardware; application mobility is supported on any hardware that supports AIX 6.

    Partition Migration
    A partition migration operation can occur either when a partition is powered off (inactive), or when a partition is providing service (active).
    The migration process can be performed either with a powered-off or a live partition. The following two available migration types are discussed in more detail in the next sections:
    Inactive migration The logical partition is powered off and moved to the destination system.
    Active migration The migration of the partition is performed while service is provided, without disrupting user activities.

    Policy-based relocation
    WPAR Manager also has the capability to monitor performance of workloads running in WPARs and relocate those workloads to different AIX systems to improve performance. For example, if CPU or memory usage in a WPAR or group of WPARs is, on average, higher than a value that you specify, then the WPAR Manager might attempt to relocate one or more WPARs to a more powerful, or less busy, server in your datacenter. The detailed mechanics of how the policy engine works are beyond the scope of this article, but will be covered in a later article.

    Encrypted File System
    The Encrypted File System (EFS) is a J2 filesystem-level encryption through individual key stores. This allows for file encryption in order to protect confidential data from attackers with physical access to the computer. User authentication and access control lists can protect files from unauthorized access while the operating system is running; however, it’s easy to circumvent the control lists if an attacker gains physical access to the computer.
    One solution is to store the encrypted files on the disks of the computer.
    In EFS, a key is associated to each user. These keys are stored in a cryptographically protected key store and upon successful login, and the user's keys are loaded into the kernel and associated with the process credentials. When the process needs to open an EFS-protected file, the system tests the credentials. If the system finds a key matching the file protection, the process is able to decrypt the file key and file content. The cryptographic information is kept in the extended attributes for each file. EFS uses extended attribute Version 2, and each file is encrypted before being written on the disk. The files are decrypted when they are read from the disk into memory so that the file data kept in memory is in clear format. The data is decrypted only once, which is a major advantage. When another user requires access to the file, their security credentials are verified before being granted access to the data even though the file data is already in memory and in clear format. If the user is not entitled to access the file, the access is refused. File encryption does not eliminate the role of traditional access permissions, but it does add more granularity and flexibility.

    Snapshot
    The snapshot functionality was provided so that administrators could create consistent, integrated snapshots of an online (or offline) JFS2 file system. A point-in-time copy of the source file system is created. The snapshot operation executes quickly and requires very little disk space. Multiple snapshots can be taken, which can be very handy if you need to perform multiple changes to a system with corresponding backups at each stage.
    The snapshot retains the same permissions as the original file system. Once the image has been created, you can use it to restore files or an entire file system to a known point in time. Or, you can mount the snapped file system and perform a backup of it to tape (via Tivoli Storage Manager for example). This could help reduce your backup time requirements.
    There are two types of JFS2 snapshots.
    The first is an external snapshot which uses a separate logical volume for the snapshot image.
    The second type allocates space out of the original file system, this is known as an internal snapshot.

    An external snapshot can be mounted separately on its own mount point. A file system can use only one type of snapshot method at a time; that is, you cannot take an external snapshot of a file system and then take an internal snapshot of the same file system while the external exists (and vice versa).
    It is also important to note that you cannot use internal snapshots unless the file system was enabled to support them at the time of file system creation. So if you want to use snapshots on an existing file system and you did not enable it for internal snapshots, you will need either to recreate the file system with the snapshot option or use an external snapshot.
    One of the big advantages to using JFS2 snapshots is that we do not need to create a complete copy of all of the data. Nor do we need to recover the entire contents of the data. Thus reducing overhead and saving time.



    Network Installation Management

    Benefits of NIM
    NIM allows system administrators to install, upgrade, back up and maintain AIX 5L systems remotely
    Multiple machines can be installed or maintained at the same time
    Multiple AIX 5L versions, Maintenance Levels (MLs) or Technology Levels (TLs) can be used in the same NIM environment
    Disaster recovery - local systems can be recovered quickly with NIM
    Systems can be installed with standard and consistent AIX 5L operating system images
    Recovering a system backup (mksysb) from tape can be difficult when booting from CD media
    Cloning systems is another excellent feature of NIM which can assist you in moving from one server to another, or even from one site to another site

    NIM Overview
    A basic NIM environment consists of several IBM System p machines (at least two) connected via a TCP/IP network. A physical network can be shared by several NIM environments, therefore the machines attached to the same
    physical network can be part of several NIM environments
    The machines (or clients) can be standalone systems, LPARs, diskless, and dataless clients (also known as “thin clients”). The machines can be controlled by one (or more) Hardware Service Console

    Master
    Master refers the machine where you set up and maintain your NIM environment. You can also initiate installations from here (push mode). It is the key piece of the NIM environment. Refer to 2.1.4, “NIM master selection” on page 26, for information about how to chose a machine to become the NIM master.
    Client
    The NIM client can be the target for NIM master-initiated operations such as installation, updates, and so forth (push mode). Also, a client can initiate its own installation or update (pull mode). A NIM client automatically becomes a
    resource server when it holds NIM resources.

    Resource server
    Any machine (the master or a standalone client) can be configured by the master as a server for a particular software resource. In most environments, the master is also resource server.
    If other machines are already reporting to the master and they are installed (AIX), you can chose one of them to act as a resource server, thus relieving the NIM master of the heavy I/O load (disk and network). In such cases, the NIM master is only used to run administrative tasks.

    Push and pull modes
    The push mode operation is initiated from the master. The pull mode operation is initiated from the client. The very first time a client is installed, only the pull mode can be used.

    NIM database
    The NIM database is stored in the AIX Object Data Management (ODM) repository on the NIM master and is divided into four classes: machines, networks, resources, groups.

    Networks class
    The network is what allows the machines in a NIM environment to communicate which each other. If the network is a simple local area network (LAN), the definition of the network object is simplified. The purpose of the network object is to depict the network topology used by the NIM environment. Each modification made on the physical network must be reflected to the NIM database.

    Resources class
    A NIM resource is a pointer to a file or a directory located on a Resource Server, and the NIM Master is the default Resource Server. For example, the bosinst_data resource points to a file that contains information used to perform
    the installation process without requiring any manual intervention. The basic characteristics of a resource are the NIM name, the location (the location of the file or directory), and the resource server hosting the resource

    Groups class
    A group is a collection of either machines or resources. Only two types are available:
    # res_group, for grouping a set of resources
    # mac_group, for grouping a set of machines

    SPOT
    SPOT, or Shared Product Object Tree, is a directory of code (installed filesets) that is used during client booting procedure. It is equivalent in content to the code that resides in the /usr file system on a system running AIX 5L
    (binary objects - executables and libraries, header files and shell scripts). This resource (directory) replaces the content of the basic ramdisk image available on installation CDs.
    The SPOT also contains the code needed to generate the boot images (kernels, which will be stored in the /tftpboot directory) that the client uses until it can manage to NFS-mount the SPOT directory.

    lpp_source
    An lpp_source is a directory similar to AIX install CDs. It contains AIX Licensed Program Products (LLPs) in Backup File Format (BFF) format and RPM Package Manager (RPM) filesets that you can install

    mksysb
    This resource is a file containing the image of the root volume group (generated with the AIX mksysb command) of a machine. It is used to restore a machine, or to install it from scratch (also known as “cloning” a client). A mksysb resource can be defined by first creating the mksysb image with SMIT or the mksysb command, and then defining the resource that references the backup image file.

    bosinst_data
    The bosinst_data resource is a flat ASCII file similar to the bosinst.data file used for restoring system backup images from tape or CD/DVD. When you start the installation of a IBM System p machine, you will be prompted to
    select the console, language, installation method, the target disk(s) for rootvg, and so on. This resource comes in handy for both push or pull installation of multiple machines at the same time, because it will automate the installation
    process without requiring interactive answers for the installation parameters.

    script
    The script resource is a file. After the BOS installation is finished on your client, you can perform customization such as file system resizing, additional user creation, and so forth. The script resource points to a file (usually a shell
    script) that contains all the commands needed to perform customization.

    To rebuild and recover a NIM client /etc/niminfo file, use the niminit command on the client:
    niminit -a master=<MASTER_HOSTNAME> -a name=<CLIENT_NIM_NAME>

    NIM improovement in AIX 5.3
    nimsh
    Alternate NIM master
    Creating an SPOT resource from a mksysb
    mksysb migration using NIM

    Removing a NIM client definition
    /tftpboot This is used for storing NIM master boot images (kernels) and info files for the NIM clients.
    /export/lpp_source This is used for storing versions of AIX base level filesets in specific directories; a NIM lpp_source resource represents a directory in which software installation images are stored
    /export/spot This is used for storing non-/usr Shared Product Object Three (SPOT) in individual directories (one per SPOT version). A SPOT is required to install or initialize all types of machine configurations, as well as to generate a
    specific kernel image used for network booting a client system. Everything that a client machine requires in a /usr
    file system, such as the AIX kernel, executable commands, libraries, and applications, should be included in the
    SPOT.
    /export/images This is used for storing system backup images. These images can be created by using the NIM mksysb operation (“pulling” a mksysb image from a NIM client), or by copying a system backup as a file from a previously created backup file, tape, or DVD. To create subdirectories during NIM mksysb creation, you must
    export the NIM_MKSYSB_SUBDIRS=yes variables before using the command
    # nim -o define -t mksysb

    Basic Naming Conventions
    lpp_source lpp<OS#><TL#><SP#>
    spot spot<OS#><TL#><SP#>
    mksysb mksysb<OS#><TL#><SP#>
    network net_<IP#NET>

    Initializing NIM master
    nimconfig command can be used to perform a nim initializing
    Configuring the NIM master using the nimconfig command creates the master, boot, the first network object, nim_script with the /export/nim directory; the master (nimesis) daemon is also started.
    root@master:/: lsnim
    master machines master
    boot resources boot
    nim_script resources nim_script
    net_10_1_1 networks ent

    To convert an old NIM master into a client:
    Create a client definition on the NIM master.
    Remove the /etc/niminfo file (this will remove also the nim command from the system).
    Uninstall the bos.sysmgt.nim.master fileset.
    Check whether the bos.sysmgt.nim.client fileset is installed.
    Use the niminit command to register with the new NIM master (client registration must be enabled for this procedure to work).

    Verify auxiliary services (inetd, bootpd and tftpd)
    root@msater:/: lssrc -ls inetd
    tftp /usr/sbin/tftpd tftpd -n active
    bootps /usr/sbin/bootpd bootpd /etc/bootptab active

    The NIM master uses the bootpd (BOOT Protocol) and tftpd (Trivial File Transfer Protocol) daemons (services). Both services are normally started by the inetd daemon (InterNET master daemon).

    SMS and Console flow during NIM client installation
    At the Initial Program Load (IPL) menu, press the 1 key to access SMS Menu.
    select 2. Setup Remote IPL (Initial Program Load

    Select the desired adapter from the NIC Adapters list,


    select 1. IP Parameters to access the IP Parameters menu


    1. From the IP Parameters menu select 1. Client IP Address, to set the Client IP Address.
    2. From the IP Parameters menu select 2. Server IP Address, to set the Server IP Address (NIM master).
    3. From the IP Parameters menu select 3. Gateway IP Address. Leave it if the NIM master and client are on the same IP subnet.
    4. From the IP Parameters menu select 4. Subnet Mask, to set the Subnet Mask for the IP network.
    5. Press ESC and return to the previous menu


    Network Parameters menu, select 3. Ping Test


    From the Network Parameters menu, select 1. Execute Ping Test



    Next, press M from the Ping Test menu to return to the Main Menu

    From the Main Menu, choose 5. Select Boot Options


    From the Multiboot menu, choose 1. Select Install/Boot Device


    From the Select Device Type menu, select 6. Network to choose a network


    From the NIC Adapters list, in this case, select 1. to access the Select Task menu


    From the Select Task menu, select 2. Normal Mode Boot


    From the exit SMS menu select 1. Yes to let the network installation begin


    First the network boot (IPL) is initiated for the LPAR.
    Next, the LPAR issues BOOTP requests to the server IP-address specified in the IP Parameters menu. It receives the BOOTP record information from /etc/bootptab on the NIM master. This record specifies the TFTP server for sending a bootable kernel. The LPAR issues TFTP requests to the server listed in the /etc/bootptab sa field from the NIM master.


    Then control is passed to the transferred and loaded AIX 5L kernel.
    In this case the installation is made using a NIM mksysb resource.
    After installing the software, the boot image is created.
    Once the installation is finished and the LPAR is restarted and The newly installed kernel is loaded and the LPAR is restarted



    LPAR and VIO Server Terminology

    Logical partitioning
    Logical partitioning (LPAR) is the ability to logically slice up a single system's CPU, memory, and other resources to create multiple and separate servers. Each LPAR has its own allocation of CPU, memory, and I/O devices. This type of partitioning is done at the firmware level, not at the physical resource level. Thus, with the IBM System p5 servers, you are able to create LPARs using Micro-Partitioning™ -- assigning less than one physical CPU per LPAR.

    Virtual I/O Server
    It is a special kind of partition called a Virtual I/O Server (VIO Server). The VIO Server provides the ability to share I/O resources among several LPARs. You define virtual Ethernet and disk devices on the VIO Server and then make them available to the other LPARs on the system. Without the ability to share I/O devices on the managed system, each LPAR would require its own dedicated devices. If you have two LPARs on a system, you would need at least two Ethernet cards and two disk controllers. However, if you have 15 LPARs, then you would need at least 15 Ethernet cards and 15 disk controllers. With the use of VIO LPARs, you could host 15 LPARs with a much smaller number of Ethernet and disk controllers.
    Of course, you are still able to dedicate I/O devices to an AIX partition if you need to for performance reasons. For example, if you have a large database LPAR that requires a large amount of throughput, you can dedicate a disk controller to that LPAR so that it won't have to compete for I/O resources with other LPARs using the VIO's disk controller.
    On the other hand, you might have several small LPARs that do not require a large amount of data throughput. They would be ideal candidates for using the VIO Server since each of them would not completely utilize the full capacity of an Ethernet card or disk controller (see Figure 3). Together they could take full advantage of the hardware's capacity

    Dynamic logical partitioning
    Not only are you able to logically carve up your hardware into multiple LPARs, but IBM provides the capability of dynamically adding, removing, or moving resources between the partitions while they are up and running. You can add or remove CPU, memory, or I/O slots to a running partition without first shutting that LPAR down.
    Dynamic logical partitioning (DLPAR) provides a great deal of flexibility to your computing environment. As conditions change in your environment, you are able to react by moving hardware resources to where they are needed without interrupting services to the LPARs. The options for using this feature are endless. A word of caution, though, not all applications are able to handle the removal of CPUs or memory gracefully, so testing is needed in your particular situation before you begin to use this feature in a production environment.
    In the rest of this article, I'll discuss the steps you need to follow to configure an installed p5 system. It is impossible, in an article of this length, to discuss all the options available to you in configuring an LPAR system. Throughout the article, I'll cover the configuration basics and then point you to documents that provide more detailed information for that particular component. I'm also assuming the hardware is installed and connected to your network.

    Micro-Partitioning
    Micro-Partitioning is the ability to distribute the processing capacity of one or more physical processors among one or more logical partitions. Thus, processors are shared among logical partitions.

    System Profile
    A system profile is a collection of one or more partition profiles. System profiles can be used to specify which partition profiles are activated at the same time.

    Partition Profile
    A partition profile represents a particular configuration for a logical partition. A partition profile contains information about the resources assigned to the partition. These resources include memory, processor processing capacity, and physical I/O slots. Partition profiles also include information about which devices provide important partition functions

    Memory in LPAR
    There are three memory values for a partition profile:
    # Minimum memory: This amount of memory is required for the partition. The profile will fail to activate if the minimum memory is not met.
    # Desired memory: This is the requested amount of memory for the partition. On profile activation, the partition will receive an amount of memory between the minimum and desired amounts depending on what is available.
    # Maximum memory: The maximum represents the upper limit for memory. Memory allocations cannot go above the maximum. In order to go above the maximum, the maximum value needs to be changed.

    Dedicated Processor
    Dedicated processor means you are working with whole processors that are dedicated to a single partition. The dedicated processor handles the processing for one specific logical partition. If you choose to assign dedicated processors to a logical partition, you must assign at least one processor to that partition.

    Shared Processor
    The shared processing pool allows you to assign partial processors to a logical partition. The physical processors are held in the shared processing pool and are shared among logical partitions. At partition creation time, the shared processing units can be defined as 0.1.
    There are two types
    Capped and
    Uncapped
    Capped prohibits the LPAR from using more processing units than is currently allocated to it. If an LPAR is allocated 1.5 processing units but the workload requires 2.2 units, the LPAR would not be allowed access to the additional units, even if they were available in the shared processor pool.
    Uncapped allows the LPAR to use more processing units than is currently allocated. In this example, the LPAR would have access to the additional 0.7 processing units needed for the workload. If Uncapped is selected, you need to provide the Weight value for this LPAR. LPARs with higher Weight values receive higher priority when shared pool resources are fully used.

    Virtual Processor
    Virtual processors tell the LPAR's OS how many physical processors it has. The OS cannot use parts of a processor, so you have to tell it in whole numbers how many processors it can use. Also, the number of virtual processors limits how many physical processors an LPAR can use. For example, if you give an LPAR two virtual processors, it cannot use more than two physical processors even though Maximum processing units is higher than two.
    Minimum number of virtual processors: You must have at least one virtual processor for every part of a physical processor assigned to the LPAR. For example, if you have assigned 2.5 processing units to the LPAR, the minimum number of virtual processors is three.
    Desired number of virtual processors: You can assign a virtual processor to every 0.1 processing unit allocated to the LPAR. There is no recommended value for this configuration setting. You need to use trial and error to determine what is the best setting for your application(s).
    Maximum number of virtual processors: You can only have 10 virtual processors per processing unit. Therefore, you cannot assign a value greater than 10 times the maximum processing units value.

    Virtual I/O Adapter
    Virtual IO adapters give a partition the flexibility to use certain types of resources with requiring the physical hardware to be present. If the partition will be using virtual IO, make sure that the Yes radio button is selected. Otherwise, select the No radio button.
    The following virtual IO adapters can be created:
    Virtual Ethernet
    Virtual Serial
    Virtual SCSI

    Virtual Ethernet
    Virtual Ethernet technology is supported on AIX® 5.3 on POWER5™ hardware. This technology enables IP-based communication between logical partitions on the same system using a VLAN capable software switch in POWER5 systems. Shared Ethernet Adapter technology (part of the optional Virtual I/O Server feature on POWER5 hardware) enables the logical partitions to communicate with other systems without assigning physical Ethernet slots to the logical partitions.
    The number of virtual Ethernet adapters for each LPAR varies by operating system. AIX 5.3 limits the number of virtual Ethernet adapters to 256 virtual Ethernet for each LPAR. Besides a PVID, the number of additional VID values that can be assigned for each virtual Ethernet adapter is 20, which indicates that each virtual Ethernet adapter can be used to access 21 networks. The HMC generates a locally administered Ethernet MAC address for the virtual Ethernet adapters so that these addresses do not conflict with physical Ethernet adapter MAC addresses. To ensure uniqueness among the virtual Ethernet adapters, the address generation is based on the system serial number, LPAR ID, and adapter ID.
    When using the Integrated Virtualization Manager, only PVID is allowed (no additional VLANs), and only the PVID may only be 1-4.


    Virtual Serial
    Virtual serial adapters provide a point-to-point connection from one logical partition to another, or from the Hardware Management Console (HMC) to each logical partition on the managed system. Virtual serial adapters are used primarily to establish terminal or console connections to logical partitions. When you create a logical partition, the HMC automatically creates two virtual server serial adapters on the logical partition. These virtual server serial adapters allow you to establish a terminal or console connection to the logical partition through the HMC.
    You can also create pairs of virtual serial adapters on logical partitions so that you can access and control one logical partition directly from another logical partition. For example, one logical partition uses the disk resources of another logical partition using virtual SCSI adapters. You can create a server serial adapter on the logical partition that uses the disk resources and a client serial adapter on the logical partition that owns the disk resources. This connection allows the logical partition that owns the disk resources to shut down the logical partition that uses the disk resources before you back up data on the logical partition that owns the disk resources.

    Virtual SCSI
    Virtual SCSI (Small Computer Systems Interface) adapters provide one logical partition with the ability to use storage I/O (disk, CD, and tape) that is owned by another logical partition.A virtual SCSI client adapter in one logical partition can communicate with a virtual SCSI server adapter in another logical partition. The virtual SCSI client adapter allows a logical partition to access a storage device being made available by the other logical partition. The logical partition owning the hardware is the server logical partition, and the logical partition that uses the virtualized hardware is the client logical partition. With this arrangement, the system can have many server logical partitions.

    NPIV (Network Port ID Virtualization)
    N_Port ID Virtualization is a fibre channel industry standerd method for virtualizing a physical fibre channel port. NPIV allows one F_Port to be associated with multiple N_Port Ids, so a physical fibre channel HBA can be shared accross multiple guest operating systems. On POWER, NPIV allows logical partitions (LPARs) to have dedicated N_Port IDs, giving the OS a unique identity to the SAN, just as if it had a dedicated physical HBA(s).

Popular Posts

Is this site helping you ?