Home About BC DR HA Support Training Download
You are here: Home/ GlobalSysAdmin/ AIX/ Please Login or Register

-
Current Location
-

js
  GlobalSysAdmin
    AIX



-
 Mt Xia Inc.
-
BOOK:Advanced Shell Scripting



 Join our LinkedIn Group
AIX Advanced Technical Experts
Contract Opportunities

www.LinkedIn.com
-
 Train By Tweet
Training courses Tweet By Tweet
Various Subjects and Products
TrainByTweet.com


AIX Admin Methodology
Global Consolidation Project
All AIX admins should join
www.aixexpert.com
-

digg Digg this page
del.icio.us Post to del.icio.us
Slashdot Slashdot it!


AIX FREE DEMO UNIX
Automated AIX Deployment
Reduce your DC costs

www.siteox.com

LPAR Leasing
Lease an AIX / i5 LPAR
Reduce your costs

www.mtxia.com

Server Leasing
Lease a Server off-site
Reduce your costs

www.mtxia.com

Data Center Automation
Business Continuity and DR
Virtualization/Consolidation

www.mtxia.com

HMC Service
Hardware Management Console
Manage Remote AIX / i5 LPARs

www.siteox.com

Business Web Site Hosting
$3.99 / month includes Tools,
Shopping Cart, Site Builder

www.siteox.com

Disaster Recovery
Small Business Oriented
Off-Site Facilities

www.mtxia.com

IBM pSeries / iSeries
Reduce your Costs
Off-Site Server Hosting

www.mtxia.com

This section is for Base Build Documents. Here you will find simple instructions to install a server. This is also where you will find the basic, common requirements for a given server and platform to build a server from scratch. Another notable item included within this section is "common customization" of servers. This is a list of specific items that are common to all servers that are not a base build but common to the Mt Xia AIX build for all environments and applications. Detail How-To documents to explain these processes are located HERE

To add anything to this section, contact Dana French.





Common Build Checklist

1.) Configure Server in HMC if P4/5-LPAR ( Hardware Management Console )

2.) Install Server via NIM ( either mxfnim or mxdnim - depending on the Datacenter the target Server resides in )

3.) NIM install via mksysb for existing configurations ( e.g. adding an eGate Server will use the eGate mksysb )

4.) NIM install via base media for new configurations ( Create mksysb for this configurations for future installs )

5.) Configure SNMP ( integrate "xmpeek -l hostname" into all systems startup procedure )

6.) Configure TSM ( Standard and HACMP configurations )

7.) Configure TEC ( Standard and HACMP configurations )

8.) Configure ProWeb ( ProWeb )

9.) Configure NMON ( Performance reporting )

10.) Add Server to CMDB ( Configuration Management Database )

11.) Set ODM Error Notify ( Notification to TEC when hardware errors are detected )

12.) Create XMON User ( user id for performance based administrative logins )

13.) Tune Memory on System ( vmtune install )

14.) Install and run invscout ( inventory management for patch and microcode revisions )

15.) Tune TCP/IP ( disable IPv6, tune specific TCP Stack components )

16.) Install DocitWeb ( Hardware and OS inventory of systems )

17.) Tune NFS ( only if server uses nfs )

18.) Install Performance Toolbox ( Client and Server binaries - configure to be seen by PerfToolbox Server )

19.) Install INQ utility ( information tool for drives )

20.) Install KSH93.att ( Install latest KSH for compatibility with existing scripts and programs )

21.) Install SUDO ( Administrative/Security Tool )

22.) Install SSH ( Administrative/Security Tool )

23.) Validate System Components ( Rootvg filesystem sizes, paging space size and placement, etc ... )

24.) Validate PCI Card Slot Placement ( NIC, SCSI, SSA, GigE, etc ... )

25.) Set SSH Keys for xmon user to mtxdoc ( for administrative activities and existing programs to work )

26.) Configure Rshell passwordless login to NIM Servers ( For NIM functionality and DR functionality )

27.) Set System Dump Space to hd7x and hd7y ( System Dump Space )

28.) Set Paging Space up on seperate spindle ( hd6 on seperate disk if large enough and seperate VG for additional paging space )

29.) Update Frame/System Spreadsheet ( Publish updated spreadsheet to mtxdoc )

30.) Validate LUN List ( SAN disks are presented correctly, HDLM Setup and rootvg and pagingvg in exclusion list for HDLM )

31.) Configure NTP ( Configure Time daemon )

32.) Configure Prompt ( Configure PS1 in /etc/profile )

33.) Configure MOTD ( Configure MOTD in /etc/motd )

34.) Configure Herald ( Configure Herald in /etc/security/login.cfg )

35.) Configure Base User Security ( Configure defaults i.e. password limits )

36.) Configure Server in PLM if P5-LPAR ( Partition Load Manager )

37.) Remove Rshell passwordless login to NIM Servers ( Upon Build completion and pior to turnover )

38.) Configure Server for Secure Access Portal ( Upon Build Completion Portal not complete yet )

39.) Generate New Secure Root Password ( Generate Password and deliver to Unix Team Lead )

40.) Create Change Ticket for LOS Turnover ( LOS Manager and Team Lead as approvers )



New system build task list assembled for virtualization follows:



System Build Tasks - Detailed Function List

  • Business Continuity
    • Disaster Recovery
      • High Availability
        • Virtualization

Facilities Preparation

  • Architecture and Design.
  • Validate technical design and implementation of supporting environment.
  • Define system concept , design , and infrastructure interfaces.
  • Blueprint of supporting infrastructure systems , and services in support of the project.
  • Engineering analyses and liaison with other team members supporting implementation.
  • Manage high severity issues arising from infrastructure issues during deployment.
  • Ensure integration into existing infrastructure global support architecture.
  • Define mangement network requirments.
  • Define threat modeling.
  • Provide Security design guidelines.
  • Provide Security architecture and design reviews.
  • Conduct Security testing.
  • Conduct Security deployment reviews.
  • Conduct vulnerability assessment , and auditing of servers.
  • Define power requirments of hardware.
  • Define cooling requirments of hardware.
  • Define Network requirments on enviroment.
  • Certify environment meets established infrastructure standards.
  • Receiving and Inventory.
  • Receive equipment and validate components.
  • Infrastructure Installation.
  • Work with network cabling contractors.
  • Work with power cabling contractors.
  • Work with cooling contractors.
  • Work with electrical contractors.
  • Work with server engineers on configuring supporting infrastructure.
  • Work with storage engineers on configuring supporting infrastructure.
  • Work with network engineers on configuring supporting infrastructure.
  • Work with inspectors certifying the additions .
  • Run new power to Racks and power up the racks.
  • Run network to each rack / frame.
  • Run network out to patch panels.
  • Run network from patch panels to switches.
  • Run SAN fabric as required.
  • Run SAN fiber as required.
  • Configure and power up the racks.
  • Install smaller customer install systems into appropriate racks.
  • Work with IBM CE to physically certify frames.
  • Allocate floor location to facilities and document in CMDB.
  • Infrastructure Configuration.

Node Names / IP Addresses / DNS Host Names

  • 15m: Identify node name for HMC
  • 15m: Identify node name for each VIO Server LPAR
  • 15m: Identify node name for each VIO Client LPAR
  • 15m: Identify TCP/IP subnet for all boot addresses associated with each frame.
  • 15m: Identify TCP/IP subnet(s) for one or more persistent addresses associated with each frame.
  • 15m: Identify TCP/IP subnet(s) for one or more service addresses associated with each frame.
  • 15m: Identify TCP/IP subnet(s) for one or more management addresses associated with each frame.
  • 15m: Identify TCP/IP subnet(s) required for HMC connections to the identified subnets.
  • 15m: Identify IP addresses for each HMC network adapter.
  • 15m: Identify a NIM server network name to use for configuring NIM resources associated with each VIO Client LPAR.
  • 15m: Identify NIM server IP address to use for remote operating system download of primary VIO server on each frame.
  • 15m: Identify NIM server IP address to use for remote operating system download of secondary VIO server on each frame.
  • 15m: Identify network gateway address to use for remote operating system download of primary VIO servers on each frame.
  • 15m: Identify network gateway address to use for remote operating system download of secondary VIO servers on each frame.
  • 15m: Identify NIM server IP address to use for remote operating system download of each VIO Client LPAR.
  • 15m: Identify network gateway address to use for remote operating system download of each VIO Client LPAR.
  • 15m: Identify service IP address for primary VIO servers on each frame.
  • 15m: Identify service IP address for secondary VIO servers on each frame.
  • 15m: Identify service DNS name for primary VIO servers on each frame.
  • 15m: Identify service DNS name for secondary VIO servers on each frame.
  • 15m: Identify management DNS name for primary VIO servers on each frame.
  • 15m: Identify management DNS name for secondary VIO servers on each frame.
  • 15m: Identify VIO Client LPAR node names and resource group names.
  • 15m: Identify boot IP addresses for each network adapter for each VIO Client LPAR.
  • 15m: Identify persistent IP address for each VIO Client LPAR.
  • 15m: Identify management name IP address for each VIO Client LPAR.
  • 15m: Identify node name service IP address for each VIO Client LPAR.
  • 15m: Identify resource group service IP address for each VIO Client LPAR.
  • 15m: Identify resource group alias name for each VIO Client LPAR.
  • 15m: Identify VIO servers that provide I/O and storage resources for each VIO Client LPAR on each frame.
  • 15m: CHANGE REQUEST: Request "boot" IP addresses
  • 15m: CHANGE REQUEST: Request "service" IP addresses
  • 15m: CHANGE REQUEST: Request "persistent" IP addresses
  • 15m: CHANGE REQUEST: Request "management" IP addresses
  • 15m: CHANGE REQUEST: Request DNS entry for "boot" host name
  • 15m: CHANGE REQUEST: Request DNS entry for "service" host name
  • 15m: CHANGE REQUEST: Request DNS entry for "persistent" host name
  • 15m: CHANGE REQUEST: Request DNS entry for "management" host name
  • 15m: CHANGE REQUEST: Request DNS entry for "resource group alias" host name

9h45m

HMC Preparation

  • 16h: Install and configure the HMC (NOTE: cannot power-on any frame until completed).
  • 6h: Upgrade HMC software to latest version.
  • 4h: Update HMC software for all required fixes.
  • 15m: Turn on remote ssh access.
  • 15m: Turn on remote web access.

26h30m

NIM Server Preparation

  • 1h: Create NIM Network object to accomodate subnets identified for operating system downloads.
  • 4h: Prepare/upload VIO Server installation NIM Resources.
  • 2h: Prepare/upload VIO Server update NIM Resources.
  • 8h: Prepare/upload Client LPAR installation NIM Resources.
  • 15m: Update "/etc/hosts" to contain all IP addresses and DNS hostnames identified for VIO Servers.
  • 15m: Update "/etc/hosts" to contain all IP addresses and DNS hostnames identified for VIO Client LPARs.
  • 30m: Create VIO Server machine objects.
  • 30m: Allocate VIO Server installation objects.

16h30m

Frame Preparation

  • 1h: Turn on frame and discover all resources through HMC.
  • 8h: Update frame microcode.
  • 4h: Update all adapter firmware.
  • 2h: Automate microcode/firmware update mechanism.
  • 2h: Document all microcode and firmware.
  • 4h: Verify connectivity of all network adapters.
  • 4h: Verify connectivity of all storage adapters.

25h0m

VIO Server LPAR Preparation

  • 1h: From HMC, create create primary VIO server for each frame.
  • 1h: From HMC, create secondary VIO server for each frame.
  • 15m: Define CPU sharing for each VIO Server.
  • 15m: Define CPU allocations for each VIO Server.
  • 15m: Define processing unit allocations for each VIO Server.
  • 15m: Define memory allocations for each VIO server.
  • 2h: Allocate application network adapters for each VIO server.
  • 2h: Allocate management network adapters for each VIO server.
  • 2h: Allocate storage adapters for each VIO server.
  • 2h: Allocate virtual ethernet adapters for application network on each VIO server. Even numbered slots should only exist on primary VIO servers , odd numbered slots should only exist on secondary VIO servers.
  • 2h: Allocate virtual ethernet adapters to host the VIO server application network IP address on each VIO server. Even numbered slots should only exist on primary VIO servers , odd numbered slots should only exist on secondary VIO servers.
  • 2h: Allocate virtual ethernet adapters for management network on each VIO server. Even numbered slots should only exist on primary VIO servers , odd numbered slots should only exist on secondary VIO servers.
  • 2h: Allocate virtual ethernet adapters to host the VIO server management network IP address on each VIO server. Even numbered slots should only exist on primary VIO servers , odd numbered slots should only exist on secondary VIO servers.
  • 6h: Allocate virtual SCSI adapters for all slots between 10 - 499. Even numbered slot should only exist on primary VIO servers , odd numbered slots should only exist on secondary VIO servers.
  • 1h: Request a boot disk from the SAN for each VIO Server.
  • 30m: CHANGE REQUEST: Allocate one LUN from SAN for each VIO Server for operating system storage

24h30m

VIO Client LPAR Preparation

  • 15m: Identify business function of each LPAR.
  • 15m: Identify support TIER of each LPAR.
  • 15m: Identify operational environment of each LPAR.
  • 15m: Identify application codes for each LPAR.
  • 15m: Identify customer code for each LPAR.
  • 2h: Identify desired CPU quantity for each LPAR.
  • 2h: Identify desired memory quantity for each LPAR.
  • 2h: Identify desired storage quantity for each LPAR.
  • 4h: Identify application support responsible parties.
  • 1h: Identify VIO Client LPAR node name and resource group names.
  • 2h: Assign a virtual SCSI adapter slot group to each VIO Client LPAR.
  • 1h: From HMC, create each VIO Client LPAR.
  • 15m: Define CPU sharing for each VIO Client LPAR.
  • 15m: Define CPU allocations for each VIO Client LPAR.
  • 15m: Define processing unit allocations for each VIO Client LPAR.
  • 15m: Define memory allocations for each VIO Client LPAR.
  • 1h: Allocate additional physical application network adapters for each VIO Client LPAR.
  • 1h: Allocate additional physical management network adapters for each VIO Client LPAR.
  • 2h: Allocate additional physical storage adapters for each VIO Client LPAR.
  • 1h: Allocate virtual ethernet adapters for application network on each VIO Client LPAR from both VIO servers on the frame.
  • 1h: Allocate virtual ethernet adapters for management network on each VIO Client LPAR from both VIO servers on the frame.
  • 1h: Allocate virtual SCSI adapters for all slots in slot group assigned to each VIO Client LPAR from both VIO servers on the frame.
  • 1h: Request a boot disk from the SAN for each VIO Client LPAR.
  • 30m: CHANGE REQUEST: Allocate one LUN from SAN for each VIO Client LPAR for operating system storage

24h45m

VIO Server Build

  • 1h: Configure the NIM server to perform a remote operating system load of the primary VIO server.
  • 1h: Configure the NIM server to perform a remote operating system load of the secondary VIO server.
  • 1h: From the HMC, boot each VIO server LPAR to it's SMS menu and configure one of the network adapters to perform a network boot from the NIM server.
  • 15m: Ping test each VIO server before begining the network boot from NIM.
  • 4h: Boot the primary VIO server LPAR from the NIM server and load the IOS.
  • 4h: Boot the secondary VIO server LPAR from the NIM server and load the IOS.
  • 15m: After the primary VIO server LPAR is loaded , login as the user "padmin" and accept the license agreement.
  • 15m: After the secondary VIO server LPAR is loaded , login as the user "padmin" and accept the license agreement.
  • 15m: Reset the date/time on each VIO Server
  • 15m: Set the "padmin" password to the current standard production value.
  • 30m: Change the hostname of the VIO server to the previously identified value and modify the "/etc/rc.net" file to reflect this hostname change.
  • 1h: Setup the password-less ssh login between the NIM server and each VIO server (primary and secondary) for the "root" user.
  • 2h: Load additional MPIO drivers necessary to communicate with the SAN system.
  • 2h: Remove all SAN disks, fscsi, and fcs adapters and reboot the VIO server in order for the new MPIO drivers to take effect.
  • 3h: Upgrade primary VIO server IOS to latest level.
  • 3h: Upgrade secondary VIO server IOS to latest level.
  • 1h: Reboot each VIO server.
  • 1h: Verify WWN's and SAN connectivity.
  • 6h: Identify each disk by PVID and allocate for VIO Client LPAR operating system usage.
  • 30m: Remove all FC attached disks.
  • 15m: Set fast fail parameter for SCSI adapters and reconfigure FC adapters.
  • 15m: Change reserve policy on all hdisks to "no_reserve"
  • 15m: Change algorithm on all hdisks to "round_robin"
  • 15m: Change hcheck_interval on all hdisks to "20".
  • 15m: Change queue_depth on all hdisks to "16".
  • 15m: Change physical volume identifier on all hdisks to "yes".
  • 2h: Set the path priority value for each hdisk to evenly distribute the load between the primary and secondary VIO servers.
  • 1h: Remove any internal SCSI hdisks from the rootvg and the system.
  • 15m: Reconfigure ethernet adapters for etherchannel configuration, begin by unconfiguring all ethernet adapters.
  • 15m: Rediscover all ethernet adapters
  • 15m: Turn off the chksum_offload parameter on all adapters
  • 15m: Identify the application ethernet adapters and create the application etherchannel adapter.
  • 15m: Identify the management ethernet adapters and create the management etherchannel adapter.
  • 2h: Build the SEA's (shared ethernet adapters). Use the virtual ethernet adapters associated with slots 500 and 850 on the primary VIO server, and slots 501 and 851 on the secondary VIO server.
  • 1h: Configure VIO server IP addresses for the application network on the virtual ethernet adapter associated with slots 510 and 511.
  • 1h: Configure VIO server IP addresses for the management network on the virtual ethernet adapter associated with slots 860 and 861.
  • Repeat this VIO server configuration process for each VIO Server.

42h0m

VIO Client LPAR Build

  • 30m: CHANGE REQUEST: Allocate one LUN from SAN for each VIO Client LPAR for operating system storage
  • 30m: CHANGE REQUEST: Allocate one or more LUNs from SAN for application storage on each VIO Client LPAR.
  • 30m: CHANGE REQUEST: Allocate one or more LUNs from SAN for paging space on each VIO Client LPAR as necessary.
  • Primary Frame

    • 1h: Discover, identify, and record VIO Client LPAR operating system LUN on primary VIO server, on primary frame.
    • 1h: Discover, identify, and record VIO Client LPAR operating system LUN on secondary VIO server , on primary frame.
    • 1h: Discover, identify, and record VIO Client LPAR application LUNs on primary VIO server on primary frame.
    • 1h: Discover, identify, and record VIO Client LPAR application LUNs on secondary VIO server on primary frame.
    • 2h: Allocate VIO Client LPAR operating system LUN on primary VIO server, on primary frame, to vhost adapter defined for OS usage associated with the slot group for this LPAR.
    • 2h: Allocate VIO Client LPAR operating system LUN on secondary VIO server, on primary frame, to vhost adapter defined for OS usage associated with the slot group for this LPAR.
    • 4h: Allocate VIO Client LPAR application LUNs on primary VIO server, on primary frame, to vhost adapter defined for VG usage associated with the slot group for this LPAR.
    • 4h: Allocate VIO Client LPAR application LUNs on secondary VIO server, on primary frame, to vhost adapter defined for VG usage associated with the slot group for this LPAR.

    Fail-over frame

    • 1h: Discover, identify, and record VIO Client LPAR operating system LUN on primary VIO server on fail-over frame.
    • 1h: Discover, identify, and record VIO Client LPAR operating system LUN on secondary VIO server on fail-over frame.
    • 1h: Discover, identify, and record VIO Client LPAR application LUNs on primary VIO server on fail-over frame.
    • 1h: Discover, identify, and record VIO Client LPAR application LUNs on secondary VIO server on fail-over frame.
    • 2h: Allocate VIO Client LPAR operating system LUN on primary VIO server, on fail-over frame, to vhost adapter defined for OS usage associated with the slot group for this LPAR.
    • 2h: Allocate VIO Client LPAR operating system LUN on secondary VIO server, on fail-over frame, to vhost adapter defined for OS usage associated with the slot group for this LPAR.
    • 4h: Allocate VIO Client LPAR application LUNs on primary VIO server, on fail-over frame, to vhost adapter defined for VG usage associated with the slot group for this LPAR.
    • 4h: Allocate VIO Client LPAR application LUNs on secondary VIO server, on fail-over frame, to vhost adapter defined for VG usage associated with the slot group for this LPAR.

  • 4h: Create the client LPAR boot image on the NIM Server
  • 30m: Define the mksysb_image resource object
  • 2h: Define the SPOT resource object
  • 30m: From the HMC, boot the VIO Client LPAR on the primary frame to SMS and define the network boot IP information.
  • 30m: Configure the NIM server to perform a remote operating system load of the VIO Client LPAR on the primary frame.
  • 2h: Boot the VIO Client LPAR on the primary frame from the NIM server and perform a "mksysb" operating system load.
  • 30m: From the HMC, boot the VIO Client LPAR on the fail-over frame to SMS and define the network boot IP information.
  • 30m: Configure the NIM server to perform a remote operating system load of the VIO Client LPAR on the fail-over frame.
  • 2h: Boot the VIO Client LPAR on the fail-over frame from the NIM server and perform a "mksysb" operating system load.
  • 2h: On each VIO Client LPAR: Set the path priority for each hdisk.
  • 15m: On each VIO Client LPAR: Configure the node name and host name.
  • 30m: On each VIO Client LPAR: Unconfigure all ethernet adapters.
  • 15m: On each VIO Client LPAR: Rediscover all ethernet adapters.
  • 15m: On each VIO Client LPAR: Turn off the chksum_offload parameter on all adapters.
  • 1h: Create etherchannels and distribute the network traffic evenly across dual VIO servers.
  • 15m: On each VIO Client LPAR: Even numbered hosts should use the even numbered virtual ethernet adapter as primary network path.
  • 15m: On each VIO Client LPAR: Odd numbered hosts should use the odd numbered virtual ethernet adapter as primary network path.
  • 30m: Create the "/etc/hosts" and insert all names and IP addresses for all cluster nodes.
  • 15m: On each VIO Client LPAR: Configure a "boot" IP address on each etherchannel adapter.
  • 15m: Configure the "persistent" IP address on the application network etherchannel adapter.
  • 15m: Configure the "management" IP address on the management network etherchannel adapter.
  • 30m: Create the "/.rhosts" and insert all names and IP addresses for all cluster nodes.
  • 30m: Create the "/usr/es/sbin/cluster/etc/rhosts" file and insert all boot hostnames, node names.
  • 30m: Create the "/usr/es/sbin/cluster/netmon.cf" file and insert all boot hostnames, node names.
  • 15m: Remove any existing cluster definitions.
  • 30m: Identify cluster wide unique volume group major numbers.
  • 30m: Create Volume groups using an enterprise wide unique naming structure.
  • 30m: Create JFS log logical volumes for JFS using an enterprise wide unique naming structure.
  • 30m: Create file system or raw logical volumes using an enterprise wide unique naming structure.
  • 30m: Create JFS2 file systems using an enterprise wide unique naming structure.
  • 15m: Change ownership of file system mount points to a user that will administer applications.
  • 15m: Change permissions of file system mount points to allow users to read and search.
  • 1h: Adjust the asynchronous I/O and system tuning parameters
  • 15m: Increase the paging space.
  • 15m: Change permission on TSM log files to allow clients to write to them.
  • 1h: Customize the "/etc/inittab" file.
  • 15m: Deny FTP access to all users.
  • 1h: Configure Tivoli TEC agent.
  • 1h: Adjust AIX tuning parameters.
  • 15m: Configure time zone.
  • 30m: Add performance storage area to list of NFS directories mounted at boot time.
  • 2h: Configure Partition Load Manager.
  • 4h: Configure Workload Manger.
  • 1h: Configure console access.
  • 2h: Adjust default security settings.
  • 1h: Adjust ulimit settings.
  • 1h: Configure NTP Time daemon.
  • 15m: Configure primary prompt string PS1 in "/etc/profile".
  • 15m: Configure MOTD in "/etc/motd".
  • 15m: Configure Herald in "/etc/security/login.cfg".
  • 1h: Configure TSM backups.

75h0m

High Availability Clustered Multi-Processing

  • 2h: Determine the critical components, functions and dependencies.
  • 4h: Determine the recovery strategies for each HA component, function and dependency.
  • 1h: Determine the HA requirements.
  • 1h: Identify specific system data, libraries, file systems, etc.
  • 1h: Identify specific application data, libraries, databases, etc.
  • 15m: Identify cluster wide unique volume group major numbers.
  • 1h: Create Volume groups using an enterprise wide unique naming structure.
  • 1h: Create JFS log logical volumes for JFS using an enterprise wide unique naming structure.
  • 1h: Create file system or raw logical volumes using an enterprise wide unique naming structure.
  • 1h: Create JFS2 file systems using an enterprise wide unique naming structure.
  • 15m: Change ownership of file system mount points to a user that will administer applications.
  • 15m: Change permissions of file system mount points to allowr users to read and search.
  • 1h: Obtain boot IP addresses for all network interfaces on all cluster nodes.
  • 1h: Obtain persistent IP addresses for each cluster node.
  • 1h: Obtain node name service IP addresses for each cluster node.
  • 1h: Obtain resource group service IP addresses for each resource group.
  • 1h: Obtain management IP addresses for each cluster node.
  • 15m: Identify enterprise wide unique boot host names.
  • 15m: Identify perisistent node name.
  • 15m: Identify resource group.
  • 15m: Identify management IP addresses.
  • 15m: Identify resource group alias names.
  • 15m: Create a cluster and identify at least one boot address of all cluster nodes.
  • 30m: Create all resource group and node name service addresses of all cluster nodes.
  • 4h: Create application server start/stop scripts customized for each resource group and application.
  • 30m: Create application servers for each resource group.
  • 30m: Create persistant addresses for each cluster node.
  • 1h: Create all resource groups configured to never fall back once a failover occurs.
  • 15m: Add the Service address to each resource group for each cluster node.
  • 15m: Add application servers and volumes to resource groups for each cluster node.
  • 2h: Synchronize and verify cluster configuration.
  • 1h: Start HACMP on all cluster nodes.
  • 1h: Perform failover testing for IP addresses.
  • 1h: Perform failover testing for volume groups.
  • 1h: Perform failover testing for file systems.
  • 1h: Perform failover testing for applications.
  • 1h: Certify all failover properly, and start/stop as designed.
  • 8h: Document HA configuration.
  • 4h: Develop the technical HA plan for each application module.
  • 8h: Develop HA test procedures.
  • 8h: Exercise and test the HA configuration.
  • 8h: Compile HA failover test results.
  • 8h: Conduct HA test post mortems with all support teams.
  • 8h: Present the HA test results to client for sign-off.
  • 8h: Develop communication/notification procedures for HA and return to service events.

95h15m

Disaster Recovery

  • 30m: CHANGE REQUEST: Allocate one LUN from SAN for each disaster recovery LPAR for operating system storage
  • 30m: CHANGE REQUEST: Allocate one or more replicated LUNs from SAN for application storage on each disaster recovery LPAR.
  • 30m: CHANGE REQUEST: Allocate one or more LUNs from SAN for paging space on each disaster recovery LPAR as necessary.
  • 1h: Discover, identify, and record disaster recovery LPAR operating system LUN on primary VIO server, on disaster recovery frame.
  • 1h: Discover, identify, and record disaster recovery LPAR operating system LUN on secondary VIO server, on disaster recovery frame.
  • 1h: Discover, identify, and record replicated disaster recovery LPAR application LUNs on primary VIO server on disaster recovery frame.
  • 1h: Discover, identify, and record replicated disaster recovery LPAR application LUNs on secondary VIO server on disaster recovery frame.
  • 2h: Allocate disaster recovery LPAR operating system LUN on primary VIO server, on disaster recovery frame, to vhost adapter defined for OS usage associated with the slot group for this disaster recovery LPAR.
  • 2h: Allocate disaster recovery LPAR operating system LUN on secondary VIO server, on disaster recovery frame, to vhost adapter defined for OS usage associated with the slot group for this disaster recovery LPAR.
  • 4h: Allocate disaster recovery LPAR application LUNs on primary VIO server, on disaster recovery frame, to vhost adapter defined for VG usage associated with the slot group for this disaster recovery LPAR.
  • 4h: Allocate disaster recovery LPAR application LUNs on secondary VIO server, on disaster recovery frame, to vhost adapter defined for VG usage associated with the slot group for this disaster recovery LPAR.
  • 4h: Create the disaster recovery LPAR boot image on the NIM Server
  • 30m: Define the mksysb_image resource NIM object
  • 2h: Define the SPOT resource object
  • 30m: From the HMC, boot the disaster recovery LPAR on the primary frame to SMS and define the network boot IP information.
  • 30m: Configure the NIM server to perform a remote operating system load of the disaster recovery LPAR on the primary frame.
  • 2h: Boot the disaster recovery LPAR on the primary frame from the NIM server and perform a "mksysb" operating system load.
  • 2h: On each disaster recovery LPAR: Set the path priority for each hdisk.
  • 15m: On each disaster recovery LPAR: Configure the node name and host name.
  • 30m: On each disaster recovery LPAR: Unconfigure all ethernet adapters.
  • 15m: On each disaster recovery LPAR: Rediscover all ethernet adapters.
  • 15m: On each disaster recovery LPAR: Turn off the chksum_offload parameter on all adapters.
  • 1h: Create etherchannels and distribute the network traffic evenly across dual VIO servers.
  • 15m: On each disaster recovery LPAR: Even numbered hosts should use the even numbered virtual ethernet adapter as primary network path.
  • 15m: On each disaster recovery LPAR: Odd numbered hosts should use the odd numbered virtual ethernet adapter as primary network path.
  • 30m: Create the "/etc/hosts" and insert all names and IP addresses for all cluster nodes.
  • 15m: On each disaster recovery LPAR: Configure a "boot" IP address on each etherchannel adapter.
  • 15m: Configure the "persistent" IP address on the application network etherchannel adapter.
  • 15m: Configure the "management" IP address on the management network etherchannel adapter.
  • 30m: Create the "/.rhosts" and insert all names and IP addresses for all cluster nodes.
  • 30m: Create the "/usr/es/sbin/cluster/etc/rhosts" file and insert all boot hostnames, node names.
  • 30m: Create the "/usr/es/sbin/cluster/netmon.cf" file and insert all boot hostnames, node names.
  • 15m: Remove any existing cluster definitions.
  • 30m: Identify cluster wide unique volume group major numbers.
  • 30m: Identify volume groups from replicated application LUNs.
  • 30m: Import volume groups from replicated application LUNs.
  • 15m: Change ownership of file system mount points to a user that will administer applications.
  • 15m: Change permissions of file system mount points to allow users to read and search.
  • 1h: Adjust the asynchronous I/O and system tuning parameters on disaster recovery LPAR.
  • 15m: Increase the paging space on disaster recovery LPAR.
  • 15m: Change permission on TSM log files to allow clients to write to them on disaster recovery LPAR.
  • 1h: Customize the "/etc/inittab" file on disaster recovery LPAR.
  • 15m: Deny FTP access to all users on disaster recovery LPAR.
  • 1h: Configure Tivoli TEC agent on disaster recovery LPAR.
  • 1h: Adjust AIX tuning parameters on disaster recovery LPAR.
  • 15m: Configure time zone on disaster recovery LPAR.
  • 30m: Add performance storage area to list of NFS directories mounted at boot time on disaster recovery LPAR.
  • 2h: Configure Partition Load Manager on disaster recovery LPAR.
  • 4h: Configure Workload Manger on disaster recovery LPAR.
  • 1h: Configure console access on disaster recovery LPAR.
  • 2h: Adjust default security settings on disaster recovery LPAR.
  • 1h: Adjust ulimit settings on disaster recovery LPAR.
  • 1h: Configure NTP Time daemon on disaster recovery LPAR.
  • 15m: Configure primary prompt string PS1 in "/etc/profile" on disaster recovery LPAR.
  • 15m: Configure MOTD in "/etc/motd" on disaster recovery LPAR.
  • 15m: Configure Herald in "/etc/security/login.cfg" on disaster recovery LPAR.
  • 1h: Configure TSM backups on disaster recovery LPAR.
  • 4h: Ensure adequate connectivity/bandwidth between data centers to support DR functionality.
  • 4h: Determine the critical components, functions and dependencies.
  • 8h: Determine the recovery strategies for each DR component, function and dependency.
  • 8h: Determine the DR Plan requirements.
  • 8h: Identify specific system data, libraries and file systems required at recovery site.
  • 8h: Identify specific application data, libraries and databases required at recovery site.
  • 24h: Develop and document the Specific DR plan for the supported environment.
  • 24h: Develop the technical DR Recovery Plans for each recovery environment.
  • 40h: Develop DR test procedures.
  • 24h: Exercise and test the DR recovery plans.
  • 8h: Compile DR test results.
  • 8h: Conduct DR test post mortems with all DR recovery teams.
  • 8h: Present the DR exercise results to client for sign-off.
  • 8h: Develop communication/notification procedures for DR & Return to Service events.
  • 4h: Publish basic DR Plan.
  • 8h: Incorporate new DR plan into the overall DR strategy for Customer.
  • 8h: Distribute updated DR plan to participants.
  • 8h: Communicate and turnover DR procedures to ongoing support team.
  • 8h: Communicate and turnover backup & restoration procedures to onging support team.
  • 8h: Communicate and turnover security scheme to onging support team.

283h15m

Configuration Management

  • Integrate new application environment into existing configuration management process.
  • Define new application code for existing configuration management tool.

Change Management

  • Integrate new application environment into existing change management process.
  • Integrate new application environment into existing change management tool.
  • Define change approvers and set up in change management tool.
  • Conduct change management process/tool training.

Availability Management

  • Integrate new application environment into existing availability management process.
  • Develop availability reporting method and procedures.
  • Integrate new application environment into existing availability management tool.
  • Develop SLA availability reporting methods and procedures for new application environment.

Incident Management

  • Integrate new application environment into existing incident management process.
  • Identify primary points of contact for incidents.
  • Define incident escalation procedures.
  • Integrate new application environment into existing incident management tool.
  • Conduct incident management process/tool training.
  • Develop SLA incident reporting methods and procedures for new application environment.

Problem Management

  • Integrate new application environment into existing problem management process.

Outage Management

  • Integrate new application environment into existing outage management process.
  • Identify primary points of contact for outage planning.
  • Define outage escalation procedures.
  • Define outage approvers and set up in outage management tool.
  • Conduct outage management process/tool training.

Project Management

  • Schedule kickoff meeting with Vendors, CGE and Customer involvement.
  • Conduct kickoff meeting.
  • Plan and staff project resources needs.
  • Build project schedule.
  • Update project schedule on a weekly bases.
  • Schedule status meetings with Customer.
  • Conduct client status meetings with Customer.
  • Schedule weekly team meetings (Vendors and CGE).
  • Conduct weekly team meetings (Vendors and CGE).
  • Build and maintain team contact lists.
  • Manage Issue/risk.
  • Communicate to client known issue/risks.
  • Build and submit storage requests.
  • Request charge code access in Replican for CGE team members.
  • Track funding and report progress to client.
  • Create HAWK form and deliver to facilities.

-
Server Build Documents
-
 

APP50 DR Doc
APP51 DR Doc
APP52 DR Doc
APP53 DR Doc
DB55 DR Doc
DB56 DR Doc
APP80 DR Doc
APP81 DR Doc
APP82 DR Doc
DB85 DR Doc
DB86 DR Doc
atlas_a DR Doc
atlas_b DR Doc
atlas_c DR Doc
delphi_e DR Doc
delphi_f DR Doc
Move Egate Preprod
Build Power 5 EGATE PreProd
Build Power 5 EGATE Prod
EGATE DR Test Procedures
SAP Sandbox Build Procedure
MXF DEFS Build Procedure
AIX Sys Build Perot


AIX FREE DEMO UNIX
Automated AIX Deployment
Reduce your DC costs

www.siteox.com

LPAR Leasing
Lease an AIX / i5 LPAR
Reduce your costs

www.mtxia.com

Server Leasing
Lease a Server off-site
Reduce your costs

www.mtxia.com

Data Center Automation
Business Continuity and DR
Virtualization/Consolidation

www.mtxia.com

HMC Service
Hardware Management Console
Manage Remote AIX / i5 LPARs

www.siteox.com

Business Web Site Hosting
$3.99 / month includes Tools,
Shopping Cart, Site Builder

www.siteox.com

Disaster Recovery
Small Business Oriented
Off-Site Facilities

www.mtxia.com

IBM pSeries / iSeries
Reduce your Costs
Off-Site Server Hosting

www.mtxia.com