hacmp for aix planning guide -...

205
Bull HACMP 4.4 Planning Guide AIX 86 A2 55KX 02 ORDER REFERENCE

Upload: buikhanh

Post on 15-Mar-2018

250 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Bull HACMP 4.4Planning Guide

AIX

86 A2 55KX 02ORDER REFERENCE

Page 2: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE
Page 3: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Bull HACMP 4.4Planning Guide

AIX

Software

August 2000

BULL CEDOC357 AVENUE PATTONB.P.2084549008 ANGERS CEDEX 01FRANCE

86 A2 55KX 02ORDER REFERENCE

Page 4: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

The following copyright notice protects this book under the Copyright laws of the United States of Americaand other countries which prohibit such actions as, but not limited to, copying, distributing, modifying, andmaking derivative works.

Copyright Bull S.A. 1992, 2000

Printed in France

Suggestions and criticisms concerning the form, content, and presentation ofthis book are invited. A form is provided at the end of this book for this purpose.

To order additional copies of this book or other Bull Technical Publications, youare invited to use the Ordering Form also provided at the end of this book.

Trademarks and Acknowledgements

We acknowledge the right of proprietors of trademarks mentioned in this book.

AIX� is a registered trademark of International Business Machines Corporation, and is being used underlicence.

UNIX is a registered trademark in the United States of America and other countries licensed exclusively throughthe Open Group.

Year 2000

The product documented in this manual is Year 2000 Ready.

The information in this document is subject to change without notice. Groupe Bull will not be liable for errorscontained herein, or for incidental or consequential damages in connection with the use of this material.

Page 5: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Contents

About This Guide xi

Chapter 1 Planning an HACMP Cluster—Overview 1-1

Design Goal: Eliminating Single Points of Failure . . . . . . . . . . 1-1Eliminating Cluster Objects as Single Points of Failure . . 1-2

Cluster Planning Worksheets . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-3Paper Worksheets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-3Online Worksheets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-3

The Planning Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-3Step 1: Draw a Cluster Diagram (Chapter 2) . . . . . . . . . . . . 1-4Step 2: Plan Your TCP/IP Networks (Chapter 3) . . . . . . . . 1-4Step 3: Plan Your Serial Networks (Chapter 4) . . . . . . . . . . 1-4Step 4: Plan Your Shared Disk Devices (Chapter 5) . . . . . 1-4Step 5: Plan Your Shared LVM Components (Chapter 6) . 1-4Step 6: Plan Your Application Servers and Resource Groups (Chapter 7) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1-5Step 7: Tailor the Cluster Event Processing (Chapter 8) . . 1-5Step 8: Plan Your HACMP for AIX Clients (Chapter 9) . . 1-5Step 9: Installing and Configuring Your HACMP Cluster 1-5

Chapter 2 Drawing the Cluster Diagram 2-1

Prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2-1Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2-1Making a First Pass At Drawing the Cluster Diagram . . . . . . . 2-1

Naming the Cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2-3Identifying the Highly Available Applications and Resource Types . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2-3Selecting the Number of Nodes . . . . . . . . . . . . . . . . . . . . . . 2-4Assigning Cluster IP Addresses . . . . . . . . . . . . . . . . . . . . . . 2-5Selecting the Method of Shared Disk Access . . . . . . . . . . . 2-5

Where You Go From Here . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2-6

Chapter 3 Planning TCP/IP Networks 3-1

Prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-1Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-1Selecting Public and Private Networks . . . . . . . . . . . . . . . . . . . . 3-1

Planning Guide Preface v

Page 6: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Contents

vi

Network Used for Cluster Lock Manager Traffic . . . . . . . . 3-2Designing a Network Topology . . . . . . . . . . . . . . . . . . . . . . . . . 3-2

Sample Network Topologies . . . . . . . . . . . . . . . . . . . . . . . . . 3-2Single Network . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-3Dual Network . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-4Point-to-Point Connection . . . . . . . . . . . . . . . . . . . . . . . . . . 3-5

Identifying Network Components . . . . . . . . . . . . . . . . . . . . . . . . 3-6Nodes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-6Network Adapters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-6Network Interfaces . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-8Networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-9

Defining a Network Mask . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-10Defining IP Addresses for Standby Adapters . . . . . . . . . . . . . . 3-10

Placing Standby Adapters on a Separate Subnet . . . . . . . . 3-11Subnet Considerations for Cluster Monitoring with Tivoli 3-12

Defining Boot Addresses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-13Boot/Service/Standby Address Requirements for Resource Groups . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-13

Defining Hardware Addresses . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-14Avoiding Network Conflicts . . . . . . . . . . . . . . . . . . . . . . . . . 3-17

Using HACMP with NIS and DNS . . . . . . . . . . . . . . . . . . . . . . . 3-18How HACMP Enables and Disables Nameserving . . . . . . 3-18

Planning for Cluster Performance . . . . . . . . . . . . . . . . . . . . . . . . 3-21Setting I/O Pacing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-21Setting Syncd Frequency . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-22Setting Failure Detection Parameters . . . . . . . . . . . . . . . . . . 3-22

Adding the TCP/IP Network Topology to the Cluster Diagram . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-25Completing the TCP/IP Networks Worksheets . . . . . . . . . . . . . 3-26

Completing the TCP/IP Networks Worksheet . . . . . . . . . . 3-26Completing the TCP/IP Network Adapter Worksheet . . . . 3-26

Where You Go From Here . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3-27

Chapter 4 Planning Serial Networks 4-1

Prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4-1Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4-1Node Isolation and Partitioned Clusters . . . . . . . . . . . . . . . . . . . 4-1Serial Network Topology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4-3

Supported Serial Network Types . . . . . . . . . . . . . . . . . . . . . . 4-4Adding the Serial Network Topology to the Cluster Diagram . 4-5Completing the Serial Networks and Serial Network Adapter Worksheets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4-6

Completing the Serial Networks Worksheet . . . . . . . . . . . . 4-6Completing the Serial Network Adapter Worksheet . . . . . 4-7

Where You Go From Here . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4-7

Planning Guide

Page 7: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Contents

Chapter 5 Planning Shared Disk Devices 5-1

Prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5-1Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5-1Choosing a Shared Disk Technology . . . . . . . . . . . . . . . . . . . . . . 5-1

SCSI Disks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5-2IBM 9333 Serial Disk Subsystems . . . . . . . . . . . . . . . . . . . . 5-5IBM Serial Storage Architecture Disk Subsystem . . . . . . . 5-6

Power Supply Considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5-6SCSI Configurations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5-6IBM 9333 Serial Disk Subsystem Configurations . . . . . . . 5-7IBM SSA Disk Subsystem Configurations . . . . . . . . . . . . . 5-7

Planning for Non-Shared Disk Storage . . . . . . . . . . . . . . . . . . . . 5-7Planning for Shared Disk Storage . . . . . . . . . . . . . . . . . . . . . . . . 5-9Planning a Shared SCSI-2 Disk Installation . . . . . . . . . . . . . . . . 5-9

Disk Adapters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5-9Cables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5-10Sample SCSI-2 Differential Configuration . . . . . . . . . . . . . 5-10Sample SCSI-2 Differential Fast/Wide Configuration . . . . 5-11Sample IBM 7135-210 RAIDiant Disk ArrayConfiguration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5-11Sample IBM 2105 Versatile Storage Server Configuration 5-13

Planning a Shared IBM 9333 Serial Disk Installation . . . . . . . . 5-14Sample Two-Node Configuration. . . . . . . . . . . . . . . . . . . . . . 5-15Sample Four-Node Configuration . . . . . . . . . . . . . . . . . . . . . 5-16Sample Eight-Node Configuration . . . . . . . . . . . . . . . . . . . . 5-17

Planning a Shared IBM SSA Disk Subsystem Installation . . . . 5-18IBM Manuals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5-18Adapters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5-18Using SSA Features for High Availability . . . . . . . . . . . . . . 5-19

9333 and SSA Disk Fencing in Concurrent Access Clusters . . 5-21Completing the Disk Worksheets . . . . . . . . . . . . . . . . . . . . . . . . . 5-23

Completing the Shared SCSI-2 Disk Worksheet . . . . . . . . . 5-24Completing the Shared SCSI-2 Disk Array Worksheet . . . 5-24Completing the IBM 9333 Serial Disk Worksheet . . . . . . . 5-24Completing the IBM Serial Storage Architecture Disk Subsystems Worksheet . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5-25

Adding the Disk Configuration to the Cluster Diagram . . . . . . 5-25Where You Go From Here . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5-25

Chapter 6 Planning Shared LVM Components 6-1

Prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-1Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-1

Planning for Non-Concurrent Access . . . . . . . . . . . . . . . . . . 6-1Planning for Concurrent Access . . . . . . . . . . . . . . . . . . . . . . 6-1TaskGuide for Creating Shared Volume Groups . . . . . . . . . 6-2

Planning Guide Preface vii

Page 8: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Contents

vii

LVM Components in the HACMP for AIX Environment . . . . 6-2Physical Volumes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-3Volume Groups . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-3Logical Volumes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-4Filesystems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-5

LVM Mirroring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-5Mirroring Physical Partitions . . . . . . . . . . . . . . . . . . . . . . . . 6-5Mirroring Journal Logs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-7

Quorum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-7Quorum at Vary On . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-7Quorum after Vary On . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-8Disabling and Enabling Quorum . . . . . . . . . . . . . . . . . . . . . . 6-8Quorum in Non-Concurrent Access Configurations . . . . . . 6-9Quorum in Concurrent Access Configurations . . . . . . . . . . 6-10

Using NFS with HACMP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-11Reliable NFS Server Capability . . . . . . . . . . . . . . . . . . . . . . . 6-11Creating Shared Volume Groups . . . . . . . . . . . . . . . . . . . . . . 6-11NFS Exporting Filesystems and Directories . . . . . . . . . . . . 6-12NFS Mounting and Fallover . . . . . . . . . . . . . . . . . . . . . . . . . . 6-12

Planning Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-15Completing the Shared LVM Components Worksheets. . . . . . 6-16

Non-Concurrent Access Worksheets . . . . . . . . . . . . . . . . . . . 6-16Completing the Non-Shared Volume Group Worksheet (Non-Concurrent Access) . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-16Concurrent Access Worksheets . . . . . . . . . . . . . . . . . . . . . . . 6-18

Where You Go From Here . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6-19

Chapter 7 Planning Applications, Application Servers, and Resource Groups 7-1

Prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7-1Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7-1

Application Servers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7-1Some Applications are Integrated with HACMP . . . . . . . . 7-2

Planning Applications and Application Servers . . . . . . . . . . . . 7-2Completing the Application Worksheet . . . . . . . . . . . . . . . . 7-3Completing the Application Server Worksheet . . . . . . . . . . 7-4

Planning for AIX Fast Connect . . . . . . . . . . . . . . . . . . . . . . . . . . 7-4Converting from AIX Connections to AIX Fast Connect . 7-5Planning Considerations for Fast Connect . . . . . . . . . . . . . . 7-5Fast Connect as a Highly Available Resource . . . . . . . . . . . 7-6Completing the Fast Connect Worksheet . . . . . . . . . . . . . . . 7-6

Planning for AIX Connections . . . . . . . . . . . . . . . . . . . . . . . . . . . 7-7AIX Connections Realms and Services . . . . . . . . . . . . . . . . 7-7Planning Notes for AIX Connections . . . . . . . . . . . . . . . . . . 7-7AIX Connections as a Highly Available Resource . . . . . . . 7-7

i Planning Guide

Page 9: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Contents

Completing the AIX Connections Worksheet . . . . . . . . . . . 7-8Planning CS/AIX Communications Links . . . . . . . . . . . . . . . . . 7-9

Completing the CS/AIX Communications Links Worksheet . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7-9

Planning Resource Groups . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7-10Guidelines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7-10Completing the Resource Group Worksheet. . . . . . . . . . . . . 7-12

Where You Go From Here . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7-14

Chapter 8 Tailoring Cluster Event Processing 8-1

Prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8-1Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8-1Customizing Cluster Event Processing. . . . . . . . . . . . . . . . . . . . . 8-1

Event Notification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8-2Pre- and Post-Event Scripts . . . . . . . . . . . . . . . . . . . . . . . . . . 8-2Event Recovery and Retry . . . . . . . . . . . . . . . . . . . . . . . . . . . 8-3

Completing the Cluster Event Worksheet . . . . . . . . . . . . . . . . . . 8-3Where You Go From Here . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8-4

Chapter 9 Planning HACMP for AIX Clients 9-1

Prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9-1Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9-1Different Types of Clients: Computers and Terminal Servers . 9-1

Client Application Systems . . . . . . . . . . . . . . . . . . . . . . . . . . 9-1NFS Servers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9-1Terminal Servers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9-1

Clients Running Clinfo . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9-2Reconnecting to the Cluster . . . . . . . . . . . . . . . . . . . . . . . . . . 9-2Tailoring the clinfo.rc Script . . . . . . . . . . . . . . . . . . . . . . . . . 9-2

Network Components and Clients Not Running Clinfo . . . . . . 9-2Where You Go From Here . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9-3

Appendix A Planning Worksheets A-1

Appendix B Using the Online Cluster Planning Worksheet Program B-1

Appendix C Single-Adapter Networks C-1

Appendix D Applications and HACMP D-1

Index X-1

Planning Guide Preface ix

Page 10: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Contents

x

Planning Guide
Page 11: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

n an

About This Guide

This guide presents information necessary to plan an HACMP cluster.

Who Should Use This Guide

This guide is intended for system administrators, network administrators, and customer engineers responsible for:

• Planning hardware and software resources for an HACMP for AIX environment

• Configuring networks

• Defining physical and logical storage

• Installing and configuring an HACMP cluster.

Before You Begin

As a prerequisite for planning your HACMP cluster, you should be familiar with:

• ESCALA components (including disk devices, cabling, and network adapters for each system). HACMP for AIX, Version 4.4 runs on ESCALA uniprocessor systems, SP systems, and PowerPC Symmetric Multiprocessors (SMP).

• The AIX operating system, including the Logical Volume Manager subsystem.

• The System Management Interface Tool (SMIT).

• Communications, including the TCP/IP subsystem.

This guide explains the topics listed above as they relate to the HACMP for AIX software. Prior knowledge of these topics can aid you in planning, installing, and configuring an HACMP cluster.

Using the Planning Worksheets

Appendix A contains blank copies of the worksheets used in this guide. These worksheets help you plan, install, and track an HACMP cluster. Make copies of each worksheet and keep the blank originals in this book. The worksheets are especially useful when making modifications and performing diagnostics.

Alternatively, you can use the online planning worksheets. See Appendix B for information.

Additionally, you will benefit from creating diagrams of your cluster.

How To Use This Guide

This document has the following chapters and appendixes.

• Chapter 1, Planning an HACMP Cluster—Overview, describes the design goals for building an HACMP cluster and lists the recommended steps you should follow to plaHACMP cluster.

Planning Guide Preface xi

Page 12: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

xii

• Chapter 2, Drawing the Cluster Diagram, describes how to draw a cluster diagram and explains essential concepts and terminology you must understand to plan an HACMP cluster.

• Chapter 3, Planning TCP/IP Networks, provides information specific to planning TCP/IP networks supported in an HACMP for AIX environment.

• Chapter 4, Planning Serial Networks, uses the diagram you drew in Chapter 2 to help you plan for serial networks in your HACMP cluster.

• Chapter 5, Planning Shared Disk Devices, discusses information you must consider before configuring shared external disks in an HACMP cluster. It specifically describes the shared disk configurations the HACMP for AIX software supports.

• Chapter 6, Planning Shared LVM Components, describes the concepts and terminology you need to understand to plan for shared volume groups in an HACMP cluster.

• Chapter 7, Planning Applications, Application Servers, and Resource Groups, describes how to plan a cluster around mission-critical applications and provides recommendations you must consider to plan for application servers and resource groups in an HACMP cluster.

• Chapter 8, Tailoring Cluster Event Processing, describes how to tailor cluster event processing for your cluster.

• Chapter 9, Planning HACMP for AIX Clients, discusses planning considerations for clients that can access nodes in an HACMP cluster.

• Appendix A, Planning Worksheets, contains worksheets you can use in planning your HACMP cluster, its network and disk architecture, and the applications, application servers, and resources to be made highly available.

• Appendix B, Using the Online Cluster Planning Worksheet Program, contains information about using the online planning worksheet program.

• Appendix C, Single-Adapter Networks, describes a program that is useful for determining service adapter failure in a cluster with single-adapter networks.

• Appendix D, Applications and HACMP, addresses some of the key issues to consider when making your applications highly available under HACMP.

Highlighting

The following highlighting conventions are used in this book:

Italic Identifies variables in command syntax, new terms and concepts, or indicates emphasis.

Bold Identifies routines, commands, keywords, files, directories, menu items, and other items whose actual names are predefined by the system.

Monospace Identifies examples of specific data values, examples of text similar to what you might see displayed, examples of program code similar to what you might write as a programmer, messages from the system, or information that you should actually type.

Planning Guide

Page 13: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

ISO 9000

ISO 9000 registered quality systems were used in the development and manufacturing of this product.

Related Publications

The following publications provide additional information about the HACMP for AIX software:

• Release Notes in /usr/lpp/cluster/doc/release_notes describe hardware and software requirements

• HACMP for AIX, Version 4.4: Concepts and Facilities, order number 86 A2 54KX 02

• HACMP for AIX, Version 4.4: Installation Guide, order number 86 A2 56KX 02

• HACMP for AIX, Version 4.4: Administration Guide, order number 86 A2 57KX 02

• HACMP for AIX, Version 4.4: Programming Locking Applications, order number 86 A2 59KX 02

• HACMP for AIX, Version 4.4: Programming Client Applications, order number 86 A2 60KX 02

• HACMP for AIX, Version 4.4: Enhanced Scalability Installation and Administration Guide Volumes I and II, order numbers 86 A2 62KX 02 and 86 A2 89KX 01

• HACMP for AIX, Version 4.4: Master Index and Glossary, order number 86 A2 65KX 02

The IBM AIX document set, as well as manuals accompanying machine and disk hardware, also provide relevant information.

The following manuals cover SSA disk subsystem hardware.

• 7133 Models 010 & 020 SSA Disk Subsystem: Installation Guide, Order Number GA33-3260

• 7133 Models 500 and 600 SSA Disk Subsystem: Installation Guide, 86 A1 93GX

• 7133 SSA Disk Subsystem: Operator Guide, Order Number 86 A1 90GX

• 7133 SSA Disk Subsystem: Service Guide, Order Number 86 A1 94GX

• 7133 SSA Disk Subsystems: Additional Installation and Service Information, Order Number 86 A1 69JX

• 7133 Hardware Technical Reference, Order Number 86 A1 91GX

The following manuals cover SSA adapter hardware.

• IBM SSA 4-Port Adapter: Installation and Reference, Order Number SC23-2775-00

• SSA Adapters User’s Guide and Maintenance Information, Order Number 86 A1 99GX

• SSA 4-Port Adapter Enhanced SSA 4-Port Adapter: Technical Reference, Order Number 86 A1 96GX

• Adapters, Devices, and Cable Information for Micro Channel Bus Systems, Order Number 86 A1 76AT

Planning Guide Preface xiii

Page 14: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

xiv

The following manual covers general SSA reference material. It includes a substantial discussion of high availability.

• A Practical Guide to Serial Storage Architecture for AIX, Order Number SG24-4599

The following manuals provide information about the IBM 2105 Versatile Storage Server.

• IBM Versatile Storage Server Introduction and Planning Guide, Order Number GC26-7223-01

• IBM Versatile Storage Server Host Systems Attachment Guide, Order Number SC26-7225-00.

• IBM Versatile Storage Server User’s Guide, Order Number SC26-7224-00.

• IBM Versatile Storage Server SCSI Command Reference 2105 Model B09, Order Number SC26-7226.

Ordering Publications

To order additional copies of this guide, use order number 86 A2 55HX 02.

Planning Guide

Page 15: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning an HACMP Cluster—OverviewDesign Goal: Eliminating Single Points of Failure

1

Chapter 1 Planning an HACMP Cluster—Overview

This chapter provides an overview of the recommended planning process and describes the paper and online cluster planning worksheets.

Design Goal: Eliminating Single Points of FailureThe HACMP for AIX software provides numerous facilities you can use to build highly available clusters. Designing the cluster that provides the best solution for your organization requires careful and thoughtful planning. In fact, adequate planning is the key to building a successful HACMP cluster. A well-planned cluster is easier to install, provides higher availability, performs better, and requires less maintenance than a poorly planned cluster.

Your major goal throughout the planning process is to eliminate single points of failure. A single point of failure exists when a critical cluster function is provided by a single component. If that component fails, the cluster has no other way of providing that function, and the service dependent on that component becomes unavailable.

For example, if all the data for a critical application resides on a single disk, and that disk fails, that disk becomes a single point of failure for the entire cluster. Clients cannot access that application until the data on the disk is restored. Likewise, if dynamic application data is stored on internal disks rather than on external disks, it is not possible to recover an application by having another cluster take over the external disks. Therefore, identifying necessary logical components required by an application, such as filesystems and directories (which could contain application data and configuration variables), is an important prerequisite for planning a successful cluster.

Realize that, while your goal is to eliminate all single points of failure, you may have to make some compromises. There is usually a cost associated with eliminating a single point of failure. For example, purchasing an additional hardware device to serve as backup for the primary device increases cost. The cost of eliminating a single point of failure should be compared against the cost of losing services should that component fail. Again, the purpose of the HACMP for AIX software is to provide a cost-effective, highly available computing platform that can grow to meet future processing demands.

Important: HACMP for AIX is designed to recover from a single hardware or software failure. It may not be able to handle multiple failures, depending on the sequence of failures. For example, the default event scripts cannot do an adapter swap after an IP address takeover (IPAT) has occurred if only one standby adapter exists for that network.

To be highly available, all cluster resources associated with a critical application should have no single points of failure. As you design an HACMP cluster, your goal is to identify and address all potential single points of failure. Questions to ask yourself include:

• What services are required to be highly available? What is the priority of these services?

• What is the cost of a failure compared to the necessary hardware to eliminate the possibility of this failure?

Planning Guide 1-1

Page 16: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning an HACMP Cluster—OverviewDesign Goal: Eliminating Single Points of Failure

1-2

• What is the required availability of these services? Do they need to be available 24 hours a day, seven days a week? Or is eight hours a day, five days a week sufficient?

• What could happen to disrupt the availability of these services?

• What is the allotted time for replacing a failed resource? What is an acceptable degree of performance degradation while operating after a failure?

• Which failures are detected as cluster events? Which failures need to have custom code written to detect the failure and trigger a cluster event?

• What is the skill level of the group implementing the cluster? And the group maintaining the cluster?

To plan, implement, and maintain a successful HACMP cluster requires continuing communication among many groups within your organization. Ideally, you should assemble representatives from the following areas (as applicable) to aid in HACMP planning sessions:

• Network administration

• System administration

• Database administration

• Application programming

• Support

• End users.

Eliminating Cluster Objects as Single Points of Failure

The table below summarizes potential single points of failure within an HACMP cluster and describes how to eliminate them.

See the HACMP for AIX Concepts and Facilities guide for an in-depth discussion on eliminating cluster objects as single points of failure.

Cluster Object Eliminated as Single Point of Failure By...

Node Using multiple nodes

Power source Using multiple circuits or uninterruptable power supplies

Network adapter Using redundant network adapters

Network Using multiple networks to connect nodes

TCP/IP subsystem Using serial networks to connect adjoining nodes and clients

Disk adapter Using redundant disk adapters

Controller Using redundant disk controllers

Disk Using redundant hardware and disk mirroring

Application Assigning a node for application takeover

Planning Guide

Page 17: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning an HACMP Cluster—OverviewCluster Planning Worksheets

Note: In an HACMP for AIX environment on an SP machine, the SP Switch adapter is a single point of failure and should be promoted to node failure. See the appendix on using the SP machine in the HACMP for AIX Installation Guide for complete information on the SP Switch adapter functionality.

Cluster Planning WorksheetsAt each step of the planning process, you can use planning worksheets to guide and organize your planning for each component of your cluster.

Both paper and online planning worksheets are provided with HACMP.

Paper Worksheets

The paper worksheets are located in Appendix A, Planning Worksheets. They are organized in the recommended sequence of planning steps. You make copies of these worksheets and write down all details of your cluster components in the appropriate spaces. You then refer to these sheets as you install and configure your cluster.

Paper worksheets can be useful when discussing cluster planning options with your team. They provide a written record of your initial cluster configuration decisions, which can help you trace problems in the future.

Online Worksheets

The online versions of the planning worksheets allow you to enter your information as you go. After you have completed all of the planning steps, HACMP will automatically apply the configuration.

Like the paper worksheets, the web-based online worksheet panels show the recommended sequence of steps you should take to plan your cluster. The user interface for the online worksheets has a different format than the paper worksheets, for example, in some cases an online panel consolidates information from several paper worksheets. Each worksheet format provides you with a framework to help you organize the appropriate information to configure each part of your cluster.

For instructions on installing and using the online planning worksheets, see Appendix B, Using the Online Cluster Planning Worksheet Program.

The Planning ProcessThis section describes the recommended steps you should follow to plan an HACMP cluster. As you plan a cluster, be aware that you will need to plan for application servers and resource groups within the cluster, and you will need to tailor event processing to allow the cluster to handle special failure situations.

Even if you are using the online planning worksheets, you should go over the information in this guide first, and continue to refer to it as you enter configuration data in the worksheets.

Planning Guide 1-3

Page 18: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning an HACMP Cluster—OverviewThe Planning Process

1-4

e and ode for

l view g steps luster

s

of the

lating

nning

ting to en es this

lating

Note: If you are planning to use one of the predefined Quick Configuration cluster configurations, you may want to skip directly to the appropriate section of the HACMP for AIX Installation Guide. However, it is still worth your while to familiarize yourself with all the issues involved in the planning process.

Steps 7 and 8 are not covered by the Quick Configuration utility, therefore you should be sure to read at least these even if you are using the Quick Configuration option.

Step 1: Draw a Cluster Diagram (Chapter 2)

In this step you plan the core of the cluster—the applications to be made highly availablthe types of resources they require, the number of nodes, shared IP addresses, and a msharing disks (non-concurrent or concurrent access). Your goal is to develop a high-leveof the system that serves as a starting point for the cluster design. In subsequent planninyou focus on specific subsystems, such as disks and networks. Chapter 2, Drawing the CDiagram, describes this step of the planning process.

Step 2: Plan Your TCP/IP Networks (Chapter 3)

In this step you plan the TCP/IP network support for the cluster. You first examine issuerelating to TCP/IP networks in an HACMP for AIX environment, and then complete the TCP/IP network worksheets. Chapter 3, Planning TCP/IP Networks, describes this stepplanning process.

Step 3: Plan Your Serial Networks (Chapter 4)

In this step you plan the serial network support for the cluster. You first examine issues reto serial networks in an HACMP for AIX environment, and then you complete the serial network worksheets. Chapter 4, Planning Serial Networks, describes this step of the plaprocess.

Step 4: Plan Your Shared Disk Devices (Chapter 5)

In this step you plan the shared disk devices for the cluster. You first examine issues reladifferent types of disk arrays and subsystems in an HACMP for AIX environment, and thdiagram the shared disk configuration. Chapter 5, Planning Shared Disk Devices, describstep of the planning process.

Step 5: Plan Your Shared LVM Components (Chapter 6)

In this step you plan the shared volume groups for the cluster. You first examine issues reto LVM components in an HACMP for AIX environment, and then you fill out worksheetsdescribing physical and logical storage. Chapter 6, Planning Shared LVM Components,describes this step of the planning process.

Planning Guide

Page 19: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning an HACMP Cluster—OverviewThe Planning Process

r

Step 6: Plan Your Application Servers and Resource Groups (Chapter 7)

In this step you plan a cluster around mission-critical applications, listing application server and resource group information specific to your HACMP cluster. Chapter 7, Planning Applications, Application Servers, and Resource Groups, describes this step of the planning process.

Step 7: Tailor the Cluster Event Processing (Chapter 8)

In this step you tailor the event processing for your cluster. Chapter 8, Tailoring Cluster Event Processing, describes this step of the planning process.

Step 8: Plan Your HACMP for AIX Clients (Chapter 9)

In this step you examine issues relating to HACMP for AIX clients. Chapter 9, Planning HACMP for AIX Clients, describes this step of the planning process.

Step 9: Installing and Configuring Your HACMP Cluster

After completing the planning steps, you are ready to install the cluster. Use the planning diagrams and worksheets you completed during the planning process to guide you through the installation process.

If you’ve used the online worksheet program, you can create an AIX file to configure youcluster for you. Instructions for this process are in Appendix B, Using the Online ClusterPlanning Worksheet Program.

See the HACMP for AIX Installation Guide to install and configure your HACMP cluster.

Planning Guide 1-5

Page 20: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning an HACMP Cluster—OverviewThe Planning Process

1-6

Planning Guide
Page 21: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Drawing the Cluster DiagramPrerequisites

llow y f you g

nning you t. You

2

Chapter 2 Drawing the Cluster Diagram

This chapter describes how to draw a cluster diagram.

PrerequisitesIt is essential that you understand the concepts and terminology necessary in planning an HACMP cluster. Read the HACMP for AIX Concepts and Facilities guide before beginning the planning process. The planning steps in this chapter assume a thorough understanding of the information presented in that guide.

OverviewTo develop a robust cluster design, you must progress through a series of steps. In each step, you add a piece to the cluster before moving forward to follow-on steps, which further define the cluster design. At the end of this process, you have designed a cluster that provides a high availability solution tailored to the particular needs of your organization.

In this chapter, you draw a diagram that shows the framework of the cluster. Your goal is to develop a high-level view of the cluster that becomes the starting point for the overall cluster design. In subsequent planning steps, you expand and refine the diagram by focusing on specific subsystems, such as disks and networks.

As you work through the planning process, you will be guided through a series of worksheets to help you organize your configuration information. The worksheets referred to in these chapters are paper worksheets that you fill out by hand and then refer to while configuring your cluster.

You may also use the online cluster planning “worksheets,” the web-based panels that ayou to enter configuration data as you plan, and then configure your cluster from there bcreating a file and transferring it from your PC-based system to your AIX cluster nodes. Iplan to use the online worksheets, refer to Appendix B, Using the Online Cluster PlanninWorksheet Program for instructions on the installation and use of this program.

Making a First Pass At Drawing the Cluster DiagramThe purpose of the cluster diagram is to combine the information from each step in the plaprocess into one drawing that shows the cluster’s function and structure. In this section,make a first pass at drawing the cluster diagram. Remember, this pass is a starting poinwill refine your diagram throughout the planning process.

Note: Even if you are using the online cluster planning worksheet program, in which you enter configuration data as you plan, it is highly recommended that you complete the entire planning process first, including drawing and refining a cluster diagram.

Planning Guide 2-1

Page 22: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Drawing the Cluster DiagramMaking a First Pass At Drawing the Cluster Diagram

2-2

The initial pass of the cluster diagram identifies:

• The cluster name and ID

• Applications to be made highly available and the types of resources used by the applications

• The number of nodes in the cluster

• IP addresses for the nodes

• The method of shared disk access.

An example of a first pass at a cluster diagram is shown in the following figure.

A First Pass of a Cluster Diagram

This diagram describes a two-node cluster that makes the Image Cataloger demo (an application supplied with the HACMP for AIX software) highly available. One node is named node A; the other node B. The cluster uses cascading resources and shares one external disk accessed by one node at a time.

Complete the following steps to begin the cluster diagram:

• Name the cluster and assign it an ID

• Identify the highly available applications and resource group types

• Select the number of cluster nodes

• Determine the IP addresses for client use (optional)

• Select the method of shared disk access.

Planning Guide

Page 23: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Drawing the Cluster DiagramMaking a First Pass At Drawing the Cluster Diagram

Naming the Cluster

Name the cluster and give it an ID. The cluster name is an arbitrary string of no more than 31 characters (alphanumeric and underscores only). The cluster ID can be any positive integer less than 99,999.

Important: Make sure that each cluster name and ID you define does not conflict with the names and IDs of other clusters at your site. See the HACMP for AIX Installation Guide for more information about defining cluster names and IDs.

As shown in the preceding diagram, the cluster is named bivalves; it has an ID of 1.

Identifying the Highly Available Applications and Resource Types

The purpose of the HACMP for AIX software is to ensure that critical applications and services are available. This guide presumes you have already identified these applications. For each application, you need to specify the type of resources the application uses, and a resource takeover strategy (based on the access mode chosen for each node).

Note: Some applications are integrated into HACMP and are configured differently than other applications. For information on planning for AIX Connections, AIX Fast Connect, and CS/AIX software, see Chapter 7, Planning Applications, Application Servers, and Resource Groups.

The HACMP for AIX software supports a maximum of up to 20 resource groups per cluster. Three types of resource groups are supported:

• Cascading, where a resource may be taken over by one or more nodes in a resource chain according to the takeover priority assigned to each node. The available node within the cluster with the highest priority will own the resource. You can also choose to set a flag so that a cascading resource will not fall back to a higher priority owner node when that node reintegrates into the cluster.

• Rotating, where a resource is associated with a group of nodes and rotates among these nodes. When a node fails, the next available node on its boot address and listed first in the resource chain will acquire the resource group. The takeover node may be currently active, or it may be in a standby state. When a detached node rejoins the cluster, however, it does not reacquire resource groups; instead, it rejoins as a standby.

• Concurrent access, where a resource that can be managed by the HACMP for AIX Cluster Lock Manager may be simultaneously shared among multiple applications residing on different nodes.

Keep the following considerations in mind when deciding which resource group type to assign to an application:

• If maximizing performance is essential, cascading resources may be the best resource group choice. Using cascading resources ensures that an application is owned by a particular node whenever that node is active in the cluster. This ownership allows the node to cache data the application uses frequently, thereby improving the node’s performance for providing data to the application.

If the active node fails, its resources will be taken over by the node with the highest priority in the resource chain.

Planning Guide 2-3

Page 24: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Drawing the Cluster DiagramMaking a First Pass At Drawing the Cluster Diagram

2-4

Note: When the failed node reintegrates into the cluster, it temporarily disrupts application availability as it takes back its resources from the takeover node, unless you have defined the resource groups to use cascading without fallback.

• If minimizing downtime is essential, rotating resources may be the best choice. Application availability is not disrupted during node reintegration because the reintegrating node rejoins the cluster as a standby node only and does not attempt to reacquire its resources.

The HACMP for AIX Concepts and Facilities guide provides more information about the various resource group types. It also presents a set of configurations that use different sample types of shared resources. Review these examples for more background on using the different resource group types in an HACMP cluster.

For each node, indicate the applications and their corresponding resource group type on the cluster diagram. Each application can be assigned to only one resource group. A single node, however, can support different resource group types. For applications using cascading resources, also specify the takeover priority for each node. The priority is determined by the order in which the nodes are listed in the resource chain.

For example, the cluster in the previous diagram makes the Image Cataloger application, which uses cascading resources, highly available. The node NodeA has the highest takeover priority in the cluster because its name is listed first in the resource chain. The node NodeB is listed next in the resource chain and therefore is designated as the next cluster node to take over resources in the event of a fallover. In this diagram, the node NodeA is first in the cluster to acquire resources, and the node NodeB is second.

Selecting the Number of Nodes

An HACMP cluster can have from two nodes to a maximum of eight nodes. Keep the following considerations in mind when determining the number of nodes in your cluster configuration:

• Clusters can have up to eight nodes. The number of nodes that can access a disk subsystem, however, varies based on the subsystem used. For example, the IBM 9333 serial disk drive subsystem can be accessed by up to eight nodes, whereas the IBM 7135-210 RAIDiant Disk Array can be accessed by up to four nodes.

• Reliable NFS server capability that allows a backup processor to recover current NFS activity should the primary NFS server fail, preserving the locks on NFS filesystems and dupcache is available for 2-node clusters only.

• Avoid grouping what could be smaller clusters into a single large cluster. Clusters that have entirely separate functions and do not share resources should not be combined in a single cluster. Several smaller clusters are easier to design, implement, and maintain than one large cluster.

• For performance reasons, it may be desirable to use multiple nodes to support the same application. To provide mutual takeover services, the application must be designed in a manner which allows multiple instances of the application to run on the same node.

For example, if an application requires that the dynamic data reside in a directory called /data, chances are that the application cannot support multiple instances on the same processor. For such an application running in a non-concurrent environment, try to partition the data so that multiple instances of the application can run, each accessing a unique database.

Planning Guide

Page 25: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Drawing the Cluster DiagramMaking a First Pass At Drawing the Cluster Diagram

ource, fail. If ess

. For

ses and

on

Furthermore, if the application supports configuration files that enable the administrator to specify that the dynamic data for instance1 of the application resides in the data1 directory, instance2 resides in the data2 directory, and so on, then multiple instances of the application are probably supported.

• In certain configurations, adding nodes to the cluster design can increase the level of availability the cluster provides; adding nodes also gives you more flexibility in planning node fallover and reintegration.

For the cluster diagram, draw a box representing each node in the cluster; then name each node. The node name can include alphabetic and numeric characters and underscores. Use no more than 31 characters. In the sample diagram, the nodes are named NodeA and NodeB; however, node names do not have to match the system’s hostname.

Assigning Cluster IP Addresses

The HACMP for AIX software lets you define an IP address as a cluster resource. As a resan IP address can be acquired by other cluster nodes should the node with this addressyou plan to use cascading resource groups, you need to plan on configuring for IP addrtakeover. For rotating resource groups, you need to designate an IP address that will bedynamically shared by the nodes in the resource chain.

One of the goals in planning the cluster is to ensure client access to a known IP addresseach type of resource group, there is a strategy for maintaining this connection.

Chapter 3, Planning TCP/IP Networks, discusses issues relating to assigning IP addreshardware address swapping.

Selecting the Method of Shared Disk Access

The HACMP for AIX software supports two methods (modes) of accessing applications shared external disk devices:

• Non-concurrent access, where only one node has access to a shared external disk at a given time. If this node fails, one of the peer nodes must take over the disk and restart applications to restore critical services to clients. Typically, takeover occurs within 30 to 300 seconds, but this range depends on the number and types of disks being used, the number of volume groups, the filesystems (whether shared or cross-mounted), and the number of critical applications in the cluster configuration.

• Concurrent access, where from two to eight processors can simultaneously access an application residing on a shared external disk, thus offering near continuous availability of resources.

If required to support a given set of applications, you can assign both concurrent and non-concurrent disks to a node.

The HACMP for AIX Concepts and Facilities guide describes shared disk access in depth. It also presents a sample set of configurations that use different shared disk access methods. Review these examples for more information on using shared disk access methods in an HACMP cluster.

Planning Guide 2-5

Page 26: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Drawing the Cluster DiagramWhere You Go From Here

2-6

pter

For the cluster diagram, draw a box representing each shared disk; then label each box with a shared disk name. Next, write descriptions for applications that will access the shared disks. Finally, indicate whether the shared disk will be accessed in concurrent or non-concurrent mode.

For example, the cluster in the sample diagram has a single shared external disk named shared_disk1. This shared disk is accessed in non-concurrent mode by the Image Cataloger application.

Chapter 5, Planning Shared Disk Devices, and Chapter 6, Planning Shared LVM Components, provide more information on accessing shared disks in HACMP clusters.

Where You Go From HereAt this point, you should have a diagram similar to the sample diagram shown in this chapter. In subsequent planning steps, you will expand and refine the diagram by focusing on specific subsystems. Next, you will plan the cluster’s TCP/IP network topology, described in Cha3, Planning TCP/IP Networks.

Planning Guide

Page 27: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksPrerequisites

pters.

3 ‘, , , , , um m mm mm hjjm mm m mm mm mm jum mm mm mm mm m mm

Chapter 3 Planning TCP/IP Networks

This chapter describes planning TCP/IP network support for an HACMP cluster.

PrerequisitesIn Chapter 2, Drawing the Cluster Diagram, you began to plan your cluster by drawing a cluster diagram. This diagram is the starting point for the follow-on planning you do in this chapter. In particular, you must have decided whether or not the cluster will use IP address takeover to maintain specific IP addresses.

OverviewIn this chapter, you plan the TCP/IP networking support for the cluster, including:

• Deciding which types of networks and point-to-point connections to use in the cluster.

• Designing the network topology.

• Defining a network mask for your site.

• Defining IP addresses (adapter identifiers) for each node’s service and standby ada

• Defining a boot address for each service adapter that can be taken over, if you are using IP address takeover or rotating resources.

• Defining an alternate hardware address for each service adapter that can have its IP address taken over, if you are using hardware address swapping.

• Tuning the cluster for HACMP performance

After addressing these issues, add the network topology to the cluster diagram. Next complete the TCP/IP Networks Worksheet and/or fill in network information on the appropriate online worksheet panel.

Selecting Public and Private NetworksIn the HACMP for AIX environment, a public network connects multiple nodes and allows clients to access these nodes. A private network is a point-to-point connection that links two or more nodes directly.

As an independent, layered component of AIX, the HACMP for AIX software works with most TCP/IP-based networks. HACMP for AIX has been tested with standard Ethernet interfaces (en*) but not with IEEE 802.3 Ethernet interfaces (et*), where * reflects the interface number. HACMP for AIX also has been tested with Token-Ring and Fiber Distributed Data Interchange (FDDI) networks, with IBM Serial Optical Channel Converter (SOCC), Serial Line Internet Protocol (SLIP), and Asynchronous Transfer Mode (ATM) point-to-point connections. HACMP for AIX has been tested with both the Classical IP and LAN Emulation ATM protocols.

Planning Guide 3-1

Page 28: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksDesigning a Network Topology

3-2

Note: ATM and SP Switch networks are special cases of point-to-point, private networks that can connect clients.

See the documentation specific to your network type for detailed descriptions of its features.

Network Used for Cluster Lock Manager Traffic

The Cluster Lock Manager (CLM) selects a private network in preference to a public network for its lock traffic. If no private network is defined, the lock manager chooses a public network. If more than one private or public network exists, the lock manager chooses randomly.

Once the lock manager chooses a network during startup, it continues to use that network, even if another network becomes available. It only switches networks if the network it is using fails.

Designing a Network Topology In the HACMP for AIX environment, the network topology is the combination of networks and point-to-point connections that link cluster nodes and clients.

The HACMP for AIX software supports a maximum of 32 networks per cluster and 24 TCP/IP network adapters on each node. These numbers provide a great deal of flexibility in designing a network configuration. The design affects the degree of system availability such that the more communication paths that connect clustered nodes and clients, the greater the degree of network availability.

When designing your network topology, you must determine the number and types of:

• Networks and point-to-point connections that connect nodes

• Network adapters connected to each node.

Sample Network Topologies

Examples of network topologies for an HACMP cluster are shown throughout the following pages.

Planning Guide

Page 29: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksDesigning a Network Topology

Single Network

In a single-network setup, each node in the cluster is connected to just one network and has only one service adapter available to clients. In this setup, a service adapter on any of the nodes may fail, and a standby adapter will acquire its IP address. The network itself, however, is a single point of failure. The following figure shows a single-network configuration.

Single-Network Setup

In the single-network setup, each node is connected to one network.Each node has one service adapter and can have none, one, or morestandby adapters per public network.

Network

Client

Planning Guide 3-3

Page 30: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksDesigning a Network Topology

3-4

both

le to

Dual Network

A dual-network setup has two separate networks for communication. Nodes are connected to two networks, and each node has two service adapters available to clients. If one network fails, the remaining network can still function, connecting nodes and providing resource access to clients.

In some recovery situations, a node connected to two networks may route network packets from one network to another. In normal cluster activity, however, each network is separate—logically and physically.

Keep in mind that a client, unless it is connected to more than one network, is susceptibnetwork failure.

The following figure shows a dual-network setup.

Dual-Network Setup

Planning Guide

Page 31: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksDesigning a Network Topology

Point-to-Point Connection

A point-to-point connection links two (neighbor) cluster nodes directly. SOCC, SLIP, and ATM are point-to-point connection types. In HACMP clusters of four or more nodes, however, use a SOCC line only as a private network between neighbor nodes because it cannot guarantee cluster communications with nodes other than its neighbors.

The following diagram shows a cluster consisting of two nodes and a client. A single public network connects the nodes and the client, and the nodes are linked point-to-point by a private high-speed SOCC connection that provides an alternate path for cluster and lock traffic should the public network fail.

A Point-to-Point Connection

Planning Guide 3-5

Page 32: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksIdentifying Network Components

3-6

ecause ion of

with that

have a

he

, or

belan r, should

Identifying Network ComponentsThis section explains terminology you must understand to complete the network worksheets referenced at the end of this chapter.

Nodes

A node is a RS/6000 uniprocessor, an SP node, or an SMP system unit that runs the server portion of the HACMP for AIX software in a cluster environment. Within an HACMP cluster environment, a node is identified by a unique node name.

In a non-HACMP for AIX environment, however, a node typically is identified by a hostname that is associated with a network interface (adapter label). If a node has only one network interface, the hostname usually also uniquely identifies the node. But in an HACMP cluster environment, a node typically can have more than one network interface, and this hostname-to-network interface association does not uniquely identify the node. Instead, an IP address association with node adapters uniquely identifies the host. At any given time, the hostname corresponds to only one of the node’s IP labels.

A node’s name should match the adapter label of the primary network’s service adapter bother applications may depend on this hostname. See the following section for the definitan adapter label. The HACMP for AIX Installation Guide provides instructions for setting thenode name.

Network Adapters

A network adapter (interface) connects a node to a network. A node typically is configuredat least two network interfaces for each network to which it connects: a service interfacehandles cluster traffic, and one or more standby interfaces. A service adapter must alsoboot address defined for it if IP address takeover is enabled.

Adapters in an HACMP cluster have a label and a function (service, standby, or boot). Tmaximum number of network interfaces per node is 24.

When configuring the SP with multiple networks and Enhanced Security, you should, toeliminate single points of failure, configure Kerberos service principals (godm, rcmd) on more than one, and preferably all, networks you plan to configure in your HACMP cluster environment. You can do this either at initial setup and installation of Kerberos on the SPlater when you are customizing the nodes. See the HACMP for AIX Installation Guide for details about how to configure an HACMP cluster for Kerberos.

Adapter Label

A network adapter is identified by an adapter label. For TCP/IP networks, the adapter la is the name in the /etc/hosts file associated with a specific IP address. Thus, a single node chave several adapter labels and IP addresses assigned to it. The adapter labels, howevenot be confused with the “hostname.”

Planning Guide

Page 33: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksIdentifying Network Components

an

to be

role

has s used kes

ger adapter apter,

. Your y

The following example entries show that nodea has two Ethernet adapters, labeled svc and stdby. The adapter labels reflect separate network interface functions, each associated with a unique IP address:

100.100.50.1 nodea_svc100.100.51.1 nodea_stdby

For boot adapters, you can simply add the suffix “boot” to the node name, as in nodea_boot. For more information on adapter functions, see the following section.

When deciding on an adapter label, keep in mind that the adapter label also can reflect adapter interface name. For example, one interface can be labeled nodea_en0 and the other labeled nodea_en1, where en0 and en1 indicate the separate adapter names.

Whichever naming convention you use for adapter labels in an HACMP cluster, be sureconsistent.

Adapter Function

In the HACMP for AIX environment, each adapter has a specific function that indicates theit performs in the cluster. An adapter’s function can be service, standby, or boot.

Service Adapter The service adapter is the primary connection between the node and the network. A nodeone service adapter for each physical network to which it connects. The service adapter ifor general TCP/IP traffic and is the address the Cluster Information Program (Clinfo) maknown to application programs that want to monitor or use cluster services.

Note: In configurations using rotating resources, the service adapter on the standby node remains on its boot address until it assumes the shared IP address. Consequently, Clinfo makes known the boot address for this adapter.

Note: In an HACMP for AIX environment on the RS/6000 SP, the Ethernet adapters can be configured as service adapters but should not be configured for IP address takeover. For the SP switch network, service addresses used for IP address takeover are ifconfig alias addresses used on the css0 network.

Note: In configurations using the Classical IP form of the ATM protocol (i.e. not ATM LAN Emulation), a maximum of 7 service adapters per cluster is allowed if hardware address swapping is enabled.

Standby Adapter A standby adapter backs up a service adapter. If a service adapter fails, the Cluster Manaswaps the standby adapter’s address with the service adapter’s address. Using a standbyeliminates a network adapter as a single point of failure. A node can have no standby ador it can have from one to seven standby adapters for each network to which it connectssoftware configuration and hardware constraints determine the actual number of standbadapters that a node can support.

Planning Guide 3-7

Page 34: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksIdentifying Network Components

3-8

ld fail, ing the

over

rvice arted, it ddress

e wever,

. The face.

Note: In an HACMP for AIX environment on the RS/6000 SP, for an IP address takeover configuration using the SP switch, standby adapters are not used.

Boot Adapter (Address) IP address takeover is an AIX facility that allows one node to acquire the network address of another node in the cluster. To enable IP address takeover, a boot adapter label (address) must be assigned to the service adapter on each cluster node. Nodes use the boot label after a system reboot and before the HACMP for AIX software is started.

Note: In an HACMP for AIX environment on the RS/6000 SP, boot addresses used in IP address takeover are ifconfig alias addresses used on the css0 network.

When the HACMP for AIX software is started on a node, the node’s service adapter is reconfigured to use the service label (address) instead of the boot label. If the node shoua takeover node acquires the failed node’s service address on its standby adapter, makfailure transparent to clients using that specific service address.

During the reintegration of the failed node, which comes up on its boot address, the takenode will release the service address it acquired from the failed node. Afterwards, the reintegrating node will reconfigure its boot address to its reacquired service address.

Consider the following scenario: Suppose that Node A fails. Node B acquires Node A’s seaddress and services client requests directed to that address. Later, when Node A is restcomes up on its boot address and attempts to reintegrate into the cluster on its service aby requesting that Node B release Node A’s service address. When Node B releases threquested address, Node A reclaims it and reintegrates into the cluster. Reintegration, hofails if Node A has not been configured to boot using its boot address.

Note: The boot address does not use a separate physical adapter, but instead is a second name and IP address associated with a service adapter. It must be on the same subnetwork as the service adapter. All cluster nodes must have this entry in the local /etc/hosts file and, if applicable, in the nameserver configuration.

Network Interfaces

The network interface is the network-specific software that controls the network adapterinterface name is a three- or four-character string that uniquely identifies a network interThe first two or three characters identify the network protocol. For example, en indicates a standard Ethernet network.

Network interfaces and their character identifiers are shown in the following table:

Interface Identifier

Standard Ethernet en

Token-Ring tr

SLIP sl

Planning Guide

Page 35: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksIdentifying Network Components

ts and ver,

r

k,

The next character is the number AIX assigns the device. For example, 0. An interface name of en0 indicates that this is the first standard Ethernet interface on the system unit.

Networks

Networks in an HACMP cluster are identified by a name and an attribute.

Network Name

The network name is a symbolic value that identifies a network in an HACMP for AIX environment. Cluster processes use this information to determine which adapters are connected to the same physical network. The network name is arbitrary, in most cases, and it must be used consistently. If several adapters share the same physical network, make sure that you use the same network name when defining these adapters.

Network Attribute

A TCP/IP network’s attribute is either public or private.

FDDI fi

SOCC so

SP Switch css

ATM at

Public A public network connects from two to eight nodes and allows cliento monitor or access cluster nodes. Ethernet, Token-Ring, FDDI, SLIP are considered public networks. Note that a SLIP line, howedoes not provide client access.

Private A private network provides point-to-point communication betweentwo nodes; it typically does not allow client access. A SOCC line oan ATM network also are private networks; however, an ATM network does allow client connections and may contain standby adapters. If an SP node is used as a client, the SP Switch networalthough private, can allow client access.

Interface Identifier

Planning Guide 3-9

Page 36: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksDefining a Network Mask

3-1

ice of

Defining a Network Mask The HACMP for AIX software uses the subnet feature of TCP/IP to divide a single physical network into separate logical subnets. In order to use subnets, you must define a network mask for your system.

An IP address consists of 32 bits. Some of the bits form the network address; the remainder form the host address. The network mask (or netmask) determines which bits in the IP address refer to the network and which bits refer to the host, as shown in the following example:

In the preceding figure, the netmask is shown in both dotted decimal and binary format. A binary 1 indicates that a bit is part of the network address. A binary 0 indicates that a bit is part of the host address. In the preceding example, the network portion of the address occupies 24 bits; the host portion occupies 8 bits. It is convenient (but not necessary) to define a subnet on an octet boundary.

See the AIX Version 4 System Management Guide: Communications and Networks manual for more information about classes of addresses. Also, ask your network administrator about the class and subnets used at your site.

Defining IP Addresses for Standby Adapters The design of the HACMP for AIX software specifies that:

• All client traffic be carried over the service adapter.

• Standby adapters be hidden from client applications and carry only internal Cluster Manager traffic.

To comply with these rules, pay careful attention to the IP addresses you assign to standby adapters. Standby adapters must be on a separate subnet from the service adapters, even though they are on the same physical network. Placing standby adapters on a different subnet from the service adapter allows HACMP for AIX to determine which adapter TCP/IP will use to send a packet to a network.

If there is more than one adapter on the same subnet with the same network address, there is no way to guarantee which of these adapters will be chosen by the IP as the transmission route. All choices will be correct, since each choice will deliver the packet to the correct network. To guarantee that only the service adapter handles critical traffic, you must limit the IP’s cho

0 Planning Guide

Page 37: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksDefining IP Addresses for Standby Adapters

of a

dress

r ts are

r client mit the h either

50.112 the

you

a transmission route to one adapter. This keeps all traffic off the standby adapter so that it is available for adapter swapping and IP address takeover (IPAT). Limiting the IP’s choicetransmission route also facilitates identifying an adapter failure.

Note: The netmask for all adapters in an HACMP network must be the same even though the service and standby adapters are on different logical subnets. See the HACMP for AIX Concepts and Facilities guide for more information about using the same netmask for all adapters.

Placing Standby Adapters on a Separate Subnet

To place a standby adapter on a different subnet from a service adapter, give it an IP adthat has a different network address portion.

Consider the following adapter IP address and network mask:

IP address:

100.100.50.121 01100100 01100100 00110010 01111001

Netmask:

255.255.255.0 11111111 11111111 11111111 00000000

In this example, the network address is 100.100.50. The host address is 121. An adapteconfigured in this way can transmit packets destined to a network whose first three octe100.100.50.

Now consider a node with the following two network adapters:

Adapter 1’s IP address:

100.100.50.121 01100100 01100100 00110010 01111001

Adapter 2’s IP address:

100.100.50.25 01100100 01100100 00110010 00011001

Netmask:

255.255.255.0 11111111 11111111 11111111 00000000

In this example, suppose that a client program on this node wanted to send data to anotheat address 100.100.50.112. You cannot predict whether adapter 1 or adapter 2 will transdatagram, because the destination IP address (100.100.50.112) can be reached througadapter.

Now consider a node with the following configuration:

Adapter 1’s IP address:

100.100.50.121 01100100 01100100 00110010 01111001

Adapter 2’s IP address:

100.100.51.25 01100100 01100100 00110011 00011001

Netmask:

255.255.255.0 11111111 11111111 11111111 00000000

In this case, you can determine that the data destined for the client at address 100.100.will be sent through adapter 1, since it is the only candidate that can send the packet to network with address 100.100.50.

Remember, pay careful attention to the IP addresses you assign to standby adapters ascomplete the networks worksheet referenced at the end of this chapter.

Planning Guide 3-11

Page 38: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksDefining IP Addresses for Standby Adapters

3-1

the

rvice,

subnet, dby

Note: Although standby adapters are on the same physical network (and must be on the same netmask) as service adapters, standby adapters must be on a different logical subnet from the service adapters. Different HACMP networks may use different netmasks, but the netmask must be the same for all adapters within a given HACMP network.

If you configure multiple standby adapters on cluster nodes, they all must be configured on the same subnet to handle cluster node failures. In addition, keep in mind that with multiple standby adapters configured, a swap_adapter event on the standby adapter routing heartbeats on a node may cause all standbys on the node to appear to fail, since only one heartbeat route exists per node for the standbys on that node.

Subnet Considerations for Cluster Monitoring with Tivoli

If you plan to monitor an HACMP node with your Tivoli management software, and you do not have a dedicated network for the Tivoli Management Region (TMR), you must create an IP address alias in order to ensure the proper functioning of IP address takeover. You must place this alias in the /etc/hosts file.

The subnet of this alias must be different than the node’s service and standby adapters, andsame as the subnet of the TMR (server) node.

Here is an example of what you might insert into the /etc/hosts file for a Tivoli-monitored cluster node named HAnode and a Tivoli server node named TMRnode. HAnode has sestandby, and alias IP addresses; TMRnode simply has a service IP address.

The netmask for this network is 255.255.255.0

You can see in this example that the alias address and the TMR address are on the sameand this subnet is in addition to the two already used for the cluster node’s service and stanadapters.

Adapter Label Address

HAnode_svc 10.10.20.88

HAnode_stby 10.50.25.88

HAnode_alias 10.50.21.89

TMRnode 10.50.21.10

2 Planning Guide

Page 39: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksDefining Boot Addresses

Defining Boot Addresses When using IP address takeover, you must define a boot address for each service adapter on which IP address takeover might occur. To define a boot address, see the HACMP for AIX Administration Guide.

The boot address and the service address must be on the same subnet. That is, the two addresses must have the same value for the network portion of the address; the host portion must be different.

Use a standard formula for assigning boot addresses. For example, the boot address could be the host address plus 64. This formula yields the following boot addresses:

NODE A service address: 100.100.50.135NODE A boot address: 100.100.50.199Network mask: 255.255.255.0NODE B service address: 100.100.50.136NODE B boot address: 100.100.50.200Network mask: 255.255.255.0

Boot/Service/Standby Address Requirements for Resource Groups

The following charts specify the boot, service, and standby address requirements for each resource group configuration in a two- and three-node cluster.

Two-Node Cluster

Resource Groups Boot Service Standby

Cascading without IP address takeover

(same for Cascading without Fallback)

none required 2 2

Cascading with IP address takeover

(same for Cascading Without Fallback)

1 per highly available service address

1 per highly available client connection

2

Rotating 2 1 per node per network minus 1

2

Concurrent not applicable not applicable not applicable

Planning Guide 3-13

Page 40: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksDefining Hardware Addresses

3-1

Three-Node Cluster

Defining Hardware Addresses

Note: You cannot use hardware address swapping on the SP Ethernet or SP Switch networks.

The hardware address swapping facility works in tandem with IP address takeover. Hardware address swapping maintains the binding between an IP address and a hardware address, which eliminates the need to flush the ARP cache of clients after an IP address takeover. This facility, however, is supported for Ethernet, Token-Ring, FDDI, and ATM adapters. It does not work with the SP Switch.

Note that hardware address swapping takes about 60 seconds on a Token-Ring network, and up to 120 seconds on a FDDI network. These periods are longer than the usual time it takes for the Cluster Manager to detect a failure and take action. Selecting an Alternate Hardware Address

This section provides hardware addressing recommendations for Ethernet, Token Ring, FDDI, and ATM adapters. Note that any alternate hardware address you define for an adapter should be similar in form to the default hardware address the manufacturer assigned to the adapter.

To determine an adapter’s default hardware address, use the netstat -i command (when the networks are active).

Using netstatTo retrieve hardware addresses using the netstat -i command, enter:

netstat -i | grep link

which returns output similar to the following (leading 0s are suppressed):

lo0 16896 link#1 186303 0 186309 0 0en0 1500 link#2 2.60.8c.2f.bb.93 2925 0 1047 0 0tr0 1492 link#3 10.0.5a.a8.b5.7b 104544 0 92158 0 0

Resource Groups Boot Service Standby

Cascading without IP address takeover

(same for Cascading without Fallback)

none required 3 3

Cascading with IP address takeover

(same for Cascading Without Fallback)

1 per highly available service address

1 per highly available client connection

3

Rotating 3 1 per node per network minus 1

3

Concurrent not applicable not applicable not applicable

4 Planning Guide

Page 41: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksDefining Hardware Addresses

new

output,

tr1 1492 link#4 10.0.5a.a8.8d.79 79517 0 39130 0 0fi0 4352 link#5 10.0.5a.b8.89.4f 40221 0 1 1 0fi1 4352 link#6 10.0.5a.b8.8b.f4 40338 0 6 1 0at0 9180 link#7 8.0.5a.99.83.57 54320 0 8 1 0at2 9180 link#8 8.0.46.22.26.12 54320 0 8 1 0

Specifying an Alternate Ethernet Hardware Address

To specify an alternate hardware address for an Ethernet interface, begin by using the first five pairs of alphanumeric characters as they appear in the current hardware address. Then substitute a different value for the last pair of characters. Use characters that do not occur on any other adapter on the physical network.

For example, you could use 10 and 20 for node A and node B, respectively. If you have multiple adapters for hardware address swapping in each node, you can extend to 11 and 12 on node A, and 21 and 22 on node B.

Specifying an alternate hardware address for adapter interface en0 in the preceding output thus yields the following value:

To define this alternate hardware address to the cluster environment, see the HACMP for AIX Installation Guide.

Specifying an Alternate Token-Ring Hardware Address

To specify an alternate hardware address for a Token-Ring interface, set the first two digits to 42, indicating that the address is set locally.

Specifying an alternate hardware address for adapter interface tr0 in the preceding output thus yields the following value:

To define this alternate hardware address to the cluster environment, see the HACMP for AIX Installation Guide.

Specifying an Alternate FDDI Hardware Address

To specify an alternate FDDI hardware address, enter the new address into the Adapter Hardware Address field as follows, without any decimal separators:

1. Use 4, 5, 6, or 7 as the first digit (the first nibble of the first byte) of the new address.

2. Use the last 6 octets of the manufacturer’s default address as the last 6 digits of theaddress.

Here’s a list of some sample alternate addresses for adapter interface fi0 in the precedingshown with decimals for readability:

40.00.00.b8.89.4f40.00.01.b8.89.4f

Original address 02608c2fbb93

New address 02608c2fbb10

Original address 10005aa8b57b

New address 42005aa8b57b

Planning Guide 3-15

Page 42: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksDefining Hardware Addresses

3-1

ing

from

50.00.00.b8.89.4f60.00.00.b8.89.4f7f.ff.ff.b8.89.4f

Specifying an Alternate ATM Hardware Address

The following procedure applies to ATM Classic IP interface only. Hardware address swapping for ATM LAN Emulation adapters works just like hardware address swapping for the Ethernet and Token-Ring adapters that are being emulated.

Note: An ATM adapter has a hardware address which is 20 bytes in length. The first 13 bytes are assigned by the ATM switch, the next 6 bytes are burned into the ATM adapter, and the last byte represents the interface number (known as the selector byte). The above example only shows the burned in 6 bytes of the address. To select an alternate hardware address, you replace the 6 burned in bytes, and keep the last selector byte. The alternate ATM adapter hardware address is a total of 7 bytes.

To specify an alternate hardware address for an ATM Classic IP interface:

1. Use a value in the range of 40.00.00.00.00.00 to 7f.ff.ff.ff.ff.ff for the first 6 bytes.

2. Use the interface number as the last byte.

Here’s a list of some sample alternate addresses for adapter interface at2 in the precedoutput, shown with decimals for readability:

40.00.00.00.00.00.0240.00.01.00.00.00.0250.00.00.00.01.00.0260.00.00.01.00.00.027f.ff.ff.ff.ff.ff.02

Since the interface number is hard-coded into the ATM hardware address, it must moveone ATM adapter to another during hardware address swapping. This imposes certain requirements and limitations on the HACMP for AIX configuration.

HACMP Configuration Requirements for ATM Hardware Address Swapping (Classical IP Only)

• If the hardware address moves to another adapter on the same machine (adapter swapping), the interface will have to be configured on that adapter as well. Likewise, when IP address takeover occurs, the interface associated with the adapter on the remote node will need to be configured on the takeover node.

• There can be no more than 7 ATM service adapters per cluster that support hardware address swapping.

• Each of these service interfaces must have a unique ATM interface number.

• On nodes that have one standby adapter, the standby adapters on all cluster nodes will use the eighth possible ATM interface (at7), so that there is no conflict with the service interface used by any of the nodes. This will guarantee that during IP address takeover with hardware address swapping the interface associated with the hardware address is not already in use on the takeover node.

• If any node has more than one standby adapter, the total number of available service interfaces is reduced by that same number. For example, if two nodes have two standby adapters each, then the total number of service interfaces is reduced to 5.

6 Planning Guide

Page 43: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksDefining Hardware Addresses

• Any ATM adapters that are not being used by HACMP for AIX, but are still configured on any of the cluster nodes that are performing ATM hardware address swapping, will also reduce the number of available ATM interfaces on a one-for-one basis.

Network Configuration Requirements for ATM Hardware Address Swapping (Classical IP and LAN Emulation)

• Hardware address swapping for ATM requires that all adapters that can be taken over for a given service address be attached to the same ATM switch.

Avoiding Network Conflicts

Each network adapter is assigned a unique hardware address when it is manufactured, which ensures that no two adapters have the same network address. When defining a hardware address for a network adapter, ensure that the defined address does not conflict with the hardware address of another network adapter in the network. Two network adapters with a common hardware address can cause unpredictable behavior within the network.

Afterwards, to confirm that no duplicate addresses exist on your network, bring the cluster up on your new address and ping it from another machine. If you receive two packets for each ping (one with a trailing DUP!), you have probably selected an address already in use. Select another address and try again. Cycle the Cluster Manager when performing these operations because the alternate address is used only after the HACMP for AIX software is running and has reconfigured the adapter to use the service IP address the alternate hardware address associates with it.

Planning Guide 3-17

Page 44: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksUsing HACMP with NIS and DNS

3-1

Using HACMP with NIS and DNSHACMP facilitates communication between the nodes of a cluster so that each node can determine whether the designated services and resources are available. Resources can include, but are not limited to, IP addresses and names and storage disks.

Subnetting the service and standby adapters (using TCP/IP) and having at least two separate networks (Ethernet and RS232) permits HACMP to determine whether a communication problem exists in the adapter, the network, or another common component such as TCP/IP software itself. The ability to subnet the adapters ensures that HACMP can direct the keepalive traffic from service to standby adapters, standby to service adapters, standby to standby adapters, and so on. For example, if there are two nodes in a cluster, and both standby and service adapters of Node A can receive and send to the standby adapter of node B, but cannot communicate with the service adapter of Node B, then it can be assumed that the service adapter is not working properly. HACMP recognizes this and performs the swap between the service and standby adapters.

Some of the commands used to perform the swap require IP lookup. This defaults to a nameserver for resolution if NIS or DNS is operational. If the nameserver was accessed via the adapter that is down, the request will time-out. To ensure that the cluster event (in this case an adapter swap) completes successfully and quickly, HACMP disables NIS or DNS hostname resolution. It is therefore required that the nodes participating in the cluster have entries in the /etc/hosts file.

How HACMP Enables and Disables Nameserving

This section provides some additional details on the logic a system uses to perform hostname resolution and how HACMP enables and disables DNS and NIS.

If a node is using either domain nameserving or NIS, the hostname is normally resolved by contacting a suitable server. At best, this causes a time delay; at worst, no response is ever returned because communication with the server is lost. For example, if NIS alone is running, hostname resolution is performed via the following logic in AIX:

8 Planning Guide

Page 45: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksUsing HACMP with NIS and DNS

Note: The following applies if the NIS configuration maps include the host map.

As shown, if the NIS host tables were built with the -b flag, the NIS Server continues its search via domain nameserving if that option is available. The key point, however, is that under certain circumstances (for example, an adapter being down, and the NIS server being unavailable), hostname resolution should be localized for quick response time. This infers that the local /etc/hosts file should be consulted, and the systems that may be involved in a cluster reconfiguration must be placed within the file in addition to loopback and the local hostname. It also means that the client portion of NIS that is running on that machine must be turned off. This is the case if the cluster node is an NIS client regardless of its status as a server (master or slave). Remember, even an NIS server uses the ypbind to contact the ypserv daemon running either on the master server or a slave server to perform the lookup within the current set of NIS maps.

Planning Guide 3-19

Page 46: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksUsing HACMP with NIS and DNS

3-2

P’s lity of ediate

ipts

, the m (using

Similarly, the logic followed by DNS (AIX) is:

In this situation, if both DNS and NIS were not disabled, a response to a hostname request might be as long as the time required for both NIS and DNS to time-out. This would hinder HACMsystem reconfiguration, and increase the takeover time required to maintain the availabidesignated resources (IP address). Disabling these services is the only reliable and immmethod of expediting the communication changes HACMP is attempting to accomplish.

The method HACMP uses to cleanly start and stop NIS and DNS is found within the scrcl_nm_nis_on and cl_nm_nis_off within the /usr/sbin/cluster/events/utils directory.

For NIS, checking the process table for the ypbind daemon lets HACMP know that NIS is running, and should be stopped. A check for the existence of the file /usr/sbin/cluster/hacmp_stopped_ypbind lets HACMP know that NIS client services mustbe restarted following the appropriate cluster configuration events. As mentioned earliercommands startsrc -s ypbind and stopsrc -s ypbind are used to start and stop this node frousing NIS name resolution. This is effective whether the node is a master or slave serverNIS client services), or simply an NIS client.

0 Planning Guide

Page 47: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksPlanning for Cluster Performance

ct the nd work

nager tted. ause

state ire the he cluster

noted

HACMP uses the AIX command namerslv to stop and start DNS on the cluster node. After checking for the existence of an /etc/resolv.conf file (the method for determining if domain name resolution is in effect), HACMP uses the namerslv -E command to stop nameserving by moving the /etc/resolv.conf file to a specified file. After reconfiguration, HACMP runs the namerslv -B command to restore the domain name configuration file from the specified file. Using this method, it is not necessary to stop and restart the named daemon.

Note: Using HACMP for AIX with nameserving requires that the “Host Uses NIS or Name Server” Run Time Parameter be set to True. For more information on Run Time Parameters, see the HACMP for AIX Installation Guide.

Planning for Cluster PerformanceHACMP 4.4 provides easier and greater control over several tuning parameters that affecluster’s performance. Setting these tuning parameters correctly to ensure throughput aadjusting the HACMP failure detection rate can help avoid “failures” caused by heavy nettraffic.

Cluster nodes sometimes experience extreme performance problems, such as large I/Otransfers, excessive error logging, or lack of memory. When this happens, the Cluster Macan be starved for CPU time. It might not reset the “deadman switch” within the time alloMisbehaved applications running at a priority higher than the cluster manager can also cthis problem.

The “deadman switch” is the AIX kernel extension that halts a node when it enters a hungthat extends beyond a certain time limit. This enables another node in the cluster to acquhung node’s resources in an orderly fashion, avoiding possible contention problems. If tdeadman switch is not reset in time, it can cause a system panic and dump under certainconditions.

Setting these tuning parameters correctly may avoid some of the performance problemsabove. You can set these parameters using HACMP SMIT screens:

• High and low watermarks for I/O pacing

• syncd frequency rate

• HACMP Failure Detection Rate (Custom)

• HACMP cycles to failure

• HACMP heartbeat rate.

Setting I/O Pacing

AIX users have occasionally seen poor interactive performance from some applications when another application on the system is doing heavy input/output. Under certain conditions I/O can take several seconds to complete. While the heavy I/O is occurring, an interactive process can be severely affected if its I/O is blocked or if it needs resources held by a blocked process.

Planning Guide 3-21

Page 48: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksPlanning for Cluster Performance

3-2

e I/O c with

e ess e

es. If

itial

ior

ed.

is ld not

ster

he tions. nd for ting

s:

Under these conditions, the HACMP for AIX software may be unable to send keepalive packets from the affected node. The Cluster Managers on other cluster nodes interpret the lack of keepalives as node failure, and the I/O-bound node is “failed” by the other nodes. When thfinishes, the node resumes sending keepalives. Its packets, however, are now out of synthe other nodes, which then kill the I/O-bound node with a RESET packet.

You can use I/O pacing to tune the system so that system resources are distributed morequitably during high disk I/O. You do this by setting high- and low-water marks. If a proctries to write to a file at the high-water mark, it must wait until enough I/O operations havfinished to make the low-water mark.

By default, AIX is installed with high- and low-water marks set to zero, which disables I/O pacing.

While enabling I/O pacing may have a slight performance effect on very I/O intensive processes, it is required for an HACMP cluster to behave correctly during large disk writyou anticipate heavy I/O on your HACMP cluster, you should enable I/O pacing.

Although the most efficient high- and low-water marks vary from system to system, an inhigh-water mark of 33 and a low-water mark of 24 provides a good starting point. These settings only slightly reduce write times and consistently generate correct fallover behavfrom the HACMP for AIX software.

See the AIX Performance Monitoring & Tuning Guide for more information on I/O pacing.

Setting Syncd Frequency

The syncd setting determines the frequency with which the I/O disk-write buffers are flushFrequent flushing of these buffers reduces the chance of deadman switch time-outs.

The AIX default value for syncd as set in /sbin/rc.boot is 60. It is recommended to change thvalue to 10. Note that the I/O pacing parameter setting should be changed first. You shouneed to adjust this parameter again unless you get frequent time-outs.

Setting Failure Detection Parameters

Each supported cluster network in a configured HACMP cluster has a corresponding clunetwork module. Each network module monitors all I/O to its cluster network.

Each network module maintains a connection to other network modules in the cluster. TCluster Managers on cluster nodes send messages to each other through these connecEach network module is responsible for maintaining a working set of service adapters averifying connectivity to cluster peers. The network module also is responsible for reporwhen a given link actually fails. It does this by sending and receiving periodic heartbeat messages to or from other network modules in the cluster.

Currently, network modules support communication over the following types of network

• Serial (RS232)

• Target-mode SCSI

• Target-mode SSA

• IP

• Ethernet

2 Planning Guide

Page 49: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksPlanning for Cluster Performance

busy

es is at the

• Token-Ring

• FDDI

• SOCC

• SLIP

• SP Switch

• ATM.

The Failure Detection Rate is made up of two components:

• cycles to fail (cycle): the number of heartbeats that must be missed before detecting a failure

• heartbeat rate (hbrate): the number of microseconds between heartbeats.

Together, these two values determine the Failure Detection Rate. For example, for a NIM with an hbrate of 1,000,000 microseconds and a cycle value of 12, the Failure Detection Rate would be 12 (1 second x 12 cycles).

The default heartbeat rate is usually optimal. Speeding up or slowing down failure detection is a small, but potentially significant area where you can adjust cluster fallover behavior. However, the amount and type of customization you add to event processing has a much greater impact on the total fallover time. You should test the system for some time before deciding to change the failure detection speed of any network module.

If HACMP for AIX cannot get enough CPU resources to send heartbeats on IP and serial networks, other nodes in the cluster will assume the node has failed, and initiate takeover of the node’s resources. In order to ensure a clean takeover, the Deadman Switch crashes thenode if it is not reset within a given time period. The Deadman Switch uses the followingformula:

N=((keepalives * missed_keepalives) -1)Where keepalives and missed_keepalives are for the slowest network in the cluster.

The table below shows the Deadman Switch Timeout for each network. Each of these timone second less than the time to trigger an HACMP event on that network. Remember thDeadman Switch is triggered on the slowest network in your cluster.

Planning Guide 3-23

Page 50: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksPlanning for Cluster Performance

3-2

Deadman Switch Timeouts in Seconds Per Network

If you decide to change the failure detection rate of a network module, keep the following considerations in mind:

• Failure detection is dependent on the fastest network linking two nodes.

• The failure rate of networks varies, depending on their characteristics. For example, for an Ethernet, the normal failure detection rate is two keepalives per second; fast is about four per second; slow is about one per second. For an SP switch network, because no network traffic is allowed when a node joins the cluster, normal failure detection is 30 seconds; fast is 10 seconds; slow is 60 seconds.

• Before altering the NIM, you should give careful thought to how much time you want to elapse before a real node failure is detected by the other nodes and the subsequent takeover is initiated.

• Faster heartbeat rates may lead to false failure detections, particularly on busy networks. For example, bursts of high network traffic may delay heartbeats and this may result in nodes being falsely ejected from the cluster. Faster heartbeat rates also place a greater load on networks. If your networks are very busy and you experience false failure detections, you can try slowing the failure detection speed on the network modules to avoid this problem.

• It is recommended that you first change the Failure Detection Rate from normal to slow (or fast) before trying to customize this rate. If you do need to customize, change the hbrate, and then adjust the cycle value as needed to reach the desired Failure Detection Rate.

• The Failure Detection Rate should be set equally for every NIM used by the cluster. The change must be synchronized across cluster nodes. The new values will become active the next time cluster services are started.

See the HACMP Installation Guide for more details on configuring these parameters.

NETWORK SLOW NORMAL FAST

ATM 63 31 15

Ethernet 11 5 3.5

FDDI 11 5 3,5

Token-Ring 11 5 3.5

RS232 17 11 5

SLIP 17 11 5

SOCC 11 5 3.5

SP Switch 63 15 7

TMSSA 17 11 5

TMSCSI 17 11 5

4 Planning Guide

Page 51: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksAdding the TCP/IP Network Topology to the Cluster Diagram

Adding the TCP/IP Network Topology to the Cluster Diagram

You can now add the TCP/IP network topology to the cluster diagram. The following cluster diagram includes the TCP/IP network topology:

A Cluster Diagram with TCP/IP Network Topology

To add the network topology information to your diagram:

1. Sketch in the networks, including any point-to-point connections between nodes.

2. Name each network and indicate its attribute. Remember, a network name is an arbitrary string.

For example, in the preceding figure, the public Ethernet network is named ether1 and the private SOCC point-to-point connection is named socc1.

3. Add the network adapters. Give each adapter a name and indicate its function. If you are using IP address takeover or rotating resources, remember to include a boot address for each service adapter that can be taken over.

For example, in the preceding figure, the service adapter for the Ethernet network on the node nodeA is named NodeA_svc, and the corresponding boot adapter is NodeA_boot.

Note that the names and IDs that you define (for example, ether1) apply only within the specific HACMP cluster and have no meaning outside of it.

Planning Guide 3-25

Page 52: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksCompleting the TCP/IP Networks Worksheets

3-2

r. To

ing

site

r to

ted to t:

Completing the TCP/IP Networks WorksheetsBecause the HACMP for AIX software supports multiple TCP/IP networks within a single cluster, installing a network topology to support HACMP can be a complex task. Use the following worksheets in Appendix A, Planning Worksheets, to make the process easier:

• TCP/IP Networks Worksheet (one for the cluster)

• TCP/IP Networks Adapter Worksheet (one for each node)

Appendix A includes examples of completed worksheets.

Note that the worksheet instructions here correspond to the paper worksheets and do not exactly match the online worksheets. In the online worksheet program, you would enter network configuration information in the “Networks” and “Adapters” panels.

Completing the TCP/IP Networks Worksheet The TCP/IP Networks Worksheet helps you organize the networks for an HACMP clustecomplete the worksheet:

1. Enter the cluster ID in the Cluster ID field.

2. Enter the cluster name in the Cluster Name field.

3. In the Network Name field, give each network a symbolic name. You use this value durthe install process when you configure the cluster.

Remember that each SLIP and SOCC line is a distinct network and must be listed separately.

4. Indicate the network’s type in the Network Type field. For example, it may be Ethernet,Token-Ring, and so on.

5. Indicate the network’s function in the Network Attribute field. That is, the network can be public or private.

6. In the Netmask field, provide the network mask of each network. The network mask is dependent and must be the same for all adapters in the HACMP network.

7. In the Node Names field, list the names of the nodes connected to each network. Refeyour cluster diagram.

Completing the TCP/IP Network Adapter Worksheet

The TCP/IP Network Adapter Worksheet helps you define the network adapters conneceach cluster node. Complete the following steps for each node on a separate workshee

1. Enter the node name in the Node Name field.

2. Leave the Interface Name field blank. You will enter values in this field after you configure the adapter following the instructions in the HACMP for AIX Installation Guide.

3. Enter the symbolic name of the adapter in the Adapter IP Label field.

4. Identify the adapter’s function as service, standby, or boot in the Adapter Function field.

6 Planning Guide

Page 53: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksWhere You Go From Here

sheet.

ning

5. Enter the IP address for each adapter in the Adapter IP Address field. Note that the SMIT Add an Adapter screen displays an Adapter Identifier field that correlates to this field on the worksheet.

6. Enter the name of the network to which this network adapter is connected in the Network Name field. Refer to the TCP/IP Networks Worksheet for this information.

7. Identify the network as public or private in the Network Attribute field. Refer to the TCP/IP Networks Worksheet for this information.

8. This field is optional. If you are using IP address takeover or rotating resources, in the Boot Address field enter the boot address for each service address that can be taken over.

9. This field is optional. If you are using hardware address swapping, in the Adapter HW Address field enter the hardware address for each service address with an assigned boot address. The hardware address is a 12 or 14 digit hexadecimal value. Usually, hexadecimal numbers are prefaced with “0x” (zero x) for readability. Do not use colons to separate the numbers in the adapter hardware address.

Note: Entries in the Adapter HW Address field should refer to the locally administered address (LAA), which applies only to the service adapter.

For each cluster node, repeat these steps on a separate TCP/IP Network Adapter Work

Where You Go From HereYou have now planned the TCP/IP network topology for the cluster. The next step in theplanning process is to lay out the serial network topology for the cluster. Chapter 4, PlanSerial Networks, describes this process.

Planning Guide 3-27

Page 54: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning TCP/IP NetworksWhere You Go From Here

3-2

8 Planning Guide
Page 55: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Serial NetworksPrerequisites

4

Chapter 4 Planning Serial Networks

This chapter describes planning serial networks for an HACMP cluster.

PrerequisitesIn Chapter 2, Drawing the Cluster Diagram, you drew a diagram showing the basic layout of your cluster. This diagram is the basis for the planning in this chapter.

OverviewIn this chapter you plan the serial networks for your cluster. The chapter first defines node isolation and partitioned clusters, discusses serial network topology, and describes the supported serial network types. You then add the serial network topology to the cluster diagram and complete the Serial Networks Worksheet in Appendix A, Planning Worksheets.

Node Isolation and Partitioned ClustersNode isolation occurs when all networks connecting two or more parts of the cluster fail. Each group (one or more) of nodes is completely isolated from the other groups. A cluster in which certain groups of nodes are unable to communicate with other groups of nodes is a partitioned cluster. The following figure illustrates a partitioned cluster:

Planning Guide 4-1

Page 56: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Serial NetworksNode Isolation and Partitioned Clusters

4-2

et.

, it is e by a

rks

a does

on hat

A Partitioned Cluster

The problem with a partitioned cluster is that each node on one side of the partition interprets the absence of heartbeats from the nodes on the other side to mean that those nodes have failed, and generates node failure events for those nodes. Once this occurs, nodes on each side of the cluster (if so configured) attempt to take over resources from nodes that are still active and, therefore, still legitimately own those resources. These attempted takeovers can cause unpredictable results in the cluster—for example, data corruption due to a disk being res

To guard against the failure of the TCP/IP subsystem and to prevent partitioned clustersstrongly recommended that each node in the cluster be connected to its neighboring nodpoint-to-point serial network, thus forming a logical “ring.” This logical ring of serial networeduces the chance of node isolation by allowing neighboring Cluster Managers to communicate even when all TCP/IP-based networks fail.

Serial networks are especially important in concurrent access configurations so that datnot become corrupted when TCP/IP traffic among nodes is lost.

It is important to understand that the serial network does not carry TCP/IP communicatibetween nodes; it only allows nodes to exchange heartbeats and control messages so tCluster Managers have accurate information about the status of peer nodes.

Node A Node B

Node C Node D

Planning Guide

Page 57: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Serial NetworksSerial Network Topology

and

Serial Network TopologyAs mentioned in the preceding section, a serial network reduces the occurrence of node isolation (and thus spurious resource takeovers) by allowing Cluster Managers to communicate even after all TCP/IP-based networks have failed. To eliminate the TCP/IP subsystem as a single point of failure within a cluster, the serial networks must form a logical ring so that all Cluster Managers can learn the status of other Cluster Managers in the cluster.

For example, the cluster in the following diagram uses serial networks to connect four nodes.

Point-to-Point Serial Networks in a Four-Node Cluster

Each node in the cluster must have a point-to-point serial network connection with its neighboring nodes. The following list indicates the number of serial networks needed to eliminate the TCP/IP subsystem as a single point of failure for this cluster:

• Node A to Node B

• Node B to Node C

• Node C to Node D

• Node A to Node D

Note that, in order to form the logical ring, Node A has two “neighboring” nodes, Nodes BD.

Planning Guide 4-3

Page 58: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Serial NetworksSerial Network Topology

4-4

Supported Serial Network Types

The HACMP for AIX software supports the following serial networks: a raw RS232 serial line, a SCSI-2 Differential or SCSI-2 Differential Fast/Wide bus using target mode SCSI, and a target mode SSA loop.

RS232 Serial Line

If you are using shared disk devices other than SSA or SCSI-2 Differential or SCSI-2 Differential Fast/Wide devices, you must use a raw RS232 serial line as the serial network between pairs of nodes. The RS232 serial line provides a point-to-point connection between nodes in an HACMP cluster and is classified as a tty device requiring a dedicated serial port at each end.

You can label the tty device using any name or characters you choose; however, it is most commonly identified by a four-character string, where the first three characters are tty and the fourth is the AIX assigned device number.

For example, if a node called clam is using a tty device as its serial line, you can label its adapter arbitrarily as clam_serial, or you can label it clam_tty1 to reflect the AIX-assigned device number.

See the HACMP for AIX Installation Guide for information on configuring the raw RS232 serial line as a serial network in an HACMP cluster.

Note: On the SP thin or wide nodes there are no serial ports available. Therefore, any HACMP/ES configurations that require a tty network need to make use of a serial adapter card (8-port async EIA-232 adapter, FC/2930), available on the SP as an RPQ.

Note: The 7013-S70, 7015-S70, and 7017-S70 do not support the use of native serial ports in an HACMP/ES RS232 serial network. Configuration of an RS232 serial network in an S70 system requires a PCI multi-port Async card.

Target Mode SCSI

You can configure a SCSI-2 bus as an HACMP for AIX serial network only if you are using SCSI-2 Differential or SCSI-2 Differential Fast/Wide devices. SCSI-1 Single-Ended and SCSI-2 Single-Ended do not support serial networks in an HACMP cluster. The advantage of using the SCSI-2 Differential bus is that it does not require a dedicated serial port at each end of the connection. It is recommended that you use a maximum of four target mode SCSI networks in a cluster.

The target mode SCSI device that connects nodes in an HACMP cluster is identified by a seven-character name string, where the characters tmscsi are the first six characters and the seventh character is the number AIX assigns to the device (for example, tmscsi1).

See the HACMP for AIX Installation Guide for information on configuring a SCSI-2 Differential or SCSI-2 Differential Fast/Wide bus as a serial network in an HACMP cluster.

Planning Guide

Page 59: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Serial NetworksAdding the Serial Network Topology to the Cluster Diagram

Target Mode SSA

You can configure a target mode SSA connection between nodes sharing disks connected to SSA on Multi-Initiator RAID adapters (FC 6215 and FC 6219). The adapters must be at Microcode Level 1801 or later.

You can define a serial network to HACMP that connects all nodes on an SSA loop. By default, node numbers on all systems are zero. In order to configure the target mode devices, you must first assign a unique node number to all systems on the SSA loop. Do not use the number 0.

The target mode SSA device that connects nodes in an HACMP for AIX cluster is identified by a six-character name string, where the characters tmssa are the first five characters and the sixth character is the unique node number you assign to the device (for example, tmssa1).

Configuring the target mode connection creates two special files in the /dev directory of each node, the /dev/tmssa#.im and /dev/tmssa#.tm files. The file with the .im extension is the initiator, which transmits data. The file with the .tm extension is the target, which receives data.

Adding the Serial Network Topology to the Cluster Diagram

You now add the serial network topology to the cluster diagram as shown in the following figure:

A Cluster Diagram with Serial Network Topology

Planning Guide 4-5

Page 60: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Serial NetworksCompleting the Serial Networks and Serial Network Adapter Worksheets

4-6

r a

To add the serial network topology to the cluster diagram:

1. Sketch in the serial networks. Remember, it is recommended that a serial network connect each pair of adjoining nodes.

2. Name each network; then indicate that its attribute is serial. A network name is an arbitrary string. For example, you could name an RS232 line serial1.

For example, the cluster in the preceding figure has a single serial network named rs2321.

3. Add the serial network adapters (if planning for a tmscsi connection) or specify a tty port connection, then give each a name. Indicate its function as service.

For example, the node Node1 is using tty1 as its serial line connection; its adapter label is Node1_tty1. If you plan for a tmscsi connection, its adapter label might be Node1_tmscsi2.

At this point, you should have a cluster diagram similar to the one in the preceding figure.

Completing the Serial Networks and Serial Network Adapter Worksheets

After completing a cluster diagram, use the following paper worksheets, provided in Appendix A, Planning Worksheets, to record the serial network topology for your cluster:

• Serial Networks Worksheet (one for the cluster)

• Serial Network Adapter Worksheet (one for each node).

Appendix A includes examples of completed serial network worksheets.

See Appendix B, Using the Online Cluster Planning Worksheet Program if you are using the online planning worksheets.

Completing the Serial Networks Worksheet

The Serial Networks Worksheet helps you organize the serial networks for an HACMP cluster. To complete the Serial Networks Worksheet:

1. Enter the cluster ID in the Cluster ID field.

2. Enter the cluster name in the Cluster Name field.

3. Give each network a symbolic name in the Network Name field. Remember, each RS232 serial line or target mode SCSI-2 Differential bus is a distinct network and must be listed separately.

4. In the Network Type field, indicate the network’s type as either an RS232 serial line otarget mode SCSI-2 Differential bus.

5. Indicate the network’s attribute as serial in the Network Attribute field.

6. List the names of the nodes connected to each serial network in the Node Names field. Refer to your cluster diagram.

Note: None of these serial networks use the TCP/IP protocol; therefore they do not require a netmask.

Planning Guide

Page 61: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Serial NetworksWhere You Go From Here

uster.

rocess pter.

Completing the Serial Network Adapter Worksheet

The Serial Network Adapter Worksheet helps you define the serial connections between cluster nodes. Complete the following steps for each node on a separate worksheet:

1. Enter the node name in the Node Name field.

2. Enter the slot number used to make the serial connection.

3. Leave the Interface Name field blank. You will enter the device name in this field after you configure the serial adapters during installation.

4. Enter the symbolic name of the adapter in the Adapter Label field.

5. Enter the name of the network to which this adapter is connected in the Network Name field. Refer to the Serial Networks Worksheet for the network name.

6. Identify the network as serial in the Network Attribute field.

7. Identify the adapter’s function as service in the Adapter Function field.

Repeat these steps on a new Serial Network Adapter Worksheet for each node in the cl

Where You Go From HereYou have now planned the serial networks for the cluster. The next step in the planning pis to lay out the shared disk configuration for your cluster, described in the following cha

Planning Guide 4-7

Page 62: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Serial NetworksWhere You Go From Here

4-8

Planning Guide
Page 63: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPrerequisites

5

Chapter 5 Planning Shared Disk Devices

This chapter discusses information you must consider before configuring shared external disks in an HACMP cluster.

PrerequisitesRead the HACMP for AIX Concepts and Facilities guide for a discussion of the shared disk access methods supported by the HACMP for AIX software.

Also, refer to AIX documentation for the general hardware and software setup for your disks.

OverviewThe HACMP for AIX software supports several shared disk configurations. Choosing the best setup for your needs requires careful thought before actually installing the disks. This chapter specifically includes information to help you:

• Choose a disk technology.

• Plan shared and non-shared storage.

• Plan to install any of the following SCSI disk configurations: SCSI-2 SE, SCSI-2 Differential, SCSI-2 Differential Fast/Wide, IBM 7137 Disk Array, IBM 7135 models 110 and 210 RAIDiant Disk Arrays, and IBM 2105 Versatile Storage Server models B09 and 100. Adapter requirements for each configuration are included in the planning considerations. The arrays can appear to the host as multiple SCSI devices. Use the general SCSI disk instructions in this chapter to plan for these arrays.

• Plan to install an IBM 9333 serial disk subsystem. Adapter requirements are included in the planning considerations.

• Plan to use an IBM Serial Storage Architecture (SSA) disk subsystem in your configuration, such as the IBM 7133 disk subsystem. Adapter requirements are included in the planning considerations.

After reading this chapter, you will be able to complete the shared disk configuration portion of your cluster diagram. Completing this diagram will make installing disks and arrays easier.

Choosing a Shared Disk TechnologyThe HACMP for AIX software supports the following disk technologies as shared external disks in a highly available cluster:

• SCSI-2 SE, SCSI-2 Differential, and SCSI-2 Differential Fast/Wide adapters and drives

• IBM 9333 serial-link adapters and serial disk drive subsystems or enclosures

• IBM SSA adapters and SSA disk subsystems.

Planning Guide 5-1

Page 64: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesChoosing a Shared Disk Technology

5-2

e. If he data

You can combine these technologies within a cluster. Before choosing a disk technology, however, review the considerations for configuring each technology as described in this section.

SCSI Disks

The HACMP for AIX software supports the following SCSI disk devices and arrays as shared external disk storage in cluster configurations:

• SCSI-2 SE, SCSI-2 Differential, and SCSI-2 Differential Fast/Wide disk devices

• The IBM 7135-110 and 7135-210 RAIDiant Disk Arrays

• The IBM 7137 Disk Array

• The IBM 2105-B09 and 2105-100 Versatile Storage Servers

These devices and arrays are described in the following section.

SCSI-2 SE, SCSI-2 Differential, and SCSI-2 Differential Fast/Wide Disk Devices

The benefit of the SCSI implementation is its low cost. It provides a shared disk solution that works with most supported RISC System/6000 processor models (including the SP and the SMP) and requires minimal hardware overhead.

In an HACMP cluster, shared SCSI disks are connected to the same SCSI bus for the nodes that share the devices. They may be used in both concurrent and non-concurrent modes of access. In a non-concurrent access environment, the disks are “owned” by only one node at a timthe owner node fails, the cluster node with the next highest priority in the resource chainacquires ownership of the shared disks as part of fallover processing. This ensures that tstored on the disks remains accessible to client applications. The following restrictions, however, apply to using shared SCSI disks in a cluster configuration:

• SCSI-2 SE, SCSI-2 Differential and SCSI-2 Differential Fast/Wide disks and disk arrays support non-concurrent shared disk access. The only SCSI-2 Differential disk devices that support concurrent shared disk access are the IBM 7135-110 and 7135-210 RAIDiant Disk Arrays, the IBM 7137 Disk Array, and the IBM 2105-B09 and 2105-100 Versatile Storage Servers.

• Different types of SCSI buses can be configured in an HACMP cluster. Specifically, SCSI-2 Differential and SCSI-2 Differential Fast/Wide devices can be configured in clusters of up to four nodes, where all nodes are connected to the same SCSI bus attaching the separate device types. (You cannot mix SCSI-2 SE, SCSI-2 Differential and SCSI-2 Differential Fast/Wide devices on the same bus.)

• You can connect the IBM 7135-210 RAIDiant Disk Array only to High Performance SCSI-2 Differential Fast/Wide adapters, while the 7135-110 RAIDiant Disk Array cannot use those High Performance Fast/Wide adapters.

• You can connect up to sixteen devices to a SCSI-2 Differential Fast/Wide bus. Each SCSI adapter and disk is considered a separate device with its own SCSI ID. The SCSI-2 Differential Fast/Wide maximum bus length of 25 meters provides enough length for most cluster configurations to accommodate the full sixteen-device connections allowed by the SCSI standard.

• Do not connect other SCSI devices, such as CD ROMs or tape drives, to a shared SCSI bus.

Planning Guide

Page 65: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesChoosing a Shared Disk Technology

• You cannot assign SCSI IDs 0, 1, or 2 to the IBM High Performance SCSI-2 Differential Fast/Wide Adapter; the adapter restricts the use of these IDs. Likewise, you cannot assign SCSI IDs 0, 1, or 8 through 15 to the IBM SCSI-2 adapters 2412, 2415, and 2416.

IBM 7135 RAIDiant Disk Array Devices

You can use an IBM 7135-110 or 7135-210 RAIDiant Disk Array to provide concurrent or non-concurrent access in HACMP cluster configurations. The benefits of using an IBM 7135 RAIDiant Disk Array in an HACMP cluster are its storage capacity, speed, and reliability. The IBM 7135 RAIDiant Disk Array contains a group of disk drives that work together to provide enormous storage capacity (up to 135 GB of non-redundant storage) and higher I/O rates than single large drives.

RAID LevelsThe IBM 7135 RAIDiant Disk Arrays support reliability features that provide data redundancy to prevent data loss if one of the disk drives in the array fails. As a RAID device, the array can provide data redundancy through RAID levels. The IBM 7135 RAIDiant Disk Arrays support RAID levels 0, 1, and 5. RAID level 3 can be used only with a raw disk.

In RAID level 0, data is striped across a bank of disks in the array to improve throughput. Because RAID level 0 does not provide data redundancy, it is not recommended for use in HACMP clusters.

In RAID level 1, the IBM 7135 RAIDiant Disk Array provides data redundancy by maintaining multiple copies of the data on separate drives (mirroring).

In RAID level 5, the IBM 7135 RAIDiant Disk Array provides data redundancy by maintaining parity information that allows the data on a particular drive to be reconstructed if the drive fails.

All drives in the array are hot-pluggable. When you replace a failed drive, the IBM 7135 RAIDiant Disk Array reconstructs the data on the replacement drive automatically. Because of these reliability features, you should not define LVM mirrors in volume groups defined on an IBM 7135 RAIDiant Disk Array.

Dual Active ControllersTo eliminate adapters or array controllers as single points of failure in an HACMP cluster, the IBM 7135 RAIDiant Disk Array can be configured with a second array controller that acts as a backup controller in the event of a fallover. This configuration requires that you configure each cluster node with two adapters. You connect these adapters to the two array controllers using separate SCSI buses.

In this configuration, each adapter and array-controller combination defines a unique path from the node to the data on the disk array. The IBM 7135 RAIDiant Disk Array software manages data access through these paths. Both paths are active and can be used to access data on the disk array. If a component failure disables the current path, the disk array software automatically re-routes data transfers through the other path.

Planning Guide 5-3

Page 66: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesChoosing a Shared Disk Technology

5-4

Note: This dual-active path-switching capability is independent of the capabilities of the HACMP for AIX software, which provides protection from a node failure. When you configure the IBM 7135 RAIDiant Disk Array with multiple controllers and configure the nodes with multiple adapters and SCSI buses, the disk array software prevents a single adapter or controller failure from causing disks to become unavailable.

The following restrictions apply to using shared IBM 7135 RAIDiant Disk Arrays in a cluster configuration:

• You can connect the IBM 7135-210 RAIDiant Disk Array only to High Performance SCSI-2 Differential Fast/Wide adapters, while the 7135-110 RAIDiant Disk Array cannot use those High Performance Fast/Wide adapters.

• You can connect an IBM 7135 RAIDiant Disk Array to up to four cluster nodes using a SCSI bus. You also can include up to two IBM 7135 RAIDiant Disk Arrays per bus.

Note: Each array controller and adapter on the same SCSI bus requires a unique SCSI ID.

• You may need to configure the drives in the IBM 7135 RAIDiant Disk Array into logical units (LUN) before setting up the cluster. A standard SCSI disk equates to a single LUN. AIX configures each LUN as a hard disk with a unique logical name of the form hdiskn, where n is an integer.

An IBM 7135 RAIDiant Disk Array comes preconfigured with several LUNs defined. This configuration depends on the number of drives in the array and their sizes and is assigned a default RAID level of 5. You can, if desired, use the disk array manager utility to configure the individual drives in the array into numerous possible combinations of LUNs, spreading a single LUN across several individual physical drives.

For more information about configuring LUNs on an IBM 7135 RAIDiant Disk Array, see the documentation you received with your disk array for the specific LUN composition of your unit.

Once AIX configures the LUNs into hdisks, you can define the ownership of the disks as you would any other shared disk.

To put RAID disks into concurrent mode, see the HACMP for AIX Administration Guide.

IBM 7137 Disk Array

The IBM 7137 disk array contains multiple SCSI-2 Differential disks. On the IBM 7137 array, these disks can be grouped together into multiple LUNs, with each LUN appearing to the host as a single SCSI device (hdisk).

The disk array can support concurrent shared disk access; however, mirroring across the SCSI devices is not supported in concurrent mode.

Keep in mind that if you configure the IBM 7137 disk array for concurrent disk access, the SCSI bus and the controller become single points of failure. To eliminate these single points of failure in a concurrent access environment, you should consider using one of the IBM 7135 RAIDiant Disk Array models.

Planning Guide

Page 67: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesChoosing a Shared Disk Technology

IBM 2105 Versatile Storage Server

The IBM 2105 Versatile Storage Server (VSS) provides multiple concurrent attachment and sharing of disk storage for a variety of open systems servers. RISC System/6000 processors can be attached, as well as other UNIX and non-UNIX platforms.

The VSS uses IBM SSA disk technology. Existing IBM 7133 SSA disk drawers can be used in the VSS. See the section, IBM Serial Storage Architecture Disk Subsystem section on page 5-6 for more information about SSA systems.

The RISC System/6000 attaches to the IBM 2105 Versatile Storage Server via SCSI-2 Differential Fast/Wide Adapter/A. A maximum of 64 open system servers can be attached to the VSS (16 SCSI channels with 4 adapters each).

There are many availability features included in the VSS. All storage is protected with RAID technology. RAID-5 techniques can be used to distribute parity across all disks in the array. Sparing is a function which allows you to assign a disk drive as a spare for availability. Predictive Failure Analysis techniques are utilized to predict errors before they affect data availability. Failover Protection enables one partition, or storage cluster, of the VSS to takeover for the other so that data access can continue.

The VSS includes other features such as a web-based management interface, dynamic storage allocation, and remote services support. For more information on VSS planning, general reference material, and attachment diagrams, see the URLs:

http://www.storage.ibm.com/hardsoft/products/vss/books/vssrefinfo.htmhttp://www.storage.ibm.com/hardsoft/products/vss/books/vsr1ag.htm

IBM 9333 Serial Disk Subsystems

The HACMP for AIX software supports IBM 9333 serial disk drive subsystems as shared external disk storage devices. These drives are part of the IBM 9333 High Performance Disk Drive Subsystem.

If you include the IBM 9333 serial disks in a volume group that uses LVM mirroring, you can replace a failed drive without powering off the entire subsystem.

In an HACMP cluster, you connect the shared IBM 9333 serial disks to the cluster nodes by cross-linking the IBM 9333 controllers. Depending on the model used, you can connect IBM 9333 serial disk subsystems to up to eight cluster nodes in concurrent or non-concurrent access configurations. The following restrictions apply to using shared IBM 9333 serial disk subsystems in a cluster configuration:

• The IBM 9333 serial disk subsystems Model 010 (drawer) and Model 500 (desk-side unit) support only non-concurrent access cluster configurations. In addition, these models can only be attached to two cluster nodes.

• The IBM 9333 serial disk subsystems Model 011 (drawer) and Model 501 (desk-side unit) support both concurrent access and non-concurrent access cluster configurations. These models can be used in cluster configurations that include up to eight nodes.

• A single serial adapter can support up to 16 disks. The number of serial adapters supported by the processor, therefore, determines the number of shared disks that can be connected to a node. The maximum number of shared disks that can be connected to a single node in an IBM 9333 serial configuration is 112, which can be divided among up to 7 adapter cards.

Planning Guide 5-5

Page 68: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPower Supply Considerations

5-6

IBM Serial Storage Architecture Disk Subsystem

The Serial Storage Architecture (SSA) offers many features for minimizing single points of failure and achieving high availability in an HACMP environment.

You can use IBM 7133 and 7131-405 SSA disk subsystems as shared external disk storage devices to provide concurrent access in an HACMP cluster configuration.

SSA is hot-pluggable. Consequently, if you include SSA disks in a volume group using LVM mirroring, you can replace a failed disk drive without powering off the entire system.

SSA Loop Configuration

An SSA loop is a cyclic network between adapters that access a disk subsystem. If one adapter fails, the other adapter can still access the disk. For more detailed information about loop configurations, refer to the SSA adapter product documentation listed in the preface.

The following figure shows the basic SSA loop configuration.

Basic SSA Loop Configuration

Power Supply ConsiderationsReliable power sources are critical for a highly available cluster. Each node and mirrored disk chain in the cluster should have a separate power source. As you plan the cluster, make sure that the failure of any one power source (a blown fuse, for example) does not disable more than one node or mirrored chain. This section discusses specific power supply considerations for supported disk types.

SCSI Configurations

If the cluster site has a multiple-phase power supply, you must ensure that the cluster nodes are attached to the same power phase. Otherwise, the ground will move between the systems across the SCSI bus and cause write errors.

Uninterruptable power supply (UPS) devices are necessary for preventing data loss. The bus and devices shared between two nodes are subject to the same operational power surge restrictions as standard SCSI systems. When power is first applied to a SCSI device, the attached bus may have active data corrupted. You can avoid such errors by briefly halting data

Planning Guide

Page 69: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning for Non-Shared Disk Storage

transfer operations on the bus while a device (disk or adapter) is turned on. For example, if cluster nodes are installed on two different power grids and one node has a power surge that causes it to reboot, the surviving node may lose data if a data transfer is active.

The IBM 7135 RAIDiant Disk Arrays and IBM 2105 Versatile Storage Servers are less prone to power supply problems because they come with redundant power supplies.

IBM 9333 Serial Disk Subsystem Configurations

Clusters with IBM 9333 serial disks are not prone to power supply problems. Nevertheless, UPS devices are still valuable for maintaining and ensuring no single point of failure.

In addition, when cross-linking several rack-mounted IBM 9333 serial disk subsystems, be sure to use the Dual Power Control to interconnect the IBM 9333 drawers in each rack to the power distribution unit in the other rack. This interconnection ensures that the IBM 9333 drawer does not power down when the node in the same rack goes down. If the drawer powered down, the fallover would fail when the cross-linked system, in a separate rack, attempted to take over the failed node’s disks.

IBM SSA Disk Subsystem Configurations

Clusters with IBM SSA disk subsystems are not prone to power supply problems because they come with redundant power supplies.

Planning for Non-Shared Disk StorageKeep the following considerations in mind regarding non-shared disk storage:

• The internal disks on each node in a cluster must provide sufficient space for:

• AIX software (approximately 320 MB)

• HACMP for AIX software (approximately 15 MB for a server node)

• Executable modules of highly available applications.

• The root volume group (rootvg) for each node must not reside on the shared SCSI bus.

• Use the AIX Error Notification Facility to monitor the rootvg on each node. Problems with the root volume group can be promoted to node failures. See the HACMP for AIX Installation Guide for more information on using the Error Notification facility.

• Consider mirroring rootvg. See the mirrorvg man command for a good explanation and information

• Because shared disks require their own adapters, you cannot use the same adapter for both a shared and a non-shared disk. The internal disks on each node require one SCSI adapter apart from any other adapters within the cluster.

• Internal disks must be in a different volume group from the external shared disks.

Planning Guide 5-7

Page 70: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning for Non-Shared Disk Storage

5-8

• The executable modules of the highly available applications should be on the internal disks and not on the shared external disks, for the following reasons:

Licensing Some vendors require a unique license for each processor or multi-processor that runs an application, and thus license-protect the application by incorporating processor-specific information into the application when it is installed. As a result, it is possible that even though the HACMP for AIX software processes a node failure correctly, it is unable to restart the application on the fallover node because of a restriction on the number of licenses available within the cluster for that application. To avoid this problem, make sure that you have a license for each processor in the cluster that may potentially run an application.

Starting Applications Some applications (such as databases) contain configuration files that you can tailor during installation and store with the binaries. These configuration files usually specify startup information, such as the databases to load and log files to open, after a fallover situation.

If you plan to put these configuration files on a shared filesystem, they will require additional tailoring. You will need to determine logically which system (node) is actually to invoke the application in the event of a fallover. Making this determination becomes particularly important in fallover configurations where conflicts in the location and access of control files can occur.

For example, in a two-node mutual takeover configuration, where both nodes are running different instances of the same application (different databases) and are standing by for one another, the takeover node must be aware of the location of specific control files and must be able to access them to perform the necessary steps to start critical applications after a fallover or else the fallover will fail, leaving critical applications unavailable to clients. If the configuration files are on a shared filesystem, a conflict can arise if the takeover node is not aware of the file’s location.

You can avoid much of the tailoring of configuration files by placing slightly different startup files for critical applications on local filesystems on either node. This allows the initial application parameters to remain static; the application will not need to recalculate the parameters each time it is invoked.

Planning Guide

Page 71: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning for Shared Disk Storage

Planning for Shared Disk Storage When planning for shared storage, consider the following when calculating disk requirements:

• You need multiple physical disks on which to put the mirrored logical volumes. Putting copies of a mirrored logical volume on the same physical device defeats the purpose of making copies. See Chapter 6, Planning Shared LVM Components, for more information on creating mirrored logical volumes.

When using an IBM 7135 RAIDiant Disk Array, do not create mirrored logical volumes. You must, however, account for the data redundancy maintained by the IBM 7135 RAIDiant Disk Array when calculating total storage capacity requirements. For example, in RAID level 1, because the IBM 7135 RAIDiant Disk Array maintains two copies of the data on separate drives, only half the total storage capacity is usable. Likewise, with RAID level 5, 20 to 30 percent of the total storage capacity is used to store and maintain parity information.

• Consider quorum issues when laying out a volume group. With quorum enabled, a two-disk volume group puts you at risk for losing quorum and data access. Either build three-disk volume groups or disable quorum.

In an IBM 7135 RAIDiant Disk Array, where a single volume group can contain multiple LUNs (collections of physical disks) that appear to the host as a single device (hdisk), quorum is not an issue because of the large storage capacity the LUNs provide, and because of the data redundancy capabilities of the array.

• Physical disks containing logical volume copies should be connected to different power supplies; otherwise, loss of a single power supply can prevent access to all copies. In practice, this can mean placing copies in different IBM 9333 subsystem drawers or in different SCSI desk-side units.

• Physical disks containing logical volume copies should be on separate adapters. If all logical volume copies are connected to a single adapter, the adapter is potentially a single point of failure. If the single adapter fails, you must move the volume group to an alternate node. Separate adapters prevent any need for this move.

Planning a Shared SCSI-2 Disk InstallationThe following list summarizes the basic hardware components required to set up an HACMP cluster that includes SCSI-2 SE, SCSI-2 Differential, or SCSI-2 Differential Fast/Wide devices as shared storage. Your exact cluster requirements will depend on the configuration you specify. To ensure that you account for all required components, complete a diagram for your system. Also, when planning a shared SCSI disk installation, consult the restrictions described in SCSI-2 SE, SCSI-2 Differential, and SCSI-2 Differential Fast/Wide Disk Devices section on page 5-2.

Disk Adapters

The HACMP for AIX software supports:

• IBM SCSI-2 Differential High-Performance External I/O Controller

• IBM High Performance SCSI-2 Differential Fast/Wide Adapter/A

• PCI SCSI-2 Differential Fast/Wide Adapter

Planning Guide 5-9

Page 72: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning a Shared SCSI-2 Disk Installation

5-1

• PCI SCSI-2 Fast/Wide Single-Ended Adapter

• PCI SCSI-2 Fast/Wide Differential Adapter

For each IBM 7135 RAIDiant Disk Array, an HACMP for AIX configuration requires that you configure each cluster node with two host adapters.

Note: Remove any SCSI terminators on the adapter card. You must use external terminators in an HACMP cluster. If you terminate the shared SCSI bus on the adapter, you lose termination when the cluster node that contains the adapter fails.

Cables

The cables required to connect nodes in your cluster depend on the type of SCSI bus you are configuring. Be sure to choose cables that are compatible with your disk adapters and controllers. For information on the specific type and length requirements for SCSI-2 Differential or SCSI-2 Differential Fast/Wide cables, see the hardware documentation that accompanies each device you want to include on the SCSI bus. Examples of SCSI bus configurations using IBM 7027-HSD and IBM 7204-315 enclosures, IBM 7137 Disk Array, and IBM 7135-210 RAIDiant Disk Arrays are shown throughout the following pages.

Sample SCSI-2 Differential Configuration

The following figure illustrates a SCSI-2 Differential configuration in which three nodes are attached to four SCSI-2 Differential disks in an IBM 7131-105 enclosure. Each adapter is connected to a SCSI-2 Differential Y-cable. Other required cables are labeled in the figure.

Shared SCSI-2 Differential Disk Configuration

0 Planning Guide

Page 73: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning a Shared SCSI-2 Disk Installation

Sample SCSI-2 Differential Fast/Wide Configuration

The following figure illustrates a SCSI-2 Differential Fast/Wide configuration in which three nodes are attached to four SCSI-2 Differential disks in an IBM 7204-315 enclosure. Each adapter is connected to a SCSI-2 Differential Fast/Wide Y-cable. Other required cables are labeled in the figure.

Shared SCSI-2 Differential Fast/Wide Disk Configuration

Sample IBM 7135-210 RAIDiant Disk Array Configuration

To take advantage of the path-switching capability of the IBM 7135-210 RAIDiant Disk Array software, you must configure each node with two adapters. In this way, the device driver can define multiple paths between the host and the disk array, eliminating the adapters, the controllers, and the SCSI bus as single points of failure. If a component failure disables one path, the IBM 7135-210 RAIDiant Disk Array device driver can switch to the other path automatically. This switching capability is the dual-active feature of the array.

Note: Although each controller on the IBM 7135-210 RAIDiant Disk Array contains two connectors, each controller requires only one SCSI ID.

Planning Guide 5-11

Page 74: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning a Shared SCSI-2 Disk Installation

5-1

SCSI-2 Differential Fast/Wide IBM 7135-210 RAIDiant Disk Array Configuration

The following figure illustrates an HACMP cluster with an IBM 7135-210 RAIDiant Disk Array connected to two SCSI-2 Differential Fast/Wide buses.

IBM 7135-210 RAIDiant Disk Array Configuration with SCSI-2 Differential Fast/Wide Buses

2 Planning Guide

Page 75: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning a Shared SCSI-2 Disk Installation

SCSI-2 Differential Fast/Wide IBM 7137 Disk Array Configuration

The following figure illustrates an HACMP cluster with an IBM 7137 Disk Array connected to a SCSI-2 Differential Fast/Wide bus.

IBM 7137 Disk Array with a SCSI-2 Differential Fast/Wide Bus

Note: SCSI-2 Fast/Wide connectors on an IBM 7137 Disk Array are positioned vertically.

Sample IBM 2105 Versatile Storage Server Configuration

Several diagrams are included in the information provided on the IBM web pages. See the list of figures in the attachment guide for the VSS:

http://www.storage.ibm.com/hardsoft/products/vss/books/vsr1ag.htm

Using VSS Features for High Availability

When using the VSS in an HACMP for AIX environment, the following is recommended for high availability:

• Use the Sparing function to assign disks as spares and reduce the exposure to data loss. When the VSS detects that a disk is failing, it transfers the data from the failing disk to a spare device. You are required to specify at least one disk as a spare per drawer; however you can specify two spares to a drawer for increased availability.

Planning Guide 5-13

Page 76: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning a Shared IBM 9333 Serial Disk Installation

5-1

• Configure the two host interface cards in a bay to device interface cards in the same bay.

• Configure the SCSI ports on the same interface card to the same partition of the VSS.

Planning a Shared IBM 9333 Serial Disk InstallationThe following list summarizes the basic hardware components required to set up an HACMP cluster that includes IBM 9333 serial disk subsystems as shared storage. Exact requirements depend on the configuration you specify. To ensure that you account for all required components, complete a system diagram. You should consider using the following hardware:

Refer to AIX documentation for the initial hardware and software setup of the disk subsystem. Keep in mind that to make logical volume mirroring effective, the mirrors should be placed on separate disk subsystems. When planning a shared IBM 9333 serial-link disk installation, consult the restrictions described earlier in this chapter on IBM 9333 Serial Disk Subsystems.

Adapters Used to connect a host to up to four disk drawers or desk-side units. Each disk subsystem can contain up to four disks per drawer or desk-side unit. Apart from cable-length restrictions, this makes it possible to attach an adapter to a disk associated with another node.

Note that you must use the High Performance Subsystem adapter with the IBM 9333 Model 011 or Model 501.

External serial-link cables Used to connect the host adapters on each cluster node to the IBM 9333 drawer or desk-side unit.

Local serial-link cables Used as jumpers on IBM 9333 Model 011 or Model 501 disk subsystems to connect the Multiple Systems Attachment cards to the IBM 9333 serial-link controller.

4 Planning Guide

Page 77: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning a Shared IBM 9333 Serial Disk Installation

Sample Two-Node Configuration

To support HACMP cluster configurations, each node requires one IBM 9333 adapter specifically dedicated to the shared IBM 9333 serial disk subsystem.

Take, for example, a two-node cluster consisting of Node A and Node B. Use one external serial-link cable to connect the adapter on Node A to one of the connectors on the IBM 9333 serial-link subsystem controller. Use another external serial-link cable to connect the adapter on Node B to the other connector on the same disk subsystem as shown in the following figure.

Two-Node Cluster with Shared IBM 9333 Serial Disk Subsystem

Node BNodeA

External

serial-link

cableExternal

serial-link

cable

9333-501 desk-side unit

Planning Guide 5-15

Page 78: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning a Shared IBM 9333 Serial Disk Installation

5-1

Sample Four-Node Configuration

The following figure shows a four-node cluster. Each node has an IBM 9333 serial-link adapter that is connected to connectors on the Multiple Systems Attachment card, which supports four nodes on the IBM 9333 serial disk subsystem. Note that you must use the IBM 9333 serial disk subsystem Model 501 or Model 011 to support more than two nodes.

Four-Node Cluster with Shared IBM 9333 Serial Disk Subsystem

6 Planning Guide

Page 79: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning a Shared IBM 9333 Serial Disk Installation

Sample Eight-Node Configuration

The following figure shows an eight-node cluster. Each node has an IBM 9333 serial-link adapter that is connected to connectors on one of two Multiple Systems Attachment cards, each of which supports four nodes on the IBM 9333 serial disk subsystem. Note that you must use the IBM 9333 serial disk subsystem Model 501 or Model 011 to support more than two nodes.

Eight-Node Cluster with Shared IBM 9333 Serial Disk Subsystem

Planning Guide 5-17

Page 80: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning a Shared IBM SSA Disk Subsystem Installation

5-1

Planning a Shared IBM SSA Disk Subsystem InstallationThe following discussions of adapters and SSA features for high availability are specific to using SSA disks with HACMP. They supplement the information in the manuals mentioned below.

IBM Manuals

Use the IBM manuals for:

• Rules for connecting SSA drives in loops to nodes

• Examples of configurations

• Planning charts.

Refer to the preface for a list of manuals covering SSA disk subsystem hardware, SSA adapter hardware, and general SSA reference material.

Adapters

All the IBM manuals listed in the preface are good sources of information on how to connect SSA disk subsystems to nodes. The following sections offer some additional information.

SSA 4-Port Adapter (Feature Code 6214, Type 4-D)

The 6214 SSA adapter (also called a two-way adapter) can support SSA loops containing up to two adapters per loop. In a high availability environment, then, an SSA loop is limited to one of these two configurations:

• One node containing two adapters

• Two nodes, each containing one adapter

Because of the limit imposed by the two-way adapter, if you configure both the two-way and the eight-way adapters in one loop, you can use only two adapters.

Enhanced SSA 4-Port Adapter (Feature Code 6216, Type 4-G)

The 6216 SSA adapter (also called an eight-way adapter) can support SSA loops containing up to eight eight-way adapters per loop. Most multi-node configurations set up with a minimal number of single points of failure require eight-way adapters.

Identifying Adapters

The two-way and eight-way adapters look the same, but differ in their microcode. The easiest way to determine which adapter you have installed is to run either of the following commands, which display identifying information about the microcode:

lsdev -Cc adapter

or

lscfg -vl ssax

where x is the adapter number.

8 Planning Guide

Page 81: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning a Shared IBM SSA Disk Subsystem Installation

Bypass Cards

The 7133 Models 020 and 600 disk subsystems contain four bypass cards. Each bypass card has two external SSA connectors. Through these, you connect the bypass cards and, therefore, the disk drive module strings to each other or to a node.

Note: The bypass cards can operate in either bypass or forced inline mode.

Bypass ModeWhen you set the jumpers on a bypass card so that it operates in bypass mode, the card monitors both of its external connections. If it determines that one of its connectors is connected to a powered-on SSA adapter or device, it switches to inline mode; that is, it connects the internal SSA links to the external connector. This effectively heals the break in the SSA loop.

If the bypass card determines that neither of its connectors is connected to a powered-on SSA adapter or device, it connects the internal disk strings and disconnects them from the external connector.

Forced Inline ModeWhen you set the jumpers on a bypass card so that it operates in forced inline mode, the card behaves like a signal card of Models 010 and 500; that is, none of its electronic switching circuits are used. Its internal SSA links connect to the external connector and never make an internal bypass connection.

Using SSA Features for High Availability

This section describes how you can use SSA features to make your system highly available.

Using SSA Loops

You can set up your configuration so that all SSA devices are in a loop, not just connected in a string. Although SSA devices will function when connected in a string, a loop provides two paths of communications to each device for redundancy. The adapter chooses the shortest path to a disk.

Using SSA Fiber-Optic Extenders

The SSA Fiber-Optic Extenders use cables up to 2.4 Km to replace a single SSA cable. The SSA Fiber-Optic Extender (Feature code 5500) is supported on all Model 7133 disk subsystems.

Using Fiber-Optic extenders, you can make the distance between disks greater than the LAN allows. If you do so, you cannot use routers and gateways. Consequently, under these circumstances, you cannot form an HACMP cluster between two LANs.

Daisy-Chaining the Adapters

In each node, for each loop including that node, you can daisy-chain all the adapters. The SSAR router device uses another adapter when it detects that one adapter has failed. You need only one bypass switch for the whole daisy chain of adapters in the node rather than a bypass switch for each individual adapter.

Planning Guide 5-19

Page 82: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesPlanning a Shared IBM SSA Disk Subsystem Installation

5-2

Bypass Cards in the 7133, Models 020 and 600 Disk Subsystems

Bypass cards maintain high availability when a node fails, when a node is powered off, or when the adapter(s) of a node fail. Connect the pair of ports of one bypass card into the loop that goes to and from one node. That is, connect the bypass card to only one node.

Avoid two possible conditions when a bypass card switches to bypass mode:

• Do not connect two independent loops through a bypass card. When the bypass card switches to bypass mode, you want it to reconnect the loop inside the 7133 disk subsystem, rather than connecting two independent loops. So both ports of the bypass card must be in the same loop.

• Dummy disks are connectors used to fill out the disk drive slots in a 7133 disk subsystem so the SSA loop can continue unbroken. Make sure that when a bypass card switches to bypass mode, it connects no more than three dummy disks consecutively in the same loop. Put the disks next to the bypass cards and dummy disks between real disks.

Configuring to Minimize Single Points of Failure

To minimize single points of failure, consider the following points:

• Use logical volume mirroring and place logical volume mirror copies on separate disks and in separate loops using separate adapters. In addition, it is a good idea to mirror between the front row and the back row of disks or between disk subsystems.

• The bypass card itself can be single point of failure. Two ways to avoid this are:

• With one loop: Put two bypass cards into a loop connecting to each node.

• With two loops: Do logical volume mirroring to disks in a second loop. Set each loop to go through a separate bypass card to each node.

• Set the bypass cards to forced inline mode for the following configurations:

• When connecting multiple 7133 disk subsystems.

• When the disk drives in one 7133 Model 010 or Model 600 are not all connected to the same SSA loop. In this type of configuration, forced inline mode removes the risk of a fault condition, namely, that a shift to bypass mode might cause the disk drives of different loops to be connected.

Configuring for Optimal Performance

The following guidelines can help you configure your system for optimal performance:

• With multiple nodes and SSA domains:

• A node and the disks it accesses make up an SSA domain. For configurations containing shared disk drives and multiple nodes, you need to minimize the path length from each node to the disk drives it accesses. Measure the path length by the number of disk drives and adapters in the path. Each device has to receive and forward the packet of data.

• With multiple adapters in a loop, put the disks near the closest adapter and make that the one that accesses the disks. In effect, try to keep I/O traffic within the SSA domain. Although any host can access any disk it is best to minimize I/O traffic crossing over to other domains.

• When multiple hosts are in a loop, set up the volume groups so that a node uses the closest disks. This prevents one node’s I/O from interfering with another’s.

0 Planning Guide

Page 83: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk Devices9333 and SSA Disk Fencing in Concurrent Access Clusters

p,

age es to

y for The

le , rent revent

mitting s of

nce

col of

• Distribute read and write operations evenly throughout the loops.

• Distribute disks evenly among the loops.

• Download microcode when you replace hardware.

• To ensure that everything works correctly, install the latest filesets, fixes and microcode for your disk subsystem.

• With SSA disk subsystems and concurrent access:

• Once you have decided to use concurrent mode, you may want to plan to be able to replace a failed drive in an SSA concurrent access volume group. To enable this capability, you must assign unique non-zero node numbers on each node of the cluster.

• If you specify the use of SSA disk fencing in your concurrent resource group, HACMP assigns the node numbers when you synchronize the resources.

• If you don’t specify the use of SSA disk fencing in your concurrent resource grouassign the node numbers with:

chdev -l ssar -a node_number=x

where x is the number to assign to that node. Then reboot the system.

Testing

Test all loop scenarios thoroughly, especially in multiple-node loops. Test for loop break(failure of one or more adapters). Test bypass cards for power loss in adapters and nodverify that they follow configuration guidelines.

RAID Enabled 7133

Certain storage adapters are available with the RS/6000 which provide RAID 5 capabilitthe 7133. See your IBM product documentation for more information on these adapters.presence of this adapter will affect how HACMP behaves in specific configurations. SeePlanning Shared LVM Components for more information.

9333 and SSA Disk Fencing in Concurrent Access Clusters

Preventing data integrity problems that can result from the loss of TCP/IP network communication is especially important in concurrent access configurations where multipnodes have simultaneous access to a shared disk. Chapter 4, Planning Serial Networksdescribes using HACMP-specific serial networks to prevent partitioned clusters. Concuraccess configurations using 9333 or SSA disk subsystems can also use disk fencing to pdata integrity problems that can occur in partitioned clusters.

The 9333 and SSA disk subsystems include fence registers (one per disk) capable of peror disabling access by each of the eight possible connections. Fencing provides a meanpreventing uncoordinated disk access by one or more nodes.

The 9333 and SSA hardware have a fencing command for automatically updating the feregisters. This command provides a tie-breaking function within the controller for nodes attempting to update the same fence register independently. A compare-and-swap proto

Planning Guide 5-21

Page 84: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk Devices9333 and SSA Disk Fencing in Concurrent Access Clusters

5-2

isters. ons.

roups ed as

the fence command requires that each node provide both the current and the desired contents of the fence register. If competing nodes attempt to update a register at about the same time, the first succeeds, but the second fails because it does not know the revised contents.

9333 and SSA Disk Fencing Implementation

The HACMP software manages the contents of the fence registers. At cluster configuration, the fence registers for each shared disk are set to allow access for the designated nodes. As cluster membership changes as nodes enter and leave, the event scripts call the cl_9333fence or the cl_ssa_fence utility to update the contents of the fence register. If the fencing command succeeds, the script continues normal processing. If the operation fails, the script exits with a failure status, causing the cluster to go into reconfiguration.

The HACMP for AIX Installation Guide describes enabling 9333 or SSA disk fencing for an HACMP cluster. The HACMP for AIX Administration Guide provides additional information about disk fencing.

Disk Fencing with SSA Disks in Concurrent Mode

You can only use SSA disk fencing under these conditions:

• Only disks contained in concurrent mode volume groups will be fenced.

• You configure all nodes of the cluster to have access to these disks and to use disk fencing.

• All resource groups with the disk fencing attribute enabled must be concurrent access resource groups.

• Concurrent access resource groups must contain all nodes in the cluster.

The purpose of SSA disk fencing is to provide a safety lockout mechanism for protecting shared SSA disk resources in the event that one or more cluster nodes become isolated from the rest of the cluster.

Concurrent mode disk fencing works as follows:

• The first node up in the cluster fences out all other nodes of the cluster from access to the disks of the concurrent access volume group(s) for which fencing is enabled, by changing the fence registers of these disks.

• When a node joins a cluster, the active nodes in the cluster allow the joining node access by changing the fence registers of all disks participating in fencing with the joining node.

• When a node leaves the cluster, regardless of how it leaves, the remaining nodes that share access to a disk with the departed node should fence out the departed node as soon as possible.

• If a node is the last to leave a cluster, whether the shutdown is forced or graceful, it clears the fence registers to allow access by all nodes. Of course, if the last node stops unexpectedly (is powered off or crashes, for example), it doesn’t clear the fence regIn this case, you must manually clear the fence registers using the proper SMIT opti

Enabling SSA Disk Fencing

The process of enabling SSA disk fencing for a concurrent resource group requires thatall volume groups containing SSA disks on cluster nodes must be varied off and the cluster must be down when the cluster resources are synchronized. Note that this means all volume gcontaining any of the SSA disks whether concurrent or non-concurrent, whether configur

2 Planning Guide

Page 85: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesCompleting the Disk Worksheets

part of the cluster or not, must be varied off for the disk fencing enabling process to succeed during the synchronization of cluster resources. If these conditions are not met, you will have to reboot the nodes to enable fencing.

The enabling process takes place on each cluster node: as follows

1. Assign a node_number to the ssar which matches the node_id of the node in the HACMP configuration. This means that any node_numbers, that were set prior to enabling disk fencing for purposes of replacing a drive or C-SPOC concurrent LVM functions, will be changed for disk fencing operations. The other operations will not be affected by this node_number change.

2. First remove, then remake all hdisks, pdisks, ssa adapter, and tmssa devices of the SSA disk subsystem seen by the node, thus picking up the node_number for use in the fence register of each disk.

This process is repeated each time cluster resources are synchronized while the cluster is down.

Disk Fencing and Dynamic Reconfiguration

When a node is added to the cluster through dynamic reconfiguration while cluster nodes are up, the disk fencing enabling process is performed on the added node only, during the synchronizing of topology. Any node_numbers that were set prior to enabling disk fencing for purposes of replacing a drive or C-SPOC concurrent LVM functions will be changed for disk fencing operations. Therefore, when initially setting SSA disk fencing in a resource group, the resources must be synchronized while the cluster is down. The other operations will not be affected by this node_number change.

Benefits of Disk Fencing

Disk fencing provides the following benefits to concurrent access clusters:

• It enhances data security by preventing nodes that are not active members of a cluster from modifying data on a shared disk. By managing the fence registers, the HACMP software can ensure that only the designated nodes within a cluster have access to shared 9333 and SSA disks.

• It enhances data reliability by assuring that competing nodes do not compromise the integrity of shared data. By managing the fence registers HACMP can prevent uncoordinated disk management by partitioned clusters. In a partitioned cluster, communication failures lead separate sets of cluster nodes to believe they are the only active nodes in the cluster. Each set of nodes attempts to take over the shared disk, leading to race conditions. The disk fencing tie-breaking mechanism arbitrates race conditions, ensuring that only one set of nodes gains access to the disk.

Completing the Disk Worksheets After determining the disk storage technology you will include in your cluster, complete all of the appropriate worksheets from the following list:

• Shared SCSI-2 Differential or Differential Fast/Wide Disk Worksheet

• Shared SCSI Disk Array Worksheet

• Shared IBM 9333 Serial Disk Worksheet

Planning Guide 5-23

Page 86: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesCompleting the Disk Worksheets

5-2

• Shared IBM Serial Storage Architecture Disk Subsystems Worksheet

Completing the Shared SCSI-2 Disk Worksheet

Complete a worksheet in Appendix A for each shared SCSI bus.

1. Enter the cluster ID and the Cluster name in the appropriate fields. This information was determined in Chapter 2, Drawing the Cluster Diagram.

2. Check the appropriate field for the type of SCSI-2 bus.

3. Fill in the host and adapter information including the node name, the number of the slot in which the disk adapter is installed and the logical name of the adapter, such as scsi0. AIX assigns the logical name when the adapter is configured.

4. Determine the SCSI IDs for all the devices connected to the SCSI bus.

5. The IBM SCSI-2 Differential High Performance Fast/Wide adapter cannot be assigned SCSI IDs 0, 1, or 2.

6. Record information about the Disk drives available over the bus, including the logical device name of the disk on every node. (This name, an hdisk name, is assigned by AIX when the device is configured and may vary on each node.)

Completing the Shared SCSI-2 Disk Array Worksheet

Complete a worksheet in Appendix A for each shared SCSI disk array.

1. Enter the cluster ID and the Cluster name in the appropriate fields. This information was determined in Chapter 2, Drawing the Cluster Diagram.

2. Fill in the host and adapter information including the node name, the number of the slot in which the disk adapter is installed and the logical name of the adapter, such as scsi0. AIX assigns the logical name when the adapter is configured.

3. Assign SCSI IDs for all the devices connected to the SCSI bus. For disk arrays, the controller on the disk array are assigned the SCSI ID.

4. Record information about the LUNs configured on the disk array.

5. Record the logical device name AIX assigned to the array controllers when it was configured. If you have configured an IBM RAIDiant disk array, you can optionally configure the REACT software that configures a pseudo-device called a Disk Array Router.

Completing the IBM 9333 Serial Disk Worksheet

Complete a worksheet in Appendix A for each shared 9133 configuration.

1. Enter the cluster ID and the Cluster name in the appropriate fields. This information was determined in Chapter 2, Drawing the Cluster Diagram.

2. Fill in the host and adapter information including the node name, the number of the slot in which the disk adapter is installed, and the controller name and number per drawer/desk model.

3. Determine the logical device names for all the devices connected to the shared drives.

4 Planning Guide

Page 87: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesAdding the Disk Configuration to the Cluster Diagram

Completing the IBM Serial Storage Architecture Disk Subsystems Worksheet

Complete a worksheet in Appendix A for each shared SSA configuration.

1. Enter the cluster ID and the Cluster name in the appropriate fields. This information was determined in Chapter 2, Drawing the Cluster Diagram.

2. Fill in the host and adapter information including the node name, the SSA adapter label, and the number of the slot in which the disk adapter is installed. Include dual-port number of the connection. This will be needed to make the loop connection clear.

3. Determine the SCSI IDs for all the devices connected to the SCSI bus.

Adding the Disk Configuration to the Cluster DiagramOnce you have chosen a disk technology, draw a diagram that shows the shared disk configuration. Be sure to include adapters and cables in the diagram. You may be able to expand on the cluster diagram you began in Chapter 2, Drawing the Cluster Diagram, or you may need to have a separate diagram (for the sake of clarity) that shows the shared disk configuration for the cluster.

Where You Go From HereYou have now planned your shared disk configuration. The next step is to plan the shared volume groups for your cluster. This step is described in the following chapter.

Planning Guide 5-25

Page 88: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared Disk DevicesWhere You Go From Here

5-2

6 Planning Guide
Page 89: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsPrerequisites

r a re.

does e

od of al of no ly of

s, a d tions

tems

ss the BM e

6ca r

Chapter 6 Planning Shared LVM Components

This chapter describes planning shared volume groups for an HACMP cluster.

PrerequisitesBefore reading this chapter, you must have decided which method of shared disk access—non-concurrent or concurrent—you are going to use. Read the HACMP for AIX Concepts and Facilities guide and Chapter 2, Drawing the Cluster Diagram, in this guide fodiscussion of the shared disk access methods supported by the HACMP for AIX softwa

This chapter discusses LVM issues as they relate to the HACMP for AIX environment. It not provide an exhaustive discussion of LVM concepts and facilities in general. Read thchapters on logical volumes, filesystems, paging spaces, and backups in the AIX Version 4 System Management Guide: Operating System and Devices for more information on the AIX LVM.

OverviewPlanning shared LVM components for an HACMP cluster differs depending on the methshared disk access and the type of shared disk device. This discussion assumes the gosingle point of failure. Mirroring should always be used, especially where the power suppa controller is an issue.

Planning for Non-Concurrent Access

Non-concurrent access configurations typically use journaled filesystems. (In some casedatabase application running in a non-concurrent environment may bypass the journalefilesystem and access the raw logical volume directly.) Non-concurrent access configurathat use SCSI disks, including IBM 7137 Disk Arrays, or the IBM 9333 serial disk subsyscan use AIX LVM mirroring.

Non-concurrent access configurations that use IBM 7135-110, 7135-210 RAIDiant DiskArrays, IBM 2105-B09, 2105-100 Versatile Storage Servers, or the 7133 SSA with RAIDenabled do not use LVM mirroring. These systems provide their own data redundancy.

Planning for Concurrent Access

Concurrent access configurations do not support journaled filesystems. Concurrent acceconfigurations that use IBM 7131-405 and 7133 SSA, 9333 serial disk subsystems, and7137 Disk Array should use LVM mirroring. Concurrent access configurations that use I7135-110, 7135-210 RAIDiant Disk Arrays or IBM 2105-B09, 2105-100 Versatile StoragServers do not use LVM mirroring. Instead, the these systems provide their own data redundancy.

Planning Guide 6-1

Page 90: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsLVM Components in the HACMP for AIX Environment

6-2

This chapter presents information necessary for both methods of shared disk access and points out differences where applicable. After presenting various planning considerations and guidelines, this chapter provides instruction for completing the shared LVM components worksheets for the appropriate method of shared disk access.

TaskGuide for Creating Shared Volume Groups

The TaskGuide is a graphical interface that can assist you in creating shared volume groups and adding nodes to existing volume groups. Its series of panels and online help screens guide you through each step of the process. The TaskGuide helps prevent errors, as it does not allow you to take any steps that conflict with the existing cluster configuration.

The TaskGuide for creating shared volume groups was introduced in HACMP 4.3.0. In version 4.4, the TaskGuide has two enhancements: it creates a JFS log automatically after creating a non-concurrent shared volume group, as you would need to do manually when creating a shared volume group without the TaskGuide. In addition, the TaskGuide now displays the physical location of available disks.

Note that you may still need to rename and mirror the JFS log after creating the shared volume group, as discussed on page 6-7.

For information about how to start and use the TaskGuide, see the HACMP for AIX Installation Guide, Chapter 6, Defining Shared LVM Components.

LVM Components in the HACMP for AIX EnvironmentThe LVM controls disk resources by mapping data between physical and logical storage. Physical storage refers to the actual location of data on a disk. Logical storage controls how data is made available to the user. Logical storage can be discontiguous, expanded, or replicated, and can span multiple physical disks. These features provide improved availability of data.

The LVM organizes data into the following components:

• Physical volumes

• Volume groups

• Logical volumes

• Filesystems

Considerations for each component as it relates to planning an HACMP cluster are discussed below.

Planning Guide

Page 91: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsLVM Components in the HACMP for AIX Environment

m that numbers n in

disk

Physical Volumes

A physical volume is a single physical disk. The physical volume is partitioned to provide AIX with a way of managing how data is mapped to the volume. The following figure shows how the physical partitions within a physical volume are conventionally diagrammed:

Physical Partitions on a Physical Volume

hdisk Numbers

Physical volumes are known in the AIX operating system by sequential hdisk numbers assigned when the system boots. For example, /dev/hdisk0 identifies the first physical volume in the system, /dev/hdisk1 identifies the second physical volume in the system, and so on.

When sharing a disk in an HACMP cluster, the nodes sharing the disk each assign an hdisk number to that disk. These hdisk numbers may not match, but refer to the same physical volume. For example, each node may have a different number of internal disks, or the disks may have changed since AIX was installed.

The HACMP for AIX software does not require that the hdisk numbers match across nodes (although a system is easier to manage if they do). In situations where the hdisk numbers must differ, make sure that you understand each node’s view of the shared disks. Draw a diagraindicates the hdisk numbers that each node assigns to the shared disks and record these on the appropriate volume group worksheets in Appendix A, Planning Worksheets. Whedoubt, use the hdisk’s PVID to verify its identity on a shared bus.

Use the lspv command to get the PVIDs.

Volume Groups

A volume group is a set of physical volumes that AIX treats as a contiguous, addressableregion. You can place from 1 to 32 physical volumes in the same volume group.

Planning Guide 6-3

Page 92: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsLVM Components in the HACMP for AIX Environment

6-4

ss does must

AIX r

d

The following figure shows a volume group of three physical volumes:

A Volume Group of Three Physical Volumes

Shared Volume Groups

In the HACMP for AIX environment, a shared volume group is a volume group that resides entirely on the external disks shared by the cluster nodes. Do not include an internal disk in a shared volume group, since it cannot be accessed by other nodes. If you include an internal disk in a shared volume group, the varyonvg command fails.

The name of a shared volume group must be unique. It cannot conflict with the name of an existing volume group on any node in the cluster.

In non-concurrent access environments, a shared volume group can be varied on by only one node at a time. In concurrent access environments, multiple nodes can have the volume group varied on at the same time.

As a general rule, the shared volume groups in an HACMP cluster should not be activated (varied on) automatically at system boot, but by cluster event scripts.

Note: In a non-concurrent access configuration, each volume group that has filesystems residing on it has a log logical volume (jfslog) that must also have a unique name.

Logical Volumes

A logical volume is a set of logical partitions that AIX makes available as a single storage unit—that is, the logical view of a disk. A logical partition is the logical view of a physicalpartition. Logical partitions may be mapped to one, two, or three physical partitions to implement mirroring.

In the HACMP for AIX environment, logical volumes can be used to support a journaledfilesystem (non-concurrent access) or a raw device (concurrent access). Concurrent accenot support filesystems. Databases and applications in concurrent access environmentsaccess raw logical volumes (for example, /dev/rsharedlv).

Shared Logical Volumes

A shared logical volume must have a unique name within an HACMP cluster. By default,assigns a name to any logical volume that is created as part of a journaled filesystem (foexample, lv01). If you rely on the system-generated logical volume name, this name coul

Planning Guide

Page 93: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsLVM Mirroring

. The .

s a

l disks rder to

logical

ing isk ere is in

cause the import to fail when you attempt to import the volume group containing the logical volume into another node’s ODM structure, especially if that volume group already existsHACMP for AIX Installation Guide describes how to change the name of a logical volume

Filesystems

A filesystem is written to a single logical volume. Ordinarily, you organize a set of files afilesystem for convenience and speed in managing data.

Shared Filesystems

In the HACMP for AIX system, a shared filesystem is a journaled filesystem that resides entirely in a shared logical volume.

For non-concurrent access, you want to plan shared filesystems to be placed on externashared by cluster nodes. Data resides in filesystems on these external shared disks in obe made highly available.

For concurrent access, you cannot use journaled filesystems. Instead, you must use rawvolumes.

LVM Mirroring

Note: This section does not apply to the IBM 7135-110,7135-210 RAIDiant Disk Arrays, IBM 2105-B09, 2105-100 Versatile Storage Servers or 7133 SSA with RAID enabled, which provide their own data redundancy.

LVM mirroring provides the ability to specify more than one copy of a physical partition. Usthe LVM mirroring facility increases the availability of the data in your system. When a dfails and its physical partitions become unavailable, you still have access to the data if tha mirror on an available disk. The LVM performs mirroring within the logical volume. Withan HACMP cluster, you mirror:

• Logical volume data in a shared volume group (non-concurrent and concurrent access)

• Log logical volume data for each shared volume group with filesystems (non-concurrent access).

Mirroring Physical Partitions

For a logical volume, you allocate one, two, or three copies of the physical partition that contains data. This allocation lets you mirror data, which improves the availability of the logical volume. If a copy is lost due to an error, the other unaffected copies are accessed, and AIX continues processing with an accurate copy. After access is restored to the failed physical partition, AIX resynchronizes the contents (data) of the physical partition with the contents (data) of a consistent mirror copy.

Planning Guide 6-5

Page 94: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsLVM Mirroring

6-6

The following figure shows a logical volume composed of two logical partitions with three mirrored copies. In the diagram, each logical partition maps to three physical partitions. Each physical partition should be designated to reside on a separate physical volume within a single volume group. This provides the maximum number of alternative paths to the mirror copies and, therefore, the greatest availability.

A Logical Volume of Two Logical Partitions with Three Mirrored Copies

The mirrored copies are transparent, meaning that you cannot isolate one of these copies. For example, if a user deletes a file from a logical volume with multiple copies, the deleted file is gone from all copies of the logical volume.

Using mirrored copies improves the availability of data on your cluster. The following considerations also improve data availability:

• Allocating three copies in a logical partition provides greater protection than allocating one or two copies.

• Allocating the copies of a logical partition on different physical volumes provides greater protection than allocating the copies on the same physical volume.

• Allocating the copies of a logical partition on different adapters provides greater protection than allocating the copies on a single adapter.

Keep in mind that anything that improves availability may increase the time necessary for write operations. Nevertheless, using mirrored copies spanning multiple disks (on separate power supplies) together with multiple adapters ensures that no disk is a single point of failure for your cluster.

Planning Guide

Page 95: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsQuorum

Mirroring Journal Logs

This section applies to non-concurrent access configurations, which support journaled filesystems.

AIX uses journaling for its filesystems. In general, this means that the internal state of a filesystem at startup (in terms of the block list and free list) is the same state as at shutdown. In practical terms, this means that when AIX starts up, the extent of any file corruption can be no worse than at shutdown.

Each volume group contains a jfslog, which is itself a logical volume. This log typically resides on a different physical disk in the volume group than the journaled filesystem. If access to that disk is lost, however, changes to filesystems after that point are in jeopardy.

To avoid the possibility of that physical disk being a single point of failure, you can specify mirrored copies of each jfslog. Place these copies on separate physical volumes.

Quorum

Note: This section does not apply to the IBM 7135-110,7135-210 RAIDiant Disk Arrays or IBM 2105-B09, 2105-100 Versatile Storage Servers, or 7133 SSA with RAID enabled, which provide their own data redundancy.

Quorum is a feature of the AIX LVM that determines whether or not a volume group can be placed online, using the varyonvg command, and whether or not it can remain online after a failure of one or more of the physical volumes in the volume group.

Each physical volume in a volume group has a Volume Group Descriptor Area (VGDA) and a Volume Group Status Area (VGSA).

Quorum at Vary On

When a volume group is brought online using the varyonvg command, VGDA and VGSA data structures are examined. If more than half of the copies are readable and identical in content, quorum is achieved and the varyonvg command succeeds. If exactly half the copies are available, as with two of four, quorum is not achieved and the varyonvg command fails.

VGDA Describes the physical volumes (PVs) and logical volumes (LVs) that make up a volume group and maps logical partitions to physical partitions. The varyonvg command reads information from this area.

VGSA Maintains the status of all physical volumes and physical partitions in the volume group. It stores information regarding whether a physical partition is potentially inconsistent (stale) with mirror copies on other physical partitions, or is consistent or synchronized with its mirror copies. Proper functioning of LVM mirroring relies upon the availability and accuracy of the VGSA data.

Planning Guide 6-7

Page 96: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsQuorum

6-8

ysical hould

sical

e

Quorum after Vary On

If a write to a physical volume fails, the VGSAs on the other physical volumes within the volume group are updated to indicate that one physical volume has failed. As long as more than half of all VGDAs and VGSAs can be written, quorum is maintained and the volume group remains varied on. If exactly half or less than half of the VGDAs and VGSAs are inaccessible, quorum is lost, the volume group is varied off, and its data becomes unavailable.

Keep in mind that a volume group can be varied on or remain varied on with one or more of the physical volumes unavailable. However, data contained on the missing physical volume will not be accessible unless the data is replicated using LVM mirroring and a mirror copy of the data is still available on another physical volume. Maintaining quorum without mirroring does not guarantee that all data contained in a volume group is available.

Quorum has nothing to do with the availability of mirrored data. It is possible to have failures that result in loss of all copies of a logical volume, yet the volume group remains varied on because a quorum of VGDAs/VGSAs are still accessible.

Disabling and Enabling Quorum

Quorum checking is enabled by default. Quorum checking can be disabled using the chvg -Qn vgname command, or by using the smit chvg fastpath.

Quorum Enabled

With quorum enabled, more than half of the physical volumes must be available and the VGDA and VGSA data structures must be identical before a volume group can be varied on with the varyonvg command.

With quorum enabled, a volume group will be forced offline if one or more disk failures cause a majority of the physical volumes to be unavailable. Having three or more disks in a volume group avoids a loss of quorum in the event of a single disk failure.

Quorum Disabled

With quorum disabled, all the physical volumes in the volume group must be available and the VGDA data structures must be identical for the varyonvg command to succeed. With quorum disabled, a volume group will remain varied on until the last physical volume in the volume group becomes unavailable. This section summarizes the effect quorum has on the availability of a volume group.

Forcing a VaryonA volume group with quorum disabled and one or more physical volumes unavailable can be “forced” to vary on by using the -f flag with the varyonvg command. Forcing a varyon with missing disk resources could cause unpredictable results, including a reducevg of the phvolume from the volume group. Forcing a varyon should be an overt (manual) action and sonly be performed with a complete understanding of the risks involved.

The HACMP for AIX software assumes that a volume group is not degraded and all phyvolumes are available when the varyonvg command is issued at startup or when a volume group resource is taken over during a fallover. The cluster event scripts provided with th

Planning Guide

Page 97: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsQuorum

e level ctual ilures ume e of

t go

se of le

d two ored

pply ume rum

data.

isk.

HACMP for AIX software do not “force” varyon with the -f flag, which could cause unpredictable results. For this reason, modifying the cluster event scripts to use the -f flag is strongly discouraged.

Quorum in Non-Concurrent Access Configurations

While specific scenarios can be constructed where quorum protection does provide somof protection against data corruption and loss of availability, quorum provides very little aprotection in non-concurrent access configurations. In fact, enabling quorum may mask faby allowing a volume group to varyon with missing resources. Also, designing logical volconfiguration for no single point of failure with quorum enabled may require the purchasadditional hardware. Although these facts are true, you must keep in mind that disablingquorum can result in subsequent loss of disks—after varying on the volume group—thaundetected.

Often it is not practical to configure disk resources as shown in the following figure becauthe expense. Take, for example, a cluster that requires 8 GB of disk storage (4 GB doubmirrored). This requirement could be met with two IBM 9333 or 9334 disk subsystems andisk adapters in each node. For data availability reasons, logical volumes would be mirracross disk subsystems.

Quorum in Non-Concurrent HACMP for AIX Configurations

With quorum enabled, the failure of a single adapter, cable, or disk subsystem power suwould cause exactly half the disks to be inaccessible. Quorum would be lost and the volgroup varied off even though a copy of all mirrored logical volumes is still available. Onesolution is to turn off quorum checking for the volume group. The trade-off is that, with quodisabled, all physical volumes must be available for the varyonvg command to succeed.

Using a Quorum Buster Disk

You may want to mirror the data between two 7133 drawers. Be careful to set up the configuration so that no adapter or enclosure is a single point of failure for access to theIn order to avoid the problem mentioned above, where a power failure results in the LVMvarying off the volume group (if quorum is enabled), consider using a “quorum buster” d

adapter adapter

Disk 1Disk 2Disk 3Disk 4

Disk 5Disk 6Disk 7Disk 8

Node A Node B

adapter adapter

Planning Guide 6-9

Page 98: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsQuorum

6-1

m d the

oes not ere two olume

ling iple r data

point disk

figure.

sks in access of data ns to s

one

Add a single additional disk to the volume group, on a separate power and FRU boundary from either of the mirrors of the data. This disk contains no data, it simply serves as a “quorubuster” so that if one enclosure fails or connectivity, to it is lost, quorum is maintained andata remains available on the other enclosure.

Using the quorum buster disk also ensures that, in the rare case where a clean fallover doccur, only one node has access to the quorum disk and thus you avoid the scenario whnodes battle over ownership of a shared disk. Only one node will be able to varyon the vgroup and access the data.

Quorum in Concurrent Access Configurations

Quorum must be enabled for an HACMP for AIX concurrent access configuration. Disabquorum could result in data corruption. Any concurrent access configuration where multfailures could result in no common shared disk between cluster nodes has the potential focorruption or inconsistency.

Take, for example a cluster with two sets of SSA disk subsystems configured for no singleof failure. In this configuration, logical volumes are mirrored across subsystems and eachsubsystem is connected to each node with separate adapters as shown in the following

An IBM SSA Concurrent Access Configuration

If multiple failures result in a communications loss between each node and one set of disuch a way that Node A can access subsystem 1 but not subsystem 2, and Node B cansubsystem 2 but not subsystem 1, both nodes continue to operate on the same baselinefrom the mirror copy they can access. However, each node will cease to see modificatiodata on disk that the other node makes. The result is that the data diverges and becomeinconsistent between nodes.

On the other hand, if quorum protection is enabled, the communications failure results inor both nodes varying off the volume group. Although this is a harsh action as far as theapplication is concerned, data consistency is not compromised.

0 Planning Guide

Page 99: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsUsing NFS with HACMP

Using NFS with HACMPIn version 4.4, functionality previously available in the High Availability for Network File System for AIX (HANFS for AIX) product has been added to the basic HACMP architecture. The following enhancements are included:

• Reliable NFS server capability that allows a backup processor to recover current NFS activity should the primary NFS server fail, preserving the locks on NFS filesystems and dupcache. This functionality is available for 2-node clusters only.

• Ability to specify a network for NFS mounting.

• Ability to define NFS exports and mounts at the directory level.

• Ability to specify export options for NFS-exported directories and filesystems.

In order for NFS to work as expected on an HACMP cluster, you must be aware of certain configuration issues, so you can plan accordingly:

• Creating shared volume groups

• Exporting NFS filesystems

• NFS Mounting and Fallover.

The HACMP for AIX scripts have only minimal NFS support. You may need to modify them to handle your particular configuration. The following sections contain some suggestions for planning for a variety of issues.

Reliable NFS Server Capability

An HACMP two-node cluster now takes advantage of AIX extensions to the standard NFS functionality that enable it to handle duplicate requests correctly and restore lock state during NFS server fallover and reintegration. This support was previously only available in the HANFS feature. More detail can be found in the /usr/lpp/cluster/doc/release_notes.

Creating Shared Volume Groups

When creating shared volume groups, normally you can leave the Major Number field blank and let the system provide a default for you. However, if you are using NFS, all nodes in your cluster must have identical major numbers, as HACMP uses the major number as part of the file handle to uniquely identify a Network Filesystem.

In the event of node failure, NFS clients attached to an HACMP cluster operate exactly the way they do when a standard NFS server fails and reboots. If the major numbers are not the same, when another cluster node takes over the filesystem and re-exports the filesystem, the client application will not recover, since the filesystem exported by the node will appear to be different from the one exported by the failed node.

Planning Guide 6-11

Page 100: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsUsing NFS with HACMP

6-1

NFS Exporting Filesystems and Directories

The process of NFS-exporting filesystems and directories in HACMP for AIX is different from that in AIX. Remember the following when planning for NFS-exporting in HACMP:

• Specifying Filesystems and Directories to NFS Export: In AIX, you specify filesystems and directories to NFS-export by using the smit mknfsexp command (which creates the /etc/exports file). In HACMP for AIX, you specify filesystems and directories to NFS-export by including them in a resource group in the HACMP SMIT NFS Filesystems/Directories to export field.

• Specifying Export Options for NFS Exported Filesystems and Directories: If you want to specify special options for NFS-exporting in HACMP, you can create a /usr/sbin/cluster/etc/exports file. This file has the same format as the regular AIX /etc/exports file.

Note: Use of this alternate exports file is optional. HACMP checks the /usr/sbin/cluster/etc/exports file when NFS-exporting a filesystem or directory. If there is an entry for the filesystem or directory in this file, HACMP will use the options listed. If the filesystem or directory for NFS-export is not listed in the file, or, if the alternate file does not exist, the filesystem or directory will be NFS-exported with the default option of root access for all cluster nodes.

NFS Mounting and Fallover

For HACMP for AIX and NFS to work properly together, you must be aware of the following mount issues.

• To NFS mount, a resource group must be configured with IPAT.

• If you want to use the Reliable NFS Server capability that preserves NFS locks and the dupcache in two-node clusters, the IPAT adapter for the resource group must be configured to use Hardware Address Takeover.

Cascading Takeover with Cross Mounted NFS Filesystems

This section describes how to set up cascading resource groups with cross mounted NFS filesystems so that NFS filesystems are handled gracefully during takeover and reintegration.

Note: Only cascading resource groups support automatic NFS mounting across servers during fallover. Rotating resource groups do not provide this support. Instead, you must use additional post events or perform NFS mounting using normal AIX routines.

Creating NFS Mount Points on Clients

A mount point is required in order to mount a filesystem via NFS. In a cascading resource group, all the nodes in the resource group will NFS mount the filesystem; thus you must create a mount point on each node in the resource group. On each of these nodes, create a mount point by executing the following command:

mkdir /mountpoint

2 Planning Guide

Page 101: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsUsing NFS with HACMP

ount

once

ill not p falls t it. ver.) red.

used ou can

en int.”

nner as tems,

’s e

s. This

B_rg.

where mountpoint is the name of the local mountpoint over which the remote filesystem will be mounted.

Setting Up NFS Mount Point Different from Local Mount PointHACMP handles NFS mounting in cascading resource groups as follows: The node that currently owns the resource group will mount the filesystem over the filesystem’s local mpoint, and this node will NFS export the filesystem. All the nodes in the resource group (including the current owner of the group) will NFS mount the filesystem over a differentmount point. Therefore the owner of the group will have the filesystem mounted twice - as a local mount, and once as an NFS mount.

Since IPAT is used in resource groups that have NFS mounted filesystems, the nodes wunmount and remount NFS filesystems in the event of a fallover. When the resource grouover to a new node, the acquiring node will locally mount the filesystem and NFS-expor(The NFS mounted filesystem will be temporarily unavailable to cluster nodes during falloAs soon as the new node acquires the IPAT label, access to the NFS filesystem is resto

All applications must reference the filesystem through the NFS mount. If the applicationsare dependent upon always referencing the filesystem by the same mount point name, ychange the mount point for the local filesystem mount (for example, change it to mount point_local, and use the previous local mount point as the new NFS mount point.

In the Change/Show Resources/Attributes for a Resource Group SMIT screen, the Filesystem to NFS Mount field must specify both mount points. Put the nfs mount point, ththe local mount point, separating the two with a semicolon: for example “nfspoint;localpoIf there are more entries, separate them with a space:

nfspoint1;local1 nfspoint2;local2

If there are nested mount points, the nfs mount points should be nested in the same mathe local mount points so that they match up properly. When cross mounting NFS filesysyou must also set the “filesystems mounted before IP configured” field of the Resource Group to true.

Server-to-Server NFS Cross Mounting: Example

HACMP/ES allows you to configure a cluster so that servers can NFS-mount each otherfilesystems. Configuring cascading resource groups allows the Cluster Manager to decidwhich node should take over a failed resource, based on priority and node availability.

Ensure that the shared volume groups have the same major number on the server nodeallows the clients to re-establish the NFS-mount transparently after the takeover.

In the example cluster shown below, you have two resource groups, NodeA_rg and NodeThese resource groups are defined in SMIT as follows:

Resource Group NodeA_rg

Participating node names Node A Node B

Filesystems/afs (filesystems to be locally mounted by node currently owning the resource group)

Planning Guide 6-13

Page 102: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsUsing NFS with HACMP

6-1

The filesystem you want the local node (Node A) in this resource group to locally mount and export is /afs, on Node A.You want the remote node (Node B) in this resource group to NFS-mount /afs, from Node A.

Setting up your cascading resource groups like this ensures the expected default server-to-server NFS behavior described above. On reintegration, /afs is passed back to Node A, locally mounted and exported. Node B mounts it via NFS again.

When the cluster as originally defined is up and running on both nodes, the filesystems are mounted as shown:

When Node A fails, Node B uses the cl_nfskill utility to close open files in Node A:/afs, unmounts it, mounts it locally, and re-exports it to waiting clients.

Filesystems to export/afs (Filesystem to NFS-export by node currently owning resource group. Filesystem is subset of filesystem listed above.)

Filesystems to NFS mount

/mountpointa;/afs (Filesystems/directories to be NFS-mounted by all nodes in the resource group. First value is NFS mount point; second value is local mount point)

Resource Group NodeB_rg

Participating node names Node B Node A

Filesystems /bfs

Filesystems to export /bfs

Filesystems to NFS mount /mountpointb;/bfs

4 Planning Guide

Page 103: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsPlanning Summary

After takeover, Node B has:

• /bfs locally mounted

• /bfs NFS-exported

• /afs locally mounted

• /afs NFS-exported

• a_svc:/afs NFS mounted over /mountpointa

• b_svc:/bfs NFS mounted over /mountpointb

See the man page in /usr/sbin/cluster/events/utils for information about the usage and syntax for the cl_nfskill command.

Caveats about Node Names and NFS

In the configuration described above the node name is used as the NFS hostname for the mount. This can fail if the node name is not a legitimate TCP/IP adapter label.

To avoid this problem do one of the following:

• Ensure that node name and the service adapter label are the same on each node in the cluster; or

• Alias the node name to the service adapter label in the /etc/hosts file.

Planning SummaryConsider the following guidelines as you plan shared LVM components:

• In general, planning for logical volumes concerns the availability of your data. However, creating logical volume copies is not a substitute for regularly scheduled backups. Backups protect against loss of data regardless of cause; logical volume copies protect against loss of data from physical access failure.

• All operating system files should reside in the root volume group (rootvg) and all user data should reside outside that group. This makes updating or reinstalling the operating system and backing up data more manageable.

• Volume groups that contain at least three physical volumes provide the maximum availability when implementing mirroring.

• When using copies, each physical volume containing a copy should get its power from a separate source. If one power source fails, separate power sources maintain the no-single-point-of-failure objective.

• Consider quorum issues when laying out a volume group. Quorum must be enabled on concurrent volume groups. With quorum enabled, a two-disk non-concurrent volume group puts you at risk for losing quorum and data access. Either build three-disk volume groups or disable quorum.

• Consider NFS filesystems issues as you plan the distribution of your resource groups.

• Keep in mind the cluster configurations that you have designed. A node whose resources are not taken over should not own critical volume groups.

Planning Guide 6-15

Page 104: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsCompleting the Shared LVM Components Worksheets

6-1

Completing the Shared LVM Components WorksheetsYou can now fill out a set of worksheets that help you plan the physical and logical storage for your cluster. Refer to the completed worksheets when you define the shared LVM components and the cluster resource configuration following the instructions in the HACMP for AIX Installation Guide.

Appendix A, Planning Worksheets, has blank copies of the worksheets discussed in the following sections. Photocopy these worksheets before continuing.

For instructions on entering your LVM data in the online planning worksheets, see Appendix B, Using the Online Cluster Planning Worksheet Program.

Non-Concurrent Access Worksheets

Complete the following worksheets to plan the volume groups and filesystems for a non-concurrent access configuration:

• Non-Shared Volume Group Worksheet (Non-Concurrent Access)

• Shared Volume Group/Filesystem Worksheet (Non-Concurrent Access)

• NFS-Exported Filesystem Worksheet

Completing the Non-Shared Volume Group Worksheet (Non-Concurrent Access)

For each node in the cluster, complete a Non-Shared Volume Group Worksheet (Non-Concurrent Access) for each volume group residing on a local (non-shared) disk:

1. Fill in the node name in the Node Name field.

2. Record the name of the volume group in the Volume Group Name field.

3. List the device names of the physical volumes comprising the volume group in the Physical Volumes field.

In the remaining sections of the worksheet, enter the following information for each logical volume in the volume group. Use additional sheets if necessary.

4. Enter the name of the logical volume in the Logical Volume Name field.

5. If you are using LVM mirroring, indicate the number of logical partition copies (mirrors) in the Number of Copies of Logical Partition field. You can specify one or two copies (in addition to the original logical partition, for a total of three).

6. If you are using LVM mirroring, specify whether each copy will be on a separate physical volume in the On Separate Physical Volumes? field.

7. Record the full-path mount point of the filesystem in the Filesystem Mount Point field.

8. Record the size of the filesystem in 512-byte blocks in the Size field.

6 Planning Guide

Page 105: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsCompleting the Shared LVM Components Worksheets

Completing the Shared Volume Group/Filesystem Worksheet (Non-Concurrent Access)

Fill out a Shared Volume Group/Filesystem Worksheet (Non-Concurrent Access) for each volume group that will reside on the shared disks. You need a separate worksheet for each shared volume group, so make sufficient copies of the worksheet before you begin.

1. Fill in the name of each node in the cluster in the Node Names field.

2. Record the name of the shared volume group in the Shared Volume Group Name field.

3. Leave the Major Number field blank for now. You will enter a value in this field when you address NFS issues in the following chapter.

4. Record the name of the log logical volume (jfslog) in the Log Logical Volume Name field.

5. Pencil-in the planned physical volumes in the Physical Volumes field. You will enter exact values in this field after you have installed the disks following the instructions in the HACMP for AIX Installation Guide.

In the remaining sections of the worksheet, enter the following information for each logical volume in the volume group. Use additional sheets as necessary.

6. Enter the name of the logical volume in the Logical Volume Name field.

7. If you are using LVM mirroring, indicate the number of logical partition copies (mirrors) in the Number of Copies of Logical Partition field. You can specify that you want one or two copies (in addition to the original logical partition, for a total of three). Note: This step does not apply to the IBM 7135 RAIDiant Disk Array, IBM 2105 Versatile Storage Server, or 7133 SSA with RAID enabled.

8. If you are using LVM mirroring, specify whether each copy will be on a separate physical volume in the On Separate Physical Volumes? field. Note: This step does not apply to the IBM 7135 RAIDiant Disk Array, IBM 2105 Versatile Storage Server, or 7133 SSA with RAID enabled.

9. Record the full-path mount point of the filesystem in the Filesystem Mount Point field.

10. Record the size of the filesystem in 512-byte blocks in the Size field.

Completing the NFS-Exported Filesystem Worksheet

Complete an NFS-Exported Filesystem or Directory Worksheet for filesystems or directories to be NFS-exported from a node. The information you provide will be used to update the /usr/sbin/cluster/etc/exports file.

1. Record the name of the resource group from which the filesystems or directories will be NFS exported in the Resource Group field.

2. In the Network for NFS Mount field record the preferred network to NFS mount the filesystems or directories.

3. In the Filesystem Mounted Before IP Configured field, specify true if you want the takeover of filesystems to occur before the takeover of IP address(es). Specify false for the IP address(es) to be taken over first.

4. Record the full pathname of the filesystem or directory to be exported in the Exported Directory field.

Planning Guide 6-17

Page 106: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsCompleting the Shared LVM Components Worksheets

6-1

n the

5. (optional) Record the export options you want to assign the directories and/or filesystems to be NFS exported. Refer to the exports man page for a full list of export options.

6. Repeat steps 4 and 5 for each filesystem or directory to be exported.

Concurrent Access Worksheets

Complete the following worksheets to plan the volume groups for a concurrent access configuration:

• Non-Shared Volume Group Worksheet (Concurrent Access)

• Shared Volume Group Worksheet (Concurrent Access)

Completing the Non-Shared Volume Group Worksheet (Concurrent Access)

For each node, complete a Non-Shared Volume Group Worksheet (Concurrent Access) for each volume group that resides on a local (non-shared) disk:

1. Fill in the node name in the Node Name field.

2. Record the name of the volume group in the Volume Group Name field.

3. List the device names of the physical volumes that comprise the volume group in the Physical Volumes field.

In the remaining sections of the worksheet, enter the following information for each logical volume in the volume group. Use additional sheets if necessary.

4. Enter the name of the logical volume in the Logical Volume Name field.

5. If you are using LVM mirroring, indicate the number of logical partition copies (mirrors) in the Number of Copies of Logical Partition field. You can specify one or two copies (in addition to the original logical volume, for a total of three).

6. If you are using LVM mirroring, specify whether each copy will be on a separate physical volume in the On Separate Physical Volumes? field.

7. Record the full-path mount point of the filesystem in the Filesystem Mount Point field.

8. Record the size of the filesystem in 512-byte blocks in the Size field.

Completing the Shared Volume Group Worksheet (Concurrent Access)

Fill out a Shared Volume Group Worksheet (Concurrent Access) for each volume group that will reside on the shared disks. You need a separate worksheet for each shared volume group, so make sufficient copies of the worksheet before you begin.

If you plan to create concurrent volume groups on SSA disk subsystems you must assign unique non-zero node numbers on each node of the cluster. Then you can use the procedure for replacing a failed drive in an SSA concurrent access volume group.

If you specify the use of SSA disk fencing in your concurrent resource group, HACMP assigns the node numbers when you synchronize the resources.

If you don’t specify the use of SSA disk fencing in your concurrent resource group, assignode numbers with:

chdev -l ssar -a node_number=x

8 Planning Guide

Page 107: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsWhere You Go From Here

where x is the number to assign to that node. Then reboot the system.

1. Fill in the name of each node in the cluster in the Node Names field.

2. Record the name of the shared volume group in the Shared Volume Group Name field.

3. Pencil-in the planned physical volumes in the Physical Volumes field. You will enter exact values in this field after you have installed the disks following the instructions in the HACMP for AIX Installation Guide.

In the remaining sections of the worksheet, enter the following information for each logical volume in the volume group. Use additional sheets as necessary.

4. Enter the name of the logical volume in the Logical Volume Name field.

5. Indicate the number of logical partition copies (mirrors) in the Number of Copies of Logical Partition field. You can specify one or two copies (in addition to the original logical partition, for a total of three). Note: This step does not apply to the IBM 7135 RAIDiant Disk Array, IBM 2105 Versatile Storage Server, or 7133 SSA with RAID enabled.

6. Specify whether each copy will be on a separate physical volume in the On Separate Physical Volumes? field. Note: This step does not apply to the IBM 7135 RAIDiant Disk Array, IBM 2105 Versatile Storage Server, or 7133 SSA with RAID enabled.

Where You Go From HereYou have now planned the shared LVM components for your cluster. Use this information when you define the volume groups, logical volumes, and filesystems during the install.

Planning Guide 6-19

Page 108: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Shared LVM ComponentsWhere You Go From Here

6-2

0 Planning Guide
Page 109: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsPrerequisites

7g in

Chapter 7 Planning Applications, Application Servers, and Resource Groups

This chapter describes how to plan a cluster around mission-critical applications. It includes recommendations you must consider to further plan for application servers and resource groups within an HACMP cluster. It also describes how certain applications such as AIX Connections and AIX Fast Connect are integrated with HACMP for easier configuration as highly available resources.

PrerequisitesComplete the planning steps in the previous chapters before reading this chapter.

OverviewThe central purpose for combining nodes in a cluster is to provide a highly available environment for mission-critical applications. For example, an HACMP cluster could run a database server program which services client applications. The clients send queries to the server program that responds to their requests by accessing a database, stored on a shared external disk.

Planning for these applications requires that you be aware of their location within the cluster, and that you provide a solution that enables them to be handled correctly should a node fail. In an HACMP for AIX cluster, these critical applications can be a single point of failure. To ensure the availability of these applications, the node configured to take over the resources of the node leaving the cluster should also restart these applications so that they remain available to client processes.

In this chapter, you complete planning worksheets that help you plan for applications, application servers, applications integrated with HACMP, and resource groups. Later, when following the instructions in the HACMP for AIX Installation Guide to configure these resources, refer to your completed worksheets.

Application Servers

To put the application under HACMP control, you create an application server cluster resource that associates a user-defined name with the names of specially written scripts to start and stop the application. By defining an application server, HACMP for AIX can start another instance of the application on the takeover node when a fallover occurs. For more information about creating application server resources, see Planning Applications and Application Servers on page 7-2.

Once you define the application server, you can add it to a resource group. A resource group is a set of resources (such as filesystems, volume groups, and raw disks) that you define so that the HACMP for AIX software can treat the resources as a single unit.

Planning Guide 7-1

Page 110: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsPlanning Applications and Application Servers

7-2

he s a

de 7-7.

Some Applications are Integrated with HACMP

Some applications do not require application servers, because they are already integrated with HACMP for AIX. You do not need to write additional scripts or create an application server for these to be made highly available under HACMP.

• AIX Connections software enables you to share files, printers, applications, and other resources between AIX workstations and PC and Mac clients. AIX Connections allows you to take advantage of AIX’s multi-user and multi-tasking facilities, scalability, file and record locking features, and other security features with clients on other platforms. TAIX Connections application is integrated with HACMP so that you can configure it aresource in your HACMP cluster, making the protocols handled by AIX Connections—IPX/SPX, Net BEUI, and AppleTalk—highly available in the event of noor adapter failure. For more information, see Planning for AIX Connections on page

• AIX Fast Connect allows client PCs running Windows, DOS, and OS/2 operating systems to request files and print services from an AIX server. Fast Connect supports the transport protocol NetBIOS over TCP/IP. You can configure AIX Fast Connect resources using the SMIT interface. See Planning for AIX Fast Connect on page 7-4.

• Communications Server for AIX (CS/AIX) is also integrated with HACMP, allowing you to designate Data Link Control (DLC) profile(s) and their associated objects as highly available resources. See Planning CS/AIX Communications Links on page 7-9 for more information.

In addition, the integration of these applications means the clverify utility verifies the correctness and consistency of your AIX Fast Connect, AIX Connections, or CS/AIX configuration.

Planning Applications and Application ServersAn application is itself a potential point of failure. To prevent a failure, you must have a thorough understanding of the application and how it behaves in a uniprocessor and multi-processor environment. Do not make assumptions about the application’s performance under adverse conditions. The following guidelines can help to ensure that your applications are serviced correctly within an HACMP cluster environment:

• Lay out the application and its data so that only the data resides on shared external disks. This arrangement not only prevents software license violations, but it also simplifies failure recovery. The Application Worksheet in Appendix A, Planning Worksheets, can help you plan for each application and its data. See Chapter 5, Planning Shared Disk Devices, for more information on this topic.

• Write robust scripts to both start up and shut down the application on the cluster nodes. The startup script especially must be able to recover the application from an abnormal termination, such as a power failure. You should verify that it runs properly in a uniprocessor environment before including the HACMP for AIX software. Be sure to include the start and stop resources on both the Application Worksheet and the Application Server Worksheet in Appendix A, Planning Worksheets. You will use this information as you install the HACMP for AIX software.

• Some vendors require a unique license for each processor that runs an application, which means that you must license-protect the application by incorporating processor-specific information into the application when it is installed. As a result, it is possible that even

Planning Guide

Page 111: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsPlanning Applications and Application Servers

r the

e ical .

or r

sheet

though the HACMP for AIX software processes a node failure correctly, it is unable to restart the application on the fallover node because of a restriction on the number of licenses available within the cluster for that application. To avoid this problem, make sure that you have a license for each system unit in the cluster that may potentially run an application.

• Verify that the application executes successfully in a uniprocessor environment. Debugging an application in a cluster is more difficult than debugging it on a single processor.

For further discussion of steps to take to make your applications highly available, see Appendix D, Applications and HACMP.

Completing the Application Worksheet

To help you plan the applications for your cluster, you will now complete the Application Worksheet in Appendix A, Planning Worksheets. Before performing the following procedure, photocopy this worksheet; you will need a copy for each application in the cluster.

To plan for applications in your cluster:

1. Enter the application name in the Application Name field.

2. Enter information describing the application’s executable and configuration files undeDirectory/Path, Filesystem, Location, and Sharing fields. Be sure to enter the full pathname of each file.

Note: You can store the filesystem for either the executable or configuration files on either an internal or external disk device. Different situations may require you to do it one way or the other. Be aware, if you store the filesystem on the internal device, that the device will not be accessible to other nodes during a resource takeover.

3. Enter information describing the application’s data and log files under the appropriatcolumns listed in Step 2. Data and log files can be stored in a filesystem (or on a logdevice) and must be stored externally if a resource takeover is to occur successfully

4. Enter the cluster name in the Cluster Name field.

5. Enter a resource group type in the Node Relationship field. Resource group types can beeither cascading, rotating, or concurrent. See the HACMP for AIX Concepts and Facilities guide for a thorough description of resource groups.

6. In the Node fields under the Fallover Strategy heading, enter the names of nodes to participate in a resource group takeover. Then in the Strategy fields under the appropriatenode, enter a letter (or letter/number combination) indicating the takeover strategy.

For example, for a cascading configuration, use P for primary and T plus a number ftakeover, where the number indicates the takeover priority. For either a concurrent orotating fallover strategy, use C or R, respectively. See the Sample Application Workin Appendix A, Planning Worksheets, for an example.

7. In the Normal Start Commands/Procedures field, enter the names of the start command/script you created to start the application after a resource takeover.

Planning Guide 7-3

Page 112: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsPlanning for AIX Fast Connect

7-4

tion ach

he ments

stop

ups

ast ups. d d with

8. In the Verification Commands/Procedures field, enter the names of commands or procedures to use to verify that the normal start command/script executed successfully.

9. Enter a node name and its associated caveat in the appropriate fields under the Node Reintegration/Takeover Caveats heading.

10. Repeat Steps 7 through 9 for the application’s normal stop commands/procedures.

Completing the Application Server Worksheet

To help you plan the application servers for your cluster, you will now complete the ApplicaServer Worksheet in Appendix A, Planning Worksheets. Photocopy this worksheet for eapplication server in your cluster before performing the following procedure.

To plan for application servers in your cluster:

1. Enter the cluster ID in the Cluster ID field.

2. Enter the cluster name in the Cluster Name field.

3. Record a symbolic name that identifies the server in the Server Name field. For example, you could name the application server for the Image Cataloger demo imagedemo.

4. Record the full pathname of the user-defined script that starts the server in the Start Script field. Be sure to include the script’s arguments, if necessary. The script is called by tcluster event scripts. For example, you could name the start script and specify its argufor starting the imagedemo application server as follows:

/usr/sbin/cluster/utils/start_imagedemo -d mydir -a jim_svc.

where the -d option specifies the name of the directory for storing images, and the -a option specifies the service IP address (label) for the server running the demo.

5. Record the full pathname of the user-defined script that stops the server in the Stop Script field. This script is called by the cluster event scripts. For example, you could name thescript for the imagedemo application server /usr/sbin/cluster/utils/stop_imagedemo.

Complete the above steps for each application server in the cluster.

After planning the application servers, you next plan the resource groups. Resource grocombine all the resources related to an application into a single unit.

Planning for AIX Fast ConnectThe Fast Connect application is integrated with HACMP already so you can configure FConnect resources, via the SMIT interface, as highly available resources in resource groHACMP can then stop and start the Fast Connect resources when fallover, recovery, andynamic resource group migrations occur. This application does not need to be associateapplication servers or special scripts.

Planning Guide

Page 113: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsPlanning for AIX Fast Connect

Converting from AIX Connections to AIX Fast Connect

You cannot have both AIX Fast Connect and AIX Connections configured at the same time. Therefore, if you previously configured the AIX Connections application as a highly available resource, and you now wish to switch to AIX Fast Connect, you should take care to examine your AIX Connections planning and configuration information before removing it from the resource group.

Keep these points in mind when planning for conversion from AIX Connections to Fast Connect:

• Be aware AIX Fast Connect does not handle the AppleTalk and NetWare protocols that AIX Connections is able to handle. Fast Connect is primarily for connecting with clients running Windows operating systems. Fast Connect uses NetBIOS over TCP/IP.

• You will need to unconfigure any AIX Connections services before configuring AIX Fast Connect services as resources.

• Take care to note the AIX Connections services configuration, so you can make sure that AIX Fast Connect connects you to all of the files and print queues you have been connected to with AIX Connections.

For additional details and instructions on converting from AIX Connections, see the chapter on changing resources and resource groups in the HACMP for AIX Administration Guide,

Planning Considerations for Fast Connect

When planning for configuration of Fast Connect as a cluster resource in HACMP, keep the following points in mind:

• Install the Fast Connect Server on all nodes in the cluster.

• If Fast Connect printshares are to be highly available, the AIX print queue names must match for every node in the cluster.

• For cascading and rotating resource groups, assign the same netBIOS name to each node when the Fast Connect Server is installed. This action will minimize the steps needed for the client to connect to the server after fallover.

Note: Only one instance of a netBIOS name can be active at one time. For that reason, remember not to activate Fast Connect servers that are under HACMP control.

• For concurrently configured resource groups, assign different netBIOS names across nodes.

• In concurrent configurations, you should define a second, non-concurrent, resource group to control any filesystem that must be available for the Fast Connect nodes. Having a second resource group configured in a concurrent cluster keeps the AIX filesystems used by Fast Connect cross-mountable and highly available in the event of a node failure.

• As stated in the previous section, you cannot configure both Fast Connect and AIX Connections in the same resource group or on the same node.

• Fast Connect cannot be configured in a mutual takeover configuration. In other words, a node cannot participate in two Fast Connect resource groups at the same time.

Planning Guide 7-5

Page 114: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsPlanning for AIX Fast Connect

7-6

Fast Connect as a Highly Available Resource

When configured as part of a resource group, AIX Fast Connect resources are handled by HACMP as follows.

Start/Stop of Fast ConnectWhen a Fast Connect server has resources configured in HACMP, HACMP starts and stops the server during fallover, recovery, and dynamic reconfigurations and resource group migrations.

Note: The Fast Connect server must be stopped on all nodes when bringing up the cluster. This ensures that HACMP will start the Fast Connect server and handle its resources properly.

Node FailureWhen a node that owns Fast Connect resources fails, the resources become available on the takeover node. When the failed node rejoins the cluster, the resources are again available on the original node (as long as the resource policy is such that the failed node reacquires its resources).

There is no need for clients to reestablish a connection to access the Fast Connect server, as long as IP and hardware address takeover were configured and occur, and the Fast Connect server is configured with the same NetBIOS name on all nodes (for non-concurrent resource groups).

Note: For switched networks and for clients not running Clinfo, you may need to take additional steps to ensure client connections after fallover. See Chapter 9, Planning HACMP for AIX Clients for more information.

Adapter FailureWhen a service adapter running the transport protocol needed by the Fast Connect server fails, HACMP performs and adapter swap as usual, and Fast Connect establishes a connection with the new adapter. After an adapter failure, clients are temporarily unable to access shared resources such as files and printers; after the adapter swap is complete, clients can again access their resources.

Completing the Fast Connect Worksheet

Now fill out the Fast Connect worksheet found in Appendix A, Planning Worksheets, to identify the resources you will enter in SMIT when you configure Fast Connect resources as part of the resource group.

To complete the planning worksheet for Fast Connect:

1. Record the cluster ID in the Cluster ID field.

2. Record the cluster name in the Cluster Name field.

3. Record the name of the resource group that will contain the Fast Connect resources.

4. Record the nodes participating in the resource group.

5. Record the Fast Connect resources to be made highly available. These resources will be chosen from a picklist when you configure resources in the resource group in SMIT.

Planning Guide

Page 115: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsPlanning for AIX Connections

ing s

I,

ndby.

ocols

ups.

6. Record the filesystems that contain the files or directories that you want Fast Connect to share. Be sure to specify these in the Filesystems SMIT field when you configure the resource group.

See the instructions on using SMIT to configure Fast Connect services as resources in the HACMP for AIX Installation Guide.

Planning for AIX ConnectionsAIX Connections software enables you to share files, printers, applications, and other resources between AIX workstations and PC and Mac clients. AIX Connections allows you to take advantage of AIX’s multi-user and multi-tasking facilities, scalability, file and record lockfeatures, and other security features with clients on other platforms. The AIX Connectionapplication is integrated with HACMP so that you can configure it as a resource in your HACMP cluster, making the protocols handled by AIX Connections—IPX/SPX, Net BEUand AppleTalk—highly available in the event of node or adapter failure.

AIX Connections Realms and Services

Three realms are available through AIX Connections:

• NB for NetBIOS clients offers services over two transport protocols, TCP/IP and NetBEUI. This lets your AIX workstation running Connections provide file, print, terminal emulation, and other network services to client PCs running DOS, OS/2, Windows, or Windows NT.

• NW for NetWare clients offers services over IPX/SPX, so your AIX workstation can provide services to NetWare-compatible client PCs.

• AT for AppleTalk clients offers services over AppleTalk, so your AIX workstation can act as a Macintosh AppleShare server and provide services to Macintosh clients on an Ethernet or token ring network.

Planning Notes for AIX Connections• In order to configure AIX Connections services in HACMP, you must copy the AIX

Connections installation information to all nodes on which the program might be active.

• You must configure a realm’s AIX Connections to use the service adapter, not the sta

AIX Connections as a Highly Available Resource

When AIX Connections services are configured as resources, HACMP handles the protduring the following events:

Start-upHACMP starts the protocols handled by AIX Connections when it starts the resource gro

Note: If you are running NetBIOS and it is attached to other non-HACMP applications, you need to restart those applications after fallover, startup, and dynamic reconfiguration events involving an AIX Connections resource group.

Planning Guide 7-7

Page 116: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsPlanning for AIX Connections

7-8

Node Failure and RecoveryIn the event of a node failure and node start, HACMP handles the AIX Connections services listed in the resource groups just like all other resources so listed. It starts and stops them just as it does volume groups and filesystems.

After resource group takeover for node failure (not adapter failure), AIX Connections services revert to allowing new connections so that HACMP/ES can swap adapters and reconnect users to the services. This allows the reconnection to take place without interruption, and without users even noticing.

Note: If you have services previously set to reject new connections, be aware that they are not automatically reset this way on takeover. You must reset these manually.

You can include a command to reject new connections in the start_server script if you do not wish to do the manual reset after fallover. However, be aware that when services are permanently set to reject new connections, users will experience an interruption in service as they will not automatically connect to the server.

Adapter FailureIn the event of an adapter failure, HACMP moves all AIX Connections protocols in use on the failed adapter to a currently active standby adapter if one is available. See the HAMCP for AIX Installation Guide for a discussion of adapter failure considerations.

Completing the AIX Connections Worksheet

Now fill out the AIX Connections worksheet to identify the realm/service pairs you will enter in SMIT when you configure AIX Connections as a resource.

In the AIX Connections worksheet:

1. Record the cluster ID in the Cluster ID field.

2. Record the cluster name in the Cluster Name field.

3. Record the name of the resource group that will contain the AIX Connections services.

4. Record the nodes participating in this resource group.

5. Record the AIX Connections realm/service pairs to be made highly available, using the following format:

• A realm is one of the following, as described earlier: NB, NW, AT

• A service is one of the following: file, print, term, nvt, and atlw.

You assign a name to the AIX Connections service and then specify realm/service pairs using the format

<realm>/<service_name>%<service_type>

For example, to specify the NetBIOS realm for a file you have named solenb, your realm/service pair would be in the following format:

NB/solenb%file

Planning Guide

Page 117: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsPlanning CS/AIX Communications Links

For full instructions on configuring AIX Connections resources in SMIT, see the HACMP for AIX Installation Guide.

Planning CS/AIX Communications LinksAn HACMP for AIX CS/AIX communication link contains CS/AIX configuration information which is specific to a given node and network adapter. This configuration information enables an RS/6000 computer to participate in an SNA network that includes mainframes, PCs and other workstations. The CS/AIX term for this configuration information is a DLC profile. Additional CS/AIX configuration is required outside of HACMP in order to use CS/AIX. See the manuals listed in the preface for information on configuring CS/AIX.

In a non-HACMP for AIX environment, you will lose your SNA network connection if the network adapter which the DLC profiles are associated with fail; or the node on which the CS/AIX software is running fails. HACMP for AIX enables you to place the CS/AIX DLC profiles in a highly available resource group. Once you add CS/AIX DLC profiles to a resource group, takeover and recovery will happen automatically if a network adapter or node fails. See the HACMP for AIX Installation Guide for more detail on the events that occur during adapter and node fallover and recovery for highly available CS/AIX communications links.

Keep the following points in mind when planning highly available CS/AIX communication links:

• This feature is supported with the following two CS/AIX products: Communications Server for AIX Version 4.2 and eNetwork Communications Server for AIX Version 5.0.

• HACMP for AIX CS/AIX communications links are supported over Token Ring, Standard Ethernet and FDDI networks.

• HACMP for AIX supports LU 2 and LU 6.2 CS/AIX logical unit protocols.

• HACMP for AIX requires that CS/AIX be installed on all nodes where takeover might occur. You must also define CS/AIX configuration information on all nodes where resource groups containing CS/AIX configurations might become active.

Completing the CS/AIX Communications Links Worksheet

To help you plan the CS/AIX communications links for your cluster, you will now complete the CS/AIX communications links Worksheet in Appendix A, Planning Worksheets. Photocopy this worksheet for additional CS/AIX communications links as needed.

Do the following for each CS/AIX communication link in your cluster:

1. Enter the cluster ID in the Cluster ID field.

2. Enter the cluster name in the Cluster Name field.

3. Enter the communications link name in the Communications Link Name field.

4. Enter the resource group in which the communications link is defined in the Resource Group field.

5. Enter nodes participating in the resource group in the Nodes field.

6. Enter the DLC name in the DLC Name field. This is the name of an existing CS/AIX DLC profile to be made highly available.

Planning Guide 7-9

Page 118: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsPlanning Resource Groups

7-1

7. Enter the names of any CS/AIX ports to be started automatically in the Port field.

8. Enter the names of the CS/AIX link stations in the Link Station field. This field is only available for eNetwork Communications Server for AIX Version 5.0.

9. Enter the name of an application start script in the Service field. This start script starts any application layer processes that use the communication link. This field is optional.

Planning Resource GroupsThe resource group is the central element of the HACMP for AIX system; it is what is being made highly available. Therefore, planning the resource group is a major part of the planning process. You made preliminary choices about the type of resource group and the takeover priority for each node in the resource chains in Chapter 2, Drawing the Cluster Diagram. Here you review your choices in light of the subsequent planning.

The general steps in planning a resource group are:

You walk through each step in planning a resource group when you complete the Resource Group Worksheet in Appendix A, Planning Worksheets. For more information on resource groups types, see Chapter 2, Drawing the Cluster Diagram.

Guidelines

The HACMP for AIX software does not restrict the number of individual resources in a resource group or the number of resource groups in a cluster. The benefit to this approach is that it gives you flexibility in designing your cluster.

Nevertheless, as a general rule you should strive to keep your design as simple as possible. Doing so makes it easier to configure and maintain resource groups; it also makes the takeover of a resource group faster.

Guidelines for Planning Resource Groups

Follow these guidelines when planning resource groups:

• Every cluster resource must be part of a resource group. If you want a resource to be kept separate, you define a group for that resource alone. A resource group may have one or more resources defined.

• A resource may not be included in more than one resource group.

• The components of a resource group must be unique down to the physical volume. If applications access the same data, put them in the same resource group. If applications access different data, put them in different resource groups.

Step What you do...

1 Decide whether you want your cluster to use cascading, rotating, or concurrent access resource groups.

2 Define the resource chain for the resource group. A resource chain consists of the nodes assigned to participate in the takeover of a given resource group.

3 Identify the individual resources that constitute the resource group.

0 Planning Guide

Page 119: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsPlanning Resource Groups

• A rotating resource group must have a service IP label defined for it.

• A cascading resource group may or may not include a service IP label. If you want to do IP address takeover, then you must include a service label in the resource group.

• Failure to use kerberos or /.rhosts with a Cascading without Fallback resource group will result in resource group failure.

• A resource group based on cascading or rotating resources cannot include any concurrent volume groups.

• Concurrent resource groups consist only of application servers and concurrent volume groups.

Guidelines for Planning Resource Chains

Follow these guidelines when planning resource chains:

• If a cluster has more than one rotating resource group or more than one cascading resource group using IP address takeover, and therefore multiple resource chains, the node with the highest priority in one resource chain (for example, the resource chain for resource group A) cannot also have the highest priority for another resource chain (for example, the resource chain for resource group B).

• If you include the same node in more than one resource chain, make sure that it has the ability to take all resource groups simultaneously.

• HACMP limits the number of nodes participating in a Cascading without Fallback resource group to two.

Special Considerations when Planning for a CWOF Resource Group

Keep in mind the following when planning for a Cascading without Fallback resource group.

Sticky and Non-Sticky Migration in a CWOF Resource Group

DARE migration is enhanced in a cascading resource group with Cascading without Fallback set to true. A cascading resource group with CWOF set to false supports a DARE migration with the sticky option only. This means that the node to which this resource group migrates becomes the highest priority node until another DARE migration changes this (until a DARE to another node, DARE to stop, or a DARE to default). Resource groups with CWOF = true support both sticky and non-sticky DARE migrations.

Resource Group “Clumping”

A Cascading without Fallback resource group tends to remain on the node which acquires it. Some nodes may therefore host many resource groups while others have no resource groups.

You can manage the uneven distribution of resource groups, or clumping, in several ways:

• Use DARE Migration to redistribute resource groups after node failure or reintegration

• Plan resource group participation in order to minimize clumping.

• Set the Inactive Takeover flag to false in order to manage clumping during cluster start-up

Planning Guide 7-11

Page 120: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsPlanning Resource Groups

7-1

m

CWOF Resource Group Down Though Highest Priority Node is Up

In a cascading resource group with Cascading without Fallback set to true, the possibility exists that while the highest priority node is up, the resource group is down. This situation can occur if you bring the resource group down by either a graceful shutdown or a cldare stop command. Unless you bring the resource group up manually, it will remain in an inactive state. For more information on this issue, see Miscellaneous Issues on page 4-25 of the HACMP for AIX Troubleshooting Guide.

Completing the Resource Group Worksheet

You now fill out a worksheet that helps you plan the resource groups for the cluster. Appendix A, Planning Worksheets, has blank copies of the Resource Group Worksheet referenced in the following procedure. Photocopy this worksheet before continuing.

For more info, see Planning Resource Groups on page 7-10 and Using NFS with HACMP on page 6-11.

To plan for resource groups:

1. Record the cluster ID in the Cluster ID field of the Resource Group worksheet.

2. Record the cluster name in the Cluster Name field.

3. Name the resource group in the Resource Group Name field.

Use no more than 31 characters. You can use alphanumeric characters and underscores. Duplicate entries are not allowed.

4. Record the node/resource relationship (that is, the type of resource for the resource group) in the Node Relationship field. Indicate whether the resource group is cascading, rotating, or concurrent.

You made a preliminary choice about the type of resource groups you want to use in your cluster in Chapter 2, Drawing the Cluster Diagram. Review this choice in light of the subsequent planning.

5. List the names of the nodes you want to be members of the resource chain for this resource group in the Participating Node Names field. List the node names in order from highest to lowest priority. Note that priority is ignored for concurrent resource groups.

Again, review your earlier choice in light of the subsequent planning.

6. List the filesystems to include in this resource group in the Filesystems field.

7. List in the Filesystems/Directories to Export field the filesystems and/or directories in this resource group that should be NFS-exported by the node currently holding the resource. These filesystems should be a subset of the filesystems listed above. The directories for export should be contained in one of the filesystems listed above.

8. List the filesystems in the resource group that should be NFS-mounted by the nodes in the resource chain not currently holding the resource in the Filesystems/Directories to NFS-Mount field.

If you are cross mounting NFS filesystems, this field must specify both the nfs and the local mount points. Put the nfs mount point, then the local mount point, separating the two with a semicolon: for example “nfspoint;localpoint.” If there are more entries, separate thewith a space:

2 Planning Guide

Page 121: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsPlanning Resource Groups

ou rough

have

en

ecify

in the

s, r. If bove.)

.

nfspoint1;local1 nfspoint2;local2

When cross mounting NFS filesystems, you must also set the “filesystems mounted beforeIP configured” field of the Resource Group to true.

The default options used by HACMP when performing NFS mounts are “soft,intr.” If ywant hard mounts or any other options on the NFS mounts, these can be specified ththe SMIT NFS panels for “Add a File System for Mounting” (the fastpath is smit mknfsmnt). When adding these filesystems, you should be sure to choose the filesystems option to the “MOUNT now, add entry to /etc/filesystems or both?” field, and take thedefault value of no to “/etc/filesystems entry will mount the directory on system RESTART.” This will add the options you have chosen into the /etc/filesystems entry created. This entry is then read by the HACMP scripts to pick up any options you mayselected.

9. List in the Volume Groups field the shared volume groups that should be varied on whthis resource group is acquired or taken over.

If you plan on using raw logical volumes in non-concurrent mode, you only need to spthe volume group in which the raw logical volume resides in order to include the rawlogical volumes in the resource group.

10. If you plan on using an application that directly accesses raw disks, list the raw disksRaw Disks field.

11. Specify realm/service pairs for AIX Connections in the AIX Connections Services field.

12. In the AIX Fast Connect Resources field, list the AIX Fast Connect resources (fileshareprint services, etc.) that you will specify when configuring Fast Connect on the serveyou include Fast Connect fileshares, be sure to define their filesystems (see step 6 a

Note: You cannot configure both AIX Connections and Fast Connect in the same resource group.

13. List the names of the application servers to include in the resource group in the Application Servers field.

14. List the names of the CS/AIX communication links in the Highly Available Communications Links field.

15. This field applies only to cascading resource groups. Indicate in the Inactive Takeover Activated field how you want to control the initial acquisition of a resource by a node

• If Inactive Takeover is true, then the first node in the resource chain to join the cluster acquires the resource group. Subsequently, for Cascading without Fallback = false, the resource group cascades to nodes in the chain with higher priority as they join the cluster. Note that this causes an interruption in service as resource ownership transfers to the node with the higher priority.

• The default in the SMIT screen for Inactive Takeover Activated is false. In this case, the first node up acquires the resource only if it is the node with the highest priority for that resource. Each subsequent node joining the cluster acquires any resource groups for which the node has a higher priority than the other nodes currently up in the cluster. For Cascading without Fallback = false, and depending on the order in which the nodes are brought up, this option may result in resource groups cascading to higher priority nodes, causing an interruption in service. This possibility, however, is not as likely as when Inactive Takeover is set to true.

Planning Guide 7-13

Page 122: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Applications, Application Servers, and Resource GroupsWhere You Go From Here

7-1

16. This field applies only to cascading resource groups. In the Cascading without Fallback field, choose whether you would like to define the Cascading without Fallback variable as true or false.

• If Cascading without Fallback is set to false, a resource group falls back to any higher priority node when such a node joins or reintegrates into the cluster, causing an interruption in service.

• When CWOF is set to true, the resource group will not fallback to any node which joins or reintegrates into the cluster.

Complete the preceding steps for each resource group in your cluster.

Note: For information about preventative maintenance planning by migrating resource groups to other nodes, see the HACMP for AIX Administration Guide.

Where You Go From HereYou have now planned the applications, application servers, and resource groups for the cluster. You next tailor the cluster event processing for your cluster, as discussed in the following chapter.

4 Planning Guide

Page 123: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Tailoring Cluster Event ProcessingPrerequisites

rmits a

itical

sing. If to do

d that ilor tion a

vent

8

Chapter 8 Tailoring Cluster Event Processing

This chapter describes tailoring cluster event processing for your cluster.

PrerequisitesSee the section in the HACMP for AIX Concepts and Facilities guide, which explains cluster event processing in depth, before completing the planning steps in this chapter.

OverviewThe Cluster Manager’s ability to recognize a specific series of events and subevents pevery flexible customization scheme. The HACMP for AIX software provides an event customization facility that allows you to tailor cluster event processing to your site. Customizing event processing allows you to provide a highly efficient path to the most crresources in the event of a failure. The level of efficiency, however, depends on your installation.

As part of the planning process, you need to decide whether to customize event procesthe actions taken by the default scripts are sufficient for your purposes, you do not needanything further to configure events during the installation process.

If you do decide to tailor event processing to your environment, it is strongly recommendeyou use the HACMP for AIX event customization facility described in this chapter. If you taevent processing, you must register user-defined scripts with HACMP during the installaprocess. The HACMP for AIX Installation Guide describes configuring event processing for cluster.

Warning: If necessary, you can modify the default event scripts or write your own, but these practices are strongly discouraged. They make maintaining, upgrading, and troubleshooting an HACMP cluster much more difficult.

You cannot define additional cluster events.

Customizing Cluster Event ProcessingYou can define multiple pre- and post-events for each of the events defined in the HACMPeODM class.

The event customization facility includes the following features:

• Event notification

• Pre- and post-event processing

• Event recovery and retry.

Planning Guide 8-1

Page 124: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Tailoring Cluster Event ProcessingCustomizing Cluster Event Processing

8-2

Complete customization for an event includes a notification to the system administrator (before and after event processing), and user-defined commands or scripts (before and after event processing) as shown in the following figure:

A Tailored Event

Event Notification

You can specify a notify command that sends mail to indicate that an event is about to happen (or has just occurred), and that an event script succeeded or failed. For example, a site may want to use a network_down notification event to inform system administrators that traffic may have to be re-routed. Afterwards, you can use a network_up notification event to inform system administrators that traffic can again be serviced through the restored network.

Event notification in an HACMP cluster can also be done using pre- and post-event scripts.

Pre- and Post-Event Scripts

You can specify commands or user-defined scripts that execute before and after the Cluster Manager calls an event script.

For example, you can specify one or more pre-event scripts that run before the node_down event script is processed. When the Cluster Manager recognizes that a remote node is down, it first processes these user-defined scripts. One such script might designate that a message be sent to all users to indicate that performance may be affected (while adapters are swapped, while application servers are stopped and restarted). Following the node_down event script, a post-processing event script for a network_up notification might be included to broadcast a message to all users that a certain system is now available at another network address.

The following scenarios are other examples of situations in which pre- and post-event processing are useful:

• If a node_down event occurs, site-specific actions might dictate that a pre-event script for the start_server subevent script be used. This script could notify users on the server about to takeover for the downed application server that performance may vary, or that they should seek alternate systems for certain applications they might be interested in.

• Due to a network being down, a custom installation might be able to re-route traffic through other machines by creating new IP routes. The network_up and network_up_complete event scripts could reverse the procedure, ensuring that the proper routes exist after all networks are functioning.

• A site might want to initiate a graceful takeover as a post event if a network has failed on the local node (but otherwise the network is functioning).

Planning Guide

Page 125: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Tailoring Cluster Event ProcessingCompleting the Cluster Event Worksheet

If you plan to create pre- or post-event scripts for your cluster, be aware that your scripts will be passed the same parameters used by the HACMP for AIX event script you specify. The first parameter passed will be the event name; other parameters passed are those used by the HACMP for AIX event. For example, for the network_down event, the following parameters are passed to your script: the event name, network_down, followed by the name of the node and network experiencing the failure.

All HACMP for AIX event scripts are maintained in the /usr/sbin/cluster/events directory. The parameters passed to each script are listed in the event script headers.

Event Recovery and Retry

You can specify a command that attempts to recover from an event script failure. If the recovery command succeeds and the retry count for the event script is greater than zero, the event script is rerun. You can also specify the number of times to attempt to execute the recovery command.

For example, a recovery command could include the retry of unmounting a filesystem after logging a user off, and making sure no one was currently accessing the filesystem.

If a condition that affects the processing of a given event on a cluster is identified, such as a timing issue, you can insert a recovery command with a retry count high enough to be sure to cover for the problem.

See Appendix E, 7x24 Maintenance in the HACMP for AIX Administration Guide for more tips on planning pre- and post-event scripts.

Completing the Cluster Event WorksheetYou now fill out a worksheet that helps you plan the cluster event processing for your cluster. Appendix A, Planning Worksheets, has blank copies of the worksheet referenced in the procedure below. Make photocopies of this worksheet to record all your cluster event scripts.

For instructions on entering cluster event data in the online worksheets, see Appendix B, Using the Online Cluster Planning Worksheet Program.

Completing the Cluster Event Worksheet

Complete the following steps to plan the customized processing for a specific cluster event. Enter values in the fields only as necessary.

1. Record the cluster ID in the Cluster ID field.

2. Record the cluster name in the Cluster Name field.

3. Record the cluster event description in the Cluster Event Description field.

4. Record the full pathname of the cluster event method in the Cluster Event Method field.

5. Fill in the name of the cluster event in the Cluster Event Name field.

6. Fill in the name of the event script in the Event Command field.

7. Record the full pathname of the event notification script in the Notify Command field.

8. Record the name of the pre-event script in the Pre-Event Command field.

Planning Guide 8-3

Page 126: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Tailoring Cluster Event ProcessingWhere You Go From Here

8-4

9. Record the name of the post-event script in the Post-Event Command field.

10. Record the full pathname of the event retry script in the Event Recovery Command field.

11. Indicate the number of times to retry in the Recovery Counter field.

Repeat steps 3 through 11 for each event you plan to customize.

Where You Go From HereYou have now planned the customized event processing for your cluster. Next, you address issues relating to cluster clients, as described in the following chapter.

Planning Guide

Page 127: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning HACMP for AIX ClientsPrerequisites

with

9

Chapter 9 Planning HACMP for AIX Clients

This chapter discusses planning considerations for HACMP for AIX clients.

PrerequisitesComplete the planning steps in the previous chapters before reading this chapter.

OverviewClients are end-user devices that can access the nodes in an HACMP cluster. In this stage of the planning process, you evaluate the cluster from the point of view of the clients. You need to consider the following:

• Different types of clients (computers and terminal servers)

• Clients with and without the Client Information Program (Clinfo)

• Network components (routers, bridges, gateways).

Different Types of Clients: Computers and Terminal Servers

Clients may include a variety of hardware and software from different vendors. In order to maintain connectivity with the HACMP cluster, you must consider the following issues.

Client Application Systems

All clients should run Clinfo if possible. If you have hardware other than RS/6000 nodes in the configuration, you may want to port Clinfo to those platforms. Clinfo source code is provided as part of the HACMP for AIX release.

You need to think about what applications are running on these clients. Who are the users? Is it required or appropriate for users to receive a message when cluster events affect their system?

NFS Servers

Issues related to NFS are discussed in Chapter 6, Planning Shared LVM Components.

Terminal Servers

If you plan to use a terminal server on the local area network, consider the following when choosing the hardware:

• Can you update the terminal server’s ARP cache? The terminal server must complythe TCP/IP protocol, including telnet.

Planning Guide 9-1

Page 128: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning HACMP for AIX ClientsClients Running Clinfo

9-2

the

to

uster d the onents

ot need tion

r

f a at you this is r the imilar

• Is the terminal server programmable, or does it need manual intervention when a cluster event happens?

• Can you download a control file from the cluster node to the terminal server that updates or handles cluster events’ effects on clients?

If your terminal server does not meet these operating requirements, you should choose hardware address swapping option when configuring the cluster environment.

Clients Running ClinfoThe Clinfo program calls the /usr/sbin/cluster/etc/clinfo.rc script whenever a network or nodeevent occurs. By default, this action updates the system’s ARP cache to reflect changesnetwork addresses. You can customize this script if further action is desired.

Reconnecting to the Cluster

Clients running the Clinfo daemon will be able to reconnect to the cluster quickly after a clevent. If you have hardware other than RS/6000s, SPs, or SMPs between the cluster anclients, you must make sure that you can update the ARP cache of those network compafter a cluster event occurs.

If you configure the cluster to swap hardware addresses as well as IP addresses, you do nto be concerned about updating the ARP cache. You simply must be aware that this opcauses a longer delay for the users.

Tailoring the clinfo.rc Script

For clients running Clinfo, you need to decide whether to tailor the /usr/sbin/cluster/etc/clinfo.rc script to do more than update the ARP cache when a clusteevent occurs. See the HACMP for AIX Installation Guide for more information. Also see the HACMP for AIX Programming Client Applications Guide for a sample tailored clinfo.rc script.

Network Components and Clients Not Running ClinfoIf you have configured the network so that clients attach to networks on the other side orouter, bridge, or gateway rather than to the cluster’s local networks, you must be sure thcan update the ARP cache of those network components after a cluster event occurs. Ifnot possible, be sure to use hardware address swapping when you configure the clusteenvironment. If hardware address swapping is not enabled, you will also need to updateARP cache of non-Clinfo clients on the same subnet and physical network segment in a smanner. Keep in mind that Clinfo can monitor a maximum of eight clusters.

See the HACMP for AIX Installation Guide for more information about updating non-Clinfo clients.

Planning Guide

Page 129: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning HACMP for AIX ClientsWhere You Go From Here

Where You Go From HereThis chapter concludes the planning process. You can now begin to install the HACMP for AIX software. See the HACMP for AIX Installation Guide for detailed instructions on installing the HACMP for AIX software.

Planning Guide 9-3

Page 130: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning HACMP for AIX ClientsWhere You Go From Here

9-4

Planning Guide
Page 131: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Worksheets

A

Appendix A Planning Worksheets

This appendix contains the following worksheets:

Worksheet Purpose Page

TCP/IP Networks Use this worksheet to record the TCP/IP network topology for a cluster. Complete one worksheet per cluster.

A-3

TCP/IP Network Adapter Use this worksheet to record the TCP/IP network adapters connected to each node. You need a separate worksheet for each node defined in the cluster, so begin by photocopying a worksheet for each node and filling in a node name on each worksheet.

A-5

Serial Networks Use this worksheet to record the serial network topology for a cluster. Complete one worksheet per cluster.

A-7

Serial Network Adapter Use this worksheet to record the serial network adapters connected to each node. You need a separate worksheet for each node defined in the cluster, so begin by photocopying a worksheet for each node and filling in the node name on each worksheet.

A-9

Shared SCSI-2 Differential or Differential Fast/Wide Disks

Use this worksheet to record the shared SCSI-2 Differential or Differential Fast/Wide disk configuration for the cluster. Complete a separate worksheet for each shared bus.

A-11

Shared IBM SCSI Disk Arrays

Use this worksheet to record the shared IBM SCSI disk array configurations for the cluster. Complete a separate worksheet for each shared SCSI bus.

A-13

Shared IBM 9333 Serial Disk

Use this worksheet to record the IBM 9333 shared disk configuration for the cluster. Complete a separate worksheet for each shared serial bus.

A-15

Shared IBM Serial Storage Architecture Disk Subsystem

Use this worksheet to record the IBM 7131-405 or 7133 SSA shared disk configuration for the cluster.

A-17

Non-Shared Volume Group (Non-Concurrent Access)

Use this worksheet to record the volume groups and filesystems that reside on a node’s internal disks in a non-concurrent access configuration. You need a separate worksheet for each volume group, so begin by photocopying a worksheet for each volume group and filling in a node name on each worksheet.

A-19

Shared Volume Group/Filesystem (Non-Concurrent Access)

Use this worksheet to record the shared volume groups and filesystems in a non-concurrent access configuration. You need a separate worksheet for each shared volume group, so begin by photocopying a worksheet for each volume group and filling in the names of the nodes sharing the volume group on each worksheet.

A-21

Planning Guide A-1

Page 132: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning Worksheets

A-

35

-41

NFS-Exported Filesystem/Directory

Use this worksheet to record the filesystems and directories NFS-exported by a node in a non-concurrent access configuration. You need a separate worksheet for each node defined in the cluster, so begin by photocopying a worksheet for each node and filling in a node name on each worksheet.

A-23

Non-Shared Volume Group (Concurrent Access)

Use this worksheet to record the volume groups and filesystems that reside on a node’s internal disks in a concurrent access configuration. You need a separate worksheet for each volume group, so begin by photocopying a worksheet for each volume group and filling in a node name on each worksheet.

A-25

Shared Volume Group (Concurrent Access)

Use this worksheet to record the shared volume groups and filesystems in a concurrent access configuration. You need a separate worksheet for each shared volume group, so begin by photocopying a worksheet for each volume group and filling in the names of the nodes sharing the volume group on each worksheet.

A-27

Application Use these worksheets to record information about applications in the cluster.

A-29

AIX Fast Connect Use this worksheet to record the Fast Connect resources you plan to configure in resource groups

A-33

AIX Connections Use this worksheet to record realm/service pairs for AIX Connections A-

Application Server Use these worksheets to record information about application servers in the cluster.

A-37

CS/AIX Communication Links

Use these worksheets to record information about CS/AIX Communications Links in the cluster.

A-39

Resource Group Use this worksheet to record the resource groups for a cluster. A

Cluster Event Use this worksheet to record the planned customization for an HACMP cluster event.

A-43

Worksheet Purpose Page

2 Planning Guide

Page 133: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsTCP/IP Networks Worksheet

TCP/IP Networks WorksheetCluster ID ____________

Cluster Name ____________

Network Name Network Type Network Attribute Netmask Node Names

________________ ________________ ________________ ________________ ________________

________________ ________________ ________________ ________________ ________________

________________ ________________ ________________ ________________ ________________

________________ ________________ ________________ ________________ ________________

________________ ________________ ________________ ________________ ________________

________________ ________________ ________________ ________________ ________________

________________ ________________ ________________ ________________ ________________

________________ ________________ ________________ ________________ ________________

________________ ________________ ________________ ________________ ________________

Planning Guide A-3

Page 134: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample TCP/IP Networks Worksheet

A-

Sample TCP/IP Networks WorksheetCluster ID 1

Cluster Name bivalves

Network Name Network Type Network Attribute Netmask Node Names

ether1 Ethernet public 255.255.255.0 clam, mussel, oyster

token1 Token-Ring public 255.255.255.0 clam, mussel, oyster

fddi1 FDDI public 255.255.255.0 clam, mussel

socc1 SOCC private 255.255.255.0 clam, mussel

atm1 ATM private 255.255.255.0 clam, mussel

4 Planning Guide

Page 135: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsTCP/IP Network Adapter Worksheet

TCP/IP Network Adapter WorksheetNode Name __________

Note: The SMIT Add an Adapter screen displays an Adapter Identifier field that correlates with the Adapter IP Address field on this worksheet.

Also, entries in the Adapter HW Address field should refer to the locally administered address (LAA), which applies only to the service adapter.

Interface Name

Adapter IP Label

Adapter Function

Adapter IP Address

Network Name

Network Attribute

Boot Address Adapter HW Address

________________________________________________________________________________________

________________________________________________________________________________________

________________________________________________________________________________________

________________________________________________________________________________________

________________________________________________________________________________________

________________________________________________________________________________________

________________________________________________________________________________________

________________________________________________________________________________________

________________________________________________________________________________________

Planning Guide A-5

Page 136: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample TCP/IP Network Adapter Worksheet

A-6

s

Sample TCP/IP Network Adapter WorksheetNode Name nodea

Note: The SMIT Add an Adapter screen displays an Adapter Identifier field that correlates with the Adapter IP Address field on this worksheet.

Also, entries in the Adapter HW Address field should refer to the locally administered address (LAA), which applies only to the service adapter.

Interface Name

Adapter IP Label

Adapter Function

Adapter IP Address

Network Name

Network Attribute

Boot Address Adapter HW Addres

len0 nodea_en0 service 100.10.1.10 ether1 public 0x08005a7a7610

en0 nodea_boot1 boot 100.10.1.74 ether1 public

en1 nodea_en1 standby 100.10.11.11 ether1 public

tr0 nodea_tr0 service 100.10.2.20 token1 public 0x42005aa8b57b

tr0 nodea_boot2 boot 100.10.2.84 token1 public

fi0 nodea_fi0 service 100.10.3.30 fddi1 public

sl0 nodea_sl0 service 100.10.5.50 slip1 public

css0 nodea_svc service hps1 private

css0 nodea_boot3 boot hps1 private

at0 nodea_at0 service 100.10.7.10 atm1 private 0x0020481a396500

at0 nodea_boot1 boot 100.10.7.74 atm1 private

Planning Guide

Page 137: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSerial Networks Worksheet

Serial Networks WorksheetCluster ID ____________

Cluster Name ____________

Note: RS232 serial lines, target mode SCSI-2 buses, and tmssa serial links do not use the TCP/IP protocol and do not require a netmask or an IP address.

Network Name Network Type Network Attribute Node Names

________________ ________________ serial ________________

________________ ________________ serial ________________

________________ ________________ serial ________________

________________ ________________ serial ________________

________________ ________________ serial ________________

________________ ________________ serial ________________

________________ ________________ serial ________________

________________ ________________ serial ________________

________________ ________________ serial ________________

Planning Guide A-7

Page 138: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Serial Networks Worksheet

A-

Sample Serial Networks WorksheetCluster ID 1

Cluster Name clus1

Note: RS232 serial lines, target mode SCSI-2 buses, and tmssa serial links do not use the TCP/IP protocol and do not require a netmask or an IP address.

Network Name Network Type Network Attribute Node Names

rs232a RS232 serial nodea, nodeb

tmscsi1 Target Mode SCSI serial nodea, nodeb

8 Planning Guide

Page 139: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSerial Network Adapter Worksheet

Serial Network Adapter WorksheetNode Name __________

Note: Serial networks do not carry TCP/IP traffic; therefore, no boot addresses, adapter identifiers (IP addresses), or adapter hardware addresses are required to maintain keepalives and control messages between nodes.

Slot Number Interface Name Adapter Label Network Name Network Attribute Adapter Function

_____________ _____________ _____________ _____________ serial service

_____________ _____________ _____________ _____________ serial service

_____________ _____________ _____________ _____________ serial service

_____________ _____________ _____________ _____________ serial service

_____________ _____________ _____________ _____________ serial service

_____________ _____________ _____________ _____________ serial service

_____________ _____________ _____________ _____________ serial service

_____________ _____________ _____________ _____________ serial service

_____________ _____________ _____________ _____________ serial service

Planning Guide A-9

Page 140: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Serial Network Adapter Worksheet

A-

Sample Serial Network Adapter WorksheetNode Name nodea

Note: Serial networks do not carry TCP/IP traffic; therefore, no boot addresses, adapter identifiers (IP addresses), or adapter hardware addresses are required to maintain keepalives and control messages between nodes.

Slot Number Interface Name Adapter Label Network Name Network Attribute Adapter Function

SS2 /dev/tty1 nodea_tty1 rs232a serial service

08 scsi2 nodea_tmscsi2 tmscsi1 serial service

10 Planning Guide

Page 141: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsShared SCSI-2 Differential or Differential Fast/Wide Disks Worksheet

Shared SCSI-2 Differential or Differential Fast/Wide Disks Worksheet

Note: Complete a separate worksheet for each shared SCSI-2 Differential bus or Differential Fast/Wide bus. Keep in mind that the IBM SCSI-2 Differential High Performance Fast/Wide adapter cannot be assigned SCSI IDs 0, 1, or 2. The SCSI-2 Differential Fast/Wide adapter cannot be assigned SCSI IDs 0 or 1.

Type of SCSI-2 Bus

Host and Adapter Information

SCSI Device IDs on Shared Bus

Shared Drives

SCSI-2 Differential ____________________ SCSI-2 Differential Fast/Wide ____________________

Node A Node B Node C Node D

Node Name ______________ ______________ ______________ ______________

Slot Number ______________ ______________ ______________ ______________

Logical Name ______________ ______________ ______________ ______________

Node A Node B Node C Node D

Adapter ______________ ______________ ______________ ______________

First Shared Drive ______________

Second Shared Drive ______________

Third Shared Drive ______________

Fourth Shared Drive ______________

Fifth Shared Drive ______________

Sixth Shared Drive ______________

Disk Size Logical Device Name

Node A Node B Node C Node D

First _____________ _____________ _____________ _____________ _____________

Second _____________ _____________ _____________ _____________ _____________

Third _____________ _____________ _____________ _____________ _____________

Fourth _____________ _____________ _____________ _____________ _____________

Fifth _____________ _____________ _____________ _____________ _____________

Sixth _____________ _____________ _____________ _____________ _____________

Planning Guide A-11

Page 142: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Shared SCSI-2 Differential or Differential Fast/Wide Disks Worksheet

A-

Sample Shared SCSI-2 Differential or Differential Fast/Wide Disks Worksheet

Note: Complete a separate worksheet for each shared SCSI-2 Differential bus or Differential Fast/Wide bus. Keep in mind that the IBM SCSI-2 Differential High Performance Fast/Wide adapter cannot be assigned SCSI IDs 0, 1, or 2. The SCSI-2 Differential Fast/Wide adapter cannot be assigned SCSI IDs 0 or 1.

Type of SCSI-2 Bus

Host and Adapter Information

SCSI Device IDs on Shared Bus

Shared Drives

SCSI-2 Differential SCSI-2 Differential Fast/Wide X

Node A Node B Node C Node D

Node Name nodea nodeb

Slot Number 7 7

Logical Name ascsi1 ascsi1

Node A Node B Node C Node D

Adapter 6 5

First Shared Drive 3

Second Shared Drive 4

Third Shared Drive 5

Fourth Shared Drive

Fifth Shared Drive

Sixth Shared Drive

Disk Size Logical Device Name

Node A Node B Node C Node D

First 670 hdisk2 hdisk2

Second 670 hdisk3 hdisk3

Third 670 hdisk4 hdisk4

Fourth

Fifth

Sixth

12 Planning Guide

Page 143: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsShared IBM SCSI Disk Arrays Worksheet

Shared IBM SCSI Disk Arrays WorksheetNote: Complete a separate worksheet for each shared SCSI disk array.

Host and Adapter Information

SCSI Device IDs on Shared Bus

Array Controller and Path Information

Node A Node B Node C Node D

Node Name ______________ ______________ ______________ ______________

Slot Number ______________ ______________ ______________ ______________

Logical Name ______________ ______________ ______________ ______________

Node A Node B Node C Node D

Adapter ______________ ______________ ______________ ______________

First Array Controller ______________ ______________ ______________ ______________

Second Array Controller ______________ ______________ ______________ ______________

Third Array Controller ______________ ______________ ______________ ______________

Fourth Array Controller ______________ ______________ ______________ ______________

Shared Drives Shared LUNs

Size RAID Level Logical Device Name

Node A Node B Node C Node D

______________ ______________ ______________ ______________ ______________ ______________

______________ ______________ ______________ ______________ ______________ ______________

______________ ______________ ______________ ______________ ______________ ______________

______________ ______________ ______________ ______________ ______________ ______________

______________ ______________ ______________ ______________ ______________ ______________

Array 1 Array 2

Array Controller Logical Name ______________ ______________

Array Controller Logical Name ______________ ______________

Disk Array Router Logical Name ______________ ______________

Planning Guide A-13

Page 144: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Shared IBM SCSI Disk Arrays Worksheet

A-

Sample Shared IBM SCSI Disk Arrays WorksheetThis sample worksheet shows an IBM 7135 RAIDiant Disk Array configuration.

Note: Complete a separate worksheet for each shared SCSI disk array.

Host and Adapter Information

SCSI Device IDs on Shared Bus

Array Controller and Path Information

Node A Node B Node C Node D

Node Name nodea nodeb

Slot Number 2 2

Logical Name ascsi1 ascsi1

Node A Node B Node C Node D

Adapter 14 15

First Array Controller 3

Second Array Controller 4

Third Array Controller

Fourth Array Controller

Shared Drives Shared LUNs

Size RAID Level Logical Device Name

Node A Node B Node C Node D

2GB 5 hdisk2 hdisk2

2GB 3 hdisk3 hdisk3

2GB 5 hdisk4 hdisk4

2GB 5 hdisk5 hdisk5

RAIDiant 1

Array Controller Logical Name dac0

Array Controller Logical Name dac1

Disk Array Router Logical Name dar0

14 Planning Guide

Page 145: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsShared IBM 9333 Serial Disk Worksheet

Shared IBM 9333 Serial Disk Worksheet

IBM 9333 Drawer/Desk Label __________

IBM 9333 Shared Drives in Node Name __________

IBM 9333 Drawer/Desk Label __________

IBM 9333 Shared Drives in Node Name __________

Node A Node B Node C Node D

Node Name ______________ ______________ ______________ ______________

Slot Number ______________ ______________ ______________ ______________

Logical Name ______________ ______________ ______________ ______________

Adapter I/O Connector ______________ ______________ ______________ ______________

Controller Logical Name ______________ ______________ ______________ ______________

Drive Size (MB) Logical Device Name

1 ______ ______________ ______________ ______________ ______________

2 ______ ______________ ______________ ______________ ______________

3 ______ ______________ ______________ ______________ ______________

4 ______ ______________ ______________ ______________ ______________

Adapter I/O Connector ______________ ______________ ______________ ______________

Controller Logical Name ______________ ______________ ______________ ______________

Drive Size (MB) Logical Device Name

1 ______ ______________ ______________ ______________ ______________

2 ______ ______________ ______________ ______________ ______________

3 ______ ______________ ______________ ______________ ______________

4 ______ ______________ ______________ ______________ ______________

Planning Guide A-15

Page 146: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample IBM 9333 Serial Disk Worksheet

A-

Sample IBM 9333 Serial Disk Worksheet

IBM 9333 Drawer/Desk Label drawer1

IBM 9333 Shared Drives in Node Name __________

IBM 9333 Drawer/Desk Label __________

IBM 9333 Shared Drives in Node Name __________

Node A Node B Node C Node D

Node Name clam mussel

Slot Number serdasda0 serdasda0

Logical Name 4 5

Adapter I/O Connector 0 1

Controller Logical Name serdasdc0 serdasdc0

Drive Size (MB) Logical Device Name

1 857 hdisk12 hdisk14

2 857 hdisk13 hdisk15

3

4

Adapter I/O Connector ______________ ______________ ______________ ______________

Controller Logical Name ______________ ______________ ______________ ______________

Drive Size (MB) Logical Device Name

1 ______ ______________ ______________ ______________ ______________

2 ______ ______________ ______________ ______________ ______________

3 ______ ______________ ______________ ______________ ______________

4 ______ ______________ ______________ ______________ ______________

16 Planning Guide

Page 147: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsShared IBM Serial Storage Architecture Disk Subsystems Worksheet

Shared IBM Serial Storage Architecture Disk Subsystems WorksheetHost and Adapter Information

SSA Logical Disk Drive

SSA Logical Disk Drive

Node A Node B Node C Node D

Node Name _____________ _____________ _____________ _____________

SSA Adapter Label _____________ _____________ _____________ _____________

Slot Number _____________ _____________ _____________ _____________

Dual-Port Number _____________ _____________ _____________ _____________

Logical Device Name

Node A Node B Node C Node D

_______________ _______________ _______________ _______________

_______________ _______________ _______________ _______________

_______________ _______________ _______________ _______________

_______________ _______________ _______________ _______________

Logical Device Name

Node A Node B Node C Node D

_______________ _______________ _______________ _______________

_______________ _______________ _______________ _______________

_______________ _______________ _______________ _______________

_______________ _______________ _______________ _______________

Planning Guide A-17

Page 148: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Shared IBM Serial Storage Architecture Disk Subsystems Worksheet

A-

Sample Shared IBM Serial Storage Architecture Disk Subsystems WorksheetHost and Adapter Information

SSA Logical Disk Drive

SSA Logical Disk Drive

Node A Node B Node C Node D

Node Name clam mussel

SSA Adapter Label ha1, ha2 ha1, ha2

Slot Number 2, 4 2, 4

Dual-Port Number a1, a2 a1, a2

Logical Device Name

Node A Node B Node C Node D

hdisk2 hdisk2

hdisk3 hdisk3

hdisk4 hdisk4

hdisk5 hdisk5

Logical Device Name

Node A Node B Node C Node D

hdisk2 hdisk2

hdisk3 hdisk3

hdisk4 hdisk4

hdisk5 hdisk5

18 Planning Guide

Page 149: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsNon-Shared Volume Group Worksheet (Non-Concurrent Access)

Non-Shared Volume Group Worksheet (Non-Concurrent Access)Node Name ________________

Volume Group Name ________________

Physical Volumes ________________

________________

________________

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Filesystem Mount Point ________________

Size (in 512-byte blocks) ________________

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Filesystem Mount Point ________________

Size (in 512-byte blocks) ________________

Planning Guide A-19

Page 150: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Non-Shared Volume Group Worksheet (Non-Concurrent Access)

A-

Sample Non-Shared Volume Group Worksheet (Non-Concurrent Access)Node Name clam

Volume Group Name localvg

Physical Volumes hdisk1

________________

________________

Logical Volume Name locallv

Number of Copies of Logical Partition 1

On Separate Physical Volumes? no

Filesystem Mount Point /localfs

Size (in 512-byte blocks) 100000

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Filesystem Mount Point ________________

Size (in 512-byte blocks) ________________

20 Planning Guide

Page 151: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsShared Volume Group/Filesystem Worksheet (Non-Concurrent Access)

Shared Volume Group/Filesystem Worksheet (Non-Concurrent Access)

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Filesystem Mount Point ________________

Size (in 512-byte blocks) ________________

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Filesystem Mount Point ________________

Size (in 512-byte blocks) ________________

Node A Node B Node C Node D

Node Names _______________ _______________ _______________ _______________

Shared Volume Group Name _______________

Major Number _______________ _______________ _______________ _______________

Log Logical Volume Name _______________

Physical Volumes _______________ _______________ _______________ _______________

_______________ _______________ _______________ _______________

_______________ _______________ _______________ _______________

Planning Guide A-21

Page 152: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Shared Volume Group/Filesystem Worksheet (Non-Concurrent Access)

A-

Sample Shared Volume Group/Filesystem Worksheet (Non-Concurrent Access)

Logical Volume Name basslv

Number of Copies of Logical Partition 3

On Separate Physical Volumes? yes

Filesystem Mount Point /bassfs

Size (in 512-byte blocks) 200000

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Filesystem Mount Point ________________

Size (in 512-byte blocks) ________________

Node A Node B Node C Node D

Node Names trout guppy

Shared Volume Group Name bassvg

Major Number 24 24

Log Logical Volume Name bassloglv

Physical Volumes hdisk6 hdisk6

hdisk7 hdisk7

hdisk13 hdisk16

22 Planning Guide

Page 153: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsNFS-Exported Filesystem or Directory Worksheet (Non-Concurrent Access)

NFS-Exported Filesystem or Directory Worksheet (Non-Concurrent Access)Resource Group __________

Network for NFS Mount __________

Filesystem Mounted Before IP Configured? __________

Filesystem/Directory to Export __________

Export Options (read-only, etc.) Refer to the exports man page for a full list of export options:

Filesystem/Directory to Export __________

Export Options (read-only, etc.) Refer to the exports man page for a full list of export options:

Filesystem/Directory to Export __________

Export Options (read-only, etc.) Refer to the exports man page for a full list of export options:

_________________________ _________________________ _________________________

_________________________ _________________________ _________________________

_________________________ _________________________ _________________________

_________________________ _________________________ _________________________

_________________________ _________________________ _________________________

_________________________ _________________________ _________________________

_________________________ _________________________ _________________________

_________________________ _________________________ _________________________

_________________________ _________________________ _________________________

Planning Guide A-23

Page 154: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample NFS-Exported Filesystem or Directory Worksheet (Non-Concurrent Access)

A-

Sample NFS-Exported Filesystem or Directory Worksheet (Non-Concurrent Access)Resource Group rg1

Network for NFS Mount tr1

Filesystem Mounted Before IP Configured? true

Filesystem/Directory to Export /fs1

Export Options (read-only, root access, etc.) Refer to the exports man page for a full list of export options:

Filesystem/Directory to Export /fs 2

Export Options (read-only, root access, etc.) Refer to the exports man page for a full list of export options:

client access:client1 _________________________ _________________________

root access: node 1, node 2 _________________________ _________________________

mode: read/write _________________________ _________________________

client access:client 2 _________________________ _________________________

root access: node 3, node 4 _________________________ _________________________

mode: read only _________________________ _________________________

24 Planning Guide

Page 155: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsNon-Shared Volume Group Worksheet (Concurrent Access)

Non-Shared Volume Group Worksheet (Concurrent Access)Node Name ________________

Volume Group Name ________________

Physical Volumes ________________

________________

________________

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Filesystem Mount Point ________________

Size (in 512-byte blocks) ________________

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Filesystem Mount Point ________________

Size (in 512-byte blocks) ________________

Planning Guide A-25

Page 156: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Non-Shared Volume Group Worksheet (Concurrent Access)

A-

Sample Non-Shared Volume Group Worksheet (Concurrent Access)Node Name clam

Volume Group Name localvg

Physical Volumes hdisk1

________________

________________

Logical Volume Name locallv

Number of Copies of Logical Partition 1

On Separate Physical Volumes? no

Filesystem Mount Point /localfs

Size (in 512-byte blocks) 100000

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Filesystem Mount Point ________________

Size (in 512-byte blocks) ________________

26 Planning Guide

Page 157: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsShared Volume Group/Filesystem Worksheet (Concurrent Access)

Shared Volume Group/Filesystem Worksheet (Concurrent Access)

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Node A Node B Node C Node D

Node Names _______________ _______________ _______________ _______________

Shared Volume Group Name _______________

Physical Volumes _______________ _______________ _______________ _______________

_______________ _______________ _______________ _______________

_______________ _______________ _______________ _______________

Planning Guide A-27

Page 158: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Shared Volume Group/Filesystem Worksheet (Concurrent Access)

A-

Sample Shared Volume Group/Filesystem Worksheet (Concurrent Access)

Logical Volume Name basslv

Number of Copies of Logical Partition 3

On Separate Physical Volumes? yes

Filesystem Mount Point /bassfs

Size (in 512-byte blocks) 200000

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Filesystem Mount Point ________________

Size (in 512-byte blocks) ________________

Logical Volume Name ________________

Number of Copies of Logical Partition ________________

On Separate Physical Volumes? ________________

Node A Node B Node C Node D

Node Names trout guppy

Shared Volume Group Name bassvg

Physical Volumes hdisk6 hdisk6

hdisk7 hdisk7

hdisk13 hdisk16

28 Planning Guide

Page 159: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsApplication Worksheet

Application WorksheetApplication Name _____________________

Normal Start Commands/Procedures:______________________________________________________________________________

______________________________________________________________________________

______________________________________________________________________________

______________________________________________________________________________

Verification Commands/Procedures:______________________________________________________________________________

______________________________________________________________________________

______________________________________________________________________________

______________________________________________________________________________

Node Reintegration/Takeover Caveats:

Key Application Files

Directory/Path Filesystem Location Sharing

Executables: _______________ _______________ _______________ ______________

Configuration Files: _______________ _______________ _______________ ______________

Data Files/Devices: _______________ _______________ _______________ ______________

Log Files/Devices: _______________ _______________ _______________ ______________

Cluster Name:

Node Relationship: (cascading/concurrent/rotating)

Fallover Strategy: (P = primary; T = takeover)

Node: _________ _________ _________ _________ _________ _________ _________ _________

Strategy: _________ _________ _________ _________ _________ _________ _________ _________

Node Reintegration/Takeover Caveats

____________ _______________________________________

____________ _______________________________________

____________ _______________________________________

____________ _______________________________________

Planning Guide A-29

Page 160: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsApplication Worksheet

A-

Normal Stop Commands/Procedures:______________________________________________________________________________

______________________________________________________________________________

______________________________________________________________________________

______________________________________________________________________________

Verification Commands/Procedures:______________________________________________________________________________

______________________________________________________________________________

______________________________________________________________________________

______________________________________________________________________________

Node Reintegration/Takeover Caveats:

____________ _______________________________________

____________ _______________________________________

____________ _______________________________________

____________ _______________________________________

____________ _______________________________________

Node Reintegration/Takeover Caveats

____________ _______________________________________

____________ _______________________________________

____________ _______________________________________

____________ _______________________________________

____________ _______________________________________

____________ _______________________________________

____________ _______________________________________

____________ _______________________________________

____________ _______________________________________

30 Planning Guide

Page 161: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Application Worksheet

Sample Application WorksheetApplication Name _____________________

Normal Start Commands/Procedures:• Verify that the app1 server group is running

• If the app1 server group is not running, as user app1_adm, execute app1 start -I One

• Verify that the app1 server is running

• If node Two is up, start (restart) app1_client on node Two

Verification Commands/Procedures:• Run the following command: lssrc -g app1

• Verify from the output that daemon1, daemon2, and daemon3 are “Active”

• Send notification if not “Active”

Node Reintegration/Takeover Caveats:

Key Application Files

Directory/Path Filesystem Location Sharing

Executables: /app1/bin /app1 internal non-shared

Configuration Files: /app1/config/one /app1/config/one external shared

Data Files/Devices: /app1lv1 NA external shared

Log Files/Devices: /app1loglv1 NA external shared

Cluster Name: tetra

Node Relationship: (cascading/concurrent/rotating)

cascading

Fallover Strategy: (P = primary; T = takeover)

Node: One Two Three Four

Strategy: P NA T1 T2

Node Reintegration/Takeover Caveats

One NA

Two NA

Three Must restart the current instance of app1 with app1start -Ione -Ithree

Four Must restart the current instance of app1 with app1start -Ione -Ifour

Planning Guide A-31

Page 162: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Application Worksheet (continued)

A-

Sample Application Worksheet (continued)Normal Stop Commands/Procedures:

• Verify that the app1 server group is running

• If the app1 server group is running, stop by app1stop as user app1_adm

• Verify that the app1 server is stopped

• If the app1 server is still up, stop individual daemons with the kill command

Verification Commands/Procedures:• Run the following command: lssrc -g app1

• Verify from the output that daemon1, daemon2, and daemon3 are “Inoperative”

Node Reintegration/Takeover Caveats:

Note: In this sample worksheet, the server portion of the application, app1, normally runs on three of the four cluster nodes: nodes One, Three, and Four. Each of the three nodes is running its own app1 instance: one, three, or four. When a node takes over an app1 instance, the takeover node must restart the application server using flags for multiple instances. Also, because Node Two within this configuration runs the client portion associated with this instance of app1, the takeover node must restart the client when the client’s server instance is restarted.

Node Reintegration/Takeover Caveats

One NA

Two May want to notify app1_client users to log off

Three Must restart the current instance of app1 with app1start -Ithree

Four Must restart the current instance of app1 with app1start -Ifour

32 Planning Guide

Page 163: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsAIX Fast Connect Worksheet

AIX Fast Connect Worksheet

Cluster ID: ______________________________________

Cluster Name: ______________________________________

Resource Group Nodes Fast Connect Resources

______________ ________________________ ___________________

___________________

___________________

___________________

___________________

___________________

___________________

___________________

___________________

Resource Group Nodes Fast Connect Resources

______________ ________________________ ___________________

___________________

___________________

___________________

___________________

___________________

___________________

___________________

___________________

Planning Guide A-33

Page 164: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample AIX Fast Connect Worksheet

A-

Sample AIX Fast Connect Worksheet

Cluster ID: 2

Cluster Name: cluster2

Resource Group Nodes Fast Connect Resources

rg1 NodeA, NodeC FS1%f%/smbtest/fs1

LPT1%p%printq

___________________

___________________

___________________

___________________

___________________

___________________

___________________

Resource Group Nodes Fast Connect Resources

rg2 Node B, Node D FS2%f%/smbtest/fs2

LPT2%p%printq

___________________

___________________

___________________

___________________

___________________

___________________

___________________

34 Planning Guide

Page 165: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsAIX Connections Worksheet

AIX Connections Worksheet

Cluster ID: _________________________________________

Cluster: _________________________________________

Resource Group NodesRealm (NB,NW,AT)

Service NameService Type (file,print,term,nvt,atls)

______________ __________________ ____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

Resource Group NodesRealm (NB,NW,AT)

Service NameService Type (file,print,term,nvt,atls)

______________ __________________ ____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

Planning Guide A-35

Page 166: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample AIX Connections Worksheet

A-

Sample AIX Connections WorksheetCluster ID: 1

Cluster Name: cluster1

Resource Group

NodesRealm (NB,NW,AT)

Service NameService Type (file,print,term,nvt,atls)

rg1 clam, mussel, oyster NB printnb print

NB term1 term

NB clamnb file

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

Resource Group NodesRealm (NB,NW,AT)

Service NameService Type (file,print,term,nvt,atls)

______________ __________________ ____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

____________ ______________ ______________

36 Planning Guide

Page 167: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsApplication Server Worksheet

Application Server WorksheetCluster ID ____________

Cluster Name ____________

Note: Use full pathnames for all user-defined scripts.

Server Name:

Start Script:

Stop Script:

Server Name:

Start Script:

Stop Script:

Server Name:

Start Script:

Stop Script:

Server Name:

Start Script:

Stop Script:

Planning Guide A-37

Page 168: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Application Server Worksheet

A-

Sample Application Server WorksheetCluster ID 1

Cluster Name clus1

Server Name: imagedemo

Start Script: /usr/es/sbin/cluster/utils/start_imagedemo

Stop Script: /usr/es/sbin/cluster/utils/stop_imagedemo

Server Name:

Start Script:

Stop Script:

Server Name:

Start Script:

Stop Script:

Server Name:

Start Script:

Stop Script:

38 Planning Guide

Page 169: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsCS/AIX Communications Links Worksheet

CS/AIX Communications Links WorksheetCluster ID _________________________

Cluster Name _________________________

.

.

Communications Link Name:

Resource Group:

Nodes:

DLC Name:

Port:

Link Station:

Service:

Communications Link Name:

Resource Group:

Nodes:

DLC Name:

Port:

Link Station:

Service:

Planning Guide A-39

Page 170: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample CS/AIX Communications Links Worksheet

A-

Sample CS/AIX Communications Links WorksheetCluster ID __1_______________________

Cluster Name __cluster1__________________

.

.

Communications Link Name:

Link1

Resource Group: rg1

Nodes: nodea, nodeb

DLC Name: profile1

Port: port1

Link Station: station1

Service: /tmp/service1.sh

Communications Link Name:

Resource Group:

Nodes:

DLC Name:

Port:

Link Station:

Service:

40 Planning Guide

Page 171: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsResource Group Worksheet

Resource Group Worksheet Cluster ID____________ Cluster Name____________

Resource Group Name ___________________________

Node Relationship ___________________________

Participating Node Names ___________________________

Filesystems ___________________________

Filesystems/Directories to Export ___________________________

Filesystems/Directories to NFS Mount ___________________________

Volume Groups ___________________________

Raw Disks ___________________________

AIX Connections Realm/Svc Pairs ___________________________

AIX Fast Connect Resources ___________________________

Application Servers ___________________________

Highly Available Communications Links ___________________________

Inactive Takeover ___________________________

Cascade without Fallback ___________________________

Planning Guide A-41

Page 172: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Resource Group Worksheet

A-

Sample Resource Group WorksheetCluster ID 1

Cluster Name clus1

Resource Group Name rotgrp1

Node Relationship rotating

Participating Node Names clam, mussel, oyster

Filesystems /sharedfs1

Filesystems/Directories to Export

Filesystems/Directories to NFS Mount

/sharedvg1

Volume Groups

Raw Disks

AIX Connections Realm/Svc pairs

Application Servers imagedemo

Inactive Takeover false

Cascade without Fallback false

42 Planning Guide

Page 173: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsCluster Event Worksheet

Cluster Event WorksheetNote: Use full pathnames for all Cluster Event Methods, Notify Commands,

and Recovery commands.

Cluster ID ____________

Cluster Name ____________

Cluster Event Description ______________________________________

Cluster Event Method ______________________________________

Cluster Event Name ___________________________

Event Command ___________________________

Notify Command ___________________________

Pre-Event Command ___________________________

Post-Event Command ___________________________

Event Recovery Command ___________________________

Recovery Counter ___________________________

Cluster Event Name ___________________________

Event Command ___________________________

Notify Command ___________________________

Pre-Event Command ___________________________

Post-Event Command ___________________________

Event Recovery Command ___________________________

Recovery Counter ___________________________

Planning Guide A-43

Page 174: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Planning WorksheetsSample Cluster Event Worksheet

A-

Sample Cluster Event WorksheetNote: Use full pathnames for all user-defined scripts.

Cluster ID 1

Cluster Name bivalves

Cluster Event Name node_down_complete

Event Command

Notify Command

Pre-Event Command

Post-Event Command /usr/local/wakeup

Event Recovery Command

Recovery Counter

Cluster Event Name ___________________________

Event Command ___________________________

Notify Command ___________________________

Pre-Event Command ___________________________

Post-Event Command ___________________________

Event Recovery Command ___________________________

Recovery Counter ___________________________

44 Planning Guide

Page 175: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Using the Online Cluster Planning Worksheet ProgramOnline Cluster Planning Worksheets—Overview

ted to y the

report panels gram ions ake an

in the an the s paper

new an

pply

B

Appendix B Using the Online Cluster Planning Worksheet Program

This appendix covers how to use the web-based online worksheets provided with the HACMP software in the /usr/lpp/cluster/samples/worksheets directory. The online worksheet program is a tool to aid you in planning your cluster topology and resource configuration and then applying the configuration directly to your cluster.

To use the online worksheets, you must transfer the worksheet program files to a PC-based system equipped with the appropriate web browser (see below).

This appendix contains instructions for using the online worksheets correctly, and references to the appropriate chapters in this book for detailed information about each stage of the cluster planning process. The worksheet program includes its own online help topics as well.

Note: Before you start, and while you are using the online worksheet program, refer to the chapters earlier in this guide for more information about how to plan each component of your cluster. Also consider drawing a diagram of your desired configuration before you get started, as suggested in Chapter 2, Drawing the Cluster Diagram.

Online Cluster Planning Worksheets—OverviewThe HACMP online cluster planning “worksheets” are actually a series of panels presenyou on a PC through a web browser. Each panel contains fields in which you can specifbasic components of your HACMP cluster topology and resource configuration. As you progress through the panels, you can save information you have entered, create an HTMLto print out or e-mail, or go back to earlier panels and make changes. The online planning present logical default choices whenever possible. When appropriate, the worksheet proprevents you from entering nonfeasible configuration choices by presenting only the optthat work based on what you have already entered, or by presenting a message if you minvalid choice.

The panels are based loosely on the paper worksheets that have always been providedHACMP documentation (see Appendix A, Planning Worksheets). The information in theonline worksheets has been consolidated and ordered in a slightly different sequence thpaper worksheets. Therefore, the specific worksheet instructions in the earlier planning chapterapply only to the paper worksheets; the conceptual information applies whether you useor online worksheets.

As soon as you complete the series of planning worksheets, you can actually apply the configuration to your cluster nodes. The worksheet program provides an option to createAIX file of your configuration data, and then transfer it via FTP to an AIX cluster node to athe configuration.

Planning Guide B-1

Page 176: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Using the Online Cluster Planning Worksheet ProgramInstalling and Viewing the Worksheet Program

B-

hoice

tor

g the

nse for ce for

pages

vide

Installing and Viewing the Worksheet ProgramThe online planning worksheet program files are packaged with your other HACMP filesets, in the samples directory. To use the worksheets, you must transfer the fileset to a PC-based system running an up-to-date version of the Microsoft Internet Explorer™ web browser.

Note: At the time of publication of this manual, Microsoft Internet Explorer, version 4.0 or higher, is the required browser for the online worksheets. Check the most recent HACMP product README file for any updates to this requirement.

To begin using the online cluster planning worksheets:

1. Install the HACMP for AIX software.

2. In the usr/lpp/cluster/samples/worksheets directory, find two files: worksheets.html and worksheets.jar.

3. Transfer these two files (via FTP or another method) to an appropriate folder of your con the PC-based system.

Note: You must specify binary mode for your FTP session.

4. Set the CLASSPATH variable to the worksheets.jar pathname, as follows:

Windows 95: You must edit the autoexec.bat file to include the CLASSPATH variable. The autoexec.bat file is usually located in the root (C:) directory. Open the file in an edisuch as notepad, or open a DOS window and type edit autoexec.bat at the prompt.

Edit the autoexec.bat file by adding the following line:

set CLASSPATH=c:\<folder name>\worksheets.jar

Windows NT: Go to Control Panel > System Properties > Environment. Add CLASSPATH as the variable, and the full pathname for worksheets.jar as the value.

5. In your browser, open worksheets.html.

You should see the first worksheet panel (shown on page B-3) with fields for enterincluster name and ID.

Note: It is recommended that you view the worksheets in the browser’s full screen mode.

The Online Worksheet FormatThe online planning worksheet panels are arranged in a sequence that makes logical seplanning your cluster topology and resources. In many cases, you must follow this sequencorrect configuration.

As you proceed through the worksheet panels, refer to the instructions on the following and on the worksheet panels, and refer to the specified chapters in the Planning Guide for important information about planning each component. In addition, online help topics prodetails about each worksheet panel.

2 Planning Guide

Page 177: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Using the Online Cluster Planning Worksheet ProgramThe Online Worksheet Format

Panels, Tabs, and ButtonsThe first panel you see when you open the worksheet program is the Cluster panel shown below. In this initial panel, you see the tabs that take you to individual worksheets for data entry. You also see several buttons. Most of the tabs and buttons are disabled until you enter the cluster name and ID or open an existing set of worksheets with data already in them.

Tabs

The tabs on the bottom and top of the panel indicate at all times which worksheet panels are enabled; the panel currently selected is indicated by bold type on the appropriate tabs. The tabs at the bottom of the screen indicate the major topics, and those at the top indicate subtopics within each major topic.

Bottom Tabs

The tabs at the bottom of your screen indicate three major planning topics. You should proceed with your configuration in the sequence of the tabs as follows:

1. Topology

2. Disk

3. Resources

The Cluster Notes tab brings to you a screen for miscellaneous user/system administrator notes.

The Help tab brings you to a directory of help topics for specific worksheets.

Planning Guide B-3

Page 178: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Using the Online Cluster Planning Worksheet ProgramThe Online Worksheet Format

B-

Top Tabs

When you select a major topic from the bottom tabs, a series of subtopic tabs at the top of your screen are enabled. Again, these tabs are presented from left to right in the order that works best for planning a cluster.

The entire set of planning sheets follows this hierarchy of topics and subtopics:

Worksheet Panel Buttons

Within each panel, buttons are used for the following actions:

Base Panel Buttons

On the base panel, which is visible at all times below the topic tabs, the following buttons are available:

Topology: Cluster, Node, Network, Global Network (active for HACMP/ES type only), Adapters

Disk: Disks, Volume Groups, Logical Volumes, NFS Exports

Resource: Cluster Events, Applications, Application Servers, Resource Groups, Resource Associations

Cluster Notes: Space for administrator notes

Help: All help topics

Add Updates the configuration with the information in the current entry field and makes the information available to subsequent panels.

Modify Allows you to make changes in any entry you have made in the current worksheet panel. First, in the lower window, select what entry you wish to change. The data for that entry is displayed in the entry fields above. Make the desired changes and press the Modify button.

Delete Deletes the selected entry field data.

Clear All Deletes all entries in the entire set of worksheets you have open

Create Report Saves to an HTML file all of the configuration data you have entered thus far for the set of worksheets currently open. You can then open this file in a browser and print it or distribute it via e-mail to others.

Save Writes the cluster configuration entered thus far to the PC disk, to a specified worksheet (.ws) file.*

Create Configuration Creates the AIX file that you will transfer to an AIX cluster node to apply the configuration data to the cluster.

4 Planning Guide

Page 179: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Using the Online Cluster Planning Worksheet ProgramEntering Cluster Configuration Data

Select the Help tab at any time to get specific information about entering data in a given worksheet panel.

*Due to differences in older and newer versions of msjava.dll, the behavior of the Save option may vary. In newer versions of msjava.dll, the .ws files can be saved only to the desktop.

Entering Cluster Configuration DataIf you are familiar with the planning information in the earlier chapters of this guide, and have your cluster configuration preferences mapped out in a diagram or other format, you are ready to begin entering data into the online worksheet panels as follows.

Note: If you are not in full screen mode, you may find that scrolling down and back up may cause the top (subtopic) tabs of the panel to disappear from view. They reappear when you press the Add button, or if you go to another panel and back again. Do not use the browser’s Refresh button, as this will clear the panel of the information you have entered.

Stage 1: Topology PlanningThe first step is to define the topology of your cluster: the framework of cluster nodes and the networks that connect them.

Cluster

When you open the worksheet program, the first thing you see is the Cluster screen.

In the Cluster screen, you name your cluster and assign it an ID. You also specify which product subsystem you are using: HACMP or HACMP/ES. After you have entered the name and ID, you can choose from the subtopics on the top tabs. The Author and Company fields are optional. You can also open the worksheets for an existing cluster you have already started and saved.

To complete the Cluster worksheet:

1. Fill in the cluster name. This can be any combination of 31 or fewer alphanumeric characters and underscores.

2. Fill in the cluster ID. The ID can be any positive integer up to 99999.

Note: Be sure the cluster name and ID do not duplicate those of another cluster at your site.

3. Check the appropriate box (HACMP or HACMP/ES) for the product type (subsystem) you are using.

4. Press the Add button.

In the Planning Guide, see: Naming the Cluster, page 2-3

Planning Guide B-5

Page 180: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Using the Online Cluster Planning Worksheet ProgramEntering Cluster Configuration Data

B-

Note: If you want to change what you have entered in the Cluster panel, just change your text and press Add again, except for Type (HACMP or HACMP/ES). If you need to change the Type designation after you have hit Add, use the Clear All button to start over.

Nodes

In the Nodes screen, you specify all nodes that will participate in the cluster. The maximum number of nodes is eight (HACMP) or 32 (HACMP/ES). The minimum is two for either case.

To specify cluster nodes, enter each node name and press the Add button. When you have finished this panel, go on to the Networks panel.

Networks

In this screen, you define the networks in your cluster. The drop-down list for attributes presents only the attributes that are possible for the network type you select. For example, if your network type is Ethernet, you can choose either public or private as an attribute. For an RS232 serial network, only serial is presented.

Global Network

This panel is enabled only if you selected the type HACMP/ES in the first panel.

For more information about global networks and the HACMP/ES subsystem, see the HACMP for AIX: Enhanced Scalability Installation and Administration Guide, Vol. 1.

Adapters

In this screen, you enter information about the adapters for each network you have defined. The Networks drop-down list presents a list of the networks you have defined.

The worksheet program allows you to enter adapter functions and hardware addresses that work with the types of networks you have defined.

Stage 2: Disk PlanningOnce you have defined the major components of your cluster topology, the next step is to plan your shared disk configuration.

In the Planning Guide, see: Selecting the Number of Nodes, page 2-4.

In the Planning Guide, see: Chapter 3, Planning TCP/IP Networks

Chapter 4, Planning Serial Networks

In the Planning Guide, see: Chapter 3, Planning TCP/IP Networks

Chapter 4, Planning Serial Networks

6 Planning Guide

Page 181: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Using the Online Cluster Planning Worksheet ProgramEntering Cluster Configuration Data

Disks

In this screen, you enter information about the disks you want to use in your cluster.

Volume Groups

In this screen, you specify the volume groups for your cluster.

Logical Volumes

In the Logical Volumes screen, you specify logical volumes and which volume groups they are associated with.

NFS Exports

Here, you specify which filesystems, if any, need to be exported, so other nodes can NFS mount them.

Stage 3: Resource Planning Now that you have planned the nodes, networks, and disk components for your cluster, you move on to specifying event scripts and configuring the cluster resources you wish to make highly available under HACMP.

Cluster Events

In this screen, you specify events and associated scripts. You type in the command or script path for each event. You can enter multiple entries per event per node. Note that you cannot enter a duplicate command for the same event.

Applications

In the Applications screen, you assign a name to each application and enter its directory path and filesystem.

In the Planning Guide, see:

Also see:

Chapter 5, Planning Shared Disk Devices

HACMP Concepts and Facilities Guide

In the Planning Guide, see: Chapter 6, Planning Shared LVM Components

In the Planning Guide, see:

Also see:

Chapter 6, Planning Shared LVM Components

AIX System Management Guide: Operating System and Devices

In the Planning Guide, see: Chapter 6, Planning Shared LVM Components

In the Planning Guide, see: Chapter 8, Tailoring Cluster Event Processing

In the Planning Guide, see: Chapter 7, Planning Applications, Application Servers, and Resource Groups

Planning Guide B-7

Page 182: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Using the Online Cluster Planning Worksheet ProgramApplying Worksheet Data to AIX Cluster Nodes

B-

lume he

er

saved nsfer

are

Application Servers

Here, you specify the application server name, and then choose from the picklist of applications you specified in the previous screen. You then specify the full path locations of your application start and stop scripts. In addition, you can write notes about your scripts in the space provided.

Resource Groups

In this screen, you assign names to your resource groups and specify the participating nodes and their position (priority), if appropriate, in the resource chain.

The Increase Priority and Decrease Priority buttons become visible when you highlight the node name. Note that for concurrent and rotating resource groups, priority information is not valid.

Resource Associations

In this screen, you specify which individual resources—such as filesystems, IP labels, vogroups, and application servers—are to be part of each resource group you defined in tprevious screen.

Cluster NotesIn this screen, you have space to note any special details about the cluster or your clustplanning process for future reference.

Applying Worksheet Data to AIX Cluster NodesWhen you have completed all the worksheets, are satisfied with your decisions, and havethe worksheets, you can create an AIX file to configure your actual cluster. You then trathe configuration file, via FTP, to your AIX cluster nodes.

PrerequisitesBefore you can transfer your new file to your AIX nodes, be sure the following conditionsin place:

• Your HACMP software must be installed.

• All hardware devices that you specified for your cluster configuration must be in place.

• If you are replacing an existing configuration, any current cluster ODM information should be retained in a snapshot.

• Cluster services must be stopped on all nodes.

In the Planning Guide, see: Chapter 7, Planning Applications, Application Servers, and Resource Groups

In the Planning Guide, see: Chapter 7, Planning Applications, Application Servers, and Resource Groups

In the Planning Guide, see: Chapter 7, Planning Applications, Application Servers, and Resource Groups

8 Planning Guide

Page 183: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Using the Online Cluster Planning Worksheet ProgramWhere You Go From Here

dapter apters,

. If

the

tion.

, ver, in the

Creating the AIX Configuration FileTo create an AIX file of your cluster configuration data:

1. Press the Create Configuration button located on the base panel.

(Note that this button is enabled only after you have defined the cluster nodes.)

The Create AIX Cluster Configuration dialog box appears.

2. Save your configuration file as the default (cluster.conf) or any name you choose.

Transferring the AIX File to the ClusterYou now transfer the configuration data file from the PC-based system to one of the AIX cluster nodes, in order to apply the configuration to the cluster.

1. FTP your AIX configuration file to one of your cluster nodes.

2. On the cluster node, run the cl_opsconfig command as follows:

/usr/sbin/cluster/utilities/cl_opsconfig <your configuration file>

The cl_opsconfig utility automatically performs a synchronization, including verification, of the configuration. During this process, you see a series of messages indicating the events taking place and any warnings or errors.

Boot Adapter Warnings

The cl_opsconfig utility adds all boot adapters before adding any service adapters. Because of this, you see a series of warnings indicating “There is no service interface” for each boot aas it is added. You can ignore these warnings if you have configured the proper service adas cl_opsconfig will add them later in the process, resolving the false error.

Viewing Error Messages

You can view the cl_opsconfig error messages on the screen, or redirect them to a log fileyou wish to redirect the standard error information to another file, add the symbols 2> and specify an output file, as in the following:

/usr/sbin/cluster/utilities/cl_opsconfig <your configuration file> 2> <output file>

Note that redirecting all output (standard output and standard error) is not recommended.

If errors are detected, go back to the worksheet program to fix the problems, and repeatprocess of creating the configuration file, transferring it to the cluster node, and running cl_opsconfig.

Where You Go From HereWhen your configuration file has been transferred and cl_opsconfig has run successfully without reporting configuration errors, your cluster has a basic working HACMP configura

You can now proceed to additional configuration and customization tasks—for exampleconfiguring AIX Error Notification, run time parameters, custom scripts, IP address takeoand so on. For more information, refer to the appropriate chapters in this guide, and alsoHACMP for AIX Installation Guide and Administration Guide.

Planning Guide B-9

Page 184: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Using the Online Cluster Planning Worksheet ProgramWhere You Go From Here

B-

10 Planning Guide
Page 185: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Single-Adapter Networksnetmon Enhancement for Single-Adapter Networks

ring nds

C

Appendix C Single-Adapter Networks

This appendix describes an enhancement to the netmon program for use in a cluster with single-adapter networks.

netmon Enhancement for Single-Adapter NetworksSingle-adapter networks (networks with no standby adapters) are not recommended for an HACMP for AIX environment. In cluster configurations where there are networks with no standby network adapters, it can be difficult for the HACMP for AIX software to accurately determine service adapter failure. This is because the Cluster Manager cannot use a standby adapter to force packet traffic over the service adapter to verify its operation. This shortcoming is less of an exposure if one or both of the following conditions is true:

• There are network devices that answer broadcast ICMP ECHO requests. This can be verified by pinging the broadcast address and determining the number of different IP addresses that respond.

• The service adapter is under heavy use. In this instance the inbound packet count will continue to increase over the service adapter without stimulation from the Cluster Manager.

An enhancement to netmon, the network monitor portion of the Cluster Manager, allows more accurate determination of a service adapter failure. This function can be used in configurations that require a single service adapter per network.

You can create a netmon configuration file, /usr/sbin/cluster/netmon.cf, that specifies additional network addresses to which ICMP ECHO requests can be sent. The configuration file consists of one IP address or IP label per line. The maximum number of addresses used is five. All addresses specified after the fifth one will be ignored. No comments are allowed in the file.

Here’s an example of a /usr/sbin/cluster/netmon.cf configuration file:

180.146.181.119steamerchowder180.146.181.121mussel

This file must exist at cluster startup. The cluster software scans the configuration file duinitialization. When netmon needs to stimulate the network to verify adapter function, it seICMP ECHO requests to each address. After sending the request to every address, netmon checks the inbound packet count before determining whether an adapter has failed.

Planning Guide C-1

Page 186: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Single-Adapter Networksnetmon Enhancement for Single-Adapter Networks

C-

2 Planning Guide
Page 187: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Applications and HACMPOverview

e

o call ons er,

This

D

Appendix D Applications and HACMP

This appendix addresses some of the key issues to consider when making your applications highly available under HACMP for AIX. The information provided here is of a general enough nature to apply to both the HACMP and HACMP/ES product subsystems.

Also see Chapter 7, Planning Applications, Application Servers, and Resource Groups, and Chapter 18, Configuring Cluster Resources, in the HACMP for AIX Installation Guide.

OverviewBesides understanding the hardware and software needed to make a cluster highly available, you will need to spend some time on application considerations when planning your HACMP environment. The goal of clustering is to keep your important applications available despite any single point of failure. To achieve this goal, it is important to consider the aspects of an application that make it recoverable under HACMP.

There are few hard and fast requirements that an application must meet to recover well under HACMP. For the most part, there are simply good practices that can head off potential problems. Some required characteristics, as well as a number of suggestions, are discussed here. These are grouped according to key points that should be addressed in all HACMP environments. This appendix covers the following application considerations:

• Automation—making sure your applications start and stop without user intervention

• Dependencies—knowing what factors outside HACMP affect your applications

• Interference—knowing that applications themselves can hinder HACMP’s functioning

• Robustness—choosing strong, stable applications

• Implementation—using appropriate scripts, file locations, and cron schedules

At the end of this appendix, you will find two examples of popular applications—Oracle Database™ and SAP R/3™—and some issues to consider when implementing these applications in an HACMP environment.

Application Automation: Minimizing Manual InterventionOne key requirement for an application to function successfully under HACMP is that thapplication be able to start and stop without any manual intervention.

Application Start Scripts

You should create a start script that completely starts the application. Configure HACMP tthis script at cluster startup to initially bring the application online. Since the cluster daemcall the start script, there is no option for interaction. Additionally, upon an HACMP fallovthe recovery process calls this script to bring the application on line on a standby node. allows for a fully automated recovery.

Planning Guide D-1

Page 188: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Applications and HACMPApplication Automation: Minimizing Manual Intervention

D-

he

NIS) the luster p files

a e

his will e, a n.

be

ation. rces

erns the

raceful te, the phase

ted

Keep in mind that this application start script may need to take additional action to prepare the cluster to bring the application on line. The start script will be called by HACMP as the “root”user. It may be necessary to change to a different user in order to start the application. Tsu command can accomplish this. Also, it may be necessary to run nohup on commands that arestarted in the background and have the potential to be terminated upon exit of the shell.

For example, an HACMP cluster node may be a client in a Network Information Service (environment. If this is the case, and you need to use the su command to change user id, theremust be a route to the NIS master at all times. In the event that a route doesn’t exist, andsu is attempted, the application script hangs. You can avoid this by enabling the HACMP cnode to be an NIS slave. That way a cluster node has the ability to access its own NIS mato validate a user ID.

Another good practice in application start scripts is to check the return code upon exitingscript. If the return code is not zero, an error may have occurred in starting that should baddressed. If a non-zero return code is passed back to HACMP, the event_error event is run and the cluster enters an error state. This check alerts administrators that the cluster is not functioning properly.

The start script should also check for the presence of required resources or processes. Tensure an application can start successfully. If the necessary resources are not availablmessage can be sent to the administration team to correct this and restart the applicatio

Keep in mind that the start script may be run after a primary node has failed. There mayrecovery actions necessary on the backup node in order to restart an application. This iscommon in database applications. Again, the recovery must be able to run without any interaction from administrators.

Application Stop Scripts

The most important aspect of an application stop script is that it completely stop an applicFailure to do so many prevent HACMP from successfully completing a takeover of resouby the backup nodes. In stopping, the script may need to address some of the same concstart script addresses, such as NIS and the su command.

The application stop script should use a phased approach. The first phase should be a gattempt to stop the processes and release any resources. If processes refuse to terminasecond phase should be used to forcefully ensure all processing is stopped. Finally, a thirdcan use a loop to repeat any steps necessary to ensure that the application has terminacompletely.

Note: Keep in mind that HACMP allows six minutes by default for events to complete processing. A message indicating the cluster has been in reconfiguration too long appears until the cluster completes its reconfiguration and returns to a stable state. This warning may be an indication that a script is hung and requires manual intervention. If this is a possibility, you may wish to consider stopping an application manually before stopping HACMP.

If desired, you can alter the time period before the config_too_long event is invoked. See the HACMP for AIX Troubleshooting Guide for more information.

2 Planning Guide

Page 189: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Applications and HACMPApplication Dependencies

nual

ent e ty

od. d

ss of any

sues. tion yond

r, if il to run

nsider uster

Application Tier Issues

Often, applications are of a multi-tier architecture. The first tier may be a database, the second tier an application/login tier and the third a client. You must consider all tiers of an architecture if one or more is made highly available through the use of HACMP.

For example, if the database is made highly available, and a fallover occurs, consider whether actions should be taken at the higher tiers in order to automatically return the application to service. If so, it may be necessary to stop and restart application or client tiers. This can be facilitated in one of two ways. One way is to run clinfo on the tiers, the other is to use the rsh command.

Using the Clinfo APIclinfo is the cluster information daemon. You can write a program using the Clinfo API to run on any tiers that would stop and restart an application after a fallover has completed successfully. In this sense, the tier, or application, becomes “cluster aware,” responding to events that take place in the cluster. For more information on the Clinfo API, see the maHACMP for AIX: Programming Client Applications.

Using Pre- and Post-Event ScriptsAnother way to address the issue of multi-tiered architectures is to use pre- and post-evscripts around a cluster event. These scripts would call the rsh command to stop and restart thapplication. Keep in mind that the use of the rsh command may require a loosening of securithat is unacceptable for some environments.

Another way to address the rsh security issue is by using Kerberos as an authentication methIf you choose this method, Kerberos must be in place prior to the HAMCP installation anconfiguration. IBM supports Kerberos on IBM Scalable POWERParallel (RS/6000 SP) systems. For more information about configuring Kerberos, see Appendix F of the HACMP for AIX Installation Guide.

Application DependenciesIn many cases, applications depend on more than data and an IP address. For the succeapplication under HACMP, it is important to know what the application should not depend upon in order to function properly. This section outlines many of the major dependency isKeep in mind that these dependencies may come from outside the HACMP and applicaenvironment. They may be incompatible products or external resource conflicts. Look bethe application itself to potential problems within the enterprise.

Locally Attached Devices

Locally attached devices can pose a clear dependency problem. In the event of a fallovethese devices are not attached and accessible to the standby node, an application may faproperly. These may include a CD-ROM device, a tape device, or an optical juke box. Cowhether your application depends on any of these and if they can be shared between clnodes.

Planning Guide D-3

Page 190: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Applications and HACMPApplication Interference

D-

ssage pped

g as ad an

rver ade

Hard Coding

Anytime an application is hard coded to a particular device in a particular location, there is the potential for a dependency issue. For example, the console is typically assigned as /dev/tty0. Although this is common, it is by no means guaranteed. If your application assumes this, ensure that all possible standby nodes have the same configuration.

Hostname Dependencies

Some applications are written to be dependent on the AIX hostname. They issue a command in order to validate licenses or name filesystems. The hostname is not an IP address label. The hostname is specific to a node and is not failed over by HAMCP. It is possible to manipulate the hostname, or use hostname aliases, in order to trick your application, but this can become cumbersome when other applications, not controlled by HACMP, also depend on the hostname.

Software Licensing

Another possible problem is software licensing. Software can be licensed to a particular CPU ID. If this is the case with your application, it is important to realize that a fallover of the software will not successfully restart. You may be able to avoid this problem by having a copy of the software resident on all cluster nodes. Know whether your application uses software that is licensed to a particular CPU ID.

Application InterferenceSometimes an application or an application environment may interfere with the proper functioning of HACMP. An application may execute properly on both the primary and standby nodes. However, when HACMP is started, a conflict with the application or environment may arise that prevents HACMP from functioning successfully.

Software Using IPX/SPX ProtocolA conflict may arise between HACMP and any software that binds a socket over a network interface. An example is the IPX/SPX protocol. When active, it binds an interface and prevents HAMCP from properly managing the interface. Specifically, for ethernet and token ring, it inhibits the hardware address takeover from completing successfully. A “device busy” meappears in the HACMP logs. The software using IPX/SPX must be either completely stoor not used in order for hardware address takeover to work.

Products Manipulating Network RoutesAdditionally, products that manipulate network routes can keep HACMP from functioninit was designed. These products can find a secondary path through a network that has hinitial failure. This may prevent HACMP from properly diagnosing a failure and taking appropriate recovery actions.

AIX Connections, AIX Fast Connect, and CS/AIX

You can reduce the problem of conflict with certain protocols, and the need for manual intervention, if you are using AIX Connections, AIX Fast Connect, or Communications Sefor AIX to share resources. The protocols handled by these applications can easily be mhighly available because of their integration with HACMP.

4 Planning Guide

Page 191: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Applications and HACMPRobustness of Application

ess

es . Fast

AIX AIX U2 during

apter e

the ese

ode is t be a

e state. nt the re it

AIX Connections is a network operating software that enables sharing of resources between AIX workstations and clients running other operating systems such as Windows NT, OS/2, and Macintosh. AIX Connections is already integrated with HACMP so the IPX/SPX, NetBEUI, and AppleTalk protocols handled by AIX Connections can be easily configured as highly available resources in the cluster. The protocols can then be taken over in the event of node or adapter failure. For example, in the case of a NetWare client using the IPX/SPX protocol, if AIX Connections is configured, HACMP steps up in the event of an adapter failure: it stops communications on the port, frees up the bind on the socket—thereby allowing the addrtakeover to proceed—and restarts the communication.

AIX Fast Connect software is integrated with HACMP in a similar way, so that it can be configured as a highly available resource. AIX Fast Connect allows you to share resourcbetween AIX workstation and PCs running Windows, DOS, and OS/2 operating systemsConnect supports the NetBIOS protocol over TCP/IP.

Communications Server for AIX enables an RS/6000 computer to participate in an SNAnetwork that includes mainframes, PCs, and other workstations. You can configure CS/data link profiles as HACMP communications links resources. HACMP handles the CS/protocol stopping and starting in the event of an adapter or node failure. The LU6.2 or Lconnections, link stations, and CS/AIX servers are automatically stopped and restarted failure and recovery.

For more information on configuring these applications as resources in HACMP, see Ch7, Planning Applications, Application Servers, and Resource Groups in this book, and thsections on configuring resources in the HACMP for AIX Installation Guide. You’ll also find worksheets for these applications in Appendix A, Planning Worksheets.

Robustness of ApplicationOf primary importance to the success of any application is the health, or robustness, of application. If the application is unstable or crashing intermittently, you should be sure thissues are resolved prior to placing it in a high availability environment.

Beyond basic stability, an application under HACMP should meet other robustness characteristics, such as the following.

Successful Restart After Hardware Failure

A good application candidate for HACMP should be able to restart successfully after a hardware failure. Run a test on an application prior to putting in under HACMP. Run theapplication under a heavy load and fail the node. What does it take to recover once the nback on line? Can this recovery be completely automated? If not, the application may nogood candidate for high availability.

Survival of Real Memory Loss

For an application to function well under HACMP it should be able to survive a loss of thcontents of real memory. It should be able to survive the loss of the kernel or processorWhen a node failure occurs, these are lost. Applications should also regularly check-poidata to disk. In the event that a failure occurs, the application will be able to pick up whelast check-pointed data, rather than starting completely over.

Planning Guide D-5

Page 192: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Applications and HACMPApplication Implementation Strategies

D-

within ble for own ts.

e on ied on, files by the d

each nses these entire r

either t not erver ent

Application Implementation StrategiesThere are a number of aspects of an application that you should consider as you plan for implementing it under HACMP. You must consider characteristics such as time to start, time to restart after failure, and time to stop. Your decisions in a number of areas, including those discussed in this section—scriptwriting, file storage, /etc/inittab file and cron schedule issues—can improve the probability of successful application implementation.

Writing Effective Scripts

Writing smart application start scripts can also help you avoid problems with bringing applications online.

A good practice for start scripts is to check prerequisite conditions before starting an application. These may include access to a filesystem, adequate paging space and freefilesystem space. The start script should exit and run a command to notify system administrators if the requirements are not met.

When starting a database it is important to consider whether there are multiple instancesthe same cluster. If this is the case, you must be careful to start only the instances applicaeach node. Certain database startup commands read a configuration file and start all kndatabases at the same time. This may not be a desired configuration for all environmen

Considering File Storage Locations

You should also give thought to where the configuration files reside. They could either bshared disk, and thus potentially accessed by whichever node has the volume group varor on each node’s internal disks. This holds true for all aspects of an application. Certainmust be on shared drives. These include data, logs, and anything that could be updatedexecution of the application. Files such as configuration files or application binaries coulreside in either location.

There are pros and cons to storing optional files in either location. Having files stored onnode’s internal disks implies that you have multiple copies of, and potentially multiple licefor, the application. This could require additional cost as well as maintenance in keepingfiles synchronized. However, in the event that an application needs to be upgraded, thecluster need not be taken out of production. One node could be upgraded while the otheremains in production. The “best” solution is the one that works best for a particular environment.

Considering /etc/inittab and cron Table Issues

You must also give thought to applications, or resources needed by an application, that start out of the /etc/inittab file or out of the cron table. The inittab starts applications upon booup of the system. If cluster resources are needed for an application to function, they willbecome available until after HACMP is started. It is better to use the HACMP application sfeature which allows the application to be a resource that is started only after all dependresources are online.

6 Planning Guide

Page 193: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Applications and HACMPExamples: Oracle Database™ and SAP R/3™

k

In the cron table, jobs are started according to a schedule set in the table and the date setting on a node. This information is maintained on internal disks and thus cannot be shared by a standby node. You must synchronize these cron tables so that a standby node can perform the necessary action at the appropriate time. You must also ensure the date is set the same on the primary node and any of its standby nodes.

Examples: Oracle Database™ and SAP R/3™Here are two examples illustrating issues to consider in order to make the applications Oracle Database and SAP R/3 function well under HACMP.

Example 1: Oracle Database

The Oracle Database, like many databases, functions very well under HACMP. It is a robust application that handles failures well. It can roll back uncommitted transactions after a fallover and return to service in a timely manner. There are, however, a few things to keep in mind when using Oracle Database under HACMP.

Starting OracleOracle must be started by the Oracle user ID. Thus, the start script should contain an su - oracleuser. The dash (-) is important since the su needs to take on all characteristics of the Oracle user and reside in the Oracle user’s home directory. The full command might loosomething like this:

su - oracleuser -c “/apps/oracle/startup/dbstart”

Commands like dbstart and dbshut read the /etc/oratabs file for instructions on which database instances are known and should be started. In certain cases it is inappropriate to start all of the instances, because they might be owned by another node. This would be the case in the mutual takeover of two Oracle instances. The oratabs file typically resides on the internal disk and thus cannot be shared. If appropriate, consider other ways of starting different Oracle instances.

Stopping OracleThe stopping of Oracle is a process of special interest. There are several different ways to ensure Oracle has completely stopped. The recommended sequence is this: first, implement a graceful shutdown; second, call a shutdown immediate, which is a bit more forceful method; finally, create a loop to check the process table to ensure all Oracle processes have exited.

Oracle File StorageThe Oracle product database contains several files as well as data. It is necessary that the data and redo logs be stored on shared disk so that both nodes may have access to the information. However, the Oracle binaries and configuration files could reside on either internal or shared disks. Consider what solution is best for your environment.

For more information about keeping your Oracle applications highly available, see the IBM Redbook #SG24-4788, Bullet-Proofing Your Oracle Database with HACMP: A Guide to Implementing AIX Databases with HACMP.

Planning Guide D-7

Page 194: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Applications and HACMPExamples: Oracle Database™ and SAP R/3™

D-

n s done

ust be M g

ered cation same uld then

P mon

Example 2: SAP R/3, a Multi-Tiered Application

SAP R/3 is an example of a three-tiered application. It has a database tier, an application tier, and a client tier. Most frequently, it is the database tier that is made highly available. In such a case, when a fallover occurs and the database is restarted, it is necessary to stop and restart the SAP application tier. You can do this in one of two ways: by using the rsh command, or by making the application tier nodes “cluster aware.”

Using the rsh CommandThe first way to stop and start the SAP application tier is to create a script that performs arsh to the application nodes. The application tier of SAP is stopped and then restarted. This ifor every node in the application tier. Use of the rsh command requires a method of allowingthe database node access to the application node. Certain methods, such as the use of /.rhosts files, pose a security risk and may not be desirable.

As mentioned under Application Tier Issues, another way to address the rsh security issue is by using Kerberos as an authentication method. If you choose this method, Kerberos min place prior to the HAMCP installation and configuration. IBM supports Kerberos on IBScalable POWERParallel (RS/6000 SP) systems. For more information about configurinKerberos, see Appendix F of the HACMP for AIX Installation Guide.

Making Application Tier Nodes “Cluster Aware”A second method for stopping and starting the application tier is to make the application tier nodes “cluster aware.” This means that the application tier nodes are aware of the clustdatabase and know when a fallover occurs. You can implement this by making the applitier nodes either HACMP servers or clients. If the application node is a server, it runs thecluster events as the database nodes to indicate a failure; pre- and post-event scripts cobe written to stop and restart the SAP application tier. If the application node is an HACMclient, it is notified of the database fallover via SNMP through the cluster information dae(clinfo). A program could be written using the Clinfo API to stop and restart the SAP application tier.

See the manual HACMP for AIX: Programming Client Applications for more detail on the Clinfo API.

8 Planning Guide

Page 195: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Index

+-*/.klogin file

adding Kerberos service principals 3-6/etc/hosts file

and adapter label 3-6and boot address 3-8

/usr/sbin/cluster/etc/exports file 6-12/usr/sbin/cluster/events/utils/cl_deactivate_nfs utility

6-15/usr/sbin/cluster/events/utils/cl_nfskill command 6-15

0,1,2...2105 Versatile Storage Server 5-56214 SSA adapter 5-186216 SSA adapter 5-187013-S70 4-47015-S70 4-47017-S70 4-47131-405 SSA disk subsystem

planning 5-67133 SSA disk subsystem

cluster support 5-6disk fencing 5-21

7135 RAIDiant Disk Arraycluster support 5-3

7137 Disk Arraysample configuration 5-13

7137 Disk Arrayscluster support 5-4

9333 disk fencingin concurrent access clusters 5-21

9333 serial diskseight-node cluster 5-17four-node cluster 5-16planning 5-14two-node cluster 5-15

Aadapter labels

for network adapters 3-6adapters

configuring for Kerberos 3-6network

definition and HACMP functions 3-6planning SSA disk subsystem 5-18SSA disk subsystems 5-18

AIXsetting I/O pacing 3-21setting syncd frequency 3-22

AIX Connectionsdefined 7-2, 7-7handling adapter failure 7-8handling node failure 7-7realms and services available 7-7reducing protocol conflict D-4

AIX Connections Worksheet A-35AIX Fast Connect

converting from AIX Connections 7-5handling adapter failure 7-6handling node failure 7-6planning 7-2reducing protocol conflict D-4

AIX Fast Connect Worksheetcompleting 7-6planning A-33

Application Server Worksheet A-37completing 7-3, 7-4

application serversplanning 7-1start script 7-4stop script 7-4

Application Worksheet A-29completing 7-3

applicationsautomation with start/stop scripts D-1dependency issues D-3HACMP implementation strategies D-6licensing 5-8multi-tier architecture issues D-3planning 2-3, 7-2planning for use with HACMP D-1

AT for AppleTalk clientsAIX Connections 7-7

ATMhardware address takeover

configuration requirements 3-16specifying an alternate address 3-16

Bboot adapters 3-8boot addresses 3-8

defined 3-8planning 3-13

Planning Guide X-1

Page 196: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

IndexC – E

X-

bypass cardsSSA 5-19

Ccascading resource groups

cascading without fallbackDARE migration issues 7-12

NFS cross mounting issues 6-12planning worksheet conventions 7-3

cl_9333fence utility 5-22cl_opsconfig B-9cl_ssa_fence utility 5-22clients

planning 9-1Clinfo 9-2clinfo.rc script

planning 9-2cluster

planningapplication servers 7-1applications 2-3applications and application servers 7-2clients 9-1cluster diagram 2-1cluster events 8-1design goals 1-1diagram 2-1disks 5-1list of steps 1-3networks 3-1, 4-1number of nodes 2-4resource groups 7-1, 7-10resources 2-3serial networks 4-3shared disk access 2-5shared IP addresses 2-5shared LVM components 6-1

planning for performance 3-21Cluster Event Worksheet A-43

completing 8-3cluster events

event customization facility 8-1notification 8-2planning 8-1post-processing 8-2pre-processing 8-2recovery 8-3retry 8-3

Cluster Managerevent customization 8-2

concurrent access modequorum 6-10

config_too_long message D-2connecting

SCSI bus configuration 5-10, 5-11

creatingshared volume groups

NFS issues 6-11using the TaskGuide 6-2

cross mountingNFS filesystems 6-12

CS/AIXcompleting worksheets 7-9planning communication links 7-9products supported for HACMP 7-9reducing protocol conflicts D-4

customizingcluster event processing 8-1

cycles to faildefinition 3-23

DDARE

Resource Migration utilityand cascading without fallback 7-12

Deadman Switch 3-21formula 3-23timeouts per network 3-23

definingboot addresses 3-13hardware addresses 3-14

dependency issuesapplications and HACMP D-3

disk fencingand dynamic reconfiguration 5-23concurrent access configurations 5-21SSA concurrent access clusters 5-21

disks2105 Versatile Storage Server 5-57135 RAIDiant Disk Array 5-37137 Disk Arrays 5-4configuring a quorum buster 6-9IBM 9333 serial disk subsystems 5-5planning 5-1SCSI 5-2SSA subsystem 5-6

DLC profile 7-9DNS

Using with HACMP 3-18documentation

SSA installation and maintenance 5-18dynamic reconfiguration

and disk fencing 5-23

Eenabling

SSA disk fencing 5-23Ethernet adapters

specifying alternate HW address 3-15event customization facility 8-1

2 Planning Guide

Page 197: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

IndexF – N

eventscluster events 8-1customization facility 8-1notification 8-2planning 8-1recovery 8-3retry 8-3

Ffailure detection rate

changing 3-23, 3-24definition 3-23

Fast Connectconverting from AIX Connections 7-5handling adapter failure 7-6planning 7-4reducing protocol conflicts D-4

FDDI adaptersspecifying alternate HW address 3-15

filesystemsas shared LVM component 6-5

HHACMP nameserving 3-18HANFS for AIX

functionality now in HACMP 6-11hardware address

defining 3-14hardware address swapping

ATM adapters 3-16Ethernet adapters 3-15FDDI adapters 3-15planning 3-14Token-Ring adapters 3-15

hdiskand physical volume 6-3

heartbeat ratedefinition 3-23

high availabilitywith SSA 5-19

high water marksetting 3-22

II/O

setting pacing 3-21I/O pacing 3-21IBM 9333 serial disk subsystems

cluster support 5-5IBM disk subsystems and arrays

specific model number 5-1inactive takeover

and cascading without fallback 7-14IP address

defining 3-10

IP address takeover 2-5defining boot addresses 3-13

IPX/SPX protocolinterference with HACMP applications D-4

Jjfslog 6-3

mirroring 6-7

Kkeepalives

tuning 3-24Kerberos

configuring adapters 3-6service principals 3-6

Llicenses

software 5-8logical partitions

mirroring 6-5logical volumes 5-9

as shared LVM component 6-4journal logs 6-7

low-water marksetting 3-22

LVMshared components

planning 6-1worksheets

completing 6-16

Mmanuals

SSA installation and maintenance 5-18mirroring

jfslog 6-7logical partitions 6-5

mountingNFS 6-12

Nnameserver configuration

and boot address 3-8nameserving

enabling and disabling under HACMP 3-18using with HACMP 3-18

namingresource groups 7-3

NB for NetBIOS clientsAIX Connections 7-7

netmonconfiguration file C-1

Planning Guide X-3

Page 198: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

IndexO – P

X-

network adaptersadapter label 3-6

defined 3-6configuring for Kerberos 3-6functions 3-7

boot 3-8service 3-7standby 3-7

network interfaces 3-8network mask

defining 3-10network modules

failure detection parameters 3-22networks

attribute 3-9private 3-9public 3-9serial 4-4

maximum per cluster 3-2name 3-9point-to-point 3-5sample topologies 3-2serial

planning 4-1single adapter C-1TCP/IP

planning 3-1supported types 3-1

topology 3-2NFS

caveats about node names 6-15creating shared volume groups 6-11cross mounting filesystems 6-12exporting filesystems and directories 6-12mount issues 6-12mounting filesystems 6-12nested mount points 6-13planning 6-11reliable server functionality 6-11setting up mount points for cascading groups 6-13takeover issues 6-12

NFS-Exported File System Worksheet 6-17, A-23NIS

using with HACMP 3-18node isolation

and 9333 or SSA disk fencing 5-21definition 4-1

node namesissues with NFS 6-15

nodesdefined 3-6maximum per cluster 2-4

non-concurrent accessquorum 6-9

Non-Shared Volume Group Worksheet A-19, A-25concurrent access 6-18non-concurrent access 6-16

NW for NetWare clientsAIX Connections 7-7

Oonline planning worksheets

installing B-2overview B-1using B-2, B-5

Oracle Databaseas HACMP application D-7

Ppartitioned clusters 4-1

and 9333 or SSA disk fencing 5-21performance 3-21physical volume

as shared LVM component 6-3planning

AIX Connections 7-7AIX Fast Connect 7-4application servers 7-2applications 2-3, 7-2CS/AIX communication links 7-9disks 5-1HACMP cluster

applications 2-3, 7-1applications and application servers 7-2clients 9-1cluster diagram 2-1cluster events 8-1design goals 1-1drawing cluster diagram 2-1list of steps 1-3networks 3-1, 4-1number of nodes 2-4resource groups 7-1, 7-10resources 2-3serial networks 4-3shared disk access 2-5shared IP addresses 2-5shared LVM components 6-1

networksserial 4-1TCP/IP 3-1

NFS 6-11resource groups 7-10serial networks 4-3shared disks 5-1

7135 RAIDiant Disk Array 5-37137 Disk Arrays 5-4IBM 9333 serial disk subsystems 5-5logical volume storage 5-9power supplies 5-6SCSI 5-2

4 Planning Guide

Page 199: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

IndexQ – S

SSA disk subsystem 5-6shared LVM components 6-1

file systems 6-5logical volumes 6-4physical volumes 6-3volume groups 6-3

SSA disk subsystem configuration 5-18worksheets A-1

post-processingcluster events 8-2

power suppliesand shared disks 5-6

pre-processingcluster events 8-2

private networks 3-9selecting 3-1

public networks 3-9selecting 3-1

Qquorum 6-7quorum buster disk 6-10

RReliable NFS server 2-4, 6-11reliable NFS server 6-11Resource Group Worksheet 7-12, A-41resource groups

IP address requirements 3-13maximum per cluster 2-3planning 7-1, 7-10

resourcesselecting type during planning phase 2-3

root volume group 5-7RS232 serial lines 4-4

SS70 systems

requirements for serial network 4-4SAP R/3

as HACMP application D-7scripts

start application server 7-4stop application server 7-4

SCSItarget mode 3-6, 4-4

SCSI devicesdisks 5-2, 5-9

serial linesRS232 4-4

Serial Network Adapter Worksheet A-9completing 4-7

serial networks 4-4planning 4-3supported types 4-4TMSSA 4-5

Serial Networks Worksheet A-7completing 4-6

Serial Storage Architecture(SSA) 5-18

service adapters 3-7service principals

Kerberos 3-6setting 3-21

I/O Pacing 3-21shared

volume groupsNFS issues 6-11

shared disks7135 RAIDiant Disk Array 5-37137 Disk Arrays 5-4IBM 9333 serial disk subsystems 5-5planning 5-1planning type of access 2-5SCSI

planning 5-9, 5-14SSA disk subsystems 5-6VSS 5-5

Shared IBM 9333 Serial Disk Worksheet A-15Shared IBM SCSI Disk Array Worksheet A-13Shared IBM SSA Disk Subsystem Worksheet A-17shared IP addresses

planning for 2-5shared LVM components

file systems 6-5logical volumes 6-4physical volumes 6-3planning 6-1volume groups 6-3

Shared SCSI-2 Differential or Differential Fast/Wide Disks Worksheet A-11

shared SSA disk subsystemsplanning 5-18

Shared Volume Group Worksheetconcurrent access 6-18

Shared Volume Group/Filesystem Worksheetconcurrent access A-27non-concurrent access 6-17, A-21

shared volume groupscreating with TaskGuide 6-2

single adapter networks C-1single points of failure

potential cluster components 1-2SNA network

planning for HACMP configuration 7-9software licenses 5-8SP Switch

network environment 3-1

Planning Guide X-5

Page 200: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

IndexT – W

X-

SSAadapters 5-18and high availability 5-19bypass cards 5-19disk fencing 5-21disk subsystems 5-6IBM documentation 5-18loop configuration 5-6

SSA disk fencingenabling 5-22

SSA Disk Subsystemsplanning 5-18

standby adapters 3-7start script

application servers 7-4stop script

application servers 7-4subnets

in Tivoli-monitored clusters 3-12placing standby adapter on 3-11

syncdsetting frequency for flushing buffers 3-22

Ttakeover

NFS issues 6-12target mode SCSI 3-6, 4-4target mode SSA 4-5TaskGuide for creating shared volume groups

defining volume groups 6-2TCP/IP Network Adapter Worksheet A-5

completing 3-26TCP/IP Networks Worksheet A-3

completing 3-26Tivoli, cluster monitoring with

subnet requirements 3-12Token-Ring

adaptersspecifying alternate HW address 3-15

tuning parameters 3-21tuning the cluster 3-21

Uutilities

cl_opsconfig B-9

VVersatile Storage Servers 5-5

and HACMP 5-13volume groups

as shared LVM component 6-3quorum 6-7shared

creating with TaskGuide 6-2

Wworksheets A-1

AIX Connections Worksheet A-35AIX Fast Connect A-33Application Server Worksheet A-37

completing 7-3, 7-4Application Worksheet A-29Cluster Event Worksheet 8-3, A-43CS/AIX Communications Links Worksheet A-39NFS-Exported Filesystem Worksheet 6-17, A-23Non-Shared Volume Group Worksheet A-19,

A-25concurrent access 6-18non-concurrent access 6-16

onlineinstalling online worksheets B-2using online worksheets B-1

Resource Group Worksheet 7-12, A-41Serial Network Adapter Worksheet A-9

completing 4-7Serial Networks Worksheet A-7

completing 4-6Shared IBM 9333 Serial Disk Worksheet A-15Shared IBM SCSI Disk Array Worksheet A-13Shared IBM SSA Disk Subsystem Worksheet

A-17Shared SCSI-2 Differential or Differential

Fast/Wide Disks A-11Shared Volume Group Worksheet

concurrent access 6-18Shared Volume Group/Filesystem (concurrent

access) A-27Shared Volume Group/Filesystem Worksheet

non-concurrent access 6-17, A-21TCP/IP Network Adapter 3-26TCP/IP Network Adapter Worksheet A-5TCP/IP Networks 3-26TCP/IP Networks Worksheet A-3

6 Planning Guide

Page 201: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Vos remarques sur ce document / Technical publication remark form

Titre / Title : Bull HACMP 4.4 Planning Guide

Nº Reférence / Reference Nº : 86 A2 55KX 02 Daté / Dated : August 2000

ERREURS DETECTEES / ERRORS IN PUBLICATION

AMELIORATIONS SUGGEREES / SUGGESTIONS FOR IMPROVEMENT TO PUBLICATION

Vos remarques et suggestions seront examinées attentivement.Si vous désirez une réponse écrite, veuillez indiquer ci-après votre adresse postale complète.

Your comments will be promptly investigated by qualified technical personnel and action will be taken as required.If you require a written reply, please furnish your complete mailing address below.

NOM / NAME : Date :

SOCIETE / COMPANY :

ADRESSE / ADDRESS :

Remettez cet imprimé à un responsable BULL ou envoyez-le directement à :

Please give this technical publication remark form to your BULL representative or mail to:

BULL CEDOC357 AVENUE PATTONB.P.2084549008 ANGERS CEDEX 01FRANCE

Page 202: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Technical Publications Ordering FormBon de Commande de Documents Techniques

To order additional publications, please fill up a copy of this form and send it via mail to:Pour commander des documents techniques, remplissez une copie de ce formulaire et envoyez-la à :

BULL CEDOCATTN / MME DUMOULIN357 AVENUE PATTONB.P.2084549008 ANGERS CEDEX 01FRANCE

Managers / Gestionnaires :Mrs. / Mme : C. DUMOULIN +33 (0) 2 41 73 76 65Mr. / M : L. CHERUBIN +33 (0) 2 41 73 63 96

FAX : +33 (0) 2 41 73 60 19E–Mail / Courrier Electronique : [email protected]

Or visit our web site at: / Ou visitez notre site web à:http://www–frec.bull.com (PUBLICATIONS, Technical Literature, Ordering Form)

CEDOC Reference #No Référence CEDOC

QtyQté

CEDOC Reference #No Référence CEDOC

QtyQté

CEDOC Reference #No Référence CEDOC

QtyQté

_ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ]

_ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ]

_ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ]

_ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ]

_ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ]

_ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ]

_ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ] _ _ _ _ _ _ _ _ _ [ _ _ ]

[ _ _ ] : no revision number means latest revision / pas de numéro de révision signifie révision la plus récente

NOM / NAME : Date :

SOCIETE / COMPANY :

ADRESSE / ADDRESS :

PHONE / TELEPHONE : FAX :

E–MAIL :

For Bull Subsidiaries / Pour les Filiales Bull :

Identification:

For Bull Affiliated Customers / Pour les Clients Affiliés Bull :

Customer Code / Code Client :

For Bull Internal Customers / Pour les Clients Internes Bull :

Budgetary Section / Section Budgétaire :

For Others / Pour les Autres :

Please ask your Bull representative. / Merci de demander à votre contact Bull.

Page 203: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE
Page 204: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

BULL CEDOC357 AVENUE PATTONB.P.2084549008 ANGERS CEDEX 01FRANCE

86 A2 55KX 02ORDER REFERENCE

PLA

CE

BA

R C

OD

E IN

LO

WE

RLE

FT

CO

RN

ER

Page 205: HACMP for AIX Planning Guide - Atossupport.bull.com/.../aix-hacmp/hacmp4.4/g/86Y255KX02/86A255KX02.… · Bull HACMP 4.4 Planning Guide AIX Software August 2000 BULL CEDOC 357 AVENUE

Utiliser les marques de découpe pour obtenir les étiquettes.Use the cut marks to get the labels.

AIX

86 A2 55KX 02

HACMP 4.4Planning Guide

AIX

86 A2 55KX 02

HACMP 4.4Planning Guide

AIX

86 A2 55KX 02

HACMP 4.4Planning Guide