© 2009 IBM Corporation zLinux Measurements and Architecture Clea Zolotow Senior Technical Staff...

61
© 2009 IBM Corporation zLinux Measurements and Architecture Clea Zolotow Senior Technical Staff Member IBM IT Delivery (borrowed extensively from the ECM z/Linux Team.) April 2009

Transcript of © 2009 IBM Corporation zLinux Measurements and Architecture Clea Zolotow Senior Technical Staff...

© 2009 IBM Corporation

zLinux Measurements and Architecture

Clea Zolotow

Senior Technical Staff Member

IBM IT Delivery(borrowed extensively from the ECM z/Linux Team.)

April 2009

© 2009 IBM Corporation

IBM Virtualization – Enterprise Data Center Journey

Agenda

• IBM Commitment/Announcement Highlights

―IBM Transformation and “Big Green”

―Internal Infrastructure Challenge, Approach and Benefits

• IBM Virtualization Update

―Virtualization Progress

―Business Case Approach and Client View of Savings

―Application and Workload Selection

―Successful Techniques and Lessons Learned

© 2009 IBM Corporation

Project ‘Big Green’

IBM to reallocate $1 billion each year

To accelerate “green” technologies and services

To offer a roadmap for clients to address the IT energy crisis while leveraging IBM hardware, software, services, research, and financing teams

To create a global “green” team of almost 1,000 energy efficiency specialists from across IBM

Re-affirming a long standing IBM commitment

Energy conservation efforts from 1990 – 2005 have resulted in a 40% reduction in CO2 emissions and a quarter billion dollars of energy savings

Annually invest $100M in infrastructure to support remanufacturing and recycling best practices

Double compute capacity with no increase in consumption or impact by 2010

IBM will consolidate and virtualize thousands of servers onto approximately 30 IBM System z™ mainframes

Substantial savings expected in multiple dimensions: energy, software and system support costs

The consolidated environment will use 80% less energy and 85% less floor space

This transformation is enabled by the System z sophisticated virtualization capability

Major proof point for Project Big Green

© 2009 IBM Corporation

IT Organizations are Challenged by Operational Issues

Challenges

Rising costs of systems and networking operations

Explosion in volume of data and information

Difficulty in deploying new applications and services

Security of your assets & your clients’ information

Landslide of compliance requirements

Systems and applications need to be available

Rising energy costs & rising energy demand

Power & thermal issues inhibit operations

Environmental compliance & governance mandates

Costs & Service Delivery

Business Resiliency

& Security

EnergyRequirements

© 2009 IBM Corporation

Data Center Efficiencies Achieved Consolidation of infrastructure, applications Enterprise architecture optimization Global resource deployment

IBM Metrics 1997 Today

CIOs 128 1

Host data centers 155 7

Web hosting centers 80 5

Network 31 1

Applications 15,000 4,700

TE

CH

NO

LO

GY

IBM’s Globally Integrated Enterprise Data Centers

IBM Strategic Delivery Model

GlobalResources

Strategic IGA Location

Strategic Web Locationfor IGA

Ethernet & Power9 Networks

Next Level of Infrastructure Challenge Floor space challenges in key facilities Underutilized assets in outdated Web infrastructure Continued infrastructure cost pressure Increase % IT spending to transformation initiatives

© 2009 IBM Corporation

Stages of Adoption: IBM Journey

Physical consolidation of data centers, networks and applications

Simple like-for-like server and storage virtualization

Service tools, energy facilities mgmt

Drives IT efficiencySignificant progress toward highly virtualized

environment to enable pooled System z, Power Systems, System x and storage

Green production and advanced data center facilities

Shared service delivery model

Rapid deployment of new infrastructure and services

IBM Research “Cloud”Business-driven service

management pilotsGlobally Integrated Enterprise

Highly responsive and business goal driven

© 2009 IBM Corporation

Enterprise Business Value – Expectations

Business case

Energy savings

Quality service

Early modeling identified significant potential for savings through zLinux virtualization Performed TCO virtualization assessment on IBM portfolio as cross-IBM effort

• System z, SW Migration Services, STG Lab Services, IBM Academy, ITO Migration FactoryIdentified substantial savings opportunity

• Energy

• Floor space

Annual energy usage to be reduced by 80%

Total floor space to be reduced by 85%

• 11,045 square feet for distributed solution

• 1,643 square feet for System z solution

Leverages maturity of System z stack products - high availability, resiliencyReduces complexity and increases stability, centralizes service mgmt Potential for faster provisioning speed (months days)Dynamic allocation of compute power Provides world-class security

• Labor

• Software

Comparison of Annual Energy Usage

for Workloads

Distributed Solution System z Solution

Kilowatt hours (K) Cost* ($K) Kilowatt hours (K) Cost* ($K)

Power 24,000 $2,400 4,796 $479

Cooling** 14,400 $1,440 2,877 $287

Total Energy 38,400 $3,840 7,673 $767* Electrical cost calculated at rate of .10 per kW ** Cooling is 60% of power cost

© 2009 IBM Corporation

IBM Virtualization – Enterprise Data Center Journey

Agenda

• IBM Commitment/Announcement Highlights

―IBM Transformation and “Big Green”

―Internal Infrastructure Challenge, Approach and Benefits

• IBM Virtualization Update

―Virtualization Progress

―Business Case Approach and Client View of Savings

―Application and Workload Selection

―Successful Techniques and Lessons Learned

© 2009 IBM Corporation

IBM System z Linux Virtualization Progress

Established phased approach for quick wins

Migrated initial servers from early ‘wave’ teams• Thousands of servers inventoried

• Multiple successful migrations delivering benefits as expected

• Decommission pipeline of hundreds of servers for reuse or removal

Comprehensive project plan and management system in place

• Integrated business priorities with transformational objectives

• ‘Work in progress’ approach to maximize server migrations

• Pipeline, process, technical, finance and communications support

Developed internal business case (RACE*) • Created detailed cash flow and labor analysis, migration expense, iterated

Technical solution, education plan and operational plan developed• Built upon IBM prior consolidation/simplification efforts, utilizing IBM offerings and capabilities

Highest level of support from IBM senior executive team

Phase 1

Phase 2

Phase 3-n

“Quick wins”

Highest Savings First

Finish Virtualization

*Formerly zRace

© 2009 IBM Corporation

IBM System z Linux Virtualization Progress

IBM implementing New Enterprise Data Center through achievements in

• Server and storage virtualization

• Energy efficiency and resiliency improvements

Benefits are on track with expectations

• Migration management key

• Business case is compelling

• Using System z10 technology, the number of machines could be cut by about half, with greater savings in energy, floor space, software and support costs

Lessons Learned, including:

• Enterprise strategy and sponsorship needed to drive business case and execution

• Compelling business imperative accelerates execution and drives support

• Enterprise view of migration managed by waves drives experience; savings for investment

IBM experience is driving Time to Value initiatives, integrated into IBM capabilities

• Dramatic reduction in labor through new processes supporting workload migrations

• Fall in/out analysis, working with business units, to close gaps in workload pipeline

• Piloting new testing strategy, processes & tools to automate

5th Year

4th Year

3rd Year

2nd Year

1st Year

z9 Cumulative

Cumulative 5-Year Cost Comparison

DistributedCum

© 2009 IBM Corporation

Virtualization Benefits are Significant; Migration Management is Key

Clients are able to leverage IBM experience and

capabilities to accelerate value

Expected Benefits of Virtualization

Substantial savings in multiple dimensions: energy, software and system support costs

80% less energy, 85% less floor space for consolidated environment

Improved inventory hygiene, including application to server mapping

Dramatically faster provisioning

Improved security and resiliency

Higher quality through reduced complexity, increased stability and availability

Large Scale Migration Challenges Exist

Decision-making: Integrating Enterprise and Business Unit view

Mindset/Culture related to distributed and mainframe worlds

Workload selection - multidimensional nature of selection process

Dated inventory records that are not centrally maintained

Detailed data required for internal business case

Project and program complexity – integrating multiple priorities

© 2009 IBM Corporation

Business Case Leveraged RACE Tool, Iterative Approach

Utilized RACE commercial modeling tool

• Foundation for internal business case, constructed specific environmental variables

.

Engaged SME’s within IBM

• Provided business case assumptions (i.e. depreciation/maintenance), modified as appropriate

Iterative Process

• Continuously engaged with core SME’s to ensure most current information

Project Metrics

• Weekly report of migrated servers and their disposition status (reuse or disposal using GARS*) and Energy Certificate status

- Working to incorporate actuals into the Business Case such that we can refresh our assumptions

Created financial plan for “known universe”

• Identified relevant sample (5-10%) of most likely servers to be migrated and gathered financial profile information for each

*IBM Global Asset Recovery Services

© 2009 IBM Corporation

Integration

• Integrated Functionality vs. Functionality to be implemented (possibly with 3rd party tools)

• Balanced System

• Integration of / into Standards

Further Availability Aspects

• Planned outages

• Unplanned outages

• Automated Take Over

• Uninterrupted Take Over (especially for DB)

• Workload Management across physical borders

• Business continuity

• Availability effects for other applications / projects

• End User Service

• End User Productivity

• Virtualization

Skills and Resources

• Personnel Education

• Availability of Resources

TCO: A Range of IT Cost Factors – Often Not ConsideredAvailability

• High availability

• Hours of operation

Backup / Restore / Site Recovery

• Backup

• Disaster Scenario

• Restore

• Effort for Complete Site Recovery

• SAN effort

Infrastructure Cost

• Space

• Power

• Network Infrastructure

• Storage Infrastructure

• Initial Hardware Costs

• Software Costs

• Maintenance Costs

Additional development/implementation

• Investment for one platform – reproduction for others

Controlling and Accounting

• Analyzing the systems

• Cost

Operations Effort

• Monitoring, Operating

• Problem Determination

• Server Management Tools

• Integrated Server Management – Enterprise Wide

Security

• Authentication / Authorization

• User Administration

• Data Security

• Server and OS Security

• RACF vs. other solutions

Deployment and Support

• System Programming

―Keeping consistent OS and SW Level

―Database Effort

• Middleware

―SW Maintenance

―SW Distribution (across firewall)

• Application

―Technology Upgrade

―System Release change without interrupts

Operating Concept

• Development of an operating procedure

• Feasibility of the developed procedure

• Automation

Resource Utilization and Performance

• Mixed Workload / Batch

• Resource Sharing―shared nothing vs. shared everything

• Parallel Sysplex vs. Other Concepts

• Response Time

• Performance Management

• Peak handling / scalabilityRoutinely Assessed Cost Factors

© 2009 IBM Corporation

Client View of TCO Comparison for Similar Distributed Workload vs. System z Linux results in Potential 60-75% Gross Costs Savings / 5 yrs

* HW Acquisition compares server/disk refresh of distributed environment to the cost of acquiring new mainframes/storage

Software45%

Connectivity2%

HW Acquisition*17%

Labor19%

Facilities6%

HW Maintenance11%

Potential Savings: Categories as a % of Gross Savings

Dramatic Simplification

Unit DistributedSystem z

Linux%

Reduction

Software Licenses

26,700 1,800 93%

Ports 31,300 960 97%

Cables 19,500 700 96%

Physical Network Connections

15,700 7,000 55%

Operating Cost: Distributed vs. Mainframe

Distributed Cost System z Linux Cost

Rel

ativ

e C

ost

Facilities

HW Acquisition

HW Maintenance

Software

ConnectivityLabor

Results will vary based on several factors including # of servers and work load types

© 2009 IBM Corporation

By formally decommissioning servers, IBM is able to demonstrate energy savings and receive energy efficiency credits (EECs)

Energy Efficiency Certificates Deliver Savings

Diagnose

Build

Cool Virtualize

Manage & MeasureGreen

Data Center

Client requirementsLower energy costs and achieve business benefit of Energy

Efficiency

Demonstrate Energy Efficiency Commitment

SolutionVirtualized workloads onto System z platform and reduced energy

consumption

Hundreds of servers in pipeline to be redeployed, sent to GARS* and/or energy efficiency certificates issued

IBM applied for EECs for eligible decommissioned servers to receive Energy Efficiency Credits

GARS for asset reuse, recycling and/or reclamation

BenefitsQuantifable energy reductions, tradable certificates

Demonstrated commitment to energy efficiency

IBM Launches World's First Corporate-Led Energy Efficiency Certificate Program

In Conjunction With Neuwing Energy, Program Will Provide Clients Documentation and Third-Party Verification of the Energy Saving Results of Their Projects. Read more

November 2, 2007 Press Release

EECs quantify, measure, verify, certify and monetize data center energy efficiency projects

What is an Energy Efficiency Credit?

A Neuwing EEC (Energy Efficiency Credit) is a measured & verified Megawatt Hour (MWh) of Energy Savings i.e., Energy Efficiency

*IBM Global Asset Recovery Services

© 2009 IBM Corporation

IBM is Using a ‘Work in Process' Approach to Manage the Migration

Management Approach and Reporting

Process approach borrowed from factory line management

Metrics for each process and sub-process

Quality measured with process fallout – tracked by cause

Daily status calls for issue resolution

Weekly status reporting for CIO and management team

© 2009 IBM Corporation

Decommission Process Overview

Server available as a result of virtualization efforts

Capture savings in business plan and business case

Tracking tool is updated to reflect disposition of the assets in the project

Request completed to

coordinate shipping

and update property

control

Complete Machine List Database and

ship to GARS*

If redeployed If not redeployed

*IBM Global Asset Recovery Services

Server Ready

Check for technical viability and asset value to determine if h/w is a

redeployment candidate

Apply to Neuwing for energy efficiency

certificates

© 2009 IBM Corporation

Application ViewBusiness Unit Partnership

Environment ViewManaged ‘Offerings’

DevelopmentIntranet

BU Environments

Location ViewBoulder

PoughkeepsiePortsmouth

RaleighRochester Southbury

Technology ViewDominoEmail

Static Web DB2

Linux on x86

Migration Candidates Sourcing

Enterprise Approach to Workload Migration

© 2009 IBM Corporation

Each Workload is Evaluated for Suitability Based on Technical Attributes

Priority Workloads for Consolidation:

WebSphere® applications

Domino® Applications

Selected tools: Tivoli®, WebSphere® and internally developed

WebSphere MQ

DB2® Universal Database™

Work

load

s with

softw

are

avai

lable

on uniq

ue sy

stem

onlyUnique

require

men

ts®

Already easily

virtualized on AIX®/UNIX®

Decision criteria/thresholds can be tailored by the client

Technical Attributes Software compatibility or

platform dependency

Workload characteristics

Not fully virtualized / optimized

apply

Grey area (not clearly one platform or another)

Other components

of app on mainframe

High and/or transactional

I/O

Proximity to other data on

mainframe

Test in support of prod on

AIX®/UNIX®

Small counts of isolated images

Already virtualized on Intel®

Default is chosen

when no other criteria apply

© 2009 IBM Corporation

Blue Harmony is an enterprise transformation program that will enable IBM’s Globally Integrated Enterprise (GIE) vision

Blue Harmony will enable the GIE by …

• Integrating core processes that are now business unit and region-specific, such as Quote to Cash, horizontally across the enterprise on a common systems platform – SAP

• More effectively integrating the global processes that are already in place

―For example, Blue Harmony will integrate financial reporting with business reporting, providing real time access to consolidated data

The significance of the Blue Harmony name

• The name reflects how we are harmonizing process, technology and people across the IBM enterprise

© 2009 IBM Corporation

Blue Harmony: Main Hardware Components

Portsmouth, U.K.

Poughkeepsie, N.Y.

oc48 oc48

dcx-4schannel extenders

subnetCisco 3750E-48T

CD 6504

ATTWAN1200 GSR

Xseries

Tape

pair of 2027-140 Ficon directors

z10 z10

DS8300A DS8300B

Metro mirror

z/OS Global Mirror

2109-m482109-m48

2498-M48

p595 p595

2498-M48

Tape

© 2009 IBM Corporation

Blue Harmony Mainframe Configuration

Hardware – z/10 - 56-way – 7,674 MIPS on general CPUs

• 11 General Purpose CPUs (CPs) for z/OS

• 6 Internal Coupling Facilities (ICFs) for sysplex sharing

• 6 Integrated Information Processor (zIIPs) for DB2

• 17 Integrated Facility for Linux (IFLs)

Connectivity

• 10 OSA 1000 Base T (4 ports)

• 10 OSA 10GbE (2 ports)

• 2 ESCON (16 port)

• 7 ESCON Channel Ports

• 2 FICON Express4 (SX, 4 ports)

• 36 FICON Express4 (4KM, LX, 4 ports)

• 8 ICB4 (data rate of 2GBps)

© 2009 IBM Corporation

Blue Harmony US Implementation

Z/OS NABA/B/C/DSAP

Z/OS NABF/G/H/ISAP

SY

SP

LE

X 1

Z/OS DB2

Z/OS DB2

ICF

8ICB-4

TCP/IP

Hypersockets(QDIO)

Z/VM

Z/L

iNU

X 1

Z/L

iNU

X 2

Z/L

iNU

X 3

Z/L

iNU

X 4

Z/L

iNU

X 5

Z/L

iNU

X 6

Z/VM

Z/L

iNU

X 1

Z/L

iNU

X 2

Z/L

iNU

X 3

Z/L

iNU

X 4

Z/L

iNU

X 5

Z/L

iNU

X 6

GDPSRESTORE

OSA

Blue Harmony Mainframe 1 – Serial Number A40D0

ICF

8ICB-4

OSA

Blue Harmony Mainframe 2– Serial Number 8D5FC

2-I

Cs

2-IC

s

FIRST Z/10 SECOND Z/10 BlueZone

© 2009 IBM Corporation24 IBM Confidential24

What is z/Linux at IBM

What is z/Linux

• Provisioned underneath a hypervisor

• Provisioned barebones (not supported within my current architecture)

Currently supported are:

z/ECM: The Enterprise Computing Model (ECM) is an IBM initiative to virtualize 25 percent of our own distributed IT infrastructure onto System z mainframes. ECM is also an IBM showcase for Project Big Green and the New Enterprise Data Center.

Blue Harmony: IBM's major transformation program to radically simplify the design and operation of three Globally Integrated Support Processes - Finance, Opportunity To Order, and Order To Cash

© 2009 IBM Corporation25

What is z/Linux the OS

z/VM is an operating system that runs on the z/Series mainframes. It functions both as a fully functional operating system as well as as hypervisor, much like VMWARE.

z/VM builds on more than three decades of virtualization experience. z/VM’s CP (Control Program) is a hypervisor which has been built with isolation in mind and extended this to in-memory network virtualization.

The System z based workload and as such Linux on z based services are ideal for low utilized services which exploit the fast context switching capabilities provided by the hardware, the z/VM hypervisor and the many hardware assist functions.

z/VM allows the virtualization of memory, CPU cycles, storage, consoles, and even can emulate non-existing hardware like for example a virtual zAAP processor to speed up JAVA workload.

The IO subsystem provides many advantages to high volume IO workloads, and for LINUX on z SCSI devices can be assigned as well as the mainframe typical CKD devices.

From ACS, “Mainframe on the Cloud”From ACS, “Mainframe on the Cloud”

© 2009 IBM Corporation

IBM’s Enterprise Architecture for ECM: Security Zones

Possible Architecture Path for Single LPAR with Multiple Security Zones and Multiple Clients is represents a possible solution to have multiple security zones and multiple customers (MSZ) on the same LPAR.

This would be used for supporting small and medium business and Clients who deploy multi-network-zone tiered architectures where firewall separation of zones happens externally.

This architecture is currently in place today supporting IBM internal businesses.

© 2009 IBM Corporation27

Enterprise Architecture

Security zone 2

SANs

VLAN Security Zone 3

Vlan

Vlan

Security zone 1 VPN

Firewalls

VLzH Operational Model - Conceptual

Tools Network

SRM GCMS TLM ? eESM Tivoli TEC & TMR IIP

Tivoli Gateway

Tivoli Gateway

backup

backup

Operations team System

Administrators

Build Server

Build Server

Build Server

Build Server

)

Build Server

Build Server

Hardware Asset Mgt Security

Management

Dynamic Resource

Prov

Other Config Mgt

Provisioning Server

Proxy Load Balancer

Internal Network

Proxy Load Balancer

Z System

LPAR

LPAR

LPAR

Internet

Single LPAR on z

Possible design path for using a single LPAR with multiple security zones or multiple Clients. This involves network

design and security approval

Net

wor

k V

irtua

lizat

ion

© 2009 IBM Corporation

How to manage and access the Linux guests

The four major components for managing and access the Linux guests are:

Hardware Monitor Console (HMC)

Integrated Communications Console (ICC)

VSWITCH

z/VM Management

The z/VM management for the LPARS can share the same physical ports while the VSWITCH(s) have dedicated ports.

http://news.cnet.com/8301-19413_3-10237274-240.htmlhttp://news.cnet.com/8301-19413_3-10237274-240.html

© 2009 IBM Corporation

How to manage and access the Linux guests: HMC

Hardware Monitor Console (HMC)

The HMC is the lowest level of management connectivity of a physical zSeries processor. Through the HMC hardware components are mapped and assigned to various parts of the box and defines the bounds of the different logical partitions inside the box. Operations use the HMC to implement the configuration of the box and IPL LPARs.

The HMC is physically connected to the processor by two Service Element ports.

These HMCs provide the interface to configure the processor. An administrator can physically access the HMC to configure the box or remotely connect to the HMC via an internal web service.

Private Network

Hardware Master Console(Linux or OS/2)

ServiceElement

ServiceElement

Rest of Network

© 2009 IBM Corporation

How to manage and access the Linux guests: ICC

Integrated Communications Console (ICC)

Operator management of the zSeries processor can also be made through the Integrated Communications Controller (ICC). The two ICC_OSA ports are connected to two different LAN segments for redundancy in case there are network issues accessing one of the ports. The port has an IP address that is routed in the network for AF/Remote Server connectivity can be established. OSA-ICC’s are shared between LPARS.

The AF/Remote Server communicates to the ICC_OSA via TN3270 traffic. Administrators utilize a software client to connect to the AF Remote Server which then downloads a list of LPARs the administrator is able to control. When the administrator opens a connection to the LPAR, the connection is proxied through the AF Remote Server before being passed to the processor.

The client on the administrator’s workstation communicates to the AF Remote Server over port 1035 (default). From there, the AF Remote Server communicates with the ICC_OSA over port 1024 (default).

Access to ICCCopper Only

OSA pairTN3270 traffic

AF Remote ServerSA IOM

(Windows Server)

AF Remote ServerSA IOM(backup)

ICC_OSA ICC_OSA

© 2009 IBM Corporation

How to manage and access the Linux guests: VSWITCH

The four major components for managing and access the Linux guests are:

VSWITCH

The z/VM management for the LPARS can share the same physical ports while the VSWITCH(s) have dedicated ports. External connectivity to a network from inside an LPAR can be established via a Layer 2 VLAN aware VSWITCH.

The OSA ports plug directly into a Layer 2 switched distribution box. This utilizes the VSWITCH as the access switch in a multi-tiered architecture. The deployments in questions however vary on how the VLANs are deployed to the given LPARs.

VMController

OSA

VSwitch

LPAR

zVM

Linux Linux

OSA

Port 1 Port 2

Port 52Port 51

Linux

Port 50

zSeriesProcessor

VMController

VSwitch

LPAR

zVM

Linux Linux

Port 1 Port 2

Port 52Port 51

Linux

Port 50

VMController

OSA

VSwitch

LPAR

zVM

Linux Linux

OSA

Port 1 Port 2

Port 52Port 51

Linux

Port 50

zSeriesProcessor

VMController

VSwitch

LPAR

zVM

Linux Linux

Port 1 Port 2

Port 52Port 51

Linux

Port 50

SiSi SiSi

Backup OSA

SD Block

SiSi SiSi

Service Core

VMController

OSA

VSwitch

LPAR

zVM

Linux Linux

OSA

Port 1 Port 2

Port 52Port 51

Linux

Port 50

zSeriesProcessor

VMController

VSwitch

LPAR

zVM

Linux Linux

Port 1 Port 2

Port 52Port 51

Linux

Port 50

HSRP Active

OSA OSA OSA OSAOSA OSA

© 2009 IBM Corporation

How to manage and access the Linux guests: z/VM TCPIP

The four major components for managing and access the Linux guests are:

z/VM Management TCPIP

Once an LPAR is created on the processor, with a TCP/IP Stack for CP/CMS access, an administrator can logon as a controller session to handle the creation and management of Linux guests inside the LPAR. The controller only manages inside the LPAR and is separated from the production traffic of the Linux Guests. The TCP/IP stack will have a VIPA IP address with two OSA interfaces for redundancy.

The first option shows OSAs pairs that are used for shared management only, across all z/VM instances that exist on a single processor. It is defined by sharing the OSA across all LPARs which is separate from production traffic. Access to these management OSAs need connectivity to the network segments where administrators sit.

OSA

VMController

VSwitch

zSeriesProcessor

zVM

Linux

Port 1 Port 2

Port 52Port 51

Linux

Port 50

Group 1

Linux

OSA

Port 49

OSA

VMController

VSwitch

LPAR

zVM

Linux

Port 1 Port 2

Port 52Port 51

Linux

Port 50

Group 2

Linux

OSA

Port 49

OSA

VMController

VSwitch

LPAR

zVM

Linux

Port 1 Port 2

Port 52Port 51

Linux

Port 50

Group 3

Linux

OSA

Port 49

OSA OSA

LPAR

© 2009 IBM Corporation

Performance and Capacity Tooling

To the right, you can see the overlap of capacity planning and performance.

Note that both team rely on the VMACCT data. Although Omegamon has an agent, it is not providing good response time in our environment.

SRM data utilizes standard SUSE10 monitors which are used illegally (as the CPU time is incorrect due to the virtualization). However, midrange people prefer the interface.

Work is ongoing to load the VMPTK hypervisor data into SRM to correct the CPU utilization data.

VMPTK does provide several linux level reports lxmemory and lxcpu logs that show memory and cpu usage at the linux level. (VM- Performance Tool Kit).

Z9 ECS38/512GB

LINUX1

LIN

UX

1

LIN

UX

2

LIN

UX

3

LIN

UX

4

LIN

UX

5

LIN

UX

6

SLES10 64bit

WAS DB2

Application

Omegamon XESAR

vmstatps

NmonITM

PTKOmegamon XE

ELAPSE

Other/Application Specific

Capacity

zRACE 1080z/VM Capacity Planner

SCPON (internal)

SRM (example)Smoothstart

Software Requirements

Other/Application Specific

Performance

Reduce Agents on GuestEach agent consume 2% MIPS

© 2009 IBM Corporation

Performance and Capacity: Separation of duties

Note the different responsibilities here between Capacity, z/VM, Performance, and z/Linux.

The interesting thing here was that z/VM retained their z/VM skills and z/Linux moved to folks who had previous mid-range UNIX skills.

Z9 ECS38/512GB

LINUX1 LINUX2 ZOS1

LIN

UX

1

LIN

UX

2

LIN

UX

3

LIN

UX

4

LIN

UX

5

LIN

UX

6

SLES10 64bit

WAS DB2

Application Application Developers

Middleware Competency

zLinux Performance

z/VM Performance

Capacity Planning

Responsibilities

© 2009 IBM Corporation

Tooling for Alerting

Z/VM LPAR Native Console and Hardware Alerting

HMF, WWAO Native Console alerts can be filtered and acted upon by HMF and ultimately focal pointed to another Z/OS system via PSM with WWAO.

Z/VM LPAR CPU PTK and Omegamon XE Performance Toolkit and/or Omegamon XE provide the ability to proactively monitor CPU utilization.

Linux Guest Linux Middleware - DB2 / MQ Series

ITM for Databases*Omegamon XE for Messaging (ITM 6.1)

"ITM for xxxx" are monitoring extensions whch are used for either standard middleware like Lotus DB2, MQ, Oracle - or industry wide applications like Notes, Siebel, SAP. A monitoring extension will contain both resource monitors checking for something specific (i.e a DB2 table space becoming full) or checking logs (i.e the MQ Series error log)

Linux Guest OS - Linux IBM Tivoli Monitoring Base - OS Agent ( Disk, CPU, OS Processes)*Logfile adapters / Log Monitors for OS Logs

This is primarily handled through some form of OS Agent monitoring. The key advantage is that no matter how many occurrences of a problem you get, there will be one alert generated until the problem is resolved. A logfile adapter can also be utilized as an adhoc monitoring aid where NO existing agent can satisfy the requirement.

© 2009 IBM Corporation

ITM6 Tooling Integration

Middleware and Applications

Customer Locations(Customer Premise/Extended

Premise)

(VM Lpars)Mainframe Customer Systems

(SuSE, Redhat, etc.)Linux Customer

Systems

Systems Management Infrastructure

IBM Locations(IBM Intranet / SNI Network)

Omegamon / ITM 6

Infrastructure

Intranet User

Problem Management

Service Management

View

Rep

orts

Situ

atio

n D

ata

Age

nt C

onfig

urat

ion,

Pol

icy

& S

ituat

ions

T/EC

Internet User

View

Portal D

ata

System AdminsOperations

ServerV

iew P

ortal Data

View

Portal D

ata

View

Portal D

ata

Configure M

onitoring

Update Portal Status

Receive Event / Perf Data

Future Integration

TDW

Situ

atio

n D

ata

Age

nt C

onfig

urat

ion,

P

olic

y &

Situ

atio

ns

Long Term Reporting

DB, SQL, Oracle

Websphere, Exchange

Situation Data

Agent Configuration, Policy & Situations

Situation Data

Agent Configuration, P

olicy & Situations

zOS Focal

z/VM LPAR

z/VM

Operations

View

Event D

ata

z/VMFo

cal P

oint

ing

© 2009 IBM Corporation

Tooling: Tivoli

• The Tivoli Enterprise Management Server (TEMS) can be seen as the central repository for all monitoring data and resources. In particularly it stores the definitions for conditions that indicate a problem with a particular resource, these are known as situations. It also controls the security for the monitoring solution. Each enterprise monitoring solution must contain one Hub TEMS, and can include multiple Remote TEMS. Unlike previous Tivoli Monitoring products where the lower levels of the Architecture (Endpoint or Gateway) interface directly with event management, it’s the HUB TEMS that controls forwarding of Events to the T/EC Component.

• The Tivoli Enterprise Portal Server (TEPS) functions as a graphical access server that displays near time and real time status indications on the health of resources in the enterprise. It also serves as a repository for all user data, such as the user IDs, user access control for the monitoring data, meaning what data each user will be able to access, and how it is displayed. It connects to the Hub TEMS and can be accessed by the TEP clients who are either desktop clients or browser clients. Data is only available in the TEPS for a maximum of 24 hours, and the component should not be considered a historical reporting capability. The TEPS Database is also used to store internal data used by the TEPS (TEP user IDs, queries, and navigator views).

• Tivoli Enterprise Console Server provides a centralized location for the management of events. The server creates an entry in the event database for each incoming event and then evaluates these events against a set of rules to determine if the event server should automatically perform any predefined tasks or modify the event. If human intervention is required, the event server notifies the appropriate operator. The operator performs the required tasks and then notifies the event server when the condition that caused the event has been resolved. In addition, there is the option of enabling a form of automated trouble ticketing to a desired Problem Management system.

• Tivoli Enterprise Monitoring Agents (TEMA) are the data collectors for the ITM monitoring solution. They are installed to gather data from one or multiple systems that are critical to the client business.

• Data that is collected by the agents is sent to the central Tivoli Enterprise Monitoring Server. Agents are type-specific (I.e. Unix, Linux, SAP, DB2, Omegamon XE on z/VM and Linux). The Omegamon XE on z/VM and Linux agent will be an integral part of this solution and helps address performance bottlenecks on z/VM resources. The three main areas where monitoring can be beneficial are:

• Monitoring processor usage at the system level

• Monitoring paging and spooling activity

• Monitoring resources and workloads across LPARs

© 2009 IBM Corporation

Monitoring Diagram for Existing ECM Farm

Existing z/os Mainframe

Existing WWAO z/OS Mainframe Focal Point

Existing z/os Mainframe

New ECMz Mainframe

Problem Ticketing system

Existing z/VM MainframeNew ECMz Mainframe

Existing Customer Servers

Existing Tivoli Enterprise Console

(TEC)

Tivoli Enterprise Management Server

(TEMS)Tivoli Enterprise Portal Server(TEPS)

Mainframe OperatorWorkstation

Distributed OperatorWorkstation

Existing Auto TicketingExisting Auto Ticketing

Existing Problem Management

Existing Distributed Operator alerting and command/response

Existing Mainframe Opertor alerting and command/response

Linux Guest Alerting

ECMz MonitoringArchitectural

Overview Diagram

Version 2.0 26/11/07 Richard Short Version 3.0 March 10, 2008 Gene Capano

Performance Monitoring data

z/VM Alerting

Existing Auto Ticketing

Tivoli Enterprise Monitoring Agent

(TEMA)

Tivoli Enterprise Monitoring Agent

(TEMA)

© 2009 IBM Corporation

Tivoli Monitoring Flow

Local Cache

Agents

TEMS Hub

TEPS

TEPS DB

TEMS Data (Built in operational DB)

Warehouse

TEP Browser or Desktop

MonitoringInfrastructure

TEMS Data Remote TEMS

Local Cache

Agents

Local CacheLocal Cache

AgentsAgents

TEMS HubTEMS Hub

TEPSTEPS

TEPS DBTEPS DB

TEMS Data (Built in operational DB)TEMS Data (Built in operational DB)

Warehouse

TEP Browser or Desktop

MonitoringInfrastructure

TEMS DataTEMS Data Remote TEMSRemote TEMS

Local CacheLocal Cache

AgentsAgents

Key:

TEP – Tivoli Enterprise PortalTEPS – Tivoli Enterprise Portal ServerTDW – Tivoli Data WarehouseTEMS – Tivoli Enterprise Monitoring ServerTEMA – Tivoli Enterprise Monitoring Agent

© 2009 IBM Corporation

z/VM and z/Linux Monitoring Flow

Here is the monitoring flow from the VMPTK into and the previous Omegamon XE solution (now lxmemory and lxcpu from VMPTK).

Our data ends up in a DB2 DB on z/OS.

Not much has really changed in 20 years!

CP Control

Blocks

Application

DataVM Events

*MONITOR System Service

MONDCSS

Segment

MONWRITE Utility

Performance

Toolkit

Raw

Monwrite

History

Files

HTTP

TCP/IPNetwork

DCSSSEGOUT

LINUXGuest

3270

Browser

OMEGAMON XE

VMRM

TEMS

VMIRA

© 2009 IBM Corporation

How Tivoli interacts with the z/Linux Guests

• This is a logical diagram, not a physical one.

• Note that all data flows from the z/VM host.

• Internal monitors are supported by SRM, but not by Tivoli.

z/VMz/VM

PTK

Extensions

CMD(Guest)

Linux(Guest)Linux

(Guest)Linux

(Guest)Linux

(Guest)Mon

VMIRA

LinuxIRA

LinuxIRA

Linux(Guest)Linux

(Guest)DCSSDCSS

LinuxIRA

LinuxIRA

LinuxIRA

LinuxIRA

TEMS

TEPSTEP TDW

Tivoli Management Services

z/VM command and response

flows

Data and synchronization flows

CPCP

© 2009 IBM Corporation

Autoprovisioning using IBM Director

IBM Director Administrator

Console

IBM Director Mgmt Server

Linux SuSE

CIMServer

z/VM MAP

NIC 1

NIC 2

VSWITCH

MAPLAN

MAPSERVE

TCPIP

VSMSERV

PORTMAP

NIC 1

OSA

DIRMAINT

VMRMSVMLinux RH

Linux S10

Linux SuSE

KickStart Server

OS ImagesSoftware Packages

NIC 1

NIC 1

NIC 1

OSA

z/VM Center Console extension

z/VM Center Server extension

z/VM Mgmt Access Point Agent extension

Director Console

Director Server

SBLIM Client

CIMServer

NIC

NIC

LAN

TPM Server

TPM for Software

Fusion

NIC

Global Delivery CentersGlobal Delivery

CentersGlobal Delivery Centers

Linux RH

Linux RH

Linux S10

Linux S10

1

1

2

3

4

2

z/VM

1

Cloning Solution(Canada)

(IGA + Commercial)

IBM Director SWD

Virtual Machine Manager

NIC

Tactical ApproachStrategic Solution

Breeder Solution(EMEA)

(IGA + Commercial)

Cloning/Breeder Solution provide subset of the strategic solution

Remote Access to IBM Director and other Consoles(via Remotely Anywhere)

© 2009 IBM Corporation

Autoprovisioning: The way it really works (cloning)

© 2009 IBM Corporation

Autoprovisioning: Why

• z/VM Center manifests itself inside the z/VM hypervisor as a zLinux service machine, making available the CIM Instrumentation (CIM is also known as Pegasus, and stands for Common Information Model) for z/VM, and creates the API by means of the z/VM MAP function. As seen above, both, IBM DIRECTOR Virtualization Manager and the IBM TIVOLI Provisioning Manager exploit the same API and CIM as both products have been designed to be platform (Operating System) agnostic.

• The flexibility by separating the GUI from the automation and the hardware line it runs on makes it also inflexible for zLinux deployments. So far it is not clear how much it can be tweaked to fulfill the above mentioned target architecture requirements. Testing needs to be performed and with a focus on exploiting Exits rather than changing the product.

© 2009 IBM Corporation45 IBM Confidential45

Backup Methodology.

zLinux Operating System (OS) image and essential filesystems and data to the zLinux OS will be maintained on FICON/ESCON attached disk. OS data and images are not backed up using the same technology as Application Data. The primary technology deployed for backup of Application Data on zLinux guests is the IBM Tivoli Storage Manager (TSM) using Automated Tape Libraries (ATLs) or newer Virtal Tape Libraries (VTSs) to store backup data.

© 2009 IBM Corporation46

Backup and Recovery Model

Security Zone 4

LAN

Security Zone 3

LAN

Security Zone 2LAN (DMZ)

ECM System Z Servers

Unix/AIX Servers

NE

TW

OR

K F

IRE

WA

LL

NE

TW

OR

K F

IRE

WA

LL

LTO drives in ATLs

359x Drives in VTSs ECM System Z

Servers

Unix/AIX Servers

FCP

Unix/AIX Servers

LTO drives in ATLs

Sec

urity

Zon

e 1

(In

tern

et)

NETWORK FIREWALL

ECM/z BZ AU

ECM/z RZAU

ECM/z RZ (Internet)Application User (AU)

Storage Services for ECM/z: Architecture Overview Diagram

MSmith IBM GTS Stor Comp 2007-11-30

FICON

Datacentre Perimeter

TSM Backup Servers

TSM Backup Servers

TSM Backup Servers

FCP

FC

P

FICON

FCP

FCP

LTO drives in ATLs

FCP

FCP

DS8300

359x Drives in VTSs

ECM/z BZ SAs

ECM System Z Servers

MSS YZ SAs &SAN Designers

DS8300

FCP

FC-SAN FCP

FCP

FC-SAN

FCP

FCP

FICON

FICON

Security Zone 3Security Zone 2Security Zone 1

Security Zone 4

FCP

FCP

© 2009 IBM Corporation47

Backup

© 2009 IBM Corporation48 IBM Confidential48

Defining cloud computing.

Cloud computing is a new consumption and delivery model inspired by consumer Internet services. Cloud computing exhibits the following 5 key characteristics:

•On-demand self-service •Ubiquitous network access•Location independent resource pooling•Rapid elasticity•Pay per use

© 2009 IBM Corporation

What does it really mean?

On-demand self-service. A consumer can unilaterally provision computing capabilities, such as server time and network storage, as needed without requiring human interaction with each service's provider.

Ubiquitous network access. Capabilities are available over the network and accessed through standard mechanisms that promote use by heterogeneous thin or thick client platforms (e.g., mobile phones, laptops, and PDAs).

Location independent resource pooling. The provider's computing resources are pooled to serve all consumers using a multi-tenant model, with different physical and virtual resources dynamically assigned and reassigned according to consumer demand. The customer generally has no control or knowledge over the exact location of the provided resources. Examples of resources include storage, processing, memory, network bandwidth, and virtual machines.

Rapid elasticity. Capabilities can be rapidly and elastically provisioned to quickly scale up and rapidly released to quickly scale down. To the consumer, the capabilities available for rent often appear to be infinite and can be purchased in any quantity at any time.

Pay per use. Capabilities are charged using a metered, fee-for-service, or advertising based billing model to promote optimization of resource use. Examples are measuring the storage, bandwidth, and computing resources consumed and charging for the number of active user accounts per month. Clouds within an organization accrue cost between business units and may or may not use actual currency.

http://news.cnet.com/8301-19413_3-10237274-240.htmlhttp://news.cnet.com/8301-19413_3-10237274-240.html

© 2009 IBM Corporation

Pay no attention to the man (person) behind the curtain!

© 2009 IBM Corporation

How does this apply to Z/Linux?

Provide on demand service levels.

Fast service provisioning.

Capacity Planning

SO must prepare to react on IBM’s marketing strategies (Cloud and Virtualization)

Cost:

Datacenter

Hardware

Management

Hypervisor Management

Author: Gerhard WidmayerAuthor: Gerhard Widmayer

© 2009 IBM Corporation

How does this apply to Z/Linux? (Service)

Provide on demand service levels.On demand service levels are for example “zero planned outage”, “scale-up” and “scale-out”, or “pay per usage”. The service levels can only be provided in a strict virtualized server resource pools infrastructure. SMB customers can not afford the hardware price for a cluster which they would utilize only at a small percentage.

Author: Gerhard WidmayerAuthor: Gerhard Widmayer

© 2009 IBM Corporation

How does this apply to Z/Linux? (Provisioning using the free pool)

Fast service provisioning

Server ordering and infrastructure adjustments is making provisioning a week long activity at best. This is also true because SO must create a new design for each service request rather than establishing a service inside a predefined infrastructure unless there is an enterprise architecture, such as ODCS. Competition is delivering on demand basic server services within 24 hours after accepting the request.

Using a virtualized free pool there will always be some resource available in the headroom in an appropriate resource pool (selected target location) if it is built to be multi-tenant enabled. Any service establishment is just a set of commands away, true for servers, storage, but also the real network. Increase of the resource server pool is only needed when thresholds are reached.

Patent 7685283 shows how to size for the free pool, “METHOD FOR MODELING ON DEMAND FREE POOL OF RESOURCES”

Author: Gerhard Widmayerwith adds by Clea

Author: Gerhard Widmayerwith adds by Clea

© 2009 IBM Corporation

How does this apply to Z/Linux? (Capacity Planning)

Capacity Planning –

Behind the curtain, define servers by actual consumption rather than by peak workload plus uplift.

Today’s distributed servers face the problem of hardware underutilization. Normal utilization values seen are 8% average in the distributed world.

Currently, the customer pays what was ordered regardless of usage. At the ODCS, there was an LPAR charge and a utilization charge.

In order to establish “on demand” pricing there is a need to put multiple services onto a single box, exploiting virtualization, like the ODCS. With the ever increasing compute power year by year the “boxes” became that big that utilizing single boxes by single network zones or SMB customers is very unlikely.

The ability to do coadunation is the ability to put each virtualized LPAR in it’s mathematically correct spot.

Author: Gerhard Widmayerwith adds by Clea

Author: Gerhard Widmayerwith adds by Clea

© 2009 IBM Corporation

Coadunation Example

Mixing workload shares headroom but you pay in response time at low utilization....workload management shifts peaks based on business priorities to use

"white space" but response time of lower priority work is traded off...

© 2009 IBM Corporation

How does this apply to Z/Linux? (Moolah)

Datacenter Cost

With the Green IT there is a drive to move towards larger boxes for network, storage, and servers in order to optimize the capacity per watt. For optimization of environmental savings (space, power, cooling) these large boxes need to perform at a high utilization level, which requires a very flexible setup with many isolated network zones on it.

Migrating pSeries hosts into z/Linux is proving to be quite cost-effective. The ECM consolidated environment will use 80% less energy and 85% less floor space.

Specific energy savings follows.

Hardware Cost

HW acquisition and maintenance cost goes down from a distributed environment to a z/Series environment. Further, SPOF failover is cheaper as well in the mainframe world.

Software45%

Connectivity2%

HW Acquisition*17%

Labor19%

Facilities6%

HW Maintenance11%

Potential Savings: Categories as a % of Gross Savings

Author: Gerhard Widmayerwith adds by Clea

Author: Gerhard Widmayerwith adds by Clea

© 2009 IBM Corporation

Ubiquitous network access: Based on the vSWITCH

Author: Gerhard Widmayerwith adds by Clea

Author: Gerhard Widmayerwith adds by Clea

© 2009 IBM Corporation

VSWITCHES and OSAs

From a security risk mitigation perspective a distinct VSWITCH and dedicated physical OSA ports and real switches is worth a consideration.

However the basic design should verify that the VSWITCH and VLAN isolation if bundled together with customer and application traffic is secure. In such a case the (S)CF firewall becomes a PF (public firewall), and the 3/4-NIC design would be preserved.

Author: Gerhard Widmayerwith adds by Clea

Author: Gerhard Widmayerwith adds by Clea

© 2009 IBM Corporation

VSwitch, Continued

VSWITCH is a z/VM established virtualization of a real switch. In this deployment guidance the VSWITCH is replacing the real access layer switch from an Ethernet hierarchy. For each of the three (or four) NIC’s in the NIC virtual server design a distinct VSWITCH is built. Each customer receives its own VLAN number, and each network zone receives also its own VLAN number. This VLAN assignment is true for all three or four NICs, so in effect always a triple or quad of VLAN numbers are assigned to each single network zone.

All VLAN’s of a type, like all Extended Premises, are combined in a single VSWITCH. Because the design is for SMB environments, this means that multiple customers and/or multiple network zones share a VSWITCH. The VSWITCH function must therefore provide the VLAN isolation and must disallow any VLAN cross-over interaction.

Remember that the basic design keeps the security control for allowed interaction flows in the firewalls outside the System z box, where it exists today.

Author: Gerhard Widmayerwith adds by Clea

Author: Gerhard Widmayerwith adds by Clea

© 2009 IBM Corporation

How does this apply to Z/OS?

This page intentionally left blank.This page intentionally left blank.

© 2009 IBM Corporation

Questions?