Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

21
© 2012 IBM Corporation Information Management IBM Big Data Platform Frank Ketelaars → Robert Hartevelt Presales IBM Big Data

description

 

Transcript of Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

Page 1: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

IBM Big Data Platform

Frank Ketelaars → Robert HarteveltPresales IBM Big Data

Page 2: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

Page 3: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

“At the World Economic Forum last month in Davos, Switzerland, Big Data was a marquee topic. A report by the forum, “Big Data, Big Impact,” declared data a new class of economic asset, like currency or gold.

“Companies are being inundated with data—from information on customer-buying habits to supply-chain efficiency. But many managers struggle to make sense of the numbers.”

“Increasingly, businesses are applying analytics to social media such as Facebook and Twitter, as well as to product review websites, to try to “understand where customers are, what makes them tick and what they want”, says Deepak Advani, who heads IBM’s predictive analytics group.”

“Big Data has arrived at Seton Health Care Family, fortunately accompanied by an analytics tool that will help deal with the complexity of more than two million patient contacts a year…”

“Data is the new oil.”Clive Humby

The Oscar Senti-meter — a tool developed by the L.A. Times, IBM and the USC Annenberg Innovation Lab — analyzes opinions about the Academy Awards race shared in millions of public messages on Twitter.”

“Data is the new Oil”

“…now Watson is being put to work digesting millions of pages of research, incorporating the best clinical practices and monitoring the outcomes to assist physicians in treating cancer patients.”

In its raw form, oil has little value. Once processed and refined, it helps power the world.

Page 4: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

The characteristics of big data

Collectively Analyzing the broadening Variety

Responding to the increasing Velocity

Cost efficiently processing the growing Volume

Establishing the Veracity of big data sources

30 Billion RFID sensors and counting

1 in 3 business leaders don’t trust the information they use to make decisions

50x PB

TB2020

80% of the worlds data is unstructured

2010

4

Page 5: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

5

Big Data will impact every aspect of your business

5

Run Zero-latency OperationsAnalyze all available operational data and react in real-time to optimize processes.

Reduce the cost of IT with new cost-effective technologies.

Know Everything about your CustomerAnalyze all sources of data to know your customers as individuals, from channel interactions to social media.

Instant Awareness of Fraud and RiskDevelop better fraud/risk models by analyzing all available data, and detect fraud in real-time with streaming transaction analysis.

Innovate New Products at Speed and ScaleCapture all sources of feedback and analyze vast amounts of market and research data to drive innovation.

Exploit Instrumented AssetsMonitor assets from real-time data feeds to predict and prevent maintenance issues and develop new products & services.

Page 6: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

Business-centric Big Data enables you to start with a critical business pain and expand the foundation for future requirements

“Big data” isn’t just a technology—it’s a business strategy for capitalizing on information resources

Getting started is crucial

Success at each entry point is accelerated by products within the big data platform

6

Build the foundation for future requirements by expanding further into the big data platform

Page 7: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

1 – Unlock Big Data

• Customer Need– Understand existing data sources– Expose the data within existing content

management and file systems for new uses, without copying the data to a central location

– Search and navigate big data from federated sources

• Value Statement– Get up and running quickly and discover

and retrieve relevant big data – Use big data sources in new information-

centric applications

• Customer examples– Proctor and Gamble – Connect employees

with a 360° view of big data sources

• Get started with: IBM Data Explorer (formerly known as Vivisimo Velocity)

7

Page 8: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

2 – Analyze Raw Data

• Customer Need– Ingest data as-is into Hadoop and derive insight from it– Process large volumes of diverse data within Hadoop– Combine insights with the data warehouse– Low-cost ad-hoc analysis with Hadoop to test new hypothesis

• Value Statement– Gain new insights from a variety and combination of data

sources – Overcome the prohibitively high cost of converting

unstructured data sources to a structured format – Extend the value of the data warehouse by bringing in new

types of data and driving new types of analysis– Experiment with analysis of different data combinations to

modify the analytic models in the data warehouse

• Customer examples– Financial Services Regulatory Org – managed additional data

types and integrated with their existing data warehouse

• Get started with: InfoSphere BigInsights

8

Page 9: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

3 – Simplify your Warehouse

• Customer Need– Business users are hampered by the poor performance of

analytics of a general-purpose enterprise warehouse – queries take hours to run

– Enterprise data warehouse is encumbered by too much data for too many purposes

– Need to ingest huge volumes of structured data and run multiple concurrent deep analytic queries against it

– IT needs to reduce the cost of maintaining the data warehouse

• Value Statement– Speed – 10-100x faster performance on deep analytic

queries– Simplicity – minimal administration and tuning of the

appliance– Up and running quickly

• Customer examples– Catalina Marketing – executing 10x the amount of

predictive workloads with the same staff

• Get started with: PureData for Analytics /Netezza

9

Page 10: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

4 – Reduce costs with Hadoop

• Customer Need– Reduce the overall cost to maintain data in the warehouse –

often its seldom used and kept ‘just in case’– Lower costs as data grows within the data warehouse – Reduce expensive infrastructure used for processing and

transformations

• Value Statement– Support existing and new workloads on the most cost effective

alternative, while preserving existing access and queries – Lower storage costs– Reduce processing costs by pushing processing onto

commodity hardware and the parallel processing of Hadoop

• Customer examples– Financial Services Firm – move processing of applications and

reports to Hadoop Hbase while preserving existing queries

• Get started with: IBM InfoSphere BigInsights

10

Page 11: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

11

What’s so Special About Open Source Hadoop?

Scalable• New nodes can be added on the fly

Affordable • Massively parallel computing on commodity

servers

Flexible • Hadoop is schema-less – can absorb any

type of data

Fault Tolerant • Through MapReduce software framework

Storage• Distributed• Reliable• Commodity gear

• Parallel Programming• Fault Tolerant

MapReduce

Page 12: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

12

InfoSphere BigInsights – A Closer Look

BigInsights Engine

Accelerators

User Interfaces

Visualization Admin Console

Text Analytics

Application Accelerators

Integration

Databases

Information Governance

Content Management

Apache Hadoop/cloudera

IndexingMap Reduce +

Workload Mgmt Security

Dev Tools• Performance & workload

optimizations

• Unique text analytic engines

• Spreadsheet-style visualization for data discovery & exploration

• Built-in IDE & admin consoles

• Enterprise-class security

• High-speed connectors to integration with other systems

• Analytical accelerators

More Than Hadoop

Page 13: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

13

Hadoop is Well Suited for Handling Certain Types of Big Data Challenges

Analyzing larger volumes may provide better results

Exploring data – a sandbox for ad-hoc analytics

Larger data volumes are cost prohibitive with existing technology

Deriving new insightsfrom combinations of data types

Page 14: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

5 – Analyze Streaming Data

• Customer Need– Harness and process streaming data sources – Select valuable data and insights to be stored for

further processing– Quickly process and analyze perishable data, and

take timely action

• Value Statement– Significantly reduced processing time and cost –

process and then store what’s valuable– React in real-time to capture opportunities before

they expire

• Customer examples– Ufone – Telco Call Detail Record (CDR) analytics

for customer churn prevention

• Get started with: InfoSphere Streams

Streams ComputingStreaming Data

Sources

ACTION

14

Page 15: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

InfoSphere Streams - Streaming Analytics for Big Data

• Built to analyze data in motion

– Multiple concurrent input streams

– Massive scalability

• Process and analyze a variety of data

– Structured, unstructured content, video, audio

– Advanced analytic operators

• Enables Adaptive Real-Time Analytics– With Data Warehousing

– With Hadoop Systems

1515

Page 16: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

Current fact finding

Analyze data in motion – before it is stored

Low latency paradigm, push model

Data driven – bring data to the analytics

Historical fact finding

Find and analyze information stored on disk

Batch paradigm, pull model

Query-driven: submits queries to static data

Traditional Computing Stream Computing

Stream Computing Represents a Paradigm Shift

Real-time Analytics

1616

Page 17: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

Use case– Neonatal infant monitoring– Predict infection in ICU 24 hours in advance

Solutions– 120 children monitored :120K msg/sec, billion msg/day– Trials expanding to include hospitals in US and China

University of Ontario Institute of Technology

SensorNetwork

Stream-based Distributed Interoperable Health care Infrastructure

Solutions (Applications)

Event Pre-processer

Analysis Framework

Page 18: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

Entry points are accelerated by products within the big data platform

BI / Reporting

BI / Reporting Exploration / Visualization

FunctionalApp

IndustryApp

Predictive Analytics

Content Analytics

Analytic Applications

IBM Big Data Platform

Systems Management

Application Development

Visualization & Discovery

Accelerators

Information Integration & Governance

HadoopSystem

Stream Computing

Data Warehouse

2 – Analyze Raw Rata

InfoSphere BigInsights

5 – Analyze Streaming Data

InfoSphere Streams

1 – Unlock Big Data

IBM Vivisimo

3 – Simplify your warehouse

PureData for Analytics/Netezza

4 – Reduce costs with Hadoop

InfoSphere BigInsights

18

Page 19: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

BigInsightsStreams

• Act on Data “In-Motion”

• Structured & Unstructured Data

• Extremely scalable Real time analytics

IBM’s Big Data Platform

• 10-100’s of TB of structured data

• Extreme Performance

• In-Database Analytics

• Extend SQL with MapReduce

Netezza

• Act on Data “At Rest”

• Structured & Unstructured Data

• Text Analytics

• Commodity Hardware

• “Cold Data” Storage

Page 20: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

The platform advantage

BI / Reporting

BI / Reporting Exploration / Visualization

FunctionalApp

IndustryApp

Predictive Analytics

Content Analytics

Analytic Applications

IBM Big Data Platform

Systems Management

Application Development

Visualization & Discovery

Accelerators

Information Integration & Governance

HadoopSystem

Stream Computing

Data Warehouse

• The platform provides benefit as you move from an entry point to a second and third project

• Shared components and integration between systems lowers deployment costs

• Key points of leverage• Reuse text analytics across streams and

Hadoop

• HDFS connectors between Streams and Information Integration

• Common integration, meta data and governance across all engines

• Accelerators built across multiple engines – common analytics, models, and visualization

20

Page 21: Frank Ketelaars & Robert Hartevelt, IBM - Big Data - BI Symposium 2012

© 2012 IBM Corporation

Information Management

THINK

21