Scality, Cloud Storage pour Zimbra
-
Upload
zimbrafr -
Category
Technology
-
view
3.220 -
download
2
description
Transcript of Scality, Cloud Storage pour Zimbra
Data and Storage Challenges
100s Millions of Users, 10s-100s PB of Dataand Billions of files to store and serve
What all these companies have in common ?
Internet/Cloud business impossible to sustain and developwith traditional IT approaches
Slide 2
Scality – Quick Facts
Founded 2009
Experienced management team
HQ in the San Francisco, Global reach
~ 50 Employees, 20 engineers in Paris
24 x 7 support team
US patents
$13M invested in Scality to-date
120% annual growth
Industry Associations
Slide 3
“Aggressive use of a scale-out architecture like that enabled
by Scality's RING architecture will become more prevalent, as
IT organizations develop best practices that boost storage
asset use, reduce operational overhead, and meethigh data availability
expectations.”
Customers in US, Europe and Japan
Slide 4
Email Service Providers
Cloud Providers (s3 Compatible, FileSync, Cloud Backup…)
Consumer Internet
Hardware Alliances
Big Data
Scality RING 4
Slide 5
Scality RING Organic Storage 4
x86
RingTopology
P2P
End-to-EndParallelism
DATA
MD
ObjectStorage
MESANewSQL DB
Replication
ARC
ErasureCoding
GeoRedundancy
Tiering
S3CDMI
Standard Management
Email File Storage StaaS Digital Media
Enterprise &Cloud EmailSystem
Big Data
Origin ServerScale Out File System
DataProcessingwith Hadoop
S3 & CDMI API
Les Défis du Stockage Mail
Slide 6
L’explosion des données et l’explosion des attentesLa gestion des volumes, la croissance et la
distributionLa complexitéLes performancesLa protection
Ne pas perdre les données Garder les données disponibles
La gestion des volumes
Slide 7
Chaque ZCS gère une liste fixe d’utilisateurs, on doit :Distribuer la charge CPU,
DBs, IOPSGérer les volumesLes volumes sont limités en
nombre de fichiers
L’explosion du stockage
Slide 8
Attentes clientsCout d’y répondre
Les migrations
Slide 9
Changement de HWServeurs trop chargéChangement de
versionsVolumes pleins…
La complexité
Les Performances
Slide 13
La combinaison de:volume importantDensité d’ IOQuantité de fichiers ~ milliards
La croissance ne peut pas réduire les performances
Webmail et IMAP exige les réponses < 100msec en tout temps
Distributed Architecture
From Servers to Storage Nodes
RING Topology, P2P Architecture
Limitless Scale-Out Storage based on Shared Nothing model
Fully Distributed Storage (Data and Meta-data)
Slide 14
P2P
Servers (6) Storage nodes(ex: 6/node, total=36)
Scality RING
Storage nodes projected on a ring
End-to-End Parallelism
Parallel Connectors accessto Storage Nodes Performance aggregation Redundant Data Path
Multiple Storage Nodesper server Minimum 6 to increase parallelism
and data independence Fast and easy rebuild
Multiple IO Daemonsper Server Control Physical and Boost IO
Slide 15
I/O DAEMON
S
STORAGE NODES
STORAGE NODES
STORAGE NODES
STORAGE NODES
I/O DAEMON
S
I/O DAEMON
S
I/O DAEMON
S
I/O DAEMON
S
I/O DAEMON
S
APPLICATIONS /
CONNECTORS
APPLICATIONS /
CONNECTORS
APPLICATIONS /
CONNECTORS
APPLICATIONS /
CONNECTORS
APPLICATIONS /
CONNECTORS
SSD TIEREDSTORAGESATA
Scality Parallelism Factor#Storage Nodes x #IO Daemons
vsSimple server node with only 1 IO Engine
Independent Performance and Capacity Scalability
« Exceptional performance »
Table 3. Content Delivery
Object TypeSimultaneously Sustained
Objects Delivered5 servers with Intel SSD
Internet Audio (MP3) 211,424Internet Image (JPG) 135,311
Internet Video (MPEG) 90,208
CD Audio (ISO) 18,877Broadcast TV (HD) 2,298
Slide 16
“ESG Lab verified exceptional performance for an object-based storage solution, which rivals block-based solutions. Aggregate throughput scaled linearly as nodes were added to a RING. Response times improved as the RING grew in size, allowing for predictability when deploying a RING.”
La protection
Slide 17
La perte des messages n’est pas acceptable
La disponibilité de la messagerie est essentielle
Souvent le cout oblige à des compromis
Data Replication
No data transformation Clear/Native data format Very fast access Simple projection
Class of Storage Up to 5 replicas (6 copies)
Rack-aware Guarantee of full independent object location
Self-healing Balance misplaced objects Transparently proxy misplaced objects Rebuild missing replicas Permanent CRC of all contents (no silent data corruption)
Slide 18
00
20
40
60
80
100
11
120
140
160
180
200
220
91
171
Scality ARC
Data fragments = Native Data (no transformation) Direct and fast read access Calculation only required when data is missing Highly Configurable eg ARC(2,4)
Slide 19
ARC
ScalityARC(14,4)
Data
Checksums
Data inputs
Data source Output to be stored
ScalityRING4
14
…
Geo Redundancy
Slide 20
Asynchronous multiple
independent RINGs
Synchronous stretched RING across 2 sites
Business Continuity with “true 99.999%” availability
Multi-site topology with Scality RING (up to 6 sites) Replication or Geo Erasure Coding implementation – Synchronous
Or Multi-RINGs on Multi-site (independent topology) – Asynchronous
2 Million Users, 1 Billion Objects: Telenet – Email Service Provider Case Study
Problem: Outdated Infrastructure, Fast Growing Customer Base, Budgetary Constraints, Exponential Growth of Storage Projected.Solution: Scality RING Unlimited number of files: Millions of files per bucket.
Up to 1B files per node. Pay as you grow model Fully integrated product includes Zimbra Connectors
“Scality’s storage tiering capability allows us to use lower cost disk systems at geographically dispersed sites to provide Tier 2 storage with true redundancy and disaster recovery, without incurring the extra hardware expense and back-up overhead associated with using NAS and SAN storage systems.
In addition, the cost risk associated with email storage has been significantly reduced with Scality as we can add lower cost storage as and when we need it, without large incremental spending chunks”
Nick De Jonghe, Manager Network Strategy & Architecture, at Telenet
Slide 21
En resuméChallenges Zimbra Solutions ScalityL'explosion du stockage coute cher en NAS/SAN
Stockage distribué sur des serveurs standards baisse le TCO
Limitation du nombre d'objets par volume nécessite de monter plusieurs volumes indépendants qui ne partagent pas leur capacité
pas de limitation (no inode): un seul volume: aggregation de la capacité de stockage
Performance L'ajout de capacité de stockage augmente la performance (aggregation CPU & IO operations)
Sauvegardes complexes sauvegardes intégrées distribuées géographiquement (géo-redondant)
Ajout de capacité sans interruption
système distribué – zéro interruption
Mercihttp://www.scality.com
Slide 23