Vortrag

Transcrição

Vortrag
TechDemo Backup Deepdive
15:15 - 16:45
Deduplication of
Big Data
&
TSM Storage:
wohin mit Big Data
[email protected]
Solutions Architect Data Protection
© 2014 NetApp. Alle Rechte vorbehalten.
1
Non Disclosure Agreement
This presentation/document is the exclusive property of
NetApp, confidential and does neither constitute a part
of nor can it be incorporated into any contractual
agreement with NetApp or its subsidiaries or affiliates.
Roadmap information contained in this
presentation/document is for informational purposes
only and is intended solely to assist you in planning for
the implementation and upgrade of the product features
described.
It is not a commitment to deliver any material, code, or
functionality, and should not be relied upon in making
purchasing decision.
The development, release, and timing of any features or
functionality described in this presentation/document
remains at the sole discretion of NetApp.
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
2
NetApp Deutschland
Die Fakten
Marktführer
28,0%* Marktanteil im Bereich
Open Systems Storage in Deutschland
Platz 1 beim Gesamtmarktanteil*
>25.000 installierte
Systeme
>4.000 Kunden
~600 Mitarbeiter in 8 Niederlassungen
*Quelle: IDC Disk Storage Systems Tracker März 2014,
FAS Marktanteil Deutschland (w/o ESCON/DAS), Umsatz 4-Quartale (R4Q)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
3
Chapters
Deduplication Technology for
Data Protection
Storage for DPR for Big Data
DP Methods / Trends for Big Data
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
4
Chapter “Deduplication Technology
for Data Protection”
Dedupe Fundamentals
Dedupe Targets for DPR
NetApp SteelStore
ONTAP Deduplication & Compression
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
5
Deduplication
Technology for
Data Protection,
Fundamentals
© 2014 NetApp. Alle Rechte vorbehalten.
“deduplizierbare” Datenbestände
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
7
How do backup apps write data to tape?
General/typical data serialization…
File
Header
File
Footer / Pad
Tape
Backup
#1
Image
Heade
r
Tarball
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
8
How do backup apps write data to tape?
What if we add slides to Pres_1.ppt?
Tape
Backup
#2
Image
Heade
r
Byte Shift
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
9
How do backup apps write data to tape?
What if we deleted Doc_1.doc?
Tape
Backup
#3
Image
Heade
r
Byte Shift
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
10
How do backup apps write data to tape?
What if we created Txt_1.doc?
Tape
Backup
#4
Image
Heade
r
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
11
How backup apps write data to tape
And also to disk (in many cases!)…
Data is packaged serially into proprietary “tarball” formats,
which are then written to tape (or disk)
Creating new files/data in an environment usually results in
“insertions” into backup data sets, and byte shifting to the
right
Removal of files/data from an environment usually results in
deletions in backup data sets, and byte shifting to the left
If data alignment happens at all, padding to 512-byte
boundaries is most common in tape formats
– Examples: Symantec NetBackup, UNIX tar(1)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
12
Alignment Sensitivity in Dedupe
Fixed block, fixed offset dedupe…
Example: Block boundaries are every 4 shapes
≠
Nothing matches anymore!
Insert One
Shape
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
13
Alignment Agnostic Dedupe
Variable block, variable offset dedupe…
Example: Block boundaries occur when there are
two triangles in sequence
Block boundaries move with the data
Insert One
Shapeof blocks
Lots
© 2014 NetApp. Alle Rechte vorbehalten.
still match!
13th November 2014 (collected by Dieter Unterseher, no guarantee)
14
GBC Dedupe Operations
Anchor Generation
– Used to do initial markup of opaque backup
streams
– Utilizes a fast rolling hash function to mark block
boundaries in a data dependent manner
– Anchors are persistent and aggressively cached
for high speed lookup
Grow by Compare (GBC)
– Used to identify duplicate sequences of data in
backup streams
– Utilizes high speed sequential byte compare
– Duplicate extent lengths are not size-constrained
by anchor boundaries
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
15
Getting started: Anchor Generation
!
A
!
B
!
C
!
D
!
E
Anchors are generated using a rolling hash algorithm
Hash is computed on a small window of data that slides through
the input stream byte-by-byte
A hash becomes an anchor when it is observed to have “special
properties”
Algorithm typically generates ~3 anchors per 64K of data
–
Anchor generation rate can be controlled/tuned
Anchor Generation can run either inline or post-processed
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
16
The dedupe Process
Unique data is added to common object store…
A
B
C
D
E
Common Object Store
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
17
The dedupe Process
The next backup arrives…
A
B
C
New data has appeared in the backup
Common Object Store
A
© 2014 NetApp. Alle Rechte vorbehalten.
B
C
D
13th November 2014 (collected by Dieter Unterseher, no guarantee)
E
18
Phase 2: Grow by Compare
A
B
C
REF A
Common Object Store
A
© 2014 NetApp. Alle Rechte vorbehalten.
B
13th November
C
D
2014 (collected by Dieter Unterseher, no guarantee)
E
19
Anchor Generation + More GBC
C
REF A
REF B
Common Object Store
A
© 2014 NetApp. Alle Rechte vorbehalten.
B
13th November
C
D
2014 (collected by Dieter Unterseher, no guarantee)
E
20
New Data for Common Object Store
REF A
REF C
REF B
Common Object Store
A
© 2014 NetApp. Alle Rechte vorbehalten.
B
13th November
C
D
2014 (collected by Dieter Unterseher, no guarantee)
E
21
The dedupe Process
And so it continues…
REF A
REF C
REF B
Common Object Store
A
© 2014 NetApp. Alle Rechte vorbehalten.
B
13th November
C
D
2014 (collected by Dieter Unterseher, no guarantee)
E
22
Skips
Keeping Grow-by-Compare efficient…
File headers within a backup stream may contain:
– Last Access Times
Modified when a file has been accessed (but not changed) since the
last backup was done
– Other file metadata that may change between backups
Filename
Owner, Group Owner(s)
Security Permissions
…
– Backup application metadata that may change between backups
Timestamps
Sequence numbers
Checksum fields
…
Skips make grow-by-compare extremely efficient in locating and
eradicating duplicate segments
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
23
Skips
Skips efficiently accommodate small variations between
otherwise duplicate data segments
A
B
Skip Repository
Common Store
Segment
A
© 2014 NetApp. Alle Rechte vorbehalten.
B
13th November 2014 (collected by Dieter Unterseher, no guarantee)
24
Skips
Skips efficiently accommodate small variations between
otherwise duplicate data segments
A
B
Skip Repository
Common Store
Segment
A
© 2014 NetApp. Alle Rechte vorbehalten.
B
13th November 2014 (collected by Dieter Unterseher, no guarantee)
25
Skips
Virtual tapes reference Common Object Store
And sequentially stored skips
REF A
© 2014 NetApp. Alle Rechte vorbehalten.
REF B
REF C
D
(etc)
13th November 2014 (collected by Dieter Unterseher, no guarantee)
26
DP target
Deduplication
one Option for
DP Generation 2
(classical D2D backup)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
27
Data Protection Deduplication
Findet die komplette Dedupe-Reduzierung
beim DP-Dedupe-Target statt, spricht man von
„target depuplication“ (auch „backend
Dedupe“ / „server side Dedupe“ genannt):
VTL / NAS / OST-Appliances mit DedupeFunktionalität
DP-Serversoftware mit DedupeFunktionalität (z.B. TSM ab 6.1)
Findet die Dedupe-Reduzierung in Absprache
zwischen dem DP-Client und dem DedupeTarget statt, spricht man von „client side
Deduplication“
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
28
DP-target Verdichtungsmethoden
Deduplizierung und Kompression
1. Dublettenerkennung
2. Komprimierung
Zum Verdichten von klassischen Backup-Datenströmen muss vor obiger
Reihenfolge zunächst ein Chunking per Variable Byte Range Logik nach
Dateninhalten (unabhängig von physischen Blockgrenzen) erfolgen.
Die Dedupe-Ratio ergibt sich durch Multiplikation der Dubletten& Komprimierungs-Faktoren (z.B.: 5:1 Dubletten-Erkennung und 2:1
Komprimierung ergeben
© 2014 NetApp. Alle Rechte vorbehalten.
(5x2=)
10:1).
13th November 2014 (collected by Dieter Unterseher, no guarantee)
29
Was beeinflußt die DP Deduplication-Ratio?
(am Beispiel tägliches Incremental- & wöchentliches Fullbackup)
Time
Growth
So
Mo
Di
Mi
Do
Fr
Sa
So
1)
2)
3)
Daten-Dubletten im ersten Full-Backupzyklus (stark datenabhängig, z.B. 1,3:1, nicht dargestellt)
Die File-Änderungsrate zwischen Filelevel Incremental-Backups (z.B. 1,3:1)
Die über 1 hinaus gehende Anzahl gehaltener Fullbackups (abzüglich Change-Rate zwischen den
Fulls) hat den größten Einfluss auf die Dedupe-Ratio (z.B. 3:1 oder 10:1).
4)
Die Komprimierbarkeit der Backups (z.B. ~3:1 bei Oracle, oder ~1,5:1 bei Officefiles)
Das Produkt der Faktoren dürfte (bei 20 Fulls mit 5% Change-Rate & 2:1 Komprimierbarkeit) grob 20:1 DedupeRatio ergeben (die Ratio verändert sich nicht linear zu der Anzahl Fullbackups).
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
30
Nominal hohe Dedupe-Ratios blenden –
die Optimierung der DP-Prozesse sind das Ziel
Dedupe-Ratios über 5:1 bringen nur wenig zusätzlichen Platzgewinn.
Deduplication selbst löst keine Backup-/Restore-Probleme
(es vermindert nur den Diskbedarf)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
31
DP Compression / Deduplication Design:
3 mögliche Verdichtungsoptionen denkbar
Compression - only
Post-ProcessDeduplication
Inline-Deduplication
Compression only ist zu bevorzugen:
wenn höchste Lesespeed gefordert ist (denn für Tapekopien / Restores ist dies nach Deduplication
aufgrund stark erhöhter Disk Random Reads nur noch beschränkt möglich)
wenn wenig Dubletten im Datenstrom sind
practical
wenn man Dedupe-Zusatzrisiken scheut
hints
Deduplication ist zu bevorzugen, wenn höchste Diskeinsparung gefordert ist & nenneswerte
Dubletten vorhanden sind & die Restore / Lesespeed zweitranging ist:
Postprocess Dedupe erlaubt (wo geboten) eine schnelle, risikoärmere Tapekopie vor dem Deduplication
(aber Postprocess Dedupe sollte hier immer binnen 24h fertig werden).
Inline-Deduplication skaliert bei Rund-um-die-Uhr Backups besser und spart die Diskkapazität der
Zwischenspeicherung.
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
32
Was ist das “Hash Key Collision Risiko”? practical
hints
Prüft eine Dedupe Implementierung bei gleichem Hash Key, ob der
(oft ~200-fach) längere Datenbereich wirklich den gleichen Inhalt hat?
Die meisten Dedupe-Implementierungen arbeiten Hash-basiert und prüfen dies
nicht.
Das erhöht die Speicherungs-Performance erheblich (viel weniger IOs) und
erlaubt den Einsatz großer und Kapazitäts-Disks –
Aber dies ist mit dem kleinen Restrisiko „silent data loss“ verbunden.
Sollte dies auftreten, droht dauerhafter Datenverlust:
über alle Backup-Generationen hinweg und
für daraus erstellte Non-Dedupe Kopien (z.B. Backup-Kopie auf Tape)
Welche Dedupe-Implementierungen arbeiten mit diesem Restrisiko?
NetApp‘s Implementierungen arbeiten ohne dieses Restrisiko (Dedupe nur nach
erfolgreichem Byte2Byte Datenbereichsvergleich).
Alle Dataprotection-Clients mit client side Dedupe Funktion arbeiten mit dem Restrisiko.
~ 85% der Dedupe Appliances am Markt verwendeten arbeiten mit dem Restrisiko:
Welche? Siehe unter: http://www.backupcentral.com/wiki/index.php/Disk_Targets,_currently_shipping
Alle, welche dort in Spalte “Dedupe Method” den Inhalt nur “hash” stehen haben, arbeiten mit dem Hash Key Collision Risiko.
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
33
Einsatz von Deduped Replication?
practical
hints
Die meisten Dedupe-Appliances bieten Deduped-Replication:
Dies kann helfen, eine DR-Kopie der Backups per IP in eine
andere Niederlassung zu bewegen.
aber ist mit hohen Gesamtkosten verbunden
und erzeugt einiges an Metadaten-IP-Traffic.
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
34
Deduplication von DP-Medien hat auch Nachteile
(gegenüber Compression-only)
Die Lesespeed von deduplizierten Backups wird geringer
(stört ein performantes Erstellen von Tape-Kopien)
VBR Deduplication ist komplex verpointert – birgt Risiken
Tägliche Housekeeping-Prozesse erzeugen viel Last
Es gibt Datentypen, für die Deduplication unvorteilhaft sind
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
35
OST Add On: Hybrid Deduplication („SPEED“)
(for NetBackup & BackupEXEC; Competitives call a similar Feature „Accent“ or „Boost“)
Normal Deduplication
Hybrid Deduplication
(all deduplication is performed on CS800)
(deduplication workload is distributed)
Media Server
Media Server
Media Server
Media Server
Media Server
Media Server
Media Server
Media Server
less bandwidth
used for
backups
all data sent
NAS
OST
VTL
H
O
ETERNUS CS800
1.) chunking &
hashkey-creation
OST
2.) check hashkey
against index
H
O
3.) for unique data only:
ETERNUS CS800
a.) compress
1.) chunking &
hashkey-creation
3a.) for unique data only:
compress
distribute workload
Gives increased
Aggregate Backup
Performance
2.) check hashkey
against index
3b.) for unique data only:
store
b.) store
same logic is also used for Client side Dedupe 
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
36
NetApp SteelStore
designed for Backup2Cloud
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
37
NetApp SteelStore, how does it work?
Dedupe & Compression offers optimized
Disk-Target usage & WAN-Traffic
SteelStore offers CIFS-Shares / NFSExports to write backup-Container-Files
or restore from them
The early & strong AES256
Encryption minimizes any
Security risks at local,
WAN/Internet-Connection
& at Cloud Provider
Filer2Server NDMP
Any
traditional
Backup Data
or any other DP Software like SnapProtect,
which is able to write backup-Files to
CIFS/NFS Shares
© 2014 NetApp. Alle Rechte vorbehalten.
SteelStore does:
1. Deduplication
2. Compression
3. AES256 Encryption
4. Store on the local Disk-Cache
(for days to weeks)
5. bring new blocks async to a
cloud Device (for months)
13th November 2014 (collected by Dieter Unterseher, no guarantee)
Relatively cheap Cloud
Storage offerings as DP
target:
Amazon S3,
Microsoft Azure
etc
38
NetApp SteelStore, for what makes it sense
It is designed for Backup 2 cloud & easy to implement & to use
It has the potential to replace some physical Tapes, because:
– Used Cloud Storage is also very cheap
– Backups get very earlier into another DC (which results in much
lower Data loss at Disaster; no manual tape handling for off-siding
needed)
It has the potential to offer optimal small RoBo Backups (in case
NetApp offers the virtual SteelStore Appliance for production):
– Last Backups (~ 1-2 weeks) are kept automatically locally in the
Backup-Cache
– Only older backups (~ months) are on Cloud-Provider side (Single
File Restore)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
39
NetApp SteelStore, Status
Since Nov 2014 on the NetApp Pricelist:
a Product- & Developer-buyout from Riverbed, which had
already with Version 3 ~165 Customers worldwide
NetApp is currently investing in SteelStore Skill to ramp up
# of models offered still under considerations (we start with the
Highend-Appliance; Virtual SteelStore Appliance are currently
usable for PoC)
In case you have some time to look for a Backup2cloud
solution, wait some weeks, until NetApp got enough
experience with the product.
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
40
ONTAP Deduplication
& Data Compression
(stark verkürzt behandelt;
das sehr etablierte Deduplication wird nur kurz behandelt)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
41
NetApp Compression & Deduplication
dropped license keys for Compression & Dedupe with ONTAP 8.1
transparent to all Applications & Protocols (CIFS, NFS, SAN-LUNs)
works with SnapMirror*, SnapVault (clustered ONTAP), OSSV-targets,
MetroCluster*, SnapLock, vfiler Multistore &
NDMP Dump/SM2T*/SMTape* (*= keep the benefit over the line).
No reduced volume size limits for Dedupe & Compr. since ONTAP 8.1
Compression / Deduplication combinations per FlexVol for 7-Mode &
Clustered ONTAP
–
Inline Compression only
–
Inline Compression with postprocess Compression & Dedupe
–
Postprocess Compression & Dedupe (mainly for existing undensed data)
–
Postprocess Dedupe only
–
(Postprocesses are scheduled together; first Compr., followed by dedupe)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
42
NetApp Deduplication
Flexibles Volume
ohne FAS Dedupe
(since ONTAP 7.2.5)
FAS Deduplication Prozeß
Metadaten
asynchron (Erstausführung, danach z.B. täglich)
Performanceimpact hier – Normalbetrieb zeigt fast
unverändertes Performanceverhalten
wirkt auf alle Blöcke des ActiveFileSystems. Der DiskEinspareffekt tritt somit zeitverzögert auf, sobald die
älteren Snapshots entfallen.
Flexibles Volume
nach FAS Dedupe
4KB Datenblöcke
© 2014 NetApp. Alle Rechte vorbehalten.
Datenböcke gleichen Inhalts
werden freigegeben (nach
Byte2Byte Vergleich, also kein
„Hash Key Kollision“
Datenverlustrisiko)
13th November 2014 (collected by Dieter Unterseher, no guarantee)
etwas mehr
Metadaten
meist deutlich weniger
Datenblöcke
43
NetApp Deduplication
NetApp® deduplication frees duplicate WAFL® data blocks by
reference a single data block by multiple pointers
Support on R200 / FAS2xxx / FAS3xxx / FAS6xxx & V-Series
Inode
Inode
Indirect
Block
DATA1
© 2014 NetApp. Alle Rechte vorbehalten.
Indirect
Block
DATA2
Indirect
Block
DATA3
13th November 2014 (collected by Dieter Unterseher, no guarantee)
Indirect
Block
Indirect
Block
DATA3
DATA3
44
ONTAP Data Compression
For all FAS and V-Series beginning with ONTAP 8.1
On FlexVol-Level in 64-Bit aggregates possible
Compressed Data Blocks are with 8.2 not cached in
the FlashCache- & FlashPool – Layers
Reduces Disk Storage needs & some physical IOs
(some experienced faster Speed with Compression)
Needs more Storage-Controller CPU- Cylces (some
saw under 8.2 in worst case up to 50% speed
reduction for some data with 100% occupied CPUs)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
45
How Does Compression Work?
32k compression groups
188k
logical
Compression
Groups
abcdeabcdeaaabcdeaaabcdeabcdeabc
uvwyyabxzzabuyxzrcuvwyyxzrcabxzz
fghijklmopqfghijrstrstopqklmrstn
32K
32K
32K
uvwyyabxzzabuyxzrcuvwyyxzrcabxzz
32K
fghijklmopqfghijrstrstopqklmrstn
32K
abcdeabcdeaaabcdeaaabcdeabcd
© 2014 NetApp. Alle Rechte vorbehalten.
24K
13th November 2014 (collected by Dieter Unterseher, no guarantee)
46
How Does Compression Work?
Immediate space savings
#!*~>abc
^@(%)/*n
with inline compression -&*+*-#@$abuy
post process is usable
Compressed
Data on Disk
for existing elder data
abcdeabcdeaaabcdeaaabcdeabcdeabc
#!*~
uvwyyabxzzabuyxzrcuvwyyxzrcabxzz
fghijklmopqfghijrstrstopqklmrstn
uvwyyabxzzabuyxzrcuvwyyxzrcabxzz
fghijklmopqfghijrstrstopqklmrstn
abcdeabcdeaaabcdeaaabcdeabcdeabc
abcdeabcdeaaabcdeaaabcde
© 2014 NetApp. Alle Rechte vorbehalten.
-&*+
>abc
*#@$ abuy
^@(% )/*n
-&*+
*#@$ abuy
^@(% )/*n
#!*~
>abc
56k on disk
13th November 2014 (collected by Dieter Unterseher, no guarantee)
47
Typical Storage Savings & Settings
80%
70%
Compression Only
Deduplication and Compression
practical
hints
Deduplication Only
60%
50%
40%
30%
20%
10%
0%
Typical efficiency settings (for primary and secondary volumes) may be:
C+D
D
C
C
C
C
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
C+D
C
48
Compression and/or Deduplication What makes sense / best practice? (1)
practical
hints
Results may very, run SSET 3.0 to prove your data,
which of both or both or none makes sense per volume
With existing snapshots, prefer to activate / deactivate
it for future Writes only: This avoids additional Disk-Space
needs, until the Snapshot before the full volume dense /
undense expires.
A typically good schedule time for postprocess
Compression & Dedupe run is daily 23:05 : realize
reduction before the midnight snapshot, which is
typically held longer.
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
49
Chapter „Storage for DPR for Big Data”
Data ONTAP Storage-Systems
NetApp E-Series (an example for a
traditional Block-Storage)
Physical Tape?
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
50
Storage Systems for Big Data
Data ONTAP
Storage-Systems
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
51
NetApp FAS Unified Storage
FAS8080 EX
Hybrid and all-flash
Cloud integrated
FAS8060
SDS enabled
Massively scalable
FAS8040
– Scale up
FAS8020
– Scale out
FAS2554
FAS2552
FAS2520
2880TB
24TB Hybrid Flash
384TB All-Flash
8640TB
144TB Hybrid Flash
7200TB
384TB All-Flash
72TB Hybrid Flash
384TB All-Flash
4320TB
48TB Hybrid Flash
384TB All-Flash
864TB
16TB Hybrid Flash
154TB All-Flash
749TB
16TB Hybrid Flash
504TB
154TB All-Flash
16TB Hybrid Flash
96TB All-Flash
Unify and simplify storage administration under Data ONTAP®
FlexArray software lets FAS8000 manage EMC, HP, HDS, and E-Series
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
Clustered ONTAP: Data Agility & Always on
Virtualized Servers
Non-virtualized
Servers
Big Data
Content Servers
Workloads & Users
A
B
C
D
E
Policy-Based Data Management
Unexpected
Performance
Unexpected
Growth
Growth
Issue
Replication
All
Backup
Performance
Actions
Copy
Copy
Occur
Creation
Issue
Creation
Without
Capacity
System
Upgrade
Expansion
Needed
Needed
Growth
Issue
Resolved
in
Workload
D
E
B
Disruption
- -Workload
Workload
Resolved
to Users
A
C
Unified Data Network
FCP FCoE iSCSI
CIFS/SMB NFS/pNFS
Clustered ONTAP
A1
© 2014 NetApp. Alle Rechte vorbehalten.
C1
13th November 2014 (collected by Dieter Unterseher, no guarantee)
53
Scalable Performance for SAN and NAS
Scale Out by Adding Controllers
(Maximum 24 Controllers)
Scale Up
Individual
Controllers
All-Flash FAS
All-Flash FAS
All-Flash FAS
All-Flash FAS
All-Flash FAS
All-Flash FAS
Hybrid FAS
Hybrid FAS
>4 million IOPS and nearly 5PB of all-flash capacity
NetApp® all-flash FAS: Scalable high performance for dedicated workloads
Mixed all-flash and hybrid FAS: One storage repository for all workloads
54
2014 NetApp,
All rights
reserved. .
© 2014© NetApp.
AlleInc.Rechte
vorbehalten
13th November 2014 (collected by Dieter Unterseher, no guarantee)
NetApp Flash & Storage Portfolio
Sweetspots for DP-targets
Sweetspots for ONTAP
snapshot-based DPR
SnapMirror/SnapVault targets
Sweetspots for Traditional DPR
sequential IO to DP targets
All Flash Storage
Hybrid Array
Sweetspots for VBR Backup
Deduplication Methods
Random-IO to DP targets
Flash-Accelerated Arrays
Traditional Storage
ONTAP
ONTAP
All Flash Aggregate
FlashRay
All Flash
E-Series
EF540
© 2014 NetApp. Alle Rechte vorbehalten.
Flash Pool
Flash Cache
(in conjunction with Capacity disks)
SATA / SAS
Ennn
13th November 2014 (collected by Dieter Unterseher, no guarantee)
55
NetApp E-Series
An extreme fast & reliable Block
Storage (with less functionality
compared to Data ONTAP)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
56
E-Series with Harddisks
Controllers + Shelf = Model Numbers
Controllers /
Disk Shelf
Enclosures
E2600
E2700
E5400
E5500
DE1600
(2U/12d)
E2612
E2712
E5412
E5512
E2624
E2724
E5424
E5524
E2660
E2760
E5460
E5560
3.5” only
DE5600
(2U/24d)
2.5” only
DE6600
(4U/60d)
2.5” or 3.5”
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
57
E-Series with Flash-only
EF550 Product Specifications
Base
models
12 drives (9.6TB)
24 drives (19.2TB)
Expandable
Up to 120 drives (96TB)
Drive
800GB, mixed-use, SSD
I/O
interface
options
Storage OS
Key
Features
(8) 16Gb FC
(8) 10Gb iSCSI
(8) 6Gb SAS
(4) 40Gb IB
SANtricity® 11.10
© 2014 NetApp. Alle Rechte vorbehalten.
Dynamic disk pools
Thin provisioning
Snapshot™ copies
Volume copy
Remote mirroring
Performance:
• Burst: 900,000 IOPS
• Sustained: over 400,000 IOPS
• Sustained: up to 12GB/s
13th November 2014 (collected by Dieter Unterseher, no guarantee)
58
Kundenbeispiel DP Diskpool für NetBackup & TSM
Lokation 1
Lokation 2
5 Media Server
5 Media Server
NetApp E-Series
NetApp E-Series
NetApp E-Series
750 TB
750 TB
750 TB
750 TB
Kunden Infrastruktur
NetApp E-Series
Total: 3 PB
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
59
Storage Systems for Big Data
Physical Tape ?
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
60
Why is Disk a growing
& Tape a shrinking backup-media?
Unloaded physical Tape-Media have the lowest price per TB, but:
Data
Data
Disk drives are cheaper
than Tape Drives
Disk positions in
Milliseconds Tape in Minutes
Disk allows parallel Access for Backups & Restores
Disk-Surfaces operate dust-free - Tape-Media not
Disk can leverage RAID for Non-Disruptive-Operations – be aware that
magnetism-level shrinks over time & risk growths
Disk fits for Incremental Restores (direct access is mandatory for
Snapshots, DP-Deduplication & file-level Incremental Restores)
Disk fits for DR-Readyness (BLIF / dedupe allow Backups over small
bandwith – no wait for media to get full, Data may be direct accessible)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
61
Where does Tape still fit?
practical
hints
for Full Tape sequential Writes - Read nearly Never
where Read Access-Times of many minutes are acceptable
where a tape gets mostly to streaming mode (~60 to ~380 MB/sec)
where partial Reads of Tape is very seldom (~ < 1 Read per 250 written
Media).
Typical remaining use cases for physical Tape:
Augmentation of disk-based 2nd-4th Gen. Data Protection (D2D2T):
Additional DR Copies of the disk stored backup / archive data
or for longtime retention (of monthly, quarterly or yearly backups)
(like NDMP Dump / SMTape with weeks of Snapshots)
Where Off-siding of DP Medias are a must – (for example: where even
Block-Level-Incremental-forever creates too much WAN-Traffic)
for fast large DB Backups (only as long RTO is acceptable).
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
62
Chapter „DP Methods / Trends
for Big Data”
General Overview
NDMP
DP Gen. 2: D2D(2T)
DP Gen. 3: D2D DP-Clients for blocklevel
Incremental Forever Backups
DP Gen. 4: DP Management-Software from IBM for
Array-snapshot-based DPR with ONTAP
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
63
How to Backup Big Data ?
DP Methods / Trends
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
64
Simplifying Data Protection
Fueling Convergence
Continuous
Availability
© 2014 NetApp. Alle Rechte vorbehalten.
Disaster
Policy Based
Backup &
Recovery
Data Availability
Recovery
13th November 2014 (collected by Dieter Unterseher, no guarantee)
Archive &
Retrieval
Data Protection Market paradigm change
2000
2010
2020 DP world market revenues
DR
Prim
Prim.
DP
Prim
DP & DR
HA
HA & DR
DP
1st Generation DP: 2nd Generation DP:
Incremental / Full
Regular full
backups to Tape Backup2Disk/VTL
Separate techniques
for DP, DR and HA
(Dedupe- and NonDedupe targets offer only
limited improvements to
1st generation solutions)
3rd Generation DP:
Block-incrementalForever Backups
4th Generation DP:
Snapshot-based DP,
avoids data-movement
(often CBTF / Clientfor backups and restores
Dedupe; reduced
with Mirroring / Replication
backup xfer rates &
(DP, DR, HA and Test/Devslower Restores & Tape
needs can be combined)
copies)
Dedupe & compression benefits
Traditional DP result in very poor DP SLAs
(too much data & files have to be transferred)
start at Primary Data and
can be kept in backup copies
4th Generation Snapshot based Data Protection is the future for the DC,
3rd Gen is the future for client DPR & may be an alternate for backups “out of Array Snapshots”

© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
66
Data to be transported with 1st Gen. DPR
Backup to & Restore from Tape
Application
/FileServers
with DPClient
DP MediaServer
Primary
Data
Storage
Physical Tape
(2 Copies usual)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
67
Data to be transported with 1st Gen. DPR
LAN-free Backup to & Restore from Tape
DP-Server
Application
-Servers
with DPClient
Primary
Data
Storage
Physical Tape
(2 Copies usual)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
68
Data to be transported with 2nd Gen. DPR
with traditional disktargets
DP MediaServer
Trad. DP-disktarget
Application
-Servers
with DPClient
Primary
Data
Storage
Physical Tape or
another DP Media
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
69
Data to be transported with 2nd Gen. DPR
with Target Deduplication
DP MediaServer
Application
-Servers
with DPClient
Deduped
DP-disktarget
Primary
Data
Storage
Physical Tape or
another DP Media
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
70
Data to be transported with 3rd Gen. DPR
by Client site Deduplication
Partial Dedupe
logic
DP DPServer
Application
-Servers
with DPClient
Deduped
DP-disktarget
Primary
Data
Storage
Physical Tape or
another DP Media
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
71
Data to be transported with 3rd Gen. DPR
by VMware CBTF?
DP MediaServer
Application
-Servers
with DPClient
Capacity reduced
DP-disktarget
Primary
Data
Storage
Physical Tape or
another DP Media
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
72
Data to be transported with 4th Gen. DPR
Array Snapshots with NDMP
DP MediaServer
Application
-Servers
with DPClient
Primary
Data
Storage
Nearly all Restores can be done
using the internal snapshots;
Deduplication/Compression may
Start on Primary
Physical Tape or
another DP Media
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
73
Data to be transported with 4th Gen. DPR
Array Snapshots with Replication &NDMP
DP MediaServer
Application
-Servers
with DPClient
Primary
Data
Storage
Nearly all Restores can be done
using the internal snapshots;
Deduplication/Compression may
Start on Primary
Secondary
Storage in a
DR-usable
format
optional
Physical Tape or
another DP Media
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
74
Data to be transported with 4th Gen. DPR
Array Snapshots with Replication
DP MediaServer
Application
-Servers
with DPClient
Primary
Data
Storage
© 2014 NetApp. Alle Rechte vorbehalten.
Secondary
Storage in a
DR-usable
format
Nearly all Restores can be done
using the internal snapshots;
Deduplication/Compression may
Start on Primary
13th November 2014 (collected by Dieter Unterseher, no guarantee)
75
Which performance degrading can be
expected with Snapshots?
CofW at NetApp E-Series: Based on internal Tests of NetApp, we saw the
following with their CoFW-Snapshots (compared to normal mode
without Snapshot):
– With full Flash Arrays (SSD only): ~15% Throughput reduction
– With physical Disks (HDD only): ~40% Throughput reduction
Some other competitive Storage with CoFW: has been tested by 3rd
Party institutes with ~64% Throughput reduction with physical HDDs
but only ~3.2% Throughput reduction with ONTAP WA-based
Snapshots with physical disks and the same mix of IOs.
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
76
Snapshot Technique Use-Cases
Full Copy Snapshot arrays,
Typical RoW arrays & Software Snapshots (VMware, MS VSS etc)
basic Array-Snapshot-Support of some DP Software
every
Snapshot backup
1
4 to ~16
h to few days
Disk
stage
Tape
plus typically
Disk
stage
RoW Arrays with large Block,
CoFW Arrays
plus typically
Tape
(disk/cloud)
up to hundreds of Snapshots WA Appliances (like NetApp ONTAP)
weeks of Snapshots on Primary
(disk/cloud)
up to hundreds of Snapshots
months of Snapshots on Secondary
Tape
(disk/cloud)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
77
4th Gen: Snapshot with Replication [+ NDMP]
without MetroCluster
practical
Block-Level-Incemental-Forever ~ 1,3% daily
hints
=für RoBo DP: bestens,
da lokale Restores extrem schnell;
Prim. Storage kleiner Niederlass.
SnapVault: Meist
werden hier
Snapshots
länger gehalten –
auf günstigen
SATAKapazitäts-Disks
Backup
DR falls Sec.-Storage noch verfügbar.: Server
D-RPO: ~1,3h bei 24 daily Backups
D-HWN: nur bis Prim.Storage / Server –
evtl. auch Aktivierung Sec.Storage möglich
D-RTO: Aktivierung oder alles zu übertragen,
aber schnelles 1:1 „Seeding“ möglich
Tape-Backups
werden optional &
sind rund um die Uhr
ohne Störung der
Produktion
möglich
SAN
Nur Differenz-blöcke
werden repliziert
Snapshot / SnapRestore
Server (evtl. mit
SnapDrive /
SnapManager)
FC-SAN /
iSCSI
Sehr geringe Netzwerklast
0%
NDMP Dump
or SMTape
Dedupe-/Compression-Vorteile
erhalten)
166 / 1,3%
69% / 0,8% densed
IP
WAN / LAN
NetApp Pri-Storage SnapVault oder SnapMirror
© 2014 NetApp. Alle Rechte vorbehalten.
100 / 18%
1,3% or 0,8% densed
(SnapMirror/SnapVault kann
NetApp Sec-Storage
13th November 2014 (collected by Dieter Unterseher, no guarantee)
Physical Tape
Library
78
4th Gen: Snapshot with Replication [+ NDMP]
with MetroCluster
practical
Block-Level-Incemental-Forever ~ 1,3% daily
hints
iSCSI
Snapshot / SnapRestore
DR falls Metro-Cluster-Hälfte verfügbar,
bleibt, da weit genug entfernt:
D-RPO: 0 Sekunden Datenverlust
D-HWN: Storage verfügbar
D-RTO
~1 Minute für Failover
FC-SAN /
SnapVault: Meist
werden hier
Snapshots
länger gehalten –
auf günstigen
SATAKapazitäts-Disks
Backup
Server
Tape-Backups
werden optional &
sind rund um die Uhr
ohne Störung der
Produktion
möglich
SAN
Sync
Mirror
Snapshot / SnapRestore
Server (evtl. mit
SnapDrive /
SnapManager)
FC-SAN /
iSCSI
Nur Differenz-blöcke
werden repliziert
Sehr geringe Netzwerklast
NDMP Dump
or SMTape
Dedupe-/Compression-Vorteile
erhalten)
166 / 1,3%
69% / 0,8% densed
IP
0%
WAN / LAN
NetApp Pri-Storage SnapVault oder SnapMirror
© 2014 NetApp. Alle Rechte vorbehalten.
100 / 18%
1,3% or 0,8% densed
(SnapMirror/SnapVault kann
NetApp Sec-Storage
13th November 2014 (collected by Dieter Unterseher, no guarantee)
Physical Tape
Library
79
NetApp Integrated Data Protection and
Business Continuity Solutions
Continuous
Operations
DR Readyness
Asynchronous
Replication
MetroCluster™
(full sync)
Fast Restores
SnapMirror®
Backup
out of
ONTAP
SnapVault®
SnapRestore®
Block-Level
Incremental
forever
Backups
Multiple
Recovery
Points per Day
NDMP backups
Cost
Availability
Snapshot™
Copies
Local Restore
Remote Restore
Remote Recovery – Low RTO
Capability
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
80
NDMP Dump und
SM2T / SMTape
(ONTAP Tapebackup)
As this chapter is a frequent subject of change, we offer our EMEA NetApp Partner community the access
to an ~3-monthly actualized version up to end of FY14 at:
https://fieldportal.netapp.com/search/?searchTerm=unterseher+NDMP# (after FY14 this will be accessible
for TechDemo Backup Deepdive attendees of Germany & Switzerland & Austria Sales-Partners only).
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
81
NDMP Modes
Local, 3-Way, Remote, filer2server
Local
NDMP
3-Way
NDMP
Remote
NDMP
(filer2filer
NDMP)
(filer2server
NDMP)
LAN
NDMP
Hosts
Data
+
Tape
Service
server2filer
NDMP
DMA +
Tape
Service
Data
Service
Tape
Service
DMA +
Data
Service
Data
Service
Tape
Service
DP
Clients
Automated
Tape Library
(PTL or VTL)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
82
Tapebackup Technique Comparism
(current functionality with Data Ontap)
NDMP Engine and Ontap-Version
DUMP
SMTape (=“SM2T” with Ontap 7.x)
recommended backup use case
Tape DP Archival
Tape Disasterprotection
Objects backed up per volume?
Files & Directories
of one Snapshot
The full block construct of a volume
(including all snapshots & clones)
Full; filelevel Incr/Diff
Full; ONTAP 8.1.1: +blockl. Incr / Diff
single file / directory restore?
yes
no (only full volume restore)
Preservers dedupe/compression
No
Yes, benefits are kept (=faster)
Yes (no WAFL pointers)
no
limited (file-level)
high
Backupspeed Volumes with medium/large files
high
high
Is Clustered ONTAP already supported?
yes
no (expected in future)
Tape format backward support (Restorability)
Longtime (10 Years)
2 Ontap-Versions only (~3 Years)
format backward support (7- <--> C-Mode)
YES (not for LUNs)
no
no
no (no SnapMirror License)
Dump
SMTape (SM2T)
NDMP-Standard
NDMP-Extension type=snapm
separate CLI-Commands each
backup modes
Logical media break (=without WAFL-P.)
Backupspeed Volumes with millions of
small files
License costs NetApp?
Tape format
Protocol (NDMP DMA / CLI-Command)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
83
supported NDMP Features
by Backup Application Vendors
a
Backup Application
D
DAR
ASG (Atempo) Time Navigator (>= 3.7)
Catalogic DPX (BEX) (>= 2.35)
Commvault Simpana (>=7.0)
CA BrightStor ARCserve (>= 16.0)
EMC / FTS NetWorker (>= 7.6.3)
HP Storage Data Protector (>=6.2)
c
d
f
g
s
t
u
3W
DDS
remote
f2s
s2f
SMTape
full
SMTape
Incr/Diff
CAB
4.3.
Soon exp.
8.0
b restartable
Backup
Oracle Secure Backup (>= 10.3.0.2)
?
?
?
?
NetApp SnapProtect (>=9.0SP3b)
expected
?
?
?
?
IBM Tivoli Storage Manager (>= 5.5.2)
Quest (BakBone) NetVault (>= 8.0)
Symantec NetBackup (>=5.1)
Symantec BackupEXEC (>= 12.5)
Zmanda Amanda Enterprise (>= 7.3.2)
Direct Access Recovery (DAR), and Local Backup is supported by all of them
a = Directory DAR
*2? *2? *1 *2 *2
*1 *2
?
?
*1*2 *2
*2? *1 *2
*2?
?
?
V10SP2b V10SP2b *3
?
V10SP2b V10SP2b *3
7.6.7+patch
?
c – 3-way NDMP (filer filer)
d = Dynamic Drive Sharing (DDS)
f – Remote NDMP (filer media server), with *1 = Tape-Multiplexing optional, with *2 = disk-Caching optional
g – Server-to-Filer Backup
s – SM2T / SMTape full backup/restore
t – SMTape Incremental or Differential backup
u=cluster aware backup *3= currently without local NDMP (soon expected)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
84
ONTAP Releases supported for NDMP
by Backup Application Vendors
ONTAP 8.0
ONTAP 8.1
ONTAP 8.2
(7-mode & Clustered
ONTAP)
(7-mode &
Clustered ONTAP)
(7-mode &
Clustered ONTAP)
Backup Application
ONTAP
7.3
ASG (Atempo) Time Navigator
4.2
4.2
Catalogic DPX (BEX)
3.4.1, 4.1
3.4.1, 4.1
3.4.1, 4.1
4.2
Commvault Simpana
8.0
8.0
9.0SP5
9.0, 10.0
CA BrightStor ARCServe
11.5, R15
R15, R16 (7-Mode only)
R16 (7-Mode only)
R16.5
EMC NetWorker
7.4
7.5 7.6 (7-mode only)
7.6.3, 8.0
8.0
HP Storage Data Protector
6.0
6.11, 6.2 (7-mode only
6.21, 7.01 (7-mode only
without DAR-Feature)
without DAR-Feature)
9.0.1 (both modes)
NetApp SnapProtect
9.0SP3b
9.0SP3b
9.0SP5
Oracle Secure Backup
10.3.0.2
IBM Tivoli Storage Manager
5.51, 6.1
6.1
Quest(BakBone) NetVault
8.2
NVBU 8.5.3
9.0
9.x (cDOT only)
Symantec NetBackup
6.0 & 7.0
6.5 & 7.0
7.1 & 7.5
7.5
Symantec BackupEXEC
12.5
2010R2(+ Hotfix)
2012
2012 SP4
Zmanda Amanda Enterprise
3.1
4.3.3
5.5.x,
6.1.x
V10 SP2b
6.3.3
NetApp IMT http://support.netapp.com/matrix is most actual
or use official certification side: http://www.netapp.com/us/solutions/a-z/backup-to-tape/backup-to-tape-ndmp.html
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
85
DP Generation 2:
D2D(2T)
(mit gelegentlichen Fullbackups)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
86
Allgemeines zu DP Gen 2: D2D(2T)
Future...
Gen 2 DP: Sie haben die Wahl zwischen folgenden DP-Disktargets:
2a) DP-disktargets ohne Verdichtung (z.B. mit der NetApp E-Series):
Ist aufgrund des hohen Diskbedarfs dann sinnvoll, wenn große Fullbackups nur wenige
Tage oder nur auf Tape gehalten werden.
NetApp‘s E-Series ist dafür ab ~130 TB Kapazität performant & wirtschaflich (~3
GB/sec sequ. Write / ~5 GB/sec sequ. Read, wenig Stromverbrauch).
2b) DP-disktargets mit Compression only (z.B. ONTAP ab 8.1):
Hier kann man (bei Reduzierung der Fulls) oft bis zu ~ 4 Wochen Backups auf Disk
vorhalten.
ONTAP sollte typischerweise bis max ~50 TB Kapazität auch für Tape-Prozesse
performant genug sein.
2c) DP-disktargets mit Deduplication (z.B. SteelStore, impliziert Compression):
Erlaubt oft bis max ~3 Monate Backups auf Disk zu halten.
Kann bei höheren Dedupe-Ratios sinnvoll sein, falls die Dedupe-Nachteile akzeptabel
bleiben und (gewünschte) Tape-Kopien noch genügend skalieren.
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
87
Allgemeines zu DP Gen 2: D2D(2T)
practical
hints
Ihre Backups sollten virensicher sein!
Wenn Backups nur auf Disk-/NAS-Targets gespeichert
werden (ohne Kopie auf PTL/VTL), kann über Viren/Hacker ein Totalverlust der Backups eintreten denn ein Virus der in den DP-Server eindringt kann sehr
einfach alle im Schreibzugriff befindlichen Files
zerstören.
Schnittstellen wie VTL bieten hier einen besseren
Schutz.
Falls Sie ONTAP als Disktarget verwenden,
könnten zusätzlich aktivierte ONTAP Snapshots
helfen, die Backup-Files für ~1-3 Tage vor Viren
etc zu schützen (=höherer Platzbedarf).
DP-target Appliances, welche auf Standard-OS
(wie RedHat Linux) basieren, können zusätzlich
ein Ziel von Hackern / Viren werden, welche
Lücken im OS nutzen.
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
88
DP Generation 3
D2D DP-Clients für Incremental
Forever Backups
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
89
DP Gen 3: Incremental Forever DP Clients,
Überblick
Gen 3 DP: Incremental Forever (Full nur zu Beginn) gibt es in 3
Ausprägungen von DP-Clients:
3a) Filelevel IF (TSM-like oder mit Synthetic Fulls des DPServers)
Nur begrenzt effizient, vor allem ungünstig bei kleinen
Änderungen großer Applikations-Files (wie DB-Files).
Also ist diese Technik eigentlich nur „DP Generation 2.5“.
3b) Client side Deduplication (ergibt auch einen BLIF Effekt)
3c) „echtes“ Blocklevel IF (z.B. Catalogic DPX / SPOS / OSSV
/ VMware with changed Block Tracking forever)
More hidden
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
90
DP Gen 3b: client side Deduplication
Funktionsweise / Stärken / Schwächen
Ein Teil der Dedupe-Logik ist auf den DPClient verlagert
arbeitet immer mit dem kleinen
HashKeyCollission Risiko
Der Backup-Datenstrom ist relativ gering
…
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
91
3c: VMware vStorage API-basierte Backups
mit Changed Block Tracking Forever
Ist nur für VMware Datastores (VMDKs) anwendbar und nur für die neuen
VMware Versionen ab ESX V5.
Merkt sich Blockänderungen (z.B. auf 16K Ebene) für den Folge-Backup.
Gibt diese Änderungen per API an eine Backup-Software
CBTF unterstützen typischerweise erst die neuesten DP-Software
Versionen seit ~Ende 2013 (z.B. TSM).
Die Backup-Software muss hier teilweise komplexe Dinge ergänzen
–
–
–
–
–
Z.B.: für logical Object Restores das komplette VMDK bereit stellen
Expiration (Löschen) ausgealterter Backups
Fragmentierung für Full VMDK-Restores sollte gering gehalten werden Speed
(oft wird dann doch wieder mal ein echter Fullbackup ausgeführt)
Störungen müssen erkannt werden (meist wird dann ein erneuter Fullbackup
ausgelöst)
etc
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
92
DP Management-Software
from IBM
for Array-snapshot-based
DPR with ONTAP
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
93
IBM TSM NAS Data Protection mit ONTAP (1)
Folgende Kombination ist für „Millions-of-Files“ NAS Volumes (im Vergleich zum
normalen TSM-Workflow) vorteilhaft:
SM2T / SMTape seit TSM Server 5.5.2: für schnelles DR Backup/Restore
(Differential seit ONTAP 8.1.1: TSM Support dafür gefordert / noch nicht absehbar).
Plus SnapDiff (7-Mode ony!): für beschleunigten TSM Scan-Prozess beim Backup
(mehr Details dazu auf Folgeseite)
Noch effizienter für „Millions-of-Files“ NAS Daten ist aber meist:
mehrwöchig gehaltene NetApp-erzeugte Snapshots,
ergänzt um NetApp-gesteuerte SnapMirror/SnapVault-Replikation
und / oder TSM-gesteuertem NDMP SMTape / Dump.
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
practical
hints
94
IBM TSM NAS Data Protection mit ONTAP (2)
SnapDiff (7-Mode): für beschleunigten TSM Scan-Prozess beim Backup:
ONTAP ab 7.3.3 / 8.0.1 & TSM Client ab 6.2.2.2 auch für deutsche Umlaute:
CIFS über Windows-Client, NFS über AIX- oder Linux-Client
seit TSM-Client 6.40 & ONTAP 8.1.1 vfiler-Multistore Support
Seit TSM-Client 6.40 auch für bestehende Snapshots / auch von
SnapVault/SnapMirror targets
noch kein Clustered ONTAP Support von Seiten IBM
noch kein V-Series Support (obwohl es funktionieren sollte)
IBM empfiehlt bei SnapDiff-Verwendung trotzdem gelegentlich (z.B: einmal im
Monat / Quartal) einen traditionellen TSM Full Incremental Backup (also ohne
SnapDiff) durchzuführen.
practical
hints
In dem Fall würden nur evtl. entstehende Lücken des SnapDiff-beschleunigten Backup
gefüllt. IBM hat bisher nicht beschrieben wieso – Dieter Unterseher hat folgende
Vermutungen:
Vielleicht fehlen Incremental Backups nach ACL Rechteänderungen auf Directory
Ebene, wenn SnapDiff angewendet wird?
Vielleicht kann es bei der TSM-Logik SnapDiff vorkommen, dass ein nicht zu Ende
gebrachter SnapDiff-Backup zu Backup-Lücken führen kann.
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
95
IBM TSM / FCM snapshot-based DP mit ONTAP
ONTAP Support für Snapshot/SnapRestore (ohne ReplikationsKontrolle) verfügbar seit 16.11.2012 mit FCM 3.2 (mit TSM 6.40; ONTAP
7-Mode only):
Oracle / DB2 / SAP(auf Oracle/DB2) unter AIX & Linux (native):
– blockdevice support (mit schnellem Full Volume SnapRestore)
Microsoft SQL & Exchange (native)
(seit 3.2.1 auch Exchange2013 supported):
– sehr schnelle Backups;
– logical Object Restores aus gemounteten Snapshot-Kopien
VMware: VMDK-Level Backups per Snapshot für block- & NFS Data Stores:
– Und Application Consistent Backups für MS SQL / Exchange etc (Object
Restores mit VMware & MS VSS Tools)
– Bei NFS Datastores sind auch sehr schnelle VMDK-Level Restores (über
Single File SnapRestore) möglich
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
96
IBM TSM / FCM snapshot-based DP mit ONTAP
IBM-Infos dazu:
http://www-01.ibm.com/software/tivoli/products/storage-flashcopy-mgr/
NetApp-Infos dazu:
FAQ „Tivoli Storage FlashCopy Manager Version 3.2”
vom NetApp Field Portal
Fundierte Umsetzungs-Erfahrung liegt z.B. vor:
EMPALIS (da kann man das auch live mit ONTAP vorführen
lassen); siehe http://www.empalis.de/index.php?id=2308&L=)
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
97
TechDemo Serien
practical
hints
www.techdemo.de
zum Thema „Backup & Recovery“ : siehe
http://techdemo.de/events/kat/backup.html
Neben der Ganztagesveranstaltung
„TechDemo Backup Deepdive“ bietet NetApp seit
Sommer 2013 für IT-Manager & Sales-Beauftragte
von Partnern die Halbtagesveranstaltung
„TechDemo Backup – Überblick“ und
gelegentlich Partner-Veranstaltungen;
Ergänzend können Sie immer mehr Einzelthemen
per TechDemo Online bei Bedarf abrufen:
http://techdemo.de/online.html
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
98
Ende
Offene Fragen?
© 2013 NetApp, Inc. All rights reserved. No portions of this document may be reproduced without prior written consent of NetApp, Inc. Specifications are
subject to change without notice. NetApp, the NetApp logo, Go further, faster, AutoSupport, Data ONTAP, OnCommand, SnapDrive, SnapManager,
SnapMirror, Snapshot, and SnapVault are trademarks or registered trademarks of NetApp, Inc. in the United States and/or other countries. Microsoft,
SharePoint, SQL Server, Windows, and Windows Server are registered trademarks and Hyper-V is a trademark of Microsoft Corporation. Linux is a
registered trademark of Linus Torvalds. Oracle is a registered trademark of Oracle Corporation. ESX and VMware are registered trademarks and ESXi is a
trademark of VMware, Inc. UNIX is a registered trademark of The Open Group. All other brands or products are trademarks or registered trademarks of their
respective holders and should be treated as such.
© 2014 NetApp. Alle Rechte vorbehalten.
13th November 2014 (collected by Dieter Unterseher, no guarantee)
99

Documentos relacionados