Anda di halaman 1dari 103

Storage Basics Storage Basics

An introduction to the fundamentals


of storage technology An introduction to the fundamentals
of storage technology
Contents
Section 1 –
The information society – saving data and knowledge at new levels
Section 2 –
Tiered storage: intelligent information management in the company
Section 3 –
Online storage: disks and reliability
Section 4 –

Storage Basics – An introduction to the fundamentals of storage technology


Storage networks – spoilt for choice
Section 5 –
Backup & Restore: an unloved compulsory exercise
Section 6 –
Storage management – making complex storage networks manageable
Section 7 –
Virtualization – some catching up is necessary regarding storage topics
Section 8 –
The storage strategy of Fujitsu Siemens Computers and its partners
Forecast: Future storage trends
Glossary
As well as articles about storage from our partners
Brocade, CA, EMC, NetApp, Sun and Symantec

Price: 14.99 Euros


Storage Basics
An introduction to the fundamentals
of storage technology
Storage Basics
An introduction to the fundamentals
of storage technology
Storage Basics
January 2009

Copyright
© Fujitsu Siemens Computers 2009
Text, editing, production: ZAZAmedia / Hartmut Wiehr
Printed in Germany.

Published by
Fujitsu Siemens Computers GmbH
Mies-van-der-Rohe-Strasse 8
80807 Munich, Germany

Contact
www.fujitsu-siemens.com/contact

All rights reserved. Subject to delivery and technical changes. The names reproduced in this docu-
ment can be brands whose use by third parties for own purposes can violate the rights of the
owners.
Contents
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7

Section 1 –
The information society – saving data and knowledge at new levels . . . . . . . . . . . . . . 9

Section 2 –
Tiered storage: intelligent information management in the company . . . . . . . . . . . . 15

Section 3 – Online storage: disks and reliability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23

Section 4 – Storage networks – spoilt for choice . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31

Section 5 – Backup & Restore: an unloved compulsory exercise . . . . . . . . . . . . . . . . . 39

Section 6 –
Storage management – making complex storage networks manageable . . . . . . . . . . 49

Section 7 –
Virtualization – some catching up is necessary regarding storage topics . . . . . . . . . . 57

Section 8 – The storage strategy of Fujitsu Siemens Computers and its partners . . . 61

Forecast: Future storage trends . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69

Remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73

Glossary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77

Partners . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
Preface

Dear Reader,

Coming to grips with the ever-growing flood of data is still the greatest challenge as far
as storage is concerned. This statement no longer only applies to major customers and
their data centers, but also affects SME businesses and even smaller companies. “Com-
ing to grip“ means secure data storage and accessibility with due regard to agreed qual-
ity standards (service levels) and at a reasonable cost (CAPEX und OPEX). Many of the
new technologies such as ILM, DeDup, SSD, virtualization, thin provisioning – regard-
less of whether they are already established or still just hype – are eager to help here.
We advise our customers to develop a storage solution that suits them from this
variety of established and new technologies. We have a comprehensive portfolio of
best-in-class products in this regard, but neither is it our intention nor are we able to do
everything ourselves. The way in which we monitor technology helps us to make the
right choice here. Together with technology leaders in storage we have developed stra-
tegic partnerships, we integrate their products in solutions and provide the appropriate
services.
If we see that our customers have problems, for which there is no suitable solution
on the market, we then develop our own products and solutions. CentricStor solutions
are prime examples of this.
In addition we have begun to not only develop and implement storage solutions at
our customers‘, but also to operate them ourselves for our customers. Fujitsu Siemens
Computers is to increase its investment in this managed storage business. Our man-
aged storage customers receive an invoice and the pertinent report every month. The
advantages are quite obvious: transparency as regards costs and performance, improved
efficiency and return on investment. This not only gives us the advantage that we are
becoming better in selecting and developing products that enable storage solutions to
be run on a cost-optimized basis and at defined quality standards.
Thanks to this strategy Fujitsu Siemens Computers has also developed into one of
the most successful providers of storage solutions in Europe.
8 Preface

The aim of this book is to provide you with an introduction to storage technologies
and storage networks, and to highlight the value added for your company. We also
introduce some of our partners in the storage sector and provide a comprehensive
overview of the current storage product portfolio of Fujitsu Siemens Computers. Our
products and associated services – delivered by us or our certified partners – are the
basis for storage solutions that help you contend with the growing flood of data!

Your Helmut Beck

Vice President Storage


Fujitsu Siemens Computers
Section 1

The information society –


saving data and knowledge at new levels
There has never been a specific culture without information. Human develop-
ment greatly depends on our ability to acquire information and to make it avail-
able to subsequent generations. Whenever there have been major inventions, for
example the invention of typeface and printing as well as computers and the
Internet, society has undergone a major change as well. Since the invention of
the computer by Konrad Zuse, automated information processing and the asso-
ciated electronic storage of information have affected an increasing number of
areas within our lives. Although generating information is much easier, it has
now also become easier to lose such information as well. Saving, archiving and
managing information is more important than ever before.

I
n 2008 the World Wide Web celebrated its 15th birthday and Google, the Internet
search machine company, became 10 years old. Such milestones represent a new ele-
ment in the history of computer technology: nowadays, anyone can just click on a
mouse and browse through a vast indescribable amount of information and entertain-
ment data at anytime and anywhere. Search machines, such as Google, bring a certain
amount of order to this knowledge chaos and it was this particular discovery by two
Stanford students that “revolutionized global knowledge within a very quick time.” [1]
Accessing data quickly assumes that the data can be read and is available some-
where in the World Wide Web. Its readability is based on file formats, classifications and
index or meta data which must be defined before being accessed by the search machine.
Its existence is based on the fact that it is stored somewhere on electromagnetic stor-
age systems. That is only possible using new and state-of-the-art technology which has
been developed over the last 60 to 70 years, i.e. information technology (IT).
10 Section 1

Original forms of information technology

S toring information and knowledge is not exactly a modern intervention Irrespective


of the social forms of human coexistence history has revealed certain constants in
handling such knowledge and in its significance: even in early times human groups
forwarded their experiences from generation to generation – separately and not just by
word-of-mouth. They piled up knowledge and traditions, first of all verbally and then
using images and systematic symbols, such as Sumerian pictogrammes and Egyptian
hieroglyphics. Our alphabet today used in European languages arose only about 3,300
years ago with the Sumerians and later the Egyptians. Basically a very short time ago
when compared with the overall history of man and homo sapiens in East Africa about
4 million years ago [2].
Human life organized itself around such information, partially written, not just daily
life but also fixed state institutions and not least religion itself. A wide range of cultures
thus existed around the world long before the Europeans “discovered” the world – and
became subject to occidental standards.
August 2008 saw the German weekly news magazine “Der Spiegel“ look closely at
the effects of such a flood of data on mankind and briefly looked at the very beginnings
of information technology. The first stages were:
“The Sumerian matchstick lines turned language into a standard: fixed in clay, they
outlived their creator and became independent of time and space. In 1300 B.C. we had
the first wooden bamboo strips in China and the first book. Comprehensive information
can now be moved. The Phoenicians develop the first text with syllables. About 300
years later the Greeks turned that principle into the original alphabet.” [3]
“Knowledge progress as a result of technology” made a jump forward some 3,000
years later when Johannes Gutenberg invented the copper matrices as the basis for the
mass creation of individual letters in about 1450. The first weekly papers appeared after
1600 in Strasburg and Wolfenbüttel and the first daily paper was published in Leipzig in
1659. Then things moved fast: the invention of the telegraph in mid 19th century, later
the telephone, then the typewriter, the LP, electromagnetic waves via Hertz and then
radio, film and television – until we finally reached the first computing machine devel-
oped by Konrad Zuse in 1938. The rest is “history”: Internet, World Wide Web and mobile
phones.
The information society – saving data and knowledge at new levels 11

Digitizing knowledge

T he means and technologies used by man to document experience and knowledge


either for a short time or indeed long-term have greatly developed throughout his-
tory, whereby the term “knowledge” is used here without any value attached to it, i.e.
without an assessment of the content. However, the reason why human or social groups
pass on information is still the same.
And it is the advance of computers and the World Wide Web which has now moved
society away from an industrial society to one based on information. The production of
goods in the classical sense is more in the background and it is now the product “serv-
ices” and “information” which has been established. Distributing information and enter-
tainment has become a separate profitable business and has constantly changed the
way in which society interacts [4].
More and more digital information is entered, saved and provided via networks. On
the one hand, this growth is based on the amount of accrued company data. Account-
ing was always at the center of automation developments. Some of the oldest clay
slates found in Persia contained book-keeping documentation. The first computers
were predominantly used for book-keeping purposes. And today IT has spread through-
out companies like wildfire. More and more operational areas are now electronic and
each new sector provides an increasing amount of digital data. Depending on the busi-
ness model, there are some companies today which just exist virtually within comput-
ers and that includes many Internet businesses. Classic companies started with solu-
tions for ERP (Enterprise Resource Planning) and moved via CRM (Customer Relation-
ship Management), SCM (Supply Change Management), Data Warehousing and Busi-
ness Intelligence to new areas in Web 2.0 and social media. It is frequently
production-technical and commercial data that is entered the most and processed in
applications from SAP, Oracle or Microsoft.
This structured data, which is systematically stored in database fields, can be easily
accessed via queries. Its evaluation and interpretation has become very complex due to
the quantity and constant expansion. This is why Data Warehousing exists where the
data collected data is sorted and edited according to business intelligence criteria. For
example, for airline companies want to know – for marketing and flight organization
purposes – how often their customers fly, their destination and also their chosen form,
i.e. business or economy. Databases alone do not provide such interpretations.
12 Section 1

Changing information forms within companies

T he Enterprise Strategy Group defines three phases:


Phase one saw the automation of core processes with database-based applications,
for example SAP R/3. Information became structured and transaction-oriented, hence
the name “transaction data”.
Phase two saw the structure of IT within a company change. The PC was invented
and it was introduced to companies on a decentralized basis. PCs were then merged
into workgroups which changed data via the first servers. That was the birth of Novell
and also the starting-point to develop Windows as a server operating system. And sud-
denly there were vast quantities of non-structured office files. The age of distributed
files had started and the required storage capacity was soon greater than that required
for transaction data.
And today we are at the beginning of the Internet data age Web 2.0 applications
such as social networks, wikis or blogs are also being used in companies. Each person is
now not only an information user but also an author of information and so the amount
of data that now has to be stored has considerably multiplied. The digital footprint in
the network is now already enormous and the end of such growth is not in sight. It can
be expected that Internet data will soon leave all other sorts of data way behind.

The growth of non-structured data (Web 2.0)

M ost of today’s data is not structured. It comes from Office or e-mail programmes.
Non-structured data has relevant and non-relevant data, all mixed up and not
easy to sort. That is the challenge: it is non-structured data that has to be managed
correctly so that important data can be archived on a long-term basis without any loss
and so that non-important data can be saved with a minimum of resources. The term
“compliance” is used to describe all the data storage requirements as defined by organ-
izations and the state.
The mountain of digital data is growing because the old analog storage media in the
consumer world is being replaced: voice, music, photos, TV and film are now digitally
recorded and stored resulting in gigantic quantities of data. And added to that comes
the continuing conversion of analog media. The improvement in Internet streaming
technology and the improvement in bandwidths will enable the Internet to presumably
replace media such as CD-ROMs or DVDs for music and film. This will generate increased
demand for hard disks and arrays.
The information society – saving data and knowledge at new levels 13

How big is an Exabyte?


Kilobyte (KB) 1,000 bytes OR 103 bytes. 2 Kilobytes: a typewritten page. 100 Kilobytes:
a low-resolution photograph.
1,000,000 bytes OR 106 bytes. 1 Megabyte: a small novel or a 3.5 inch
floppy disk. 2 – 10 Megabytes: a high-resolution photograph taken by
a digital camera. 3 – 4 Megabytes: a song compressed with MP3.
Megabyte (MB) 5 Megabytes: the complete works of Shakespeare. 10 Megabytes:
a minute of high-fidelity sound. 10 – 20 Megabytes: a digital chest
X-ray. 100 Megabytes: 1 meter of shelved books or two hours of com-
pressed radio. 650 Megabytes: a CD-ROM.
1,000,000,000 bytes OR 109 bytes. 1 Gigabyte: a pickup truck filled
with books. 3.2 Gigabytes: one hour of HDTV. 5 Gigabytes: the size of a
typical movie stored on a single DVD. 20 Gigabytes: a good collection
Gigabyte (GB) of Beethoven works. 50 Gigabytes: capacity of a Blue ray disc. 100
Gigabytes: a library floor of academic journals or about 1,200 hours of
downloaded MP3 music. 500 Gigabytes: the native capacity of the
largest tape cartridge in 2005.
1,000,000,000,000 bytes OR 1012 bytes. 1 Terabyte: 50,000 trees made
into paper and printed. 2 Terabytes: an academic research library. 10
Terabyte (TB) Terabytes: the print collections of the U.S. Library of Congress or pro-
jected capacity of a magnetic tape cartridge in 2015. 600 Terabytes:
National Climactic Data Center (NOAA) database.
1,000,000,000,000,000 bytes OR 1015 bytes. 1 Petabyte: 3 years of EOS
data (Earth Observing system). 2 Petabytes: all U.S. academic research
Petabyte (PB)
libraries. 20 Petabytes: production of hard disk drives in 1995. 200
Petabytes: all printed material.
1,000,000,000,000,000,000 bytes OR 1018 bytes. 2 Exabytes: total vol-
ume of information generated in 1999. 5 Exabytes: all words ever spo-
ken by human beings. 9.25 Exabytes: the amount of capacity needed to
Exabyte (EB)
hold all U.S. phone calls in one year.
90 Exabytes: estimated worldwide available digital storage capacity in
2010 for all media (disk, tape, optical).

Source: Horison Information Strategies, UC Berkeley Study “How Much Information”, IDC

The so-called Web 2.0 with its new interaction options for network participants, for
example YouTube, MyFace, LinkedIn or Xing, will also result in huge data quantities
being stored by the providers responsible. Most of today’s Blade servers and storage
arrays are being sent to such companies. This development will increase as new tech-
nologies expand, such as leased software (Software as a Service / SaaS) or Cloud Com-
puting where the user accesses programmes and data which is stored in giant data
centers “somewhere in the Internet cloud“. Medium-sized companies and start-ups will
enjoy low-priced options that enable them to use such a sophisticated infrastructure.
14 Section 1

Amazon with its large data centers is renting out computing and storage capacity for
its external customers. Of course, the appropriate network bandwidths must exist and
the provider must be 100% reliable. It is clear that new technologies which are used
first of all in the consumer environment will expand into the world of business IT. Risk
analyses for particular situations are essential, especially when looking at security top-
ics and cost savings.
The options in our “Information society” have simply not yet been exhausted. New
processes for transferring knowledge and storing information have joined the existing
procedures [5]. Information technology has enormous potential – but, just like all the
other technical progress before, it is simply a tool for specific purposes. It all depends
on how it is used and for which objectives. First of all, it has to be ensured that data
backup itself becomes more reliable. The storage example shows both the technological
opportunities as well as the restrictions. And that is why – particularly in this environ-
ment – there is a whole range of new fundamental inventions and gradual improve-
ments.

Storage in such a world is thus gaining in significance:

1) The ever-increasing mountains of data first of all has to be saved first – and where
possible using optimized resources.
2) As data has to be available anytime, anywhere, the storage systems used must pro-
vide the data worldwide via the Internet on a 24 hour-a-day basis. Data has to be
saved several times on a redundant basis in order to ensure that it is not lost due to
unforeseen circumstances.
3) An increasing amount of data with even less structure will dominate everyday data
storage processes: data management must thus become more intelligent so that
data is suitably saved according to its value (Information Lifecycle Management).
This is the only way to meet compliance requirements.
Section 2

Tiered storage: intelligent information


management in the company
Companies have looked at the topic of saving data but with various degrees of
intensity. If something goes wrong, i.e. data disappears, is stolen or the data
backup media cannot be recovered when hardware is faulty, then people tend to
keep quiet about it. It is rare for this type of situation to become public knowledge
as the company will try to protect its reputation or brand name. But almost all IT
departments follow some sort of storage strategy even if it is not well-known nor
follows specific regulations: they move their data from A to B or C in order to save
costs or they don’t need the data for a while. Most companies thus have a basis
and a need for a clear strategy when it comes to saving their digital treasures.

A
lmost all the forecasts and estimates about fast data and storage growth have so
far proven to be true despite the comments of a lot of analysts and market
observers. Such estimates have often proven to be even too conservative. In par-
ticular, recent years have shown that, in addition to the company IT and their storage
requirements, other groups in society now digitize their analog information. These
include film and video, voice and music recordings, medical x-rays (medical imaging),
TV cameras in major cities or at security-conscious locations (security & surveillance) as
well as the conversion from analog to digital radio in police, fire brigade and rescue
services. An additional factor is the so-called social networks (social communities), such
as YouTube or Facebook with their enormous amount of data photos and videos.
Digital information is being saved everywhere at a great extent including data cent-
ers which are expanding daily. But is all this data really worth it? Do all those holiday,
birthday and family snapshots – which used to bore us all at family slideshow evenings
really have to be saved on state-of-the-art technology for ever and ever? Does a com-
pany really have to save everything electronically despite legal specifications?
According to IDC the growth of all non-structured data (file-based) that is being
increasingly collected outside the company will exceed structured data (block-based)
for the first time in 2008. The balance in the company between structured data entered
in databases or ERP applications and non-structured data resulting from e-mails
16 Section 2

World wide File and Block Disk Storage Systems, 2005–2008


9000
8000
7000 Block based
File based
Petabytes

6000
5000
4000
3000
2000
1000
0
2005 2006 2007 2008
Source: IDC 2007

According to IDC, in 2008 file-based data will for the first time experience of stronger growth than
block-based data.

including attachments, Office files, presentations, videos and so on, has also shifted.
The problem of retrieving structured data via database queries and interpreting data via
Data Warehousing or Business Intelligence has today basically been solved although
there is no solution for non-structured data [1]. However, this cannot be said for the
actual storage of such data quantities: the main type of storage i.e. fast access to
expensive hard disks is limited and only assigned to really critical business applica-
tions.

Moving data

M oving data records and data in servers and applications to less performant storage
areas is necessary for several reasons:
• New up-to-date data is materializing every day, every hour in business-critical infor-
mation. Booking, purchasing and selling data must all remain in direct access, yet it
becomes “unimportant” after a certain period and even “out-of-date”. It has to be
moved for space reasons, i.e. from primary online storage to slower and lower-priced
disk systems (Nearline or Secondary Storage).
• Other files, such as presentations, are not necessary business critical yet have to be
stored nearby (Nearline) as such files are often modified and used again and again.
Such data is a typical example as it is saved several times and used or read by numer-
Tiered storage: intelligent information management in the company 17

ous employees. The analysts in the Enterprise Strategy Group (ESG) have developed
the “Iceberg” model. There are two main types of data: dynamic data which is con-
tinually changed (= visible part of the data iceberg) and permanent data that is static
or fixed and will not be changed any more (= invisible part) [2].
• Legal regulations or just careful commercial thinking requires long-term data stor-
age without having to have the contents constantly available during daily business.
It can move to an archive in whichever manner. In early days it was stored to tape
(known nowadays as so-called Virtual Tape Libraries) which simulate tape storage in
disk systems. This is a third type of storage – on tape but the data is a long way from
the original starting-point, such as servers and primary disk storage units.
• Last but not least: data loss must be avoided depending on the value of the data by
making immediate or delayed copies which can be recovered as required (Backup,
Restore, Snapshot, CDP/Continuous Data Protection). These procedures are based on
up-to-date duplication concepts which fish out the copies during the backup proc-
esses. But this is only from the backup media and not from the primary or secondary
disk storage units. More information about this complex topic is in section 5.

Even if a company only uses one aspect of these procedures, it is still using a tiered
data backup system as data is being moved, even it is from the server to a backup
medium, and even though this would be inadmissible from a professional storage view-
point.

HSM and ILM: classic concepts with long-term effects

T wo classical concepts, apart from anything else, prove that a tiered storage process
was always somehow operated within a company’s IT. Both look at a fundamental
question: what must be saved, how, for how long and on which medium? In the world
of mainframes the Hierarchical Storage Management (HSM) method was used where
the automatic storage of data is on the lowest-priced storage devices according to the
performance required for each application and according to the description offered by
Fred Moore [3]. This process is not visible for the users who can access all the data they
want without any restrictions irrespective of the storage hierarchy level currently
involved. A special HSM software looks after the various storage levels. IBM introduced
this procedure for the first time in mainframes in 1975.
Once the Internet bubble burst, StorageTek and other providers set up their strategy
on Information Lifecycle Management (ILM) so as to guide the user to a content view of
his stored data. A hardware manufacturer had a wide range of different devices on
offer; there was also more cooperation with those manufacturers who concentrated on
18 Section 2

What is Tiered Storage?


Tiering means establisching a hierarchy
of storage systems based on service
requirements (performance, business
continuity, security, protection, retention,
compliance, etx.) and cost.

Tiering storage requires some mechanism


to place data:
앫 Static – applications assigned to
specific tiers
앫 Staged – batched data movement
(e.g. archive)
앫 Dynamic – some active data mover
(e.g. HSM or ILM policy serice)

Source: SNIA

The stored data is increasingly moving towards non-access on account of the various levels
involved: from servers and quick primary storage and slower storage to backup mechanisms and
archiving.

using software to classify data. EMC bought Documentum – a manufacturer of a docu-


ment management solution (DMS) – in order to set up packages for both their old and
new customers. ILM can be seen as a continuation of HSM in the world of Unix and
Windows. ILM manages the data from its origin to its archiving or deletion and stores
the data on various fast and powerful data media depending on its individual value.
This type of storage management is based on HSM technology and uses company poli-
cies to establish an optimal match between data values and the respective storage
subsystems. Even if companies are often not aware of it they all practice some form of
ILM. Even those who keep their data for a year or even longer on high-performance and
expensive online storage, have made a decision regarding the assumed value of their
data. But whether such a decision can be justified is not clear as in the meantime the
data could have been saved on cheaper data media.
A similar approach is Tiered Storage which is basically the same as HSM and ILM, but
which looks more at the combination of IT infrastructure (hardware basis) and data
contents. The stored data is increasingly moving towards non-access on account of the
various levels involved: from servers and quick primary storage (online storage for data
that requires immediate business access) and slower storage (nearline storage for data
that is only required sometimes) to backup mechanisms and archiving. Such a structure
based on the value of the data exists in every company whatever they may call it.
Tiered storage: intelligent information management in the company 19

ILM Implementation Roadmap

앫 Instrument & manage service to ILM Deploy ILM practices across the enterprise
practices across sites

앫 Automate with ILM Management tools


앫 Capture overall savings & benefits Automate ILM-based policies & services

앫 Begin operating policy-based services


앫 Refine ILM practices & benefits Pilot ILM-based Solution Stacks

앫 Set data & info polices across domains


앫 Tier storage and protection into Standardize Information,
standard Service Levels Data & Security Services
앫 Deploy Configuration Mgmt tools

앫 Begin collaborating on requirements Identify information


앫 Identify value, lifecycle & classification assets & infrastructure
of information for each business unit
앫 Use SRM tools to identify and track resources/services

Source: SNIA

The recommendation of the SNIA Data Management Forum is to add intelligence to the tiering
efforts by integrating it into a broader ILM-based practice. More informations at www.snia.org/
dmf.

HSM and ILM can be regarded as a high-level, indeed “tactile“ strategy which estab-
lishes the data backup stages and criteria in a justified sequence. Many manufacturers
selling HSM or ILM promised their customers that they would, above all, reduce their
storage costs. This refers to the classic storage processes such as data save, backup,
restore and disaster recovery which must be classified as such so that the data can be
saved on the appropriately priced and powerful storage medium according to its value
for the company. That is more easily said than done: how can someone decide which
data should be stored at which stage and on which medium and for how long? [4]
The traditional data hierarchy was split in two: data saved on hard disks with direct
and fast access and backup or archive data saved on cheap tape and which is not in
direct access and thus partially bunkered somewhere. Those who select HSM or ILM as
their strategy want to move away from this old concept and now wish to save data
according to usage, i.e. its direct significance for the business process. Those that plan
such a step with specific criteria can save money immediately [5].
Even if this approach is not always accepted [6], HSM and ILM have had an effect:
Tiered Storage today is seen by companies to be quite normal. A real hierarchy now
almost completely dominates the world of storage. The two tiers have now become four
20 Section 2

or five and, in an ideal situation, they reflect exactly the value of the data on each tier
based on the corresponding costs. In other words, expensive primary storage (fibre
channel and SAS disks) down to less expensive secondary storage (disk-to-disk (D2D))
on SATA disks which still have to be accessed via servers or applications, and then down
to different forms of backup: either as backup storage on cheap SATA disks which have
the function of the older type of tape backup (Virtual Tape Libraries / VTL) or as classic
backup and archive on magnetic tapes.

Tiered storage and data classes

T he classic version only had two storage levels: the first step was to have the data
saved on server-related hard disks (Direct Attached Storage = DAS) and keep it there
for a while for fast data access and then the second step was to move the data to a Tape
Library. If the data had to remain accessible, enterprises could only use powerful tape
libraries from StorageTek or IBM with high-performance robotics that loaded or
unloaded the cartridges quickly.
Backup and archive data was also saved on tapes and kept at a safe place usually
away from the company premises – at least that was the theory. But if the data was
required again due to a recovery situation, it first had to be transported back and then
loaded into the productive system – a process which could often last for hours or even
days.
Further developments in disk and array technology have today resulted in a tiered
storage model which comprises at least three or four classes:
Tier 0: Fast data storage (Flash Memory or (Enterprise Class) Solid State Disk) is used
to ensure that data can be accessed very quickly. For example, Solid State Disks (SSD)
– as very expensive cache storage – have been on offer for years from specialist
companies such as Texas Memory Systems. The main customers are state-owned
organizations in the USA, banks or companies, which are actively involved in crude
oil exploration with enormous amounts of data which even when using online stor-
age would be too far away from the applications.
Tier 1: Mission critical data (such as revenue data), making up about 15% of all data,
very fast response time, FC or SAS disk, FC-SAN, data mirroring, local and remote
replication, automatic failover, 99.999% availability, recovery time objective: imme-
diate, retention period: hours.
Tier 2: Vital data, approx. 20% of data, less critical data but fast response time, FC or
SAS disk, FC-SAN or IP-SAN (iSCSI), point-in-time copies, 99.99 % availability, recov-
ery time objective: seconds, retention period: days.
Tiered storage: intelligent information management in the company 21

Tier 3: Sensitive data, about 25% of data, moderate response times, SATA disk, IP-
SAN (iSCSI), virtual tape libraries, MAID, disk-to-disk-to-tape periodical backups,
99.9% availability, recovery time objective: minutes, retention period: years.
Tier 4: Non-critical data, ca. 40% of the data, tape FC-SAN or IP-SAN (iSCSI), 99.0%
availability, recovery time objective: hours/days, retention period: unlimited.

HSM, ILM or Tiered Storage require clear-cut and continuous data classification. This
can only be handled manually, especially with non-structured data, which is again
much too expensive. The price of the equivalent software in the market, such as Data

Tiered Storage and the Data Lifecycle


Tier 1 Tier 2 Tier 3
Data type Operational Application Reference, archive
Applicatons Mission-critical, OLTP Vital, sensitive Reference
Availability 99.999% 99.99% 99.0%–99.9%
in percent
I/O, throughput Very high High Moderate, low
Scheduled None < 5 hours/year => 10 hours/year
downtime
Recovery Disk Disk, tape libraries Tape
technology 100 Probab
i lity o
f re
u
Primary storage se Secondary storage Long-term retention
앫 Enterprise class disk 앫 SATA disk and virtual tape 앫 Fixed content
앫 Mirroring and replication, 앫 Fixed content, backup/ 앫 Video, medical,
Probability of reuse (%)

CDP recorvery, reference data government regulations

Amount of data
앫 Synchronous and 앫 Point-in-time, snapshot, 앫 Tape libraries, deep archive
asynchronous (remote) deduplication 앫 Offsite vaults
V
alu
eo
fd

a
at

VTLs
SATA/JBOD
ata
Amount of d MAID
0
Average days ∞
since creation 0 days 30+ days 90+ days 1+ years forever
Recovery Time
Objective (RTO) milliseconds seconds minutes hours days
Key components Data mover (remastering)
(ILM) Policy engine
Tiered storage hierarchy

Source: Horison Information Strategies

It is increasingly important to understand that the value of data changes throughout its lifetime.
Therefore, where data should optimally reside and how it should be managed changes during its
lifespan.
22 Section 2

Movers or Policy Engines, is such that an investment has to be carefully calculated


beforehand. The result is that many customers regard ILM or Tiered Storage as a good
idea but shy away from the corresponding investment involved. In practice, ILM has
only been accepted as a comprehensive concept if it could be obtained as an integrated
product (such as CentricStor from Fujitsu Siemens Computers).
It would be a serious mistake to believe that the importance of data once stored does
not change according to the business process, or related to the moment of data entry
or other criteria. For example, immediate access to customer flight booking data is
essential before, during and after the flight but a week later it is only of interest for
those running statistics or customer behavior evaluations. Decisions about data storage
and archive locations and periods must thus be taken. Even when “nearline” tape librar-
ies were introduced at the beginning of the Nineties, it was still thought that archiving
data was the last phase before it was deleted. A life span of more than one or two years
for most data was pretty inconceivable in those days.
But that has certainly changed. On the one hand, state regulations throughout the
world now decree that IT data must be stored for longer periods; on the other hand,
new hard disk technologies have resulted in various disk storage phases. Nowadays, the
proportion of data that must still be retained towards the end of its lifecycle is increas-
ing and not reducing as in earlier days.
There are more “transit stations” than before and more computing power and server
performance must be used when moving data from the one tiered storage to the next.
Many start-up companies have been involved in classifying data and in automating the
processes involved and have since been followed by the “giants” in the branch. Fujitsu
Siemens Computers has been working together with Kazeon in this respect.
Section 3

Online storage: disks and reliability


This section looks at the extremely sensitive area of direct data storage – during
or immediately after its creation within the business process. Nothing must get
lost! The added safety harness in the form of backup on other media comes at a
later stage. In this situation, costs play a less important role. What are the fea-
tures of expensive fibre-channel and SAS disks as used in online storage? Why
are RAID systems used? And how fast must direct data access actually be? These
are just some of the questions to be looked at more closely.

I
n the early days of electronic data processing which only took place on mainframes,
punch cards were originally used to save data. This was followed in 1952 by mag-
netic tapes. Both methods were based on binary data (consisting of two figures: “1”
or “0”) either by punching or not punching a hole on paper cards or by magnetizing or
not magnetizing the tapes. This form of storage is still necessary today as computers
can only handle information which has been reduced or converted using such a binary
system. In other words: this numbering system just consists of a “1” and a “0” because
the heart of the computer, namely the processor, can only operate on such a basis. This
means in turn that data backup is basically pretty uncertain.
Magnetic tapes were fast and could save what was in those days a large amount of
data, namely five megabytes (= 5 million characters, corresponds to several books or
the complete works of Shakespeare – see the overview in section 1). But as early as
1956 alternative data media started to appear: the magnetic disk, the forerunner of the
modern hard disk. The first data media of this type consisted of a stack of 51 disks with
a diameter of 60 cm (IBM RAMAC). Hard disks as known today have several, rotating
disks which are arranged above each other in an airtight housing. In contrast to mag-
netic tapes, the data is no longer written and read sequentially to the data media, which
slows down access and retrieval, but to coated disks. A read/write head is driven by a
motor above these disks and can skip to all the positions required. In contrast to sequen-
tial storage, this type is known as random access and is much faster.
24 Section 3

Lifespan of hard disks

H ard disks consist of mechanical parts which are constantly mobile and require
energy. This is also the case even when there are no read/write accesses. The advan-
tage of magnetic tapes is they are not constantly moving and that the data is also
stored in a non-power status. The lifespan of a hard disk is basically limited as a result
of its mechanisms and thus possible error sources; they should be replaced every three
to four years on average even if some disk manufacturers claim a longer lifespan. Tapes
used professionally are said to last for even thirty or more years.
Relating to the tier storage concept, the use of tape today is moving from backup to
archiving (see section 5), while powerful hard disks have since taken on the main job of
data storage and even some of the backup activity. Disks can be directly accessed by
servers for business critical data where very fast, yet expensive FC and SAS disks are
used. Lower-priced and slower disks are used for backup to disk systems or as an inter-
mediate stage in the form of virtual tape libraries (VTL) on which data is retained for
application access (nearline storage) before it is “finally“ moved to tape according to
specific regulations or periods of time.
It was 1980 before Seagate (a company founded in 1979 und now the world‘s larg-
est disk manufacturer) launched a hard disk suitable for IBM PCs with a capacity of
5 MB. These disks and their immediate successor did not have any intelligence of their

The disk market in terms of interfaces

2007 2009

23,3 26,4 23,3


30,4

16,4
1,4
33,9 44,9

Parallel SCSI SAS Fibre Channel ATA/SATA Source: Gartner Dataquest

According to Gartner Dataquest the SAS interface is evolving into the number one disk technology.
The previosly dominant Parallel SCSI is sinking into insignificance.
Online storage: disks and reliability 25

own and were completely managed by an external controller. Today‘s standard hard
disks are based on standards, such as IDE/EIDE (Integrated Drive Electronics and
Enhanced Integrated Drive Electronics) and ATA (Advanced Technology Attachment)
which come from the consumer sector or are known as SCSI disks (SCSI (Small Compu-
ter Systems Interface) specially developed for enterprises. Many different devices could
be connected to a SCSI controller – from hard disk to scanner. The parallel data transfer
rate was much higher than the previous sequential transport methods. Since 2001 the
ATA development SATA (Serial Advanced Technology Attachment) has become more
widespread – the data is no longer transferred in parallel but in serial.
SATA hard disks now provide competition to fibre-channel hard disks as today they
have a higher degree of reliability and have fallen in price. Fibre channel technology as a
whole is regarded as particularly powerful for enterprises since the introduction of stor-
age area networks (SANs). They read reliably and quickly. SAS disks (Serial Attached SCSI)
today play a significant part in this professional sector as they are gradually replacing the
SCSI disks. As they are compatible to SATA, they can be installed together in a joint array
which can result in tier 1 and tier 2 being connected within one single device.

The performance of various disk types


Fibre Channel SAS SCSI SATA
Specification Online storage Online storage Online storage Low-end file
and transaction and transaction and storage
data data transaction data
Rotation speed 10k, 15k rpm * 10k, 15k rpm * 10k, 15k rpm * 7,200 rpm *
Seek time 3 – 4.5 ms ** 3 – 4.5 ms ** 3 – 4 ms ** 8 – 10 ms **
Typical average
5.5 – 7.5 ms 5.5 – 7.5 ms 5.5 – 7.5 ms 13 – 15 ms
access time
Power-on time
24 x 7 24 x 7 24 x 7 10 x 5
(hours x days)
I/O duty cycle High High High Low
> 1.4 million > 1.4 million > 1.4 million
MTBF 600,000 hours
hours hours hours
Maximum bus
4 gbps 3 gbps 3.2 gbps 1.5, 3.0 gbps
speed
Interactive error
Yes Yes Yes No
management
* rpm = Rotations per minute
** ms = Milli seconds

Source: Horison Information Strategies


26 Section 3

The technical options of the various hard disk types have not yet been fully exploited
and SATA will probably expand further regarding professional storage but will in turn be
replaced by SAS in nearline storage. The advantage of fibre channel is that, in addition
to better equipment with internal microprocessors for mechanic and error control, it
can be positioned away from other devices in the storage network (up to 10 kilometers,
whereas SCSI is only 25 meters). This was decisive in setting up Storage Area Networks
(SANs) since the end of the Nineties as decentralized locations, such as buildings on
extensive factory premises or within a town, could be connected to each other via the
storage network. The use of IP protocols for storage networks has since extended the
range of FC and SCSI/SAS hard disks to cover very large global distances.
Only powerful FC, SAS and SCSI disks are used in online storage as part of the data
and storage hierarchy [1]. Solid State Disks (SSD) already installed by some manufactur-
ers in their storage systems are already significant as a kind of second cache (RAM) due
to their high access rates. As they have no mechanical parts, they have a lifespan that is
longer than that of classic hard disks. But they too are nearing their end as the SSD
lifecycle comes to a conclusion after 10,000 to 1,000,000 write accesses [2] according
to manufacturer specifications.

Online storage customer requirement:


reliability, availability
• What is most difficult to replace if a company loses everything,
e.g. as a result of disaster:
– Its buildings?
– Its computers?
– Its data?
• “50 % of all companies who lose their data as a result of a fire or flood are bank-
rupt within one year.”
• Possible solutions:
– More reliable disks (migration from SATA to SAS)
– Redundant disks (RAID, mirroring)
– Regular backups

Source: Fujitsu Siemens Computers


Online storage: disks and reliability 27

Hard disks – a contradiction in terms regarding


long-term data backup

E ven if you decide to use high-value hard disks in online storage, one problem still
remains, namely the lifespan of the hard disks – and thus your data – is extremely
limited. The manufacturer’s warranty has usually expired after three to four years and
new investments are made usually for accounting purposes.
The actual reason for this situation is the strange contradiction that exists in our
so-called digital age. An increasing amount of information is saved on magnetic media,
but that is certainly anything but long-term as a power failure, head crash, material
damage, theft or the sudden “death” of the system hardware are all enough to send
your data – so carefully saved and at considerable cost – all off into Nirvana.
The correct moment for replacing hard disks also depends on conditions, such as
temperature and the amount of time the hard disks have been running. The lifecycle (or
duty cycle) is also reduced by frequent power-ups and switch-offs – more so than if a
disk runs on a round-the-clock basis; this can be compared to the stress situations that
arise when a plane takes off and lands [3].

Annualized failure rates (AFR) broken down by age groups


10

6
AFR (%)

0
1 Year
6 Month
3 Month

2 Years

3 Years

5 Years
4 Years

Source: E. Pinheiro/W.-D. Weber/L. A. Barroso, Failure Trends in a Large


Disk Drive Population, February 2007 (Google)

Not only the age of the disks is accounted for in the annual error rates but also the different
disk types.
28 Section 3

Even if the disks last somewhat longer than a three-year cycle in particular situa-
tions, it is advisable to make the change and recopy the data in good time. How impor-
tant is the price of new disks (constantly falling) and the administration hours involved
in comparison to a data catastrophe that you yourself have caused and the costs of
which could damage a company beyond belief?

Reliability with RAID

S o-called RAID solutions exist to protect data on a basically unreliable hard disk and
they are used as a standard in most of today’s disk systems. RAID, Redundant Array
of Independent (earlier known as Inexpensive) Disks started their development at the
University of Berkeley in the mid Nineties. There are two important functions in a RAID.
First of all, many small and inexpensive disks are combined to one group LUN (Logical
Unit Number) so that they provide more performance for an application. RAID 0 com-
bines several disks in order to increase performance. However, this situation increases
the risks involved (in contrast to the general belief that RAID provides greater data
security) as the statistical probability increases that the whole system will no longer
function when one of the disks in the array goes down.
The operating system sees a RAID as one single logical hard disk (consisting of many
different physical disks). The RAID function known as striping means that the capacities
of each disk in the array can be split into partitions which in turn can be addressed by
one or more LUNs: application data can thus be distributed across several disks whereby
they use an early “type” of storage virtualization.
Today, RAIDs are mainly involved in providing protection against disks failing. A RAID
controller can do more: it takes on the disk administration, modifies the configuration
and size of the system cache according to the respective application requirements. The
common factor in RAID levels (apart from RAID 0) is that there is a balance between
performance and redundancy according to the system or application requirements. The
redundancy versions require additional investments in disks which then standby for the
emergency of disk failure (so-called hot spares).
Disk systems can be reset to their original state should a disk or a LUN fail by using
parity calculations which take up a lot of computing time and nowadays are fixed in
chips. RAID 1 saves everything twice (this requires double disk capacity), while RAID 5
requires about 10% more disk space in order to ensure redundancy: One disk covers for
several disks and is ready to jump in and help should there be an emergency.
The widespread RAID 6 was developed in order to eliminate a possible RAID 5 error:
Specialists from different manufacturers and research groups wondered what would
happen if, during processor-intensive reconstruction of the original parity status after
Online storage: disks and reliability 29

Milestones in the hard disk drive industry

Capacity Company Model Year Formatted Capacity


5 MB IBM 350 Ramac 1956 4.4 MB
10 MB IBM 1301 1962 21.6 MB
100 MB IBM 2302-3 1965 112 MB
500 MB STC 8800 Super Disk 1975 880 MB
1 GB IBM 3380 1981 1,260 GB
100 GB Seagate Barracuda 180 2001 181.6 GB
500 GB HGST 7K500 2005 500 GB
1 TB HGST 7K1000 2007 1 TB
1.5 TB Seagate Barracuda 7200.11 2008 1.5 TB

Source: Storage Newsletter, edition July 2008

a disk downtime, the spare disk or another disk in the array were to fail? If such a failure
were to occur, RAID 6 has a second parity calculation ready which becomes active when
a second disk fails. In such a situation, the performance of the controller drops by more
than 30% compared with a simple RAID 5 downtime.
Most manufacturers recommend a specific RAID configuration for their systems or
applications. For example, Oracle recommends a combination of RAID 1 and 5 for its
database in order to increase performance. RAID 3 is more suitable for video streaming
and NetApp has selected RAID 4 for its NAS-Filer as very fast read/write actions can
occur on the disks [4].

How much data protection do you need?

I T manufacturers have always been fairly inventive when introducing additional pro-
tective functions for data storage around disk arrays. All these procedures are based
on the idea of redundancy: keep everything twice or more where possible. This includes
at a hardware level clusters and grids which means that specific hardware is available
several times so that the second device can run with an identical configuration and
identical data should there be any problems. The transfer between clusters and grids is
based on scaling and also subject to heated expert debate.
International companies have also networked their data storage units and have to
go to great lengths to protect their storage media, networks and backups against any
30 Section 3

misuse. Today‘s SAN and NAS infrastructures unfortunately have only low-rate security
mechanisms both at a fibre channel level as well as on an iSCSI basis. They thus fre-
quently do not meet the security policy requirements for the company‘s IT.
Zoning switches in FC-SANs ensure that access control is permitted for each storage
system. This zoning can be run on a hardware or software basis. Soft zoning means that
devices just get information about those systems with which they are to exchange data.
Hard zoning means that hardware checks all the packages and forwards them only to
the permitted addresses. LUN masking is the function in an FC-SAN which makes only
those storage areas visible to an application which the latter needs to implement its
tasks.
With IP-SANs on iSCSI basis, IPSec is used for authentication and saving data
streams, for example, via encryption [5].
Mirroring (a RAID 1 principle where i.e. one disk is an exact mirror of the other) can
also be applied to mirroring an entire storage system. An identical server can be posi-
tioned at a second location possibly several kilometers away and a storage landscape
can be set up based on a contingency data center. All the data is constantly transferred
from location A to location B so that the same data is at both locations. In the event of
a catastrophe, the productive IT including the stored data is transferred from A to B. As
everything is redundant and mirrored, IT operations can be continued.
The software elements in data backup are logically based on processes for Backup
and Restore, Continuous Data Protection (CDP) as well as Snapshots. (See section 5 for
more details.) All these processes have their origin in the basic problem of storing data
on electromagnetic media, such as hard disk or tape (and likewise DVD or Blu Ray).
Despite all the advantages of such technology the stored data could suddenly disappear
into thin air! An old medium such as paper can be longer lasting and proven methods
exist against dangers such as fire or natural catastrophes. However, protecting elec-
tronic data media on a permanent basis against downtime or damage remains a tricky
and never-ending story to which IT managers must give their full attention. There is no
patent recipe.
Section 4

Storage networks – spoilt for choice


Today storage networks are denoted as “state of the art”. At least all large compa-
nies use this technology. However, it exists in various versions, which makes its
understanding complicated – particularly for beginners. Therefore, this section
deals with the description of some fundamental architectural features of the
various approaches and the explanation of important terms. This is especially
necessary since small to medium-sized companies in particular are currently
also able to set up own storage networks for their purposes. However, which one
should you choose in practice?

H
ard disks that are installed in servers and PCs or are directly connected to the
servers in storage arrays are still the most wide-spread structure in small to
medium-sized companies – known in this case as Direct Attached Storage (DAS).
Small to medium-sized businesses have discovered the productive forces of IT and use
them for their business processes. However, at the same time their financial resources
limit their investments in an own IT infrastructure. Furthermore, there are fewer experts

EMEA market share for DAS, NAS and SAN in 2007

NAS
17 %

60 %
23 % SAN

DAS

Source: IDC, 2008

Even although SAN determines the topology of storage systems as a whole, the share of DAS is still
large, particulary in small and medium sized business.
32 Section 4

available, who in addition are not able to specialize in the same way as their colleagues
in large-scale companies, who only have to provide support for a small part of the IT.
Consequently, small to medium-sized businesses do not follow every trend and concen-
trate on the basics. Although keeping a DAS structure does not meet the state of the art
of current storage technology, it can be used as the starting point for a gradual trans-
formation.
But what does DAS really mean? Users who connect one or more storage arrays per
server have a dedicated, exclusive storage array for precisely the application that is
installed on the server. This silo structure may become somewhat complex with time –
and use a great deal of space and energy in the server room or data center – but it is
ultimately easy to monitor and manage. The disadvantage is obvious: If not split into
several partitions, each directly attached storage unit is in the same way as the indi-
vidual server not at full capacity. The superfluous capacity and computing power is
used in the server for peaks for special occasions, such as internal monthly billing or
external accesses for web-based purchase orders in the pre-Christmas period. And for
storage it is “reserved” for corresponding write and read operations – in other words
there is a gap between the investment made and benefit achieved. According to ana-
lysts individual servers only run at about 15 to 20 % capacity and for storage the value
is on average about 40%.
In other companies, where the client/server infrastructure was implemented via the
internal network, there were provided storage structures, in which several server units
had access to the same storage arrays – but separate for mainframes and Open Sys-
tems (Unix computers, and later also for Windows servers). However, the amounts of
data to be moved in the local area network (LAN) became increasingly large, which was
to the detriment of the transfer speed and caused data losses. Networks on the basis of
the Internet protocol (IP), which had originally only been developed for the transport of
messages [1], thus reached the limits of their capacity.

Why storage networks make sense

T he need for a separated network for storage purposes only was reflected toward the
end of the nineties in a separate technology for Storage Area Networks (SANs). The
new infrastructure consisted of own cabling and a further development of the SCSI
protocol, which was already used for the connection of various devices, such as storage
arrays or printers to a server, and bears the name Fibre Channel (FC). The Fibre Channel
protocol was specially developed for the transport of files. It is said to be reliable, and
most recently with 8 Gbit/sec achieved a transport speed that even outperformed the
Ethernet.
Storage networks – spoilt for choice 33

PC Client PC Client PC Client

Ethernet LAN

Server Server

Storage Area Network Fibre Channel SAN

Disk Storage Disk Storage Disk Storage Disk Storage Tape Storage
Source: Fujitsu Siemens Computers

A storage area network (SAN) constitutes a separate storage infrastructure which is only intended
for data transport.

In an FC network special switches are given the task of connecting storage arrays
with servers and also with each other. A switch works as a kind of multiple socket, to
which various devices can be connected. [2] In contrast to the wide-spread image of
Fibre Channel as “difficult to set up and manage”, specialists describe it as easy to han-
dle. This is for example the opinion of Mario Vosschmidt, Technical Consultant with the
American IT manufacturer LSI.
34 Section 4

During its time of origin this Fibre Channel architecture was particularly linked with
the Californian company Brocade, which was founded in 1995. Today, this company is
the market leader in FC switches, which work as the nerve center in an SAN and were
equipped with more “intelligence” in the course of their development. This means that
such switches can take on tasks within the network, such as zoning or virtualization.
With their help it is possible to set up a “Fabric”, a structure that forms the core of a
SAN.
One particular aspect is the configuration of different storage zones (zoning). The
administrator can define which devices and data are and which ones are not to be con-
nected with each other. This serves to protect against unauthorized access both in and
also outside a company. If a SAN is extended, additional switches and zones can be set
up, depending on the existence of ports (connection for cable). The name “Director” has
become widely accepted for larger FC switches with at least 128 ports. Brocade has
taken over several providers (McData, CNT and Inrange), who were greatly involved with
directors [3]. The intention of the manufacturer with these purchases was to strengthen
it market position vis-à-vis Cisco.
Cisco, the worldwide leader in Ethernet switches, has also had Fibre Channel solu-
tions in its portfolio for several years and has thus positioned itself as a competitor
against Brocade for Fibre Channel. Not for the first time in the history of information
technology have the cards been re-shuffled between the companies involved – a recur-
ring development that is enhanced by a forthcoming new technology: Currently, Fibre
Channel over Ethernet (FCoE) is being used to attempt to bring together the separate
networks of message transport (Ethernet or TCP/IP) and data storage (Fibre Channel and
iSCSI) to again form a common network. IP-SANs on an iSCSI basis would already be in
a position to do this, but the communication and storage transport networks are mostly
kept separate for performance reasons.
A new FCoE network calls for standards and an agreement between the various pro-
viders. However, before this is finally the case, hard-fought conflicts rage to decide
market positioning. Every manufacturer wants to be involved in FCoE, even if it has to
switch over almost completely to new products. Some providers have obviously still not
forgotten that a previous rival technology of Ethernet, named “Token Ring”, also lost
the race because the manufacturers behind it concentrated too much on their core
product and thus ultimately did not keep up with the competitors [4].
The historical service of FC-SANs, which are the prevailing storage infrastructure
today in large companies and institutions, consists in providing efficient, fast transport
services that are less susceptible to errors. Although the technology is simple in com-
parison with a classic network, problems frequently still occur in practice because server
and network administrators are too unfamiliar with data storage. Compared with Eth-
ernet, Fibre Channel has ultimately remained a niche technology, in which there is still
a lack of standards in many places. And over the last few years false expectations have
Storage networks – spoilt for choice 35

in part been raised, because the obstacles (and prices) for FC training courses were set
too high. The storage arrays attached in the SAN are mostly managed via the tools sup-
plied by the manufacturers, which in turn only requires a short familiarization period
and is directly supported by the suppliers [5].

Consolidate file services

A pproximately at the same time as the FC-SANs, an alternative network structure


came into being for the storage of data within the company network, which is par-
ticularly associated with the name of Network Appliance (today NetApp). A Network
Attached Storage (NAS) denotes an integrated overall solution, which combines servers,
operating system, storage units, file system and network services. For this purpose
NetApp offers so-called filers, which support the file services NFS (= Network File Sys-
tem, originally developed by Sun) and CIFS (= Common Internet File System under Win-
dows) and are especially suited for unstructured data.
Whereas the files are saved in blocks in an FC-SAN, i.e. in small data blocks of about
four to 128 KB, with the saving of file services we are dealing with related files. This
makes handling easier for the administrator and direct access to the file contents is also
possible. When storing in blocks, which derives from the physical splitting of hard disks
into sectors and blocks, no data can for the time being be accessed via the start and end

DAS NAS SAN


Direct Attached Storage Network Attached Storage Storage Area Network

Application Application Application

Network
File System
File System
File System
Network

Disk Storage Disk Storage Disk Storage

Source: Fujitsu Siemens Computers

Every topology pursues a different concept but the goal is the same: protecting application data.
36 Section 4

The difference between SAN and NAS


SAN NAS
Network Fibre channel – complex, IP – simple implementation, cost-effec-
expensive, closed system, tive, open system, bear security in mind
secure
TCP/IP – fast but very high overhead
Protocol FC, fast – up to 200 Mb/s
(up to 40% net)
Simple implementation, open and fast
Optimized for Rapid data transport
communication over long distances
Types of data All data Only files
Split Drives, resources Files, stored content
Storage for Server (data center) Clients (workgroups)
Drives All Only disks

Source: Fujitsu Siemens Computers

of the files and their contents and structures cannot be directly accessed, either. All PC
users know that their data is archived in certain files and folders and thus have a logical
structure. They also know that their data are ultimately scattered over the hard disk –
after lengthier use they are “fragmented” (distributed, torn apart), because with each
storage process free blocks are occupied first, regardless of the context of the file con-
tent. Consequently, the operating system needs an increasingly long amount of time to
open files. First of all the various blocks have to be on the physical level and consoli-
dated into one entity that is visible for the user. By using the command “Defragment”
the suffering Windows user puts things in order again on the hard disk – at least for a
while.
In an NAS the focus is placed on the network functions [6] and less on the perform-
ance of the hard disks used. Many users consider it a lower-cost alternative to an SAN.
Whoever decides in favor of which version depends on a great many factors that are
perhaps also very individual. David Hitz, one of the founders and now Executive Vice
President of Engineering at NetApp, expressed a frank opinion in an interview: “NAS
and SAN are like two flavors of the same ice cream. NAS is chocolate-flavored and SAN
is strawberry-flavored. And everything the customer needs to know about the two
technologies is only that both systems can be used at any time for data storage. What
intelligent person would be disturbed by the fact that the someone does not like choc-
olate-flavored ice cream, but prefers strawberry-flavored ice cream.” [7] This somewhat
flippant statement can also be interpreted in such a way that companies with SAN and
NAS have two storage architectures to choose from, which can be individually adapted
depending on their requirements. No-one needs to have any reservations.
Storage networks – spoilt for choice 37

Comparing the three topologies


DAS NAS SAN
Based on network technology No Yes Yes
Maximum number of supported
15 / 126
appliances / HBA
Vulnerability to downtime
Yes (copper) Yes (copper) No (glass)
with external influences
Price level Low High Very high
Scalability Bad Relative Very good
Maximum distance to server 25 m / 10 km
Base protocol SCSI Ethernet FCP

Source: Fujitsu Siemens Computers

A third version has been under discussion for some years now: iSCSI networks for
storage units (also known as IP-SAN) obviously overcame the lengthy introductory
phase a year ago and have achieved significant sales figures. The attraction of this
architecture is its ability to use the existing TCP/IP infrastructure for data storage. This
makes the installation and maintenance of a second infrastructure (only set up for stor-
age) superfluous, and the administrators can fall back upon their existing IP know-how.
In practice, however, there have been greater obstacles in the integration of the various
tasks of the LAN and iSCSI storage network. Nevertheless, new prospects are the result
of the new transfer speed of 10Gbit/sec for the Ethernet, because this technology is
currently faster than Fibre Channel with only 8 Gbit/sec at present. However, customers
incur additional costs due to the new cabling that becomes necessary. In the meantime,
it is generally assumed that an iSCSI infrastructure is mainly suited for small to medium-
sized companies and has found its true position there.
Section 5

Backup & Restore: an unloved


compulsory exercise
Data backup is the same story as with many necessary tasks, but which are ulti-
mately only performed half-heartedly or not at all because they require addi-
tional time or more money. A motorcar or a bicycle is often only used until it no
longer works – based on the motto “Nothing will go wrong”. And if something
happens, for example the brakes or the steering suddenly fail or a tire bursts, then
it is usually too late to avert disaster. Modern IT is similar: “In actual fact,” copies
should be made at all times of every piece of stored information and kept ready
for emergencies. However, companies that behave according to this ideal are the
exception. Since everything that does not appear to be directly necessary for the
day-to-day business or cannot be immediately converted into hard cash, is not
given the attention that would actually be appropriate to the matter. Corporate
IT and the need for data backup have been in existence now for about 40 years,
but surveys and analyses still result in glaring errors in this basic discipline. That
ought not to be so: creating an efficient and automated remedy is no great con-
juring trick.

T
he penetration of society and the economy with IT has only just begun. Increas-
ingly more parts of daily life – from communication and information procurement
right through to healthcare – are dominated by IT systems, and economic proc-
esses today depend on electronic support in almost all branches of industry and in all
sizes of companies. This interdependence of business processes and electronically gen-
erated and processed information makes it absolutely necessary for companies with
requirements of all magnitudes to ensure secure data storage.
40 Section 5

Fundamental backup terms

Backup:
Regular data backup to enable access to data in the event of data loss.

Restore:
• Recovery of data via backups (irrespective of the medium) to the time when the
backup was originally created.
• The result provides a physically intact volume / file system.
• The result does not necessarily provide data, with which applications can be
started (particularly with databases).
• Basis for recovery.

Recovery:
• Restartable restore of data and systems.
• For databases retrace with the help of redo logs to the most current version pos-
sible.
• Reset to the last possible point of consistency.

Distinction in backup volumes:


• A full backup (complete data backup) is on account of the time required only
performed at greater intervals and constitutes the basis for the following back-
ups, which only save the extensions and changes to the original database that
have taken place in the meantime.
• Incremental backup denotes a backup of the data that have arisen anew or been
changed since the last backup (regardless of whether it is an incremental or a full
backup).
• A differential backup always specifies all the changes after the last full backup. In
this respect, a differential backup needs more storage space than an incremental
one. Further specifications exist depending on the backup software used [1].
Backup & Restore: an unloved compulsory exercise 41

Backup and recovery strategies

G artner analysts assume that a fundamental change in significance of the data


backup is currently taking place [2]. The most important factors to trigger a re-
alignment of backup and recovery strategies can be specified as follows:
Local threats, such as fire, extreme weather conditions and failures in hardware,
software, networks and media endanger business continuity. Terrorist attacks and
threats have featured highly on the agenda in many countries since September 11,
2001. Power failures of a greater scope, such as those that have occurred in many
countries over the last few years, can also have an impact on data backup.
In the age of globalization and the Internet most companies cannot afford any inter-
ruptions and downtimes in the infrastructure. At the same time, the time slots for
backup and recovery are becoming increasingly tight, because work is performed on a
24/7 basis in various locations.
Unstructured data (e-mails, files) grow in a particularly disproportional manner,
while classically structured data (databases, business applications) only have compara-
bly modest growth. However, according to surveys performed by BITKOM, the German
Association for Information Technology, about 50% of the data saved is not used. This
calls for measures to move these unused data to less expensive levels of the storage
hierarchy or to remove them from a productive context and archive them earlier.
In many branches of industry, such as banking or healthcare, retail trading and
mechanical or automotive engineering, efficiently organized data backup processes
have always been decisive for a company’s market image. Data loss in banks, insurance
companies or airlines, for example, will have a relatively fast impact on the business
success of these companies. This is why data backup always has to be performed with
state-of-the-art methods and improved so as to continue to remain competitive.
And last but not least, IT infrastructure has to be equipped to meet rapidly changing
business and technical requirements. Therefore, backup and recovery processes should
be planned on a long-term and flexible basis and not at the expense of the productive
processes of IT.
42 Section 5

How does data loss come about?

Human
Hardware Error
or System 32 %
Malfunction
44 %

14 %
Software Corrupton or
3 % 7% Program Malfunction
Natural Disasters Computer Viruses
Source: Fujitsu Siemens Computers

Software errors and viruses are relatively rarely the cause of data loss. Most faults are caused by
the hardware or system, followed by human error.

I n addition to the aforesaid external influences, the direct reason for data loss is soft-
ware errors, which impair data integrity or can even cause entire systems to fail, as
well as various hardware errors, which can go from the power supply units and proces-
sors [3], via hard disks [4] to other components and even to redundant assemblies (on a
double or multiple basis) such as hard disk arrays. Added to these are user and operat-
ing errors, which even proficient administrators can make, whereby in this regard a
great deal disappears under a dense veil of silence. Which IT department and which
company gladly admits to having done something wrong?
Regardless of the imminent disasters it is frequently these technical errors of every-
day IT or simply the end of the life-span of the components and media used, whose
sudden expiry can also mean the demise of the data stored upon them. A separate
branch of industry, which has become greatly centralized over the course of the last
few years, looks after the recovery of data storage systems of various types [5].
In order to specify the durability of hard disks their capacity and expected life-span
are expressed in the term MTBF (Mean Time Between Failure), which is an extrapolated
value for the probable downtime of a drive. Many manufacturers specify values between
one and a million hours for high-end drives, which would mean a biblical life expect-
Backup & Restore: an unloved compulsory exercise 43

ancy of 114 years (= one million hours). The underlying tests assume a very high number
of parallel disks, on the basis of which possible failure rates are calculated. However, the
implied optimal conditions are in practice the exception so that real failure rates can be
very high. Redundancy through disk system arrays (see section 3) and sophisticated
backup mechanisms have to prevent this. Since the term MTBF has on account of its
inaccuracy increasingly come under criticism, other units of measurements, for exam-
ple AFR (Annualized Failure Rate), are used today. It is established in the same way as
the MTBF, but specifies the anticipated annual failure rate as a percentage of the
installed number of disks. If 8.7 out of 1000 disks fail a year, the annual failure rate or
AFR is 0.87% [6].
The average life-span of hard disks is at present three to five years, in individual
cases even longer. Companies should only rely on longer time periods if they use auto-
matic error controls with direct notification of the manufacturer’s service, which
depending on the stipulated Service Level Agreement (SLA) ensures a replacement
before the disk finally fails. As capacity increases, disks become more and more inex-
pensive (in summer 2008 Seagate announced a 1.5 TB disk), and the magnetic tapes
that are still used in storage are becoming more efficient. For example, their throughput
has in the meantime risen to more than 500 MB/sec, while capacities are also clearly on
the increase and now lie at about 1 TB (LTO 4). The life-span of magnetic tapes is speci-
fied as up to 30 years for DLT/SDLT and LTO and is thus clearly beyond that of hard disks
and solid state disks [7].

Backup architectures / Storage locations for backup data

T he consequence of this development in the price/performance ratio for the backup


architecture is that tape continues to move further toward archiving at the end of
the tiered-storage chain. An ultimately futile dispute has even developed as to which is
the better solution for backup – disk or tape. This dispute is futile because it is the par-
ticular requirements of each company that matter. If the proportion of data that has to
remain permanently accessible for a longer period of time is very high, it is advisable to
use special forms of a disk backup before the data is finally moved to tape.
For the backup there are various forms how physical data backup can take place:
• Backup to Disk / Disk Libraries:
Storage on disks is faster than write accesses on tapes. The data is usually saved on
low-price SATA disks and not on cost-intensive Fibre-Channel disks.
• Backup to Tape / Tape Libraries:
Data backup on tapes is a low-cost option of storing data. Compared with disks,
tapes have the advantage of a longer life-span. However, access is slower. [8]
44 Section 5

• Disk to Disk to Tape (D2D2T):


This version takes various requirements into account by also considering data backup
on magnetic tapes in addition to the short-term backup on disk as the next stage. If
two disk levels are used within a storage array, which combines several expensive
and cheaper hard disks, such as Fibre Channel or SAS, on the primary level and SATA
on secondary level, price advantages can be achieved because only one array needs
to be procured. Merely the time of the backup on tape has been delayed.
• Virtual Tape Libraries (VTL):
VTLs represent the ultimate in storage. Here we have a combination of fast access
times from disks with low-priced tapes. A virtual tape library (VTL) is a storage sys-
tem on the basis of a disk array, which outwardly emulates a tape library. As a result,
it is possible to integrate backup-to-disk concepts in existing data backup environ-
ments, which are usually based on tape drives.
For connected computers a VTL presents itself like one or more tape libraries. First
the data is temporarily stored on disk in order to keep the backup window to a mini-
mum. The data backup is subsequently effected on the lower-cost tapes.

To define Service Level Agreements (SLAs) for backups the following terms are often
used today:
• RPO (Recovery Point Objective)
The term RPO (Recovery Point Objective) is used to describe how much data loss the
company can afford at most. The interval between the individual backups takes its
bearings from this. For example, it follows from this that banks have zero tolerance
toward failure, whereas other branches of industry can cope with this better.

RPO RTO
Disaster

Weeks Days Hours Mins Secs Secs Mins Hours Days Weeks

The gap is
narrowing

RPO = Recovery Point Objective, RTO = Recovery Time Objective


RPO = Data Loss RTO = Downtime Source: Fujitsu Siemens Computers

RPO = The amount of data which has to be recovered after a data outage in order to be able to
resume business as usual (measured in time). RTO = The maximum recovery time which can be
tolerated before business must be resumed.
Backup & Restore: an unloved compulsory exercise 45

• RTO (Recovery Time Objective)


RTO concerns the question as to how long it takes to recover lost data or restore
them in the system? The scope may be restricted by statutory or institutional speci-
fications. Taken as an example here is healthcare, in which access to electronic
patient data has to be fully guaranteed – particularly in emergencies.
• Continuous Data Protection (CDP)
CDP means that every change in the data is also simultaneously made in the backup.
In CDP the RPO (Recovery Point Objective) is set to zero, because each change imme-
diately triggers a storage process. In this regard, CDP is the realization of the backup
ideal: Everything is saved – immediately and in its entirety.

Extended backup options

S ince the backup windows are becoming increasingly small, there are various techni-
cal approaches to meet these requirements.
Using
• points in time or
• snapshots
enables a backup to take place in a relatively small time slot. Here the data is pre-
pared as a copy at certain times during ongoing operation in order to run the backup
from this copy – irrespective of productive operations. However, a snapshot is also
stored on the primary storage system – and if this fails, this form of data backup is also
no longer effective.
The backed-up data volume can be reduced through compression factors, which
saves space and costs. More recent procedures, such as Single Instance and Data Dedu-
plication, make a direct search of the data volume to be stored and sort out double or
multiple datasets during the backup process. Such findings are replaced by pointers,
which can entail savings figures of far more than 50%. A pointer is only a reference to
the originally saved file so that it should not be saved twice or repeatedly.

• Single Instance Storage


Single Instance Storage describes how double files are detected and saved once
only.
• Data Deduplication
is already classed as one of the great inventions in storage technology [9]. Redun-
dancies are not only detected at file level here, data deduplication also takes place at
segment level. This can take place in two ways [10]:
46 Section 5

– On the target
Deduplication takes place on the storage medium itself, which helps keep the data
volume to be stored to a minimum, but the entire storage process is extended as a
result.
– At source
In addition to the low data volume the other advantage here is that the reduced
quantity of data can be transferred more quickly. This is relevant for branch offices,
because in part only analog data links are available here.

Backup versus archiving

C ompared with backup, the archiving of data is often restricted by the parameter
time – permanently versus temporarily. That may be applicable in many cases,
because weekly backup tapes possibly end up in the data bunker marked with a label
saying “Archive”. Nevertheless, such a distinction – and in particular the associated
practice – is inadequate, because both cases could not be more contrary. Backup is
about protection against data loss, whereas archiving deals with long-term storage in
order to meet e.g. statutory requirements. Backup is normally used in a 1:1 ratio to

FSC Business Continuity Model Requirements


Data Class 1 2 3 4
Description Mission Critical Business Critical Business Non Critical
Important
Availabilty > 99.99 % 99.9 % 99 % 97 %
Unscheduled
< 1 h/year < 10 h/year < 100 h/year No commitment
downtime
Planned
< 1 h/month < 2 h/month < 8 h/month Intermittent
downtime
RTO (Downtime) < 15 Min. 1 hour 8 hours 24 hours
RPO (Data loss) < 1 hour 12 hours 48 hours 96 hours
Archive Access
Seconds Seconds < 4 hours 24–48 hours
Time
Backup Success
97 % 95 % 90 % 90 %
Rate
Offline import
< 30 Min. < 45 Min. < 2 hours Not specified
response time

Source: Fujitsu Siemens Computers


Backup & Restore: an unloved compulsory exercise 47

restore, in an archive you should be able to find data relatively easily and quickly with-
out having to perform a restore.
The copies of the original data mirrored on the backup media are usually only stored
for a shorter period of time and are replaced on a permanent basis by more up-to-date,
write-related copies. As they are actually not intended for use, only specific require-
ments arise, such as currency or completeness. In the event of a disaster all information
before the oldest existing backup and after the last one is irretrievably lost.
However, these data backups must also not be kept for a long time – because the
copies are only needed in the event of a disaster. You will seldom have to restore a data
backup that is a few months old.
Archiving is an altogether different matter. This storage procedure – when informa-
tion leaves the productive process – is about preparing data on a separate medium and
making it available for later use. Archiving media are not put to one side like backup
tapes, because their use is only taken into consideration in extreme cases. Whoever
archives, has reuse as his objective – at some later point in time.
Archiving can be done of its own free will – the bandwidth ranges from precaution-
ary storage for as yet unforeseeable purposes right through to acquisitiveness – or
because legislation, the banks or other institutions have issued mandatory regulations
or at least recommend archiving. Whoever no longer wants to use the archived data
later and also does not have to comply with regulations concerning any possible resub-
mission, should also consider deletion – which saves resources, space and costs.
Section 6

Storage Management – making complex


storage networks manageable
One would think that after over 40 years of server and storage technology the
handling of such infrastructure on a daily basis and their specific applications
would be easy. That’s the theory. However, the fact is that data quantities are con-
stantly increasing and the storage landscape has become more complex and com-
prehensive. The numerous company takeovers have resulted in users operating
various IT islands with different operating systems, device parks and applications.
In the meantime there are at least four major storage architectures which require
different levels of know-how: DAS (Direct Attached Storage), SAN (Storage Area
Network), NAS (Network Attached Storage) and IP-SAN (iSCSI), whereby the last
two are based on TCP/IP protocol. The next architecture is on the horizon – one that
is only partially compatible with the above installations, namely Fibre Channel
over Ethernet (FCoE). This could return the world of storage to a more standardized
environment. But management does not become any easier as a result – indeed to
the contrary. So far each manufacturer has provided his own management tools
for his products yet generally without any standards except for SMI-S.

T
he confusion in storage management software can possibly be overcome by keep-
ing to the tools supplied by the manufacturers. Integrated configuration tools, web
tools or component managers help you at the start but do not cover the entire
planning of the overall architecture and its elements. The large storage management
suites and the administration of the entire IT infrastructure as offered by some manu-
facturers requires a great deal of experience and thus are only possible for large-scale
companies on account of its complexity.
Planning and monitoring storage installations should be set up systematically right
from the beginning and includes a constant list of all the phases, stages and modifica-
tions. This is necessary as otherwise chaos would occur when staff change their jobs.
The IT Infrastructure Library (ITIL) with its standard requirements and procedures is an
important aid in documenting such processes. Such standards are also of assistance in
the discussions between manufacturers and customers or during company mergers
50 Section 6

when different IT worlds have to be united. ITIL has issued a series of publications which
are aimed at helping a company to structure its IT and business processes.
For example, Change Management describes the procedure for the continuous
management of modifications which replaces simple but irregularly updated Excel
tables and which is based on ITIL and the management tools supplied by the respective
manufacturer. It is thus possible to avoid – right from the start – any storage wilderness
with all types of products that are only flimsily interconnected. Such a lack of clarity
results in errors and failures and thus extra overnight work for employees. However, if
standard solutions or “Storage out of the Box“ are used, the overall management is
much simpler due to the use of software that is already suitable.
Medium-sized companies have fewer financial resources which can result in them
using quicker yet not fully tested implementations for long periods, longer than is
financially viable. It is not a coincidence that it is these companies which choose tech-
nologies that they know or at least seem to know: DAS is well known today with SMEs,
and iSCSI is widespread here thanks to its close relationship with LAN architecture. As
long as these are well-tested and proven solutions, then such cautious behavior is cer-
tainly not wrong; it just makes the company less flexible than their larger competitors.
The latter can afford to have well-trained employees and longer planning phases which
enables them to text new and more effective technologies for a longer period and then
apply them productively. This particularly applies to information management which
requires both investment and know-how. Medium-sized customers can usually not
afford most of the tools used in this sector. Their purchase would also hardly be sensible
when compared to the data quantities that have to be managed.
Data management is thus frequently without good planning and basically chaotic.
As a range of Windows licenses already exists together with unused servers, non-struc-
tured data, such as Office documents tends to be saved in a non-systematic manner –
with all the “ugly” consequences regarding classification, indices and search options for
such documents. Larger companies are one step further and provide dedicated servers
in a SAN. As a SAN has already been configured, it probably also has data which could
be saved more cheaply on other storage levels with somewhat less performance – that
would be chaos at a higher level. When the IT department realizes the mix-up, it usually
adds its own negative element, i.e additional NAS filers are now created so that the
non-structured data can be saved “suitably”. And then the question arises at some time
about merging or integrating the various stand-alone storage solutions where some-
times the block level (SAN) and sometimes the file level (NAS) is the dominating ele-
ment. They also have their own solutions which in turn require more investment and
basically add yet another complexity level to the overall storage architecture.
Storage Management – making complex storage networks manageable 51

Storage Management

(1) Storage Resource Management: “SRM (Storage Resource Management) initia-


tives began in earnest in the late 1990’s. This was a Unix, Windows and later a Linux
market which held great promise 2-4 years ago but has faded in recent years. There
were several reasons why the 20+ SRM companies faded and lost momentum:
1. SRM products had a hard time moving from a reactive, reporting tool to a proac-
tive tool that could make decisions and take actions based on user-defined
policies,
2. SRM products were mainly homogeneous, thus failing to provide support for
heterogeneous environments, and
3. SRM products only dealt with disk space allocation and lacked any insight into
disk performance issues. SRM users were worn down with all the alerts and
decisions that they had to perform manually.

Today’s reality is that organizations will need to integrate a variety of vendor and
homegrown tools. Storage organizations must accept that the structure of storage
is going to be split up by vendor and type of array and that organizationally, mini-
mizing the number of vendors and storage pools is one way to reduce storage
administration overheads. “
(Fred Moore, Horison Information Strategies)

(2) Information Management


“(Information Lifecycle Management) The discipline and function of oversight and
control of information resources.
Information management services: The processes associated with managing infor-
mation as it progresses through various lifecycle states associated with a Business
Process. These services exploit information about data content and relationships in
making decisions. Examples include records management and content management
applications.” (SNIA Dictionary)

The consequence is that, due to such structures, storage management becomes dif-
ficult as problems, failures, hardware and software faults cannot be immediately identi-
fied. Counter-measures are usually taken too late in such situations. However, monitor-
ing, reporting and constant error analysis could be implemented with software support
whereby such tools are usually supplied by the manufacturers. If many different com-
ponents are used, storage management is often faced with the problem of controlling
such heterogeneity.
52 Section 6

IDC’s Storage Taxonomy

Storage

Storage Storage Storage


Hardware Software Services

Data Other
Storage Storage Storage File
Storage Protection Storage
Mechanisms Management Replication System
Systems and SW
Recovery

Disk Systems HDDs Storage


Archive and Consulting
Infra-
Infra-
HSM
structure
Tape Libraries Tape Drives Implementation

Optical Storage
Optical Drives
Jukeboxes Device Management

Storage Removable Management


Infrastructure Media Support

Source: IDC

According to IDC, storage can be devided into three main groups, which various fields are assigned
to. Storage Management is only one of many other areas.

Management elements are sufficient supply of storage for users, departments and
applications as well as provisioning. The latter is understood as providing and if neces-
sary procuring on time the storage required: sufficient storage arrays and hard disks
must be available to meet all these requirements. If, for example, seasonal peaks, such
as Christmas business, must be taken into account, this would mean low system usage
during the other periods which could be 20 to 30% below actual capacity. The techno-
logy known as Thin Provisioning has been developed to counteract this problem. These
are procedures which plan the flexible, changing assignment of storage for various
applications: capacities are assigned and removed depending on capacities [1].
“Stranded Storage” is storage which was originally provided for specific applications or
users but not used after all. The objective of Thin Provisioning is basically to use
“stranded storage capacities” again. Thin Provisioning also uses virtualization [2] via
which different physical hard disks and disk arrays are combined to logical or virtual
units. This enables a very exact assignment of storage space and you no longer have to
consider the maximum and minimum physical disk limits.
The IT department should remain independent in order to free itself of any blame
and be able to refer to the other manufacturer who has forced the user to take on a
Storage Management – making complex storage networks manageable 53

specific SAN, NAS or iSCSI infrastructure. A Fujitsu Siemens customer has gone his own
way and has deliberately selected and configured separate stand-alone solutions for
different architectures: a SAN for mainframes including mirroring for a second (backup)
data center, a virtualization solution for SQL and mail server based on Windows and
Linux and finally a NAS for FlexFrame for SAP with Fujitsu Siemens servers and NetApp
storage. All three sectors are separately managed whereby the required extra effort
guarantees that other areas are not automatically touched in an error situation. In
other words: the often vilified silo solutions can certainly make sense in certain cus-
tomer situations just like DAS systems.
This approach can also be seen in the use of software: it is often the case that too
many tools from various sources are used instead of relying on a single provider.
Whereas storage systems such as SAN, NAS or IP-SAN can be configured by the sup-
plied software, a central tool should be used for the subsequent administration which
thus enables end-to-end management. Almost all the storage manufacturers today
offer the appropriate programs.

Heterogeneous storage networks –


standardized management

I n a SAN (Storage Area Network) on Fibre Channel basis (FC) the main components in
the network are FC switches whose protocol services are easy to configure and only
have a few error sources. The second part of the SAN is the storage systems which are
usually monitored and controlled via the supplied administration tools. However, the
problem in many SAN environments is the know-how and experience of the network
administrators, who come from the classic Local Area Network (LAN) sector and do not
know enough about data storage topics.
The alternative network NAS (Network Attached Storage) was developed with the
aim of offering classic file servers and storage systems in one package. A management
software is also integrated which, above all, handles the transport of files from the
servers to the storage systems. Such an environment frequently has less value placed
on the performance of data access and the hard disks involved. If the NAS servers are
connected to powerful FC storage systems in a SAN, the amount of administration
effort rises.
The third version is IP-SANs which today are usually used as iSCSI storage networks.
Such networks which are based on known LAN technology, are – after a while – more
popular especially in medium-sized companies, but also in international companies
with more than 1,000 employees. Of course, IT departments have a wealth of experi-
ence due to their many years of work with TCP/IP networks that are in use everywhere,
54 Section 6

but it would be a mistake to believe that storage data can simply be moved around in
addition to the existing network paths. A separate IP infrastructure with the mounted
storage systems must be set up in order to have a working IP-SAN [3].
Storage management itself is above these transfer techniques and consists, first of
all, of the tools supplied by the manufacturer, also known as element manager. A web
server is installed for easier operation so that the administrators can access a browser.
More complex networks are monitored and controlled via monitoring services which
connect many storage components such as HBAs (Host Bus Adapter), switches and
storage systems via interfaces. Some examples are Control Center from EMC, BrightStor
SRM from CA or Sanpoint Control from Symantec.
In the past it was frequently the case that storage equipment from the various manu-
facturers did not “understand” each other in the network as they either do not have
suitable interfaces (APIs = Application Programming Interface) or they were not com-
patible with those from other manufacturers. In order to make storage networks user-
friendly for the companies, the SNIA (Storage Networking Industry Association which
includes almost all the storage manufacturers) organized a so-called management
initiative in order to create a standard for all devices. Many years of work were spent
by the SNIA boards before they submitted a standardization proposal known as the
Storage Management Initiative Specification (SMI-S). The manufacturers now work
much closer together, swap information about their interfaces and ensure mutual
licensing. The communication interfaces covered by the SMI-S now often have priority
over the proprietary APIs.
According to Frank Bunn, SNIA Europe and storage specialist at Symantec, SMI-S has
an important long-term influence on storage management: “SMI-S is not an overnight
solution. It is a constant process which began as early as 2002. SMI-S creates better
interoperability. It enables the standardized management of different products and
provides a consistent view of the SAN and NAS environment. Users are often thus very
enthusiastic as they can finally see their entire storage environment reflected in SMI-S.
Customers often do not even know which storage equipment they have, let alone the
quantity. But that is just the first step. The second step can greatly facilitate the man-
agement of storage systems.” [4]
As far as Bunn was concerned, the subject of SAN was previously predominantly
controlled by the larger companies. Small and average companies were more skeptical
according to the motto “too complex, too expensive and doesn’t work anyway”. Bunn:
“And they were right more often than not. However, SMI-S does make SAN manage-
ment much easier. Partners can thus implement and support storage networks who are
not complete SAN specialists.” The SMI-S versions 1.1 and 1.2 take Fibre Channel SAN as
well as NAS and iSCSI into consideration which greatly expands the environment for
integrators.
Storage Management – making complex storage networks manageable 55

However, the standardization process has not been completed despite the many
years of effort. Not every manufacturer is implementing the adapted interfaces in their
devices which would be compatible to the other providers. The mutual licensing process
often takes much longer than is really necessary as far as the topic itself is concerned.
Furthermore, various SMI-S versions are in use. These circumstances have resulted in
user acceptance not being particularly high.
Section 7

Virtualization – some catching up


is necessary regarding storage topics
Everyone worldwide is talking about virtualization. This technology is meanwhile
said to be truly “fantastic“ and the corresponding product announcements in
this sector are immense. However, many users are still wary regarding the use of
virtualization in their own company. IDC market researchers determined in a
study carried out in 2008 that only 33% of companies in Western Europe have
introduced server virtualization from VMware or Citrix XenSource and 67% have
done nothing yet in this sector. These percentages are even lower in storage; only
15 to 20% have already looked at virtualization. The utilization of storage sys-
tems is normally very low. The degree of utilization can be increased very quickly
with virtual systems where several devices or hundreds of hard disks are linked to
logical units. The potential is enormous and a great deal of money could be
saved.

A
ccording to IDC, companies are holding back from using virtualization in their
storage environment as they do not regard this as absolutely essential. Storage is
often regarded as unavoidable for data storage, saving, backup or archiving
which does not contribute in any way towards increasing productivity or improving
operational procedures. Yet the experience gathered as a result of successful server
virtualization can also apply for storage. Many users follow exactly this procedure in
practice: they expand virtualization to other sectors on a step-by-step basis.
Virtual storage space can help companies to use existing storage resources effi-
ciently as well as centralize and simplify administration. It was the x86 sector which
drove virtualization forwards as the average system utilization was very bad. To Dr.
Joseph Reger, CTO (Chief Technology Officer) at Fujitsu Siemens Computers, this is all
very obvious: “Approximately 10% utilization is simply a bad value compared to existing
systems such as mainframes or Unix platforms.” [1] Of all the options available to
improve the situation it is hardware virtualization which is the best: “The reason is
because the superior layers no longer need to take care of this,” says Reger. By “superior
58 Section 7

layers” we mean everything beyond the hardware layer, i.e. operating systems and
applications that also benefit from virtualization.
Reger continues his explanation: “By simulating various machines – pieces of hard-
ware that do not even exist – the result is a peculiar layer of transparency. With this
technology operating systems and applications need not even know that they are run-
ning in a virtualized environment. Consequently, the average degree of utilization was
dramatically increased and a great deal of money was saved.” [2]
According to Reger there are in principle three large areas, in which virtualization
technologies can be applied today: hardware virtualization, operating system virtuali-
zation and application virtualization: “With the first group it is a matter of pretending
as if we have more hardware than is actually available. This applies to server and stor-
age in the same way. You simulate virtual levels of hardware that do not even exist.
Physically, the hardware exists once only; virtually, however, more is made of it. This
means that the operating system does not know it is running on virtual instances –
from the view of the operating system there really are ten different servers or storage
arrays. If the operating system is virtualized, this means that the application thinks
there are several instances of the operating system, whereas in actual fact only one is
running.” [3] By means of thin provisioning it is for example possible to have more
logical address space available in a storage array than is physically available.
Storage virtualization is in part misused as a term by the IT industry. It originally
meant the mapping of storage resources for the servers or applications involved, i.e. the
consolidation or regrouping of physical storage units to form logical ones. Today it is
usually used for the allocation of random storage resources, including data replication
mechanisms. This covers terms and technologies, such as volume management, virtual
disk volumes, file systems, virtual tape, virtual ports or virtual SANs: something they all
have in common is the approach of separating the physical view from the logical one,
i.e. splitting physical memory into partitions or consolidating several physical hard disks
into one or even various logical units.
When dealing with virtualization within the storage network (SAN device virtualiza-
tion) itself, i.e. about virtualization at switch infrastructure level, a distinction should be
made between three different approaches:
1) With so-called in-band virtualization the control entity for the data connection,
meta data and data transport itself is on the same appliance. The scaling of these
solutions is determined by the transport performance of this appliance. The provid-
ers FalconStor and DataCore were among the first manufacturers to have offered
such solutions.
2) With out-of-band virtualization a single appliance takes care of only the meta data
and of controlling the data path, while the host and server respectively organize the
transport of the storage data to and from the storage devices.
Virtualization – some catching up is necessary regarding storage topics 59

3) The third approach consists in the separation of the control entity and data path,
which is done by an intelligent network device. This technology is known as “Split
Path Architecture for Intelligent Devices” (SPAID). Switch manufacturers like Bro-
cade and Cisco provide suitable devices for this purpose. The separation of the
instances as performed here results in increased data transport speed and enables
scaling of the concept.

Such virtualization solutions normally have two goals. The first goal is to remove the
constraints of a storage array and/or a manufacturer. The second goal is to provide
manufacturer-independent data services, such as pool building (consolidation of stor-
age capacities of the same service quality), data replication such as data snapshots,
remote mirroring or disaster recovery.
Something all these virtualization solutions have in common is that they permit a
coordinated selection of the optimal storage arrays for a certain task. In this way, stor-
age resources can be made available at will and dynamically changed independent of
the storage array. Only the basic configuration of the elements to be virtualized is still
performed by the proprietary management applications. Therefore, the users have to
use this element manager together with the selected virtualization software.
The majority of the users of virtualization solutions use them in the sense of improved
storage management. Virtualization increases the freedom of choice of the users. You
can use solutions from several storage array manufacturers together in a pool. Storage
systems are frequently already completely partitioned during first installation and
afterwards are only managed via virtualization: “The approach has proved itself to be
successful for years, particularly for very dynamic environments, for example with serv-
ice providers or users with a large number of small application islands, such as in public
administration. Wherever high scaling is required for thousands of LUNS and hundreds
of servers the use of split-path technology is absolutely necessary.” [4]
In contrast to server virtualization, storage virtualization has not yet made the
breakthrough. Neither is a market standard becoming apparent. This is certainly also
due to the fact that with the LUN concept every SAN storage array already has rudi-
mentary hard disk virtualization. However, importance is increasingly attached to online
storage virtualization in virtualized server environments [5].
It is a different story entirely with file-based storage systems, the NAS systems. There
are a number of very promising approaches for file virtualization, but a market standard
has not established itself here, either [6].
The most progress can be seen in storage virtualization with magnetic tapes. When
its comes to backup on tape, Virtual Tape Libraries (VTL) are at present best practice.
CentricStor VT is currently the leading virtual tape product in data centers.
Section 8

The storage strategy of Fujitsu Siemens


Computers and its partners
Fujitsu Siemens Computers is Europe’s largest storage provider and offers its cus-
tomers everything from one source – all that they need for their storage infra-
structure. In addition to a wide range of products from the leading manufactur-
ers, such as EMC, Brocade, NetApp, CA, Sun and Symantec, Fujitsu Siemens
Computers has developed its own range of storage products designed to meet
the particular requirements of their users. The offer also includes a range of
sophisticated services – from maintenance to Managed Storage.

U
ntil recently every application had its fixed allocated infrastructure, a collection
of servers, network components and storage systems. But now the aim is to
allocate only those infrastructure resources that the applications actually
require. This objective is achieved via virtualization: Large resource pools exist which
can then be used dynamically depending on the requirements involved. Fujitsu Siemens
Computers is moving in this direction as part of its Dynamic Infrastructure strategy of
which storage is an integral part.

CentricStor VT: intelligent tape storage


virtualization

O ne example for this process is CentricStor VT. CentricStor VT (Vir-


tual Tape) enables intelligent backup of all company data. The
dual target storage solution integrates the advantages of hard disk
and magnetic tape as the backup data can be stored on a wide range
of storage media based on rules and autonomously. Working with
tape is thus fit for ILM (Information Lifecycle Management), and
corporate data backup can be based in a flexible manner on different
Service Level Agreements (SLAs). True Tape Virtualization (TTV) from
62 Section 8

CentricStor VT represents above-average connection opportunities and also enables a


clear reduction in costs as it consolidates all the backup storage media used. The Cen-
tricStor Grid Architecture (CGA) has excellent scaling features and offers far-reaching
reliability for each data center.
The report “The Forrester Wave: Enterprise Open Systems Virtual Tape Libraries” pub-
lished on 31st January 2008 by Forrester Research looked closely at 12 providers based
on 58 criteria. The result is that Fujitsu Siemens Computers was found to be one of the
market leaders for Virtual Tape Libraries. The Forrester Report awards top marks to the
company for its strategy in the Virtual Tape Appliances sector: “Fujitsu Siemens Com-
puters is the leader in host support, architecture and tape integration”. And continues:
“As the leading system provider in Europe the company has a solid basis of more than
300 customers and over 500 installed VTLs whereby most of the installations are with
major companies. Fujitsu Siemens Computers has solid growth regarding new custom-
ers and has one of the highest growth values for existing customers.”
Version 4 of CentricStor was specifically developed in order to solve the following
problems for data backup supervisors
• data quantities are continuously increasing.
• archiving periods for data are longer.
• a second data center must be provided in case of a possible catastrophe.

The positive assessment by Forrester is also confirmed by a study carried out by the
ex-Gartner analyst Josh Krischer: ”In 1999 Fujitsu Siemens Computers launched its
CentricStor virtual tape solution, which became the most open, flexible virtual tape
library supporting most existing mainframe and major Unix and Windows operating
systems and major tape libraries. CentricStor was conceptually designed as an appli-
ance to be used with all types of tape software, servers, tape libraries and tape technol-
ogy as well. Initially it was released by Fujitsu Siemens Computers to support the
BS2000/OSD, MVS mainframe platforms and ADIC tape libraries; however, over the
years with constant certifications for major operating systems, backup applications,
tape library systems and tape drive technologies it evolved into the most versatile VTL
appliance, which can be deployed in almost any environment while protecting previous
investments in tape automation.“ [1]
The storage strategy of Fujitsu Siemens Computers and its partners 63

CentricStor FS: consolidated File Services

C entricStor FS is an innovative, scale-out storage system which


offers file services at high speed and is very reliable. It offers
practically unlimited scaling regarding storage capacity, data
throughput and access times.
The administration of CentricStor FS is simple – only one single
file system must be managed irrespective of the size. The storage
resources can thus be used more efficiently. Additional storage
capacity can be easily provided and CentricStor FS can also be
installed very easily. If a component fails, automatic failover and
fallback ensures that the system keeps running.

Terri McClure, analyst working for the Enterprise Strategy Group (ESG), con-
cluded the following in a report:
“The recently launched CentricStor FS from Fujitsu Siemens Computers is a high-end
storage system on file basis via which scaling is possible at a very fine level, thus
fulfilling capacity, availability and performance requirements as defined by file server
consolidation initiatives and Web 2.0 applications. The use of standard components
and the excellent cluster features make CentricStor FS a scalable easy-to-manage
file storage solution with a low starting price that has been specially designed for
the real world of increased file quantities.“ [2]

FibreCAT SX series: fast storage systems for the data center

T he FibreCAT SX40, SX60, SX80, SX80


iSCSI, SX88 and SX100 storage sys-
tems offer quality and reliability as
required in data centers. FibreCAT SX
systems are very fast storage systems,
in particular, the FibreCAT SX88, which is up to 50% faster than the FibreCAT SX80, and
the FibreCAT SX100, which has twice the speed and performance of the FibreCAT SX80.
This is due to Fibre Channel technology with 4 Gigabit per second, FibreCache and
state-of-the-art RAID controller technology. Snapshots are integrated as a standard.
The revolutionary FibreCap technology protects the system in case of power failure.
64 Section 8

FibreCAT SX systems are user-friendly and easy to put into operation. The adminis-
tration is not complicated thanks to the intuitive web interface. The systems are suitable
for a wide range of applications.
The analyst Hamish Macarthur from Macarthur Stroud International said the follow-
ing regarding the FibreCAT SX series: “Managing and protecting the information assets
of an organization is critical in today’s markets. The systems in which the data is resi-
dent must be secure, reliable and easy to manage. The FibreCAT SX range supports reli-
able primary storage as well as the need for faster backup and recovery. The new arrays,
with management tools included, will be a sound investment to meet the business and
compliance requirements of small, medium and large organizations.”

FibreCAT NX40 S4: reliable file and print server and


SAN/NAS gateway for medium-sized customers

T he FibreCAT NX40 S4 is a file


and print server and a SAN/
NAS gateway for medium-sized
customers. The system‘s quality
and reliability provides a wide
range of application usages in
departments and branch offices.
The FibreCAT NX40 S4 is a rack model with only 2 height units (2 HE). The system can
be equipped with SATA-II or powerful SAS hard disks and provides, when fully config-
ured, internal storage capacity of 4.5 terabytes. Additional interfaces (SAS, Fibre Chan-
nel) increase the system capacity via external storage systems and it then operates as
SAN/NAS gateway. Expansion with the FibreCAT SX/CX family is particularly simple as
the required software is already pre-installed. The FibreCAT NX40 S4 is also available as
an option with iSCSI functionality.
The system uses the operating system Windows Storage Server 2003 R2 with addi-
tional functions for the file and print server sector. The FibreCAT NX40 S4 systems are
pre-tested, pre-installed and pre-configured and thus ready for operation within min-
utes.
The storage strategy of Fujitsu Siemens Computers and its partners 65

FibreCAT TX series: LTO for medium-sized customers

T he FibreCAT TX series consists of the


entry model TX08 and the two models
TX24 S2 and TX48 S2 designed for
medium-sized customers. The FibreCAT
TX08 is a compact and reliable tape auto-
mation unit and enables a low-priced
entry into the FibreCAT TX family.
The FibreCAT TX08 is equipped with LTO technology which is synonymous for large
capacity, high speed and very low media costs. It can be equipped with a LTO-2, LTO-3
or LTO-4 drive which is a half height unit and provides storage capacity of 12.8 tera-
bytes (LTO-4 compressed). This corresponds to an automated data backup for approxi-
mately 2 weeks. The offer price includes a full version of the backup software ARCserve
Backup as well as one year‘s on-site service.
The second generation of the magnetic tape systems FibreCAT TX24 S2 and FibreCAT
TX48 S2 means that backup to tape, in comparison to hard disk, is usually more energy-
efficient as power is only used when data is read or written to tape. Both systems pro-
vide SAS and Fibre Channel interfaces to the server.

FibreCAT Solution Lab

F ujitsu Siemens Computers not only offers its own products but also has lots of
storage competence within Europe. In addition to the numerous storage experts
supporting projects on-site, the FibreCAT Solution Lab in Paderborn is a center of
expertise.
The main task of the FibreCAT Solution Lab is to constantly monitor quality with
regards to a reliable system and component availability. This also includes coordinating
the seamless integration of the various FibreCAT systems into the comprehensive solu-
tion portfolio offered by Fujitsu Siemens Computers. Other tasks include adhering to all
legal standards regarding environment protection and DIN ISO Standards and imple-
menting them on time. On request the FibreCAT Solution Lab can also carry out cus-
tomer-specific tests and create realistic application scenarios. Fujitsu Siemens custom-
ers can thus see the FibreCAT systems operating in live demonstrations.
66 Section 8

Managed Storage

G rowth in storage means above all an immense increase in data quantities which
must all be managed, saved, provided and stored. The demand for online storage
capacity increases as does the demand for backup storage volumes. Any limits to such
growth rates are not appearing on the horizon.
Against this backdrop SAP asked itself whether they wanted to continue managing
the required storage volumes themselves or to practice what their own hosting experts
recommend to their customers, namely outsource the work that is not part of their core
competence and concentrate on the important elements essential for their core busi-
ness.
SAP managers would thus place their operations and support regarding process-
supporting storage infrastructure into the skilled hands of external people. SAP has
found such a competent partner in Fujitsu Siemens Computers.
Fujitsu Siemens Computers took on the role of general contractor for SAP and the
entire responsibility for providing online storage capacity for data backup – 4 petabytes
monthly at the start and then moving up to more than 200 terabytes daily. Further-
more, appropriate reserve capacities were provided in order to meet any additional
requirements in time.
Fujitsu Siemens Computers thus supports one of the largest Managed Storage
projects in Europe and also manages the cooperation activities with the strategic part-
ners involved, namely EMC and NetApp which provide products in the SAN and NAS
environment and whose specialists are involved in the corresponding service sector.

Storage – one part of the whole picture

I t is not enough in the storage world to simply move a few levers to control this vast
flow of data. A paradigm change is required instead of individual steps. The job
involved is to connect up the various technologies into end-to-end solutions making
them easy to manage as well as operate. Storage is not a solitary topic but certainly an
important element in the overall dynamic data center. Fujitsu Siemens Computers has
thus developed a strategy known as Dynamic Infrastructure.
The storage solutions from Fujitsu Siemens Computers are a fixed part of this strat-
egy and are components used in implementing the objective of providing end-to-end IT
solutions that generate a maximum business contribution with minimum total costs.
FlexFrame for mySAP is one of the most successful and best known examples of such
solutions for a virtualized infrastructure consisting of servers and storage systems.
The storage strategy of Fujitsu Siemens Computers and its partners 67

Customers benefit even more from the fact that Fujitsu Siemens Computers work
together very closely with many partners on storage matters. One example of such
excellent cooperation can be seen in the quotes from the CEOs of our partners EMC and
NetApp:
Joe Tucci, EMC President & CEO: “The combination of EMC’s networked storage solu-
tions and the server-based solutions from Fujitsu Siemens Computers creates a wide-
ranging offer of end-to-end infrastructure solutions that meet the requirements of our
customers. EMC solutions play a central role in the Fujitsu Siemens Computers vision of
a Dynamic Data Center, and we will continue to concentrate our joint operations on
offering our customers the most comprehensive solution portfolio available on the
market.“
Dan Warmenhoven, CEO at NetApp: “The strategic partnership with Fujitsu Siemens
Computers has contributed a lot to our success and is still growing in the EMEA region.
Our jointly developed solution FlexFrame for mySAP Business Suite, the implementation
of a fast backup solution for Oracle and the Center of Excellence which has been set up
with Oracle are just some of the excellent examples that have resulted from our coop-
eration so far.“
Forecast

Future storage trends

C
ompanies are facing many challenges regarding storage systems. In particular,
the continuous increase in data results in demands for new ideas on how to make
storage more efficient. In contrast to PCs and servers where the main market
players define the standards, the storage market has innovative companies joining in
the fray which have unusual new ideas that move the market in new directions. Having
looked at the main storage technologies in the preceding sections, let us look at current
storage trends and see which ones have the potential to make a major breakthrough in
the way that a company stores its data. We have selected five relevant trends:

1. Fibre Channel over Ethernet (FCoE)


2. Massive Array of Idle Disks (MAID)
3. Flash-based Solid State Disks (SSDs)
4. Security (Encryption of Data at Rest)
5. New Storage Architectures (Grid Storage)

1. Fibre Channel over Ethernet (FCoE)

A lthough two of the three leading storage technologies (NAS, iSCSI) are based on
Ethernet, it is the third, SAN, which by far dominates the market. SANs are typically
implemented on a fibre channel basis. Fibre Channel networks require dedicated HBA
(host bus adapters, i.e. server plug-in cards) and switches. These network components
are clearly more expensive than Ethernet cards and switches and they also require spe-
cific fibre channel know-how. The question thus arises whether SAN storage networks
can also be set up on an Ethernet basis so that only one type of network is needed as
this would greatly reduce costs. Instead of 4 cables (2x Ethernet, 2x Fibre Channel) the
servers would only have 2 FCoE cables. iSCSI indeed started with this promise but fibre
channel SAN purists were never really convinced for the following reasons. In contrast
to FCoE where fibre channel tools can still be used, iSCSI changes these levels and Eth-
ernet is not deterministic. Leading network providers are now competing to merge stor-
age networks and server networks in new ways; the Ethernet standard is to be extended
so that it can provide the same quality of service as a fibre channel network. The result
is the so-called DCE (Data Center Ethernet) via which previous TCP/IP data traffic can be
70 Forecast

handled in exactly the same way as fibre channel (FCoE). In short: the 10Gb Ethernet is
available and 40Gb as well as 100Gb is on the horizon. However, there are plans to
extend fibre channel beyond 8Gb/s even up to 16Gb/s. A conversion to FCoE requires a
high amount of infrastructure investment which in due course will be balanced by sav-
ings in operating similar networks.

2. Massive Array of Idle Disks (MAID)

T he Enterprise hard disks used today run without any interruption on a 24x7 basis.
This not only wears out the mechanical parts but continually consumes energy. With
an alternative storage array on MAID basis only one small part of the hard disk is always
active. Hard disks not accessed for some period of time are switched off, but this is only
sensible if data really is not accessed for a longer period as frequently switching on and
off would use up even more power and negatively affect the disk lifespan. MAID is thus
mainly used in the important archive and backup sector where the price per saved tera-
byte is important. SATA disks are thus typically used in MAID systems. MAID is a major
step forward in significantly saving energy and towards Green IT. However, there are a
few obstacles to be surmounted. The disks should be periodically powered up and
checked to make sure that the data can still be read. You can also reduce their rotational
speed instead of switching them off completely which will significantly save power and
reduce the danger of unknown disk errors.

3. Flash-based Solid State Disks (SSDs)

F lash memory based SSDs, based on price per capacity, are more expensive than con-
ventional hard disks but, thanks to their much faster read speed and faster access
speed, SSDs that are optimized for enterprise use are already better than today’s hard
disks with regard to price per IOPS. Furthermore, they need less energy, hardly generate
any heat and have no problems regarding rotational vibration as there are no moving
parts. As the price of flash chips has fallen by an annual average of 50% in recent years,
such disks are of great interest for a variety of situations. A large number of hard disks
are used today to achieve similar speeds. The speed comes from the parallel access on
many spindles. For example, such scenarios which have only 2 SSD (two in order to
ensure data redundancy via mirroring) could be sued to replace a dozen hard disks in
RAID 5 or RAID 6 arrays. This results in a great deal of energy savings.
Future storage trends 71

4. Security (Encryption of Data at Rest)

D ata management and data storage has always gone hand-in-hand with security:
there would be serious consequences if sensitive data fell into the wrong hands.
Encryption helps to avoid misuse: the data cannot be read nor used by anyone who is
not authorized to do so. This also applies if data media is faulty or supposedly faulty
(e.g. hard disks) and has to be replaced. All data on any faulty data media must be made
non-readable. Unfortunately, there are many ways to approach encryption. It can be
carried out at various levels; at application, file system, network and switch level as well
as in devices, such as one’s own encryption appliances or tape libraries and disk drives.
Key management for an entire data center is currently a problem as most manufactur-
ers offer their own key management. Standardization does not yet exist. And without a
specific standard there is no guarantee that data users can also read the data and
exchange it as required without any problems.

5. New Storage Architectures (Grid Storage)

T oday’s storage systems are usually built up in a monolithic manner and were spe-
cially designed as a storage system. As fewer storage systems are produced than PCs
or standard servers, their production is thus expensive. And as with all monolithic sys-
tems, the maximum configuration is limited for design reasons.
A paradigm change will revolutionize the storage industry: let’s take standard serv-
ers and the large number of manufactured simple standard RAID systems and turn
them into one storage system thanks to an intelligent software layer. Such scale-out
systems can be greatly extended by just adding more and more standard components.
Growth will have no limits and the systems will offer more capacity and more access
speed as they increase in size. This architecture gives systems the scaling and flexibility
urgently required in today’s file-driven world and is also the backbone for Internet
infrastructure (Cloud Computing). CentricStor FS is one of the first incarnations of this
new type of storage generation.
Remarks

Section 1
[1] The triumph of large numbers; 10 years Google; Neue Züricher Zeitung, April 25, 2008.
[2] Details about typeface development from Charles Panati, The Browser’s Book of Beginnings
– Origins of Everything Under, and Including, the Sun, New York 1998, Page 67 ff.
[3] “Der Spiegel”, edition dated 11. 8. 2008, cover feature “Addicted to data“, Page 88.
[4] Many authors have written about the positive and negative effects of this development. For
example, Neil Postman, Joseph Weizenbaum, Nicholas Negroponte or Nicholas Carr. See the
interview in Hendrik Leitner / Hartmut Wiehr, Die andere Seite der IT – Business-Transforma-
tion durch Services und dynamische Infrastruktur, Munich 2006, Seite 215 ff.
[5] Cf. The following passage from an interview in the German weekly “Die Zeit“.
Question: “Which development has changed the most the way we handle knowledge in
recent years?”
Answer: “Two”. Firstly: hard disks now hardly cost anything. It is now no longer an utopian
idea to have man’s entire published works on disk. Secondly, nobody in the world is more
than a day’s walk from an Internet café. We now have the communication infrastructure to
provide the world’s great libraries to youngsters in Uganda or the poorer areas in the USA or
Germany.“ (Interview with Brewster Kahle, Director of the Internet Archive in San Francisco;
Die Zeit, 17. 1. 2008)

Section 2
[1] The fact that major storage manufacturers have been buying up companies, such as Cognos,
Business Objects, Hyperion, Documentum or FileNet, that developed software for document
management (DMS) or business intelligence (BI) proves that storage hardware and the crite-
ria for stored data are merging. It can also be seen as an attempt to integrate classic storage
equipment with ILM or HSM.
[2] Steve Duplessie, File Infrastructure Requirements in the Internet Computing Era, Enterprise
Strategy Group (ESG), July 2008, Page 5.
[3] Fred Moore / Horison Information Strategies, Storage Spectrum (2007), Page 76.
[4] “Every fourth user has implemented ILM at a certain point or in a certain area but only
3.5% use such solutions throughout the company. Users, who are seriously looking at such
a topic, must realize that ILM cannot be simply bought as a single product nor implement-
ed via a one-off project“. Wolfram Funk, Experton Group (quoted in the German magazine
“Computerwoche” 46/2006, Page 40)
[5] See also the interview with Michael Peterson in http://www.searchstorage.de/topics/
archiving/e-mail/articles/107136/.
[6] Cf. for example Dan Warmenhoven, NetApp CEO: “My view (regarding ILM) is that a work-
intensive data management process should not be set up if there is an automated one. And
74 Remarks

ILM is very work-intensive. (…). The user wants to migrate his data from expensive storage
forms to cheaper ones. And he requires an online archive in order to retrieve the data quickly
and also for compliance reasons. NetApp talks of data archiving and migration. Calling all
that ILM is confusing.“ (Interview in “Computerwoche”, 9. 11. 2007).

Section 3
[1] Cf. Hartmut Wiehr, “Disk systems – technology and products, in iX extra, edition 12/2007
(free download: http://www.heise.de/ix/extra/2007/ie0712.pdf).
[2] Cf. the forecast in this book, page 69.
[3] There is an early warning system in professional disk systems. Disks report during running
operations as to whether they are going to fail. Drives, which have not been in use for a
while, are addressed periodically to make sure that they still work. Do they react immediately
to a signal or only with the retries? The latter would suggest that the magnetic interface is
no longer 100%. All the corresponding information is collected and one of the spare disks is
triggered by the system if certain threshold values are exceeded. Or else a message is sent to
the service team.
[4] ICP Vortex has provided a White Paper about the various RAID-levels. The document is in
German and can be downloaded from: http://vortex.de/NR/rdonlyres/82BA6504-885D-
444E-AC71-7AC570CF56A3/0/raid_d.pdf.
[5] See Günter Thome / Wolfgang Sollbach, Fundamentals of Information Lifecycle Manage-
ment, Berlin Heidelberg 2007, page 214 ff.

Section 4
[1] For the history of the Ethernet protocol and the Internet, which were developed on behalf
of the US Army and were to ensure the transportation of messages in the event of war see
http://www.informatik.uni-bremen.de/grp/unitel/referat/timeline/timeline-2.html.
[2] For more information about SANs and Fibre Channel see http://www.snia.org/education/
tutorials/ and http://www.fibrechannel.org/.
[3] Seen from a historical viewpoint, CNT first took over its competitor Inrange, but was then
itself purchased by McData shortly afterwards. Brocade now unites all the former FC switch
providers against Cisco, the company with the largest market force in the network sector.
Seen globally, Brocade is still the market leader for FC switches, whereas Cisco has continu-
ously extended its market share, particularly in USA.
[4] Good interpretations of the development of the IT industry are to be found in Paul E. Ceruz-
zi’s, A History of Modern Computing, Cambridge and London 2003, and in Clayton M. Chris-
tensen’s, The Innovator’s Dilemma, New York 2000.
[5] Cf. Mario Vosschmidt/Hartmut Wiehr, Storage networks and their instruments of adminis-
tration; in: iX, vol. 8/2008, page 122 ff.
[6] A good overview is to be found in the brochure “NetApp Products & Solutions”.
Remarks 75

[7] Interview with David Hitz in “project 57 – Journal for Business Computing and Technology”,
Special 01/05, page 39 ff. Cf. also the interview with NetApp CEO Dan Warmenhoven in
“Computerzeitung”, issue 38/2008, page 10.

Section 5
[1] Many storage manufacturers supply their own backup programs together with their hard-
ware. EMC, a partner of Fujitsu Siemens Computers, has after the takeover of Legato in
2003 a complete solution (NetWorker) in its portfolio. Of the independent providers CA with
ARCserve and Symantec with NetBackup are particularly worth mentioning. CommVault and
BakBone can frequently be found in the Windows environment.
[2] See in this respect Dave Russell / Carolyn DiCenzo, MarketScope for Enterprise Backup/Re-
covery Software 2008, April 2008.
[3] An example of error corrections is ECC (Error Correction Code). The correction is used to
detect errors during the storage and transfer of data. The errors are then automatically
remedied in the second step.
[4] Hard disks with S.M.A.R.T. (Self-Monitoring, Analysis and Reporting Technology) have a
function that constantly monitors them, with technical values, such as temperature, startup
time or track reliability of the read and write heads being controlled.
[5] Specialist service providers, such as Kroll Ontrack or Seagate Services, can often recover hard
disks damaged by fire and water and thus save the lost data.
[6] See Hartmut Wiehr, Begrenzte Lebensdauer (= Limited life-span): Disks – Status Quo and
Trends; in iX extra, issue 12/2007; download: http://www.heise.de/ix/extra/2007/ie0712.pdf.
Other terms and contexts relating to this topic are explained there.
[7] For more details see www.speicherguide.de: Focus on tape drives and tapes (http://www.
speicherguide.de/magazin/bandlaufwerke.asp?mtyp=&lv=200). The durability of DVDs and
Blu Ray Disc (BD) is subject to very large fluctuations. The market is wide-spread, particularly
in the consumer segment. A good overview of the durability of these optical media with due
regard to their suitability for archiving purposes can be found in “c’t – Magazin für Compu-
tertechnik”, issue 16/2008, page 116 ff.
[8] As early as 1987 StorageTek presented the 4400 Automated Cartridge System (ACS), which
was the basis for the 4400 Tape Libraries (Nearline) which were introduced in 1988. The large
PowderHorn Libraries introduced in 1993 were a further development of this technology,
with which very fast access to tape cartridges is possible. A great many PowderHorns are
still in use today, and after StorageTek was taken over by Sun in 2005 the new owner had
to repeatedly extend the maintenance cycles on account of the pressure exerted by major
customers. These customers saw no reason to phase out their tried-and-trusted tape librar-
ies and replace them with follow-up models.
[9] Cf. “The Top 10 Storage Inventions of All Time”, in Byteandswitch, June 16, 2008.
[10] Hartmut Wiehr, Dedup turns backup inside out, in “Computerzeitung”, issue 28/2008,
page 10.
76 Remarks

Section 6
[1] Fujitsu Siemens Computers has developed two dynamic infrastructure solutions FlexFrame
for SAP and FlexFrame for Oracle which combine storage, server and network resources on a
single platform. Resources can thus be assigned and moved in run mode in a virtual environ-
ment and depending on current requirements.
[2] See section 7 for more about virtualization.
[3] See Mario Vosschmidt/Hartmut Wiehr, Gut eingebunden – Speichernetze und ihre Verwal-
tungsinstrumente (Well incorporated – Storage networks and their administration tools), in:
iX – Magazin für professionelle Informationstechnik, Heft 8/2008, Page 123.
[4] Published in tecchannel, 26th January 2006. For more information, see the article “SMI-S
is holding the storage networking industry together“ (Mit SMI-S hat sich die Storage-Net-
working-Industrie ihren inneren Zusammenhalt gegeben) www.searchstorage.de (Septem-
ber 5, 2007).

Section 7
[1] “Virtualization drives IT industrialization forward”: Interview by Hartmut Wiehr with Dr. Jo-
seph Reger, at www.searchstorage.de, April 10, 2004: http://www.searchstorage.de/topics/
rz-techniken/virtuelle-systeme/articles/117015/.
[2] Ibid.
[3] Ibid.
[4] Mario Vosschmidt/Hartmut Wiehr, Gut eingebunden – Speichernetze and ihre Verwal-
tungsinstrumente (Well incorporated – Storage networks and their administration tools), in:
“iX – Magazin for professionelle Informationstechnik”, issue 8/2008, page 124.
[5] LUN masking (also LUN mapping) means that only the storage area it needs to perform its
work is allocated and made visible to an application. As a result of this segmentation general
access to certain storage areas is prohibited, which at the same time increases the security
of all the applications. With SAN zoning the same principle is applied to the division of a
network into virtual subnetworks so that servers of the one zone cannot access storage
systems of another zone.
[6] Steve Duplessie, File Infrastructure Requirements in the Internet Computing Era, Enterprise
Strategy Group (ESG), July 2008.

Section 8
[1] Josh Krischer, Krischer & Associates, CentricStor Virtual Tape: the Swiss Army Knife for data
protection, September 2008
[2] Terri McClure, Enterprise Strategy Group (ESG), CentricStor FS von Fujitsu Siemens Comput-
ers, July 2008
Glossary
A C
Array (Storage Array) Compliance
A subsystem which houses a group of disks (or In data storage terminology, the word compli-
tapes), together controlled by software usually ance is used to refer to industry-wide govern-
housed within the subsystem. ment regulations and rules that cite how data
is managed and the need for organizations to
Asynchronous Replication be in compliance with those regulations. Com-
After data has been written to the primary pliance has become a major concern for
storage site, new writes to that site can be organizations and businesses, due largely in
accepted, without having to wait for the sec- part to increasing regulatory requirements
ondary (remote) storage site to also finish its which often require organizations to invest in
writes. Asynchronous Replication does not new technologies in order to address compli-
have the latency impact that synchronous rep- ance issues.
lication does, but has the disadvantage of
incurring data loss, should the primary site fail Continuous Data Protection (CDP)
before the data has been written to the sec- Refers to backing up computer data by saving
ondary site. as an automated function a copy every time
changes are made to that data. This event
B driven backup works during productive com-
puting time, and could slow down the per-
Backup/Restore formance of the IT infrastructure.
A two step process. Information is first copied
Cluster
to non-volatile disk or tape media. In the event
A group of servers that together act as a single
of computer problems (such as disk drive fail-
system, enabling load balancing and high
ures, power outages, or virus infection) result-
availability. Clustering can be housed in the
ing in data loss or damage to the original data,
same physical location (basic cluster) or can be
the copy is subsequently retrieved and restored
distributed across multiple sites for disaster
to a functional system. recovery.
Block Data
Raw data which does not have a file structure D
imposed on it. Database applications such as
Microsoft SQL Server and Microsoft Exchange Data Deduplication
Server transfer data in blocks. Block transfer is Deduplication technology segments the
the most efficient way to write to disk. incoming data stream, uniquely identifies
these data segments, and then compares them
Business Continuity to segments previously stored. If an incoming
The ability of an organization to continue to data segment is a duplicate of what has already
function even after a disastrous event, accom- been stored, the segment is not stored again
plished through the deployment of redundant but a reference is created for it (pointer). This
hardware and software, the use of fault toler- process operates at a very low level of granu-
ant systems, as well as a solid backup and larity or atomic level to identify as much
recovery strategy. redundancy as possible. The trade-offs in this
78 Glossary

filtering process pertain to efficiency, speed Fault–Tolerance


and data size. Fault–tolerance is the ability of computer
hardware or software to ensure data integrity
DAS (Direct Attached Storage) when hardware failures occur. Fault-tolerant
DAS is storage that is directly connected to a features appear in many server operating sys-
server by connectivity media such as parallel tems and include mirrored volumes, RAID–
SCSI cables. This direct connection provides volumes, and server clusters.
fast access to the data; however, storage is
only accessible from that server. DAS include File Data
the internally attached local disk drives or Data which has an associated file system.
externally attached RAID (redundant array of Fibre Channel (FC)
independent disks) or JBOD (just a bunch of A high–speed interconnect used in storage
disks). Although Fibre Channel can be used for area networks (SANs) to connect servers to
direct attached, it is more commonly used in shared storage. Fibre Channel components
storage area networks. include HBAs, hubs, switches, and cabling. The
term Fibre Channel also refers to the storage
Disaster Recovery
protocol.
The ability to recover from the loss of a com-
plete site, whether due to natural disaster or Fibre Channel over Ethernet (FCoE)
malicious intent. Disaster recovery strategies A technology that encapsulates Fibre Channel
include replication and backup/restore. frames in Ethernet frames, allowing FC traffic
to be transported over Ethernet networks.
E Standards are in work in different standardiza-
tion committees. Products are announced for
Ethernet 2009 or 2010. FCoE could be an alternative to
Local area network (LAN) topology commonly classical Fibre Channel technology.
operating at 10 megabits per second (mbps)
over various physical media such as coaxial G
cable, shielded or unshielded twisted pair, and
fiber optics. Future plans call for 1, 10 and 100 Global File System
gigabit Ethernet versions. Ethernet standards In some configurations, as with clusters or
multiple NAS boxes, it is useful to have a means
are maintained by the IEEE 802.3 committee.
to make the file systems on multiple servers or
devices look like a single file system. A global
F or dispersed file system would enable storage
Failover administrators to globally build or make
changes to file systems. To date this remains
In the event of a physical disruption to a net-
an emerging technology.
work component, data is immediately rerouted
to an alternate path so that services remain
uninterrupted. Failover applies both to cluster-
H
ing and to multiple paths to storage. In the High Availability
case of clustering, one or more services (such A continuously available computer system is
as Exchange) is moved over to a standby server characterized as having essentially no down-
in the event of a failure. In the case of multiple time in any given year. A system with 99.999%
paths to storage, a path failure results in data availability experiences only about five min-
being rerouted to a different physical connec- utes of downtime. In contrast, a high availabil-
tion to the storage. ity system is defined as having 99.9% uptime,
Glossary 79

which translates into a few hours of planned ices, and on the accommodation and environ-
or unplanned downtime per year. mental facilities needed to support IT. ITIL has
been developed in recognition of organisa-
HBA (Host Bus Adapter)
The HBA is the intelligent hardware residing on tions‘ growing dependency on IT and embodies
the host server which controls the transfer of best practices for IT Service Management. ITIL
data between the host and the target storage is often implemented when different enter-
device. prises work together, and can also facilitate
mergers and acquisitions.
Hierarchical Storage Management (HSM)
A data storage system that automatically J
moves data between high-cost and low-cost
storage media. HSM systems exist because JBOD (Just a Bunch of Disks)
high-speed storage devices, such as enterprise As the name suggests, a group of disks housed
disk drives (Fibre Channel, SAS), are more in its own box; JBOD differs from RAID in not
expensive (per byte stored) than slower devices, having any storage controller intelligence or
such as low-speed discs (SATA, desktop discs) data redundancy capabilities.
or magnetic tape drives. While it would be
ideal to have all data available on high-speed L
devices all the time, this is prohibitively expen-
sive for many organizations. Instead, HSM sys- LAN
tems store the bulk of the enterprise‘s data on Local Area Network. Hardware and software
slower devices, and then copy data to faster involved in connecting personal computers
disk drives when needed. and peripherals within close geographic con-
fines, usually within a building, or adjacent
I buildings.
ILM (Information Lifecycle Management) Load Balancing
The process of managing information growth, Referring to the ability to redistribute load
storage, and retrieval over time, based on its (read/write requests) to an alternate path
value to the organization. Sometimes referred between server and storage device, load bal-
to as data lifecycle management. ancing helps to maintain high performance
iSCSI (Internet SCSI) networking.
A protocol that enables transport of block data
LTO
over IP networks, without the need for a spe-
Linear Tape Open. The LTO family, a half-inch
cialized network infrastructure, such as Fibre
Channel. open technology with Ultrium format – a car-
tridge targeted at ultra-high capacity require-
ITIL (Information Technology Infrastructure ments.
Library)
ITIL refers to a documentation of best practice LUN (Logical Unit Number)
for IT Service Management. Used by many A logical unit is a conceptual division (a sub-
hundreds of organisations around the world, a unit) of a storage disk or a set of disks. Logical
whole ITIL philosophy has grown up around units can directly correspond to a volume drive
the guidance contained within the ITIL books (for example, C: can be a logical unit). Each
and the supporting professional qualification logical unit has an address, known as the logi-
scheme. ITIL consists of a series of books giving cal unit number (LUN), which allows it to be
guidance on the provision of quality IT serv- uniquely identified.
80 Glossary

LUN Masking Port


A method to restrict server access to storage The physical connection point on computers,
not specifically allocated to that server. LUN switches, storage arrays, etc, which is used to
masking is similar to zoning, but is imple- connect to other devices on a network. Ports
mented in the storage array, not the switch. on a Fibre Channel network are identified by
their Worldwide Port Name (WWPN) IDs; on
M iSCSI networks, ports are commonly given an
iSCSI name. Not to be confused with TCP/IP
MAN ports, which are used as virtual addresses
Metropolitan Area Network. A network capable assigned to each IP address.
of high-speed communications over distances
up to about 80 kilometers. R
Metadata RAID (Redundant Array of Independent
The information associated with a file but sep- Disks)
arate from the data in the file; required to A way of storing the same data over multiple
identify data in the file and its physical loca- physical disks to ensure that if a hard disk fails
tion on a disk. a redundant copy of the data can be accessed
instead. Example schemes include mirroring
Mirroring and RAID 5.
A disk data redundancy technique in which
data is recorded identically and either syn- Redundancy
chronously or asynchronously on multiple The duplication of information or hardware
separate disks to protect data from disk fail- equipment components to ensure that should
ures. When the primary disk is off-line, the a primary resource fail, a secondary resource
alternate takes over, providing continuous can take over its function.
access to data. Normally used for mission- Replication
critical data, mirroring is classified as RAID 1 Replication is the process of duplicating mis-
configuration and doubles disk costs. sion critical data from one highly available site
to another. The replication process can be syn-
N chronous or asynchronous; duplicates are
known as clones, point-in-time copies, or
NAS (Network Attached Storage) snapshots, depending on the type of copy
A NAS device is a server that runs an operating being made.
system specifically designed for handling files
(rather than block data). Network-attached S
storage is accessible directly on the local area
network (LAN) through LAN protocols such as SAN (Storage Area Network)
TCP/IP. Compare to DAS and SAN. A storage area network (SAN) is a specialized
network that provides access to high perform-
P ance and highly available storage subsystems
using block storage protocols. The SAN is made
Partition up of specific devices, such as host bus adapt-
A partition is the portion of a physical disk or ers (HBAs) in the host servers, switches that
LUN that functions as though it were a physi- help route storage traffic, and disk storage
cally separate disk. Once the partition is cre- subsystems. The main characteristic of a SAN
ated, it must be formatted and assigned a drive is that the storage subsystems are generally
letter before data can be stored on it. available to multiple hosts at the same time,
Glossary 81

which makes them scalable and flexible. Com- on SSDs should include anything that creates
pare with NAS and DAS. bottlenecks, such as databases, swap files,
library and index files, and authorization and
SAS/SATA
login information.
SAS: Serial Attached SCSI. While SATA (Serial
ATA) is designed for desktops, making it a good Storage Controller
choice in storage environments requiring con- Providing such functionality as disk aggrega-
figuration simplicity or optimal cost/capacity, tion (RAID), I/O routing, and error detection
SAS delivers the high performance, scalability and recovery, the controller provides the intel-
and reliability required for mainstream servers ligence for the storage subsystem. Each stor-
and enterprise storage. age subsystem contains one or more storage
controllers.
SCSI (Small Computer System Interface)
A set of standards allowing computers to com- Storage Resource Management (SRM)
municate with attached devices, such as stor- Refers to software that manages storage from
age devices (disk drives, tape libraries etc) and a capacity, utilization, policy and event-man-
printers. SCSI also refers to a parallel intercon- agement perspective. This includes bill-back,
nect technology which implements the SCSI monitoring, reporting and analytic capabilities
protocol. SCSI is available in two flavours: Par- that allow you to drill down for performance
allel SCSI and Serial Attached SCSI. Parallel and availability. Key elements of SRM include
SCSI has been the standard in connectivity for asset management, charge back, capacity
more than 20 years, and is known for its stabil- management, configuration management,
ity and reliability. Serial Attached SCSI (SAS) is data and media migration, event management,
the newest generation of SCSI, with both Serial performance and availability management,
ATA (SATA) and SAS drives. policy management, quota management, and
media management.
Snapshot
A virtual copy of a device or filesystem. Snap- Switch
shots imitate the way a file or device looked at An intelligent device residing on the network
the precise time the snapshot was taken. It is responsible for directing data from the source
not a copy of the data, only a picture in time of (such as a server) or sources directly to a spe-
how the data was organized. Snapshots can be cific target device (such as a specific storage
taken according to a scheduled time and pro- device) with minimum delay. Switches differ in
vide a consistent view of a filesystem or device their capabilities; a director class switch, for
for a backup and recovery program to work example, is a high end switch that provide
from. advanced management and availability fea-
tures.
Solid State Disk (SSD)
A solid state disk is a high-performance plug- Synchronous Replication
and-play storage device that contains no mov- In synchronous replication, each write to the
ing parts. SSD components include either primary disk and the secondary (remote) disk
DRAM or EEPROM memory boards, a memory must be complete before the next write can
bus board, a CPU, and a battery card. Because begin. The advantage of this approach is that
they contain their own CPUs to manage data the two sets of data are always synchronized.
storage, they are a lot faster than conventional The disadvantage is that if the distance
rotating hard disks; therefore, they produce between the two storage disks is substantial,
highest possible I/O rates. SSDs are most effec- the replication process can take a long time
tive for server applications and server systems, and slows down the application writing the
where I/O response time is crucial. Data stored data. See also asynchronous replication.
82 Glossary

T kept for regulatory purposes is archived to


lower cost disks.
Tape Library
In data storage, a tape library is a collection of V
magnetic tape cartridges and tape drives. An
automated tape library is a hardware device VTL (Virtual Tape Library)
that contains multiple tape drives for reading Refers to an intelligent disk-based library that
and writing data, access ports for entering and emulates traditional tape devices and tape for-
mats. Acting like a tape library with the per-
removing tapes and a robotic device for
formance of modern disk drives, data is depos-
mounting and dismounting the tape cartridges
ited onto disk drives just as it would onto a
without human intervention. To mount means
tape library, only faster. Virtual tape backup
to make a group of files in a file system struc- solutions can be used as a secondary backup
ture accessible to a user or user group. stage on the way to tape, or as their own
Target standalone tape library solution. A VTL gener-
A target is the device to which the initiator ally consists of a Virtual Tape appliance or
sends data. Most commonly the target is the server, and software which emulates tradi-
storage array, but the term also applies to tional tape devices and formats.
bridges, tape libraries, tape drives or other Virtualization
devices. In storage, virtualization is a means by which
TCP/IP multiple physical storage devices are viewed as
a single logical unit. Virtualization can be
Transmission Control Protocol/Internet Proto-
accomplished in–band (in the data path) or
col. A set of transport and network layer pro-
out-of-band. Out–of–band virtualization does
tocols developed under the auspices of the U.S.
not compete for host resources, and can virtu-
Department of Defense. Has emerged as the alize storage resources irrespective of whether
de-facto standard for communications among they are DAS, NAS or SAN.
Unix systems, particularly over Ethernet.
Volume
Thin Provisioning A volume is an area of storage on a hard disk. A
Thin provisioning is most commonly used in volume is formatted by using a file system, and
centralized large storage systems such as SANs typically has a drive letter assigned to it. A sin-
and also in storage virtualization environments gle hard disk can have multiple volumes, and
where administrators plan for both current volumes can also span multiple disks.
and future storage requirements and often
over-purchase capacity, which can result in Z
wasted storage. Since thin provisioning is
designed to allocate exactly what is needed, Zoning
exactly when it is needed, it removes the ele- A method used to restrict server access to stor-
ment of “paid for but wasted“ storage capacity. age resources that are not allocated to that
Additionally, as more storage is needed addi- server. Zoning is similar to LUN masking, but is
tional volumes can be attached to the existing implemented in the switch and operates on
consolidated storage system. the basis of port identification (either port
numbers on the switch or by WWPN of the
Tiered Storage attached initiators and targets).
Data is stored according to its intended use.
For instance, data intended for restoration in (Sources: Adaptec, Fujitsu Siemens Computers,
the event of data loss or corruption is stored Horison Information Strategies, Microsoft,
locally, for fast recovery. Data required to be SNIA, Webopedia, ZAZAmedia)
Information infrastructures
in enterprises

T
he principle of profitable trading increasingly demands efficient handling of infor-
mation in the enterprise, especially when you consider that the volume of informa-
tion is growing by an average of 60 percent per year. With its solutions, EMC is
striving to make optimal use of this capital as well as to protect, manage, store and
archive it. EMC thereby rings in a paradigm change by moving the information itself –
rather than the applications – into the center of the infrastructure. As a result, the
demands on the infrastructure must be focused on the paths the information takes in
the enterprise: from creation, capture and utilization, through to archiving and dele-
tion. The optimum strategy for setting up an information infrastructure includes intel-
ligent data storage, protection against data loss or misuse, optimizing the infrastruc-
ture of IT management and services and utilizing the value added potential of informa-
tion. Alongside the top priority of cost reduction, enterprises mainly want to improve
their compliance with all legal requirements and enhance support of their business
processes. Business demands on IT are therefore given noticeably higher priority than
technological goals such as “better data security” or “better structured data”.

EMC hardware and software to build up your information


infrastructure

Y our enterprise is changing continuously – and EMC will grow and change with you.
From hard disk libraries to content management systems and SANs, our products
have been the choice for many enterprises for years.
EMC offers the software, systems, security and services that you need to develop a
more intelligent storage concept that is effective, economical and user-friendly and
also provides you with the necessary flexibility for expansion together with your enter-
prise.
But more intelligent storage is only half the solution: a better concept is needed for
information management. This can be achieved by implementing multiple key initia-
tives. With its years of experience, EMC has identified six important areas that can help
you to store information more intelligently.
84 EMC

Six approaches to storing more intelligently.

1. Classify data and deploy tiered storage


Not all data is created equal. In fact, each byte of data used in your organization varies
in terms of how often it’s accessed and how quickly it must be recovered in the event of
downtime. Therefore, it’s vital to begin the journey toward storing more intelligently by
classifying data based on access requirements. As soon as you begin assessing your
data, you’ll find that a “one size fits all” approach to storing data no longer makes sense.
Rather, deploying tiered storage provides the needed framework and processes to
maintain information on the “right storage type” at the “right time,” thereby aligning
capabilities to requirements and ultimately lowering total cost of ownership (TCO).
EMC offers a range of scalable, easy-to-use storage platforms, software, and serv-
ices to help organizations like yours classify data and establish a tiered storage infra-
structure. Our portfolio includes EMC® Symmetrix®, EMC CLARiiON®, and EMC Celerra.
All EMC storage platforms can be deployed with a combination of high-performance
and low-cost/high-capacity disk drives within the same array, and offer feature-rich
software functionality to simplify movement, management, and protection of valuable
data. So, whether your organization requires a small-scale solution with tiered storage
deployed within one physical system, or has a large-scale environment where multiple
tiers of storage systems are needed, the ultimate benefit is the same—lower TCO.

Before After

Tier 1

Production
Tier 2
Data

Tier 3

"One size fits all" „Tiered Storage“

Durch die Klassifizierung von Daten und die Bereitstellung von


By classifying
Tiered Storagedata and Organisationen
können deploying tieredihre
storage,
TCO inorganizations
der Regel
typically
um 25%see a 25 percent improvement in TCO.
verbessern.
Information infrastructures in enterprises 85

EMC is the market leader in the area of IP storage solutions


The wide proliferation of IP storage technologies, such as iSCSI and NAS for accessing
storage systems, offers enterprises of all sizes the option of economically setting up
and operating efficient and reliable storage networks. Enterprises can now use their
existing IP technologies and IP-trained personnel to provide and operate their storage
networks. All-in-all, the methods offered by today’s mature, IP-based storage technol-
ogy make it significantly simpler and more cost-effective to achieve the desired storage
consolidation results. With the storage families Celerra NS, CLARiiON CX (FibreCAT CX)
and Symmetrix DMX, EMC is the market leader in the field of IP storage solutions.
According to IDC Research, no other supplier has provided more IP-based storage solu-
tions than EMC with its wide network of EMC Velocity2 partners. The reason for this:
EMC offers a wide range of flexible options to allow enterprises and organizations of all
sizes to profit from cost-effective and user-friendly IP storage solutions – regardless of
whether you are providing a whole new network storage system or consolidating serv-
ers and applications that are not in an existing SAN.

2. Create an active archive


When information is needed, it must be online and readily accessible—whether it’s
required daily or to meet longer-term audit cycles. Active archiving gives you the ability
to move data that is used infrequently from primary storage into a more cost-effective
storage infrastructure while still allowing quick and easy access.
EMC Centera® is the first active archive solution in the market that is purpose-built
for storage and retrieval of fixed content. By migrating static or infrequently accessed
information from the production storage environment to an EMC Centera active
archive, valuable “production storage” space is freed up and backup media require-
ments lessen as static content is removed from daily or weekly backup processes. For
compliance purposes, EMC Centera also offers guaranteed content authenticity to meet
the emerging set of regulatory requirements.
86 EMC

Before After

Tier 1
Tier 1

Active Archive
Tier 2
Tier 2

Tier 3
Tier 3

InInmany
vielen E-Mail-
e-mail oder
or file Dateisystemumgebungen
system werden
environments, greater than mehr
75 percent
ofalsthe
75%
datader
is Daten nicht verändert
not modified, – damit
which makes it ansind diese
ideal Daten for
candidate
idealearchive.
active Kandidaten für die aktive Archivierung.

3. Reduce / eliminate redundant data


One of the key drivers impacting storage costs is the amount of production data that
needs to be stored and backed up. Reducing or eliminating duplicate data in backup
environments is a key step to storing more intelligently. Consider a simple example: an
e-mail attachment is distributed and then saved by 20 different users. The impact?
Storage capacity, network bandwidth, and backup media are unnecessarily consumed
by a factor of 20. With data de-duplication, storage requirements lessen, operations are
streamlined, backup times are shortened, and overall TCO is improved.
EMC technology proactively helps to reduce or eliminate the amount of redundant
data that an organization creates, stores, and ultimately backs up. EMC Centera com-
putes a content address every time an object is requested to be stored. For example,
when two or more people try to store the same e-mail attachment, EMC Centera com-
putes the same content address each time. With this intelligence, EMC Centera easily
determines that multiple people are trying to store the same file and gives each user a
pointer to the file instead of storing multiple copies. This minimizes information redun-
dancy and the amount of storage required.
EMC Avamar® backup and recovery software can achieve the same results for your
backup data. EMC Avamar leverages global de-duplication technology to eliminate the
unnecessary transmission of redundant backup data over the network and saved to
secondary storage. By de-duplicating across sites and servers, you can dramatically
shrink the amount of time required for backups, network utilization, and the growth of
secondary storage.
Information infrastructures in enterprises 87

Before After

EMC Avamar erzielt eine Reduzierung der täglichen


EMC Avamar realizes
Backup-Daten bis zumupFaktor
to 300300
times
undreduction in zu
bietet bis daily
zehnbackup
dataschnellere
Mal and up to 10 times faster backups.
Backups.

4. Speed backup and recovery


By extending the tiered storage methodology to backup/recovery operations, organiza-
tions can dramatically improve the speed and reliability of their backup and recovery
processes. For example, disk-based solutions deliver five-fold improvement in backup
and recovery times compared with tape-based solutions.
EMC offers a complete solution for backup-to-disk needs, including assessment
services, backup software applications, and a wide range of LAN-based, SAN-based, and
virtual tape library (VTL) solutions so you can realize better performance and reliability
in your backup and recovery environment—no matter how you manage your backups
today.
All EMC storage platforms can be deployed with low-cost/high-capacity disk drive
technologies that are ideal for backup-to-disk applications.
88 EMC

Before After

Backup Backup
Data Data
Tier 3

Festplattenbasierte Lösungen bieten eine Verbesserung


Disk-based solutions provide a 5x improvement in backup /
der Backup- und Recovery-Zeiten um den Faktor 5.
recovery times.

5. Use space-saving snapshots to reduce capacity requirements


Creating full copies of production data has been a common practice to assist with
backup, recovery, and testing of application environments. As information continues to
grow, the storage requirements and costs associated with this practice have escalated,
as have the capacity requirements to support local replication activities. Today, the use
of space-saving snapshots provides a much smarter, more affordable alternative for
protecting and re-purposing production data. Specifically, the space-saving character-
istics of logical snapshot copies allow organizations to more quickly and effectively
leverage replication capabilities while minimizing the traditional costs and capacity
requirements that go along with them.
All EMC storage platforms offer the flexibility to deploy array-based full copy clone
and local snapshot replication—you can choose which to leverage based on your appli-
cation requirements and use cases. In addition, both types are supported and integrated
with key applications, such as Microsoft Exchange, SQL Server, Oracle, and SAP.
Information infrastructures in enterprises 89

Before After

Snaps

Snaps
Clones Snaps

Snapshots provide
Snapshots bieten up to bis
eine 10 times reductionReduzierung
zu zehnfache in capacity required
der
for die
für locallokale
replication.
Replikation erforderlichen Kapazität.

6. Deploy server and file virtualization


IT departments today face a number of challenges in their physical server infrastruc-
tures, beginning with underutilization and server sprawl. According to IDC, a typical x86
server deployment uses only 10 to 15 percent of its total capacity, and yet organizations
spend considerable resources provisioning, maintaining, and securing growing num-
bers of physical servers. For every $1 spent on new servers, again according to IDC, $8
is spent in ongoing management. Organizations can overcome these challenges and
implement significant efficiencies by using virtualization to optimize the infrastructure
at the server level.
Virtualization with VMware offers a fundamentally simpler and more efficient way
to manage server infrastructures. In fact, the time to provision a new server can be
decreased by up to 7o percent, new applications can be provisioned in minutes, and
time previously spent on manual, time-consuming tasks can be redirected toward stra-
tegic initiatives. What’s more, data center capacity can be scaled non-disruptively and
availability can be ensured for all applications at a reasonable cost.
When you combine VMware server virtualization with EMC Rainfinity® Global File
Virtualization- you get the added benefits of virtualizing your unstructured data envi-
ronments. The Global File Virtualization capabilities allow you to simplify management,
move data non-disruptively, address performance bottlenecks, and lower TCO by maxi-
mizing utilization in heterogeneous NAS, CAS, and file server environments
90 EMC

Addressing your key priorities


At EMC, we’re focused on the things that are most important to you, such as making
sure your information is easy to manage, taking the right steps to conserve energy, and
protecting your valuable information from security threats and breaches.

EMC solutions simplify storage management


EMC has invested heavily in developing new and improved tools that meet user require-
ments to streamline storage management and improve ease of use across all platforms.
These capabilities help simplify.

EMC solutions can be deployed quickly, and are easy to use and manage. For exam-
ple, Symmetrix Management Console, Celerra Manager, and CLARiiON Navisphere®
Manager all allow users to configure and provision a terabyte of storage in less than
one minute. And EMC Centera includes self-configuring, self-healing, and self-manag-
ing capabilities that enable an administrator to manage up to 50 times greater quantity
of content.

Energy-efficient solutions improve cost management

T oday’s data centers face critical energy issues power and cooling limitations, high
energy demands and costs, and even outages from overburdened power grids.
Advanced tools and services from EMC help optimize energy efficiency in your data
center. For example, EMC storage platforms consume less energy per terabyte than alter-
native solutions, while the EMC Power Calculator enables EMC experts to advise on ways
to reduce power and cooling costs through more efficient data management. Powerful
EMC virtualization solutions, such as VMware, can help you consolidate servers to
increase capacity utilization and eliminate unnecessary infrastructure. And finally, EMC
Energy Efficiency Services can help you make your data center as power efficient as pos-
sible. Continue to look to EMC for further innovations resulting from our investment in
research and development programs aimed at reducing your energy and cooling costs.

Storage solutions with security built-in

M ost organizations readily agree that information is their most important asset. In
fact, in 2006 organizations spent $45B on security products; however studies
show that only one in five feel protected.
Information infrastructures in enterprises 91

At EMC, we’ve taken an information-centric approach by building security into our


products, not bolting it on. This strategic approach provides our customers with
improved access control, a more secure infrastructure, and expanded compliance and
auditability—so that their information infrastructure is guarded from security threats.

Take the Next Step


With its flexible and cost-effective storage consolidation solutions, EMC provides
organizations with the information management framework they need to store more
intelligently. But storage is only part of what we do. Our focus is on improving your
business from the ground up. As such, we help organizations like yours with solutions
focused on:
• Establishing backup, recovery, and archive operations
• Ensuring effective and affordable information protection
• Virtualizing information infrastructure
• Automating data center operations
• Securing critical assets
• Leveraging content for competitive advantage
• Accelerating business value for business-critical applications, such as Microsoft
Exchange, SQL Server, Oracle and SAP.

Partnership between EMC and Fujitsu Siemens Computers

“For nearly a decade, EMC and Fujitsu Siemens Computers have worked to build a strong
relationship built on mutual trust and technological innovation. A key to our success is
the combined expertise of both our sales and technical teams who have helped our
mutual customers meet some of their biggest challenges. The combination of EMC net-
worked storage systems and software with servers from Fujitsu Siemens Computers
helps to provide a broad offering of end-to-end solutions. EMC technology plays a key
role in Fujitsu Siemens Computers‘ vision of the Dynamic Data Center and we continue
to focus our combined efforts on providing customers with the industry‘s most com-
prehensive set of solutions.“ 2007, Joe Tucci, EMC President & CEO.

Common strengths of Fujitsu Siemens Computers and EMC:


• EMC Storage integrated with Fujitsu Siemens Computers FlexFrame infrastructure
• Fujitsu Siemens Computers CentricStor is available in the EMC Select Reseller
program
• Fujitsu Siemens Computers’ Infrastructure Services is an accredited member of the
EMC Authorized Service Network (ASN) program
92 EMC

• EMC Storage integration with Microsoft is implemented via Fujitsu Siemens Com-
puters BladeFrame technology
• Common solution for Grid Computing based on Fujitsu Siemens Computers
PRIMERGY server systems
• OEM and reseller agreements for EMC Networker

For more information on EMC solutions, visit www.emc.com


DATA CENTER FABRIC

Leveraging 8 Gbit/sec Fibre Channel


End to End in the Data Center

B
rocade is a leading provider of networked storage solutions that help organiza-
tions connect, share, and manage their information. Organizations that use Bro-
cade products and services are better able to optimize their IT infrastructures and
ensure compliant data management. All of Brocade’s products are tested and certified
by Fujitsu Siemens Computers which ensures seamless integration into the comprehen-
sive and leading edge infrastructure solutions of Fujitsu Siemens Computers.
The continuing expansion of data centers and the introduction of new technologies
such as server and fabric virtualization are driving the need for higher storage network-
ing performance and greater capabilities from the data center fabric. The new Brocade
Fibre Channel (FC) switch family meets current and near-future storage networking
needs by doubling the current standard FC speed of 4 Gbit/sec to 8 Gbit/sec. New Bro-
cade 8 Gbit/sec Host Bus Adapters work in concert with Brocade 8 Gbit/sec fabrics to
deliver high performance and advanced functionality end to end.

Introduction

I n January 2008, Brocade® introduced 8 Gbit/sec capabilities for the Brocade 48000
Director and the new Brocade DCX Backbone platform. Brocade is expanding this
leadership position with the introduction of an entire family of 8 Gbit/sec switch prod-
ucts targeting a range of data center environments—from the enterprise to Small and
Medium Business (SMB). In addition Brocade is launching 8 Gbit/sec Host Bus Adapters
(HBAs), providing the industry’s first end-to-end 8 Gbit/sec solution for SMB to enter-
prise customers. These high-performance solutions are driven by a new family of Bro-
cade 8Gbit/sec ASICs, which process and route data with much higher levels of effi-
ciency. In addition to doubling performance throughput, these new ASICs offer new
capabilities that align with growing data center requirements for IT process automa-
tion, energy efficiency, and reduced Operating Expenses (OpEx). Steady increases in per-
formance and functionality have been the hallmark of Fibre Channel evolution over the
past decade. With the periodic doubling of transport speed from 1 to 2 Gbit/sec and
from 2 to 4 Gbit/sec, storage administrators have quickly exploited the new perform-
94 Brocade

ance capabilities and advanced features to build more optimized storage networks.
With the introduction of Brocade 8 Gbit/sec switches and HBAs, it is now possible to
fully integrate advanced functionality that extends from the fabric all the way to the
server platform. In trying to decide where enhanced performance and capabilities can
be applied in your own environment, consider the following:
• Storage Growth. Storage Area Network (SAN) storage capacity has dramatically
increased year over year in almost all data centers. As SAN storage grows, so do the
fabrics that interconnect storage with servers.
• Large Fabrics. As fabrics grow, more Inter-Switch Links (ISLs) are used to keep pace
with storage and server scaling.
• Higher Levels of Performance. In large-scale data centers, moving SAN bandwidth-
intensive hosts to 8 Gbit/sec connectivity enables the servers to achieve higher levels
of performance using fewer HBAs and a smaller cabling infrastructure.
• Server Virtualization. Hosting multiple operating system instances on a single host
platform dramatically increases storage I/O demands, which in turn drives up host
SAN throughput.
• Tiered Services. In a shared environment, in which IT may be using chargeback to
serve internal customers, a tiered services model requires the ability to specify serv-
ice levels for hosted applications and to monitor and manage these services end to
end—all capabilities of Brocade 8 Gbit/sec solutions.
• Backup. Large amounts of traffic to tape or disk during backups require the fastest
SAN speeds possible to fit within backup windows.
• Operational Flexibility. While not all hosts, storage, and ISLs currently require maxi-
mum speed capability, it is much easier to architect data center fabrics when high-
speed ports are available.
• Investment Protection. Existing SANs can be significantly enhanced with new capa-
bilities enabled by 8 Gbit/sec port speed. Integrated Routing and Adaptive Network-
ing services are compatible with legacy SAN equipment, extending their Return on
Investment (ROI) as data center fabrics scale.

Data centers may have some or all of these needs today. Although meeting these
needs may not require an immediate upgrade to 8 Gbit/sec for all storage applications,
future plans for expansion, virtualization, and fabric scaling will make acquiring 8 Gbit/
sec capabilities today a safe and well-founded decision. As fabrics scale, for example,
only half the number of ISLs is required with 8 Gbit/sec links than with 4 Gbit/sec links.
Likewise, the ISL oversubscription ratio is halved by upgrading from 4 to 8 Gbit/sec, ISLs
using the same number of links.
At long distances, 8 Gbit/sec can earn a very fast ROI compared to 4 Gbit/sec, due to
the high cost of dark fiber or WDM links. Almost all of these native FC extension links
support 8 Gbit/sec speeds, so utilization can be doubled on links that usually cost thou-
Leveraging 8 Gbit/sec Fibre Channel End to End in the Data Center 95

sands, if not tens of thousands, of dollars per month. This can quickly justify the equip-
ment cost for the increased speed capability. Building a high-performance foundation
that provides the flexibility to selectively deploy 8 Gbit/sec as needed simplifies data
center fabric management and accommodates the inevitable growth in applications
and data over time.

Evolving Data Center Virtualization

V irtualization of server platforms can dramatically increase the need for higher-speed
capability in the SAN. Some virtualized hosts have 10, 20, or even 30 operating sys-
tems, which can exceed the capacity of a 4 Gbit/sec HBA. Brocade’s 8 Gbit/sec end-to-
end solutions can prevent this saturation and increase the ROI on server hardware and
virtualization software investments.
Brocade recognizes these three phases for evolving virtualization in the data
center:
• Phase 1. The primary business driver for this phase is the reduction of Capital Expen-
ditures (CapEx), as a result of server consolidation and flexible test and develop-
ment.
• Phase 2. The challenge for Phase 2 is characterized by growth and deployment of
Disaster Recovery (DR) solutions and the need for high availability and automated
server provisioning. The primary business drivers are reduction of OpEx and the
requirement for Business Continuity (BC). Here are some typical use cases:
• Automated server provisioning and applications deployment using pre-built Virtual
Machine (VM) “templates”
• Data center architectures and products that provide High Availability (HA) and no
interruption of service during server maintenance or failure
• Storage replication and automated restoration of service to support DR goals
• Phase 3. We are now moving into this phase, in which business drivers are flexible IT,
variable cost, and further OpEx reductions. Phase 3 will provide data centers with
policy-driven utility computing, service-level management, and end-to-end service
provisioning.

Virtualization has fundamentally transformed the traditional relationship between


servers, storage, and the fabric interconnect. Running many VMs on a single server
hardware platform can dramatically increase the requisite Input/Output (I/O) load and
mandate offloading as much I/O processing as possible, so that CPU cycles can be
devoted more productively to application processing. Fully leveraging server virtualiza-
tion therefore requires more powerful, high-performance storage adapters.
96 Brocade

Brocade’s new family of 8 Gbit/sec switches supports the rapidly growing data center
by delivering 8 Gbit/sec performance on every port with no oversubscription. A com-
pletely non-oversubscribed switching architecture enhances server scalability by ena-
bling the rapid growth of virtual servers without compromising data center perform-
ance.
Integrated Routing (IR) fabric service is a new option on the Brocade DCX Backbone
and Brocade 5300 and 5100 Switches with the release of Fabric OS® (FOS) 6.1. As of FOS
6.1, IR can be activated on FC8 port blades with up to 128 IR ports per Brocade DCX
chassis. (When there are two Brocade DCX chassis connected via Inter-Chassis Links, a
total of 256 IR ports are available.) No additional hardware is required to enable per-
port Fibre Channel Routing; only an optional IR software license is required. IR can be
enabled on the maximum number of ports on the Brocade 5300 (80 ports) and Brocade
5100 (40 ports) via user configuration. Brocade 8 Gbit/sec HBA ASICs support a maxi-
mum of 500k I/O per Second (IOPS) per port ( >1M IOPS on a dual-port HBA) to free up
the host processors and meet virtualization productivity goals. In the future, two 8
Gbit/sec HBA ports will be able to be combined into a single, ultra-high-speed 16 Gbit/
sec connection using Brocade ISL Trunking technology, which balances traffic flows at
the frame level. Currently, the benefits of Brocade 8 Gbit/sec switching technology are
extended directly to VMs via N_Port ID Virtualization (NPIV), so that special Brocade
features, such as Top Talkers and QoS Traffic Prioritization, can be applied to individual
VMs. This end-to-end fabric and host integration is unique to Brocade and offers the
industry’s highest I/O performance for virtualized environments.
Brocade 8 Gbit/sec HBAs complement industry-leading performance with advanced
storage functionality to further streamline virtualized server operations. To meet regu-
latory compliance requirements, for example, Brocade 8 Gbit/sec HBAs implement the
industry standard Fibre Channel Security Protocol (FC-SP) and will support in-flight
data encryption for secure network transactions.
In addition, the new Brocade fabric service, Adaptive Networking, provides config-
urable Quality of Service (QoS) for each VM. With the increasing use of VM mobility to
shift application workloads from one hardware platform to another, conventional net-
working methods are no longer sufficient. Brocade meets the needs of more dynamic
virtualized environments by providing an integrated fabric and HBA solution that can
selectively deploy security and QoS to VM-hosted applications as required.

Brocade 8 Gbit/sec Products

T he Brocade DCX Backbone, Brocade’s first 8 Gbit/sec platform, with 16-, 32-, and
48-port blades, was released in early 2008. In May 2008, Brocade completed the
Leveraging 8 Gbit/sec Fibre Channel End to End in the Data Center 97

transition to 8 Gbit/sec with the release of Fabric OS 6.1 and a full family of new
switches and HBAs for end-to-end connectivity in the data center:
• Brocade 815 (single port) and 825 (dual port) HBAs
• Brocade 300 Switch with 8, 16, and 24 ports
• Brocade 5100 Switch with 24, 32, and 40 ports
• Brocade 5300 Switch with 48, 64, and 80 ports
• FC8-16, FC8-32, and FC8-48 port blades for the Brocade 48000 Director

Brocade 8 Gbit/sec switches comply with industry standards; and fabrics with 4 and
8 Gbit/sec devices interoperate seamlessly. Visit the Brocade Web site for data sheets
describing these products: www.brocade.com

Conclusion

S peed increase in Brocade switching platforms is one of many advantages from Bro-
cade’s next generation ASIC family. Higher speed in the data center brings the imme-
diate benefit of higher-performing ISLs and increased scalability; since ISL performance
is doubled, more ports can be used for servers and storage. In addition, 8 Gbit/sec is
needed for server virtualization, scaling of fabrics, backups, and high-performance
computing requirements. New capabilities, such as Adaptive Networking and Integrated
Routing, plus the enhanced power efficiencies of the new switch platforms are also
important drivers for adoption of 8 Gbit/sec technology. Every data center user has or
will have these needs in the future, and as data center plans are developed, Brocade’s
integrated end-to-end 8 Gbit/sec solution provides the broadest choice of capabilities
with the highest performance and efficiency.

Further information may be found on our website at: www.brocade.com


Contact:
Brocade Communications GmbH
Ralf Salzmann
Altkönigstrasse 6
64546 Mörfelden Walldorf
ralf.salzmann@brocade.com
NetApp
Innovative solutions for storage
and data management

N
etApp embodies innovative storage and data management with excellent cost
efficiency. The commitment to simplicity, innovation and the success of its cus-
tomers has enabled the company to become one of the fastest growing storage
and data management manufacturers. The wide-ranging solution portfolio for server-
to-storage virtualization, business applications, data security and much more has per-
suaded customers worldwide to opt for NetApp.
NetApp ensures that your business-critical data is constantly available and can also
simplify your business processes. Based on the motto “Go further, faster”, NetApp helps
companies to be successful. The storage requirement for company data will continue to
grow fast in coming years. This presents IT managers with the challenge of purchasing
an ever increasing quantity of storage equipment yet also having to manage such
devices. With the help of its powerful unified storage architecture NetApp helps com-
panies overcome these challenges efficiently: Extremely low operating costs (TCO), very
fast backup and restore processes, high availability levels, consolidation and virtualiza-
tion options as well as simplified and easy management of the entire storage environ-
ment are behind the NetApp slogan “Go further, faster“.

Solutions
Microsoft, VMware, Oracle and SAP are important strategic NetApp partners. NetApp
has developed a wide range of tools for its database and application software.

The NetApp concept

Data ONTAP
A small, very effective operating system kernel, optimized to meet storage process
requirements, provides top performance for storage applications and has a wide range
of usage: Data ONTAP permits universal data access via block or file level protocols.
Data access protocols, such as NFS, CIFS, iSCSI, FC, http and FTP permit access at file
level.
100 NetApp

Extremely flexible storage allocation


Data ONTAP means that all the hard disks available in a system can be used for all appli-
cations. A specific storage area no longer needs to be assigned to certain applications.
Each application can use all the drives installed, regardless of the overall storage space
required by the application.

Snapshot
Using “Snapshot” technology developed by NetApp, 255 snapshots can be created for
each volume created without any loss in performance. Snapshots are read-only copies
of the data on the respective volume (i.e. the data in the snapshot is stored on the same
disks as the productive data).

Flexvol & Flexclone


NetApp FlexCloneTM technology enables real cloning, an immediate replication of the
data volumes and data sets without any additional storage space.
NetApp FlexClone provides notable savings in storage space and has low overheads.

Metrocluster
If the two storage controllers of a FAS cluster are at different locations, MetroCluster
synchronously mirrors the data (see SyncMirror) in both data centers. An automatic
transfer is made if an FAS system is not available.

The NetApp unified storage concept

N etApp has designed a completely standardized storage architecture which seam-


lessly integrates all the storage products and which can be managed with one
standardized operating system (Data ONTAP). This results in a number of business ben-
efits:

Less administration effort


Fewer employees can manage more storage and more applications.

Better use of existing storage resources


As all the devices are connected in a network, data and applications can be distributed,
as required, irrespective of the physical storage space (virtualization). Storage reduc-
tions of up to 30% can thus be obtained for the same amount of data and operational
applications.
Innovative solutions for storage and data management 101

Long-term investment protection


When requirements change, the storage systems can be extended at any time to opti-
mally meet new requirements. Investment in other dedicated systems is avoided.

Flexible scalability
When storage requirements drastically increase due to more data or new business
applications, additional storage equipment can be integrated instantly.

No migration costs
The end-to-end NetApp product range enables fast and smooth data transfer when
other systems are added.

NetApp FAS systems

T he powerful, scalable and reliable NetApp FAS systems (fabric attached storage) for
simplified data management meet company requirements whatever their size –
starting from the Global 1000 firms down to medium-sized companies and small
departments.
The installation, configuration and administration processes for these systems are
extremely simple. The stable and flexible microkernel operating system Data ONTAP
enables the simultaneous and combined usage in FC SAN, iSCSI and NAS environ-
ments.
In addition to Windows, Unix and Linux, web data is also supported. The FAS product
range comprises the high-end FAS6000 series for data consolidation in large and high-
performance applications, the Midrange series FAS3100 with exceptional price/per-
formance ratio for SAN and NAS, and the FAS2000 series for data backup in branches
and medium-sized companies.
The FAS systems were designed to consolidate and provide data for a wide range of
applications, including business applications, e-mail, Enterprise Content Management,
technical applications, file shares, home directories and web content.

Maximize storage utilization and performance via virtualization


Server virtualization in the data center improves server usage and simplifies server
administration for the customer. The demands on storage systems increase, as they
must support higher I/O rates, more capacity and the fast, non-interrupted provision of
storage space. If storage optimization is not taken into consideration during server
virtualization, then you have probably only moved the costs from the servers to the
storage systems.
102 NetApp

Together with providers, such as VMware, NetApp offers solutions and best practices
for developing a virtualized infrastructure – from servers to storage – that provide a
number of advantages:
• Scalable and consistent I/O performance for all ESX protocols (NFS, iSCSI and FC)
• Flexible, fast, simple and low-priced provision and data management solutions
• First-class virtualized storage solution for thin provisioning in heterogeneous stor-
age environments

NetApp deduplication in the Esx environment


The NetApp deduplication is one of the fundamental components of our data ONTAP®
operating system. The elimination of redundant data objects and exclusive referencing
to the original object permits more efficient use of the existing storage.

SnapManager for virtualized infrastructures (VI)


SnapManager for VI provides customers with an automated solution for backing up and
restoring virtual machines in a VMware ESX environment. The two main advantages of
this solution are:
• The backups created using NetApp snapshot technology only use a fraction of the
storage space that traditional systems would require.
• The system performance of the ESX environment and thus the applications are
hardly impaired by the SnapManager backup and restore processes. More than 5000
customers worldwide (March 2008) already benefit from the advantages of a VMware
solution with NetApp storage.

Support for a virtual desktop infrastructure


In addition to server virtualization, the VMWare Virtual Desktop Infrastructure (VDI)
offers additional resource-saving virtualization technology. Application environments
no longer run on the user’s desktop processor but in virtual machines in the data center.
NetApp FlexClone can configure thousands of such virtual machines within minutes.
Data deduplication enables storage capacity savings of approx 90%.

NetApp solutions for SAP


As a worldwide technology partner for SAP, NetApp has a successful history in develop-
ing solutions which significantly simplify SAP data management. As one of the found-
ing members of the “Adaptive Computing Initiative for SAP”, NetApp has been awarded
numerous certificates for the compatibility of its storage solutions and is in the SAP
compliance list for the SAP “Adaptive Computing Services” for Unix, Linux® and Win-
dows® platforms.
NetApp won the SAP Pinnacle Award for Technical Innovation and cooperation in
the Adaptive Computing “Netweaver Innovation” category for the FlexFrameTM for
Innovative solutions for storage and data management 103

mySAP Business SuiteTM joint development with Fujitsu Siemens Computers. Integra-
tive components are NetApp system cloning and backup/recovery scenarios.
The NetApp Unified Storage model provides SAN / IP SAN and NAS connections with
block and file access methods within a single storage architecture.
Data management solutions, such as FlexCloneTM, are used to clone SAP productive
systems within a few minutes and without affecting performance and without any
additional initial storage requirement. This thus significantly simplifies the addition and
management of systems for QA, test, development, reporting, interfaces and training.
The combination of NetApp SnapShotTM and SnapRestore® provides SAP customers
with fast, simple backup and restore processes for several TB of SAP data as well as
efficient, simple upgrades and migrations of SAP systems. NetApp Adaptive Computing
solutions enable SAP customers to react dynamically, flexibly and economically to busi-
ness requirements.

NetApp also offers the following for those companies using SAP:
• A comprehensive range of products for Windows, Unix and Linux environments with
unified NAS/SAN storage solutions.
• ILM solutions: storage consolidation, backup and recovery, archiving and compliance
via ArchiveLinkTM and/or WebDAV
• High-availability and disaster recovery solutions for data encryption
• SnapManager for SAP: the solution certified by SAP simplifies the creation of appli-
cation-consistent snapshot copies, automated error-free data restores and permits
application-specific disaster recovery.
• Clones of the SAP database can also be automatically created.
• There is a worldwide support agreement between NetApp and SAP which ensures
that the customer has 7x24 SAP infrastructure support.

Further information about this and other NetApp solutions can be found at www.
netapp.com
CA RECOVERY MANAGEMENT

Data protection and availability

Solution

C
A Recovery Management offers you the functions and services which you really
need: comprehensive protection and reliable recovery in a simple easy-to-use
solution.
It includes application-related, powerful backup with replication, interrupt-free data
protection, automated failover and interrupt-free disaster recovery tests so that you
can provide the protection which you need for specific company data. A simplified
interface and new central management functions enable you to organize, monitor and
configure your entire backup environment. This results in lower operating costs and
improved administration. Thanks to the modular approach this solution is suitable for
every company whatever its size. It links up CA ARCserve Backup with CA XOsoft Repli-
cation and CA XOsoft High Availability so that you can define a multi-level data protec-
tion concept to meet your corporate objectives.

Benefits

T he solution can be seamlessly integrated into existing IT management solutions so


that Enterprise IT Management is simplified and extended.

CA Recovery Management:
a complete data protection solution

C A Recovery Management offers comprehensive and integrated data protection and


recovery functions which your company requires. Robust and proven technologies
are used which are connected via one simplified interface. These technologies provide
multi-level data protection which can be aligned to your company targets, require-
ments and guidelines and covers numerous hardware and software platforms.
106 CA Recovery Management

CA Recovery Management combines proven, powerful solutions — CA ARCserve®


Backup, CA XOsoft Replication and CA XOsoft™ High Availability.

CA ARCserve Backup Release 12

C A ARCserve Backup provides the most comprehensive data protection currently


available on the market. The solution enables central management and offers
enhanced functions to meet your company’s changing requirements.
Functions, such as FIPS-certified 256 encryption, improve security. Reports provide
useful information and optimize backup procedures as well as the catalog database
based on SQL Express with improved indexing for fast recovery. A combination of these
functions gives companies more control, higher operational efficiency, more protection
and reduced costs. This solution enables reliable data protection in the “Enterprise class”
for a wide range of operational environments. The fact that CA ARCserve Backup also
contains integrated virus protection and encryption tools makes it the most secure
backup solution currently on the market and one that can be immediately imple-
mented.

CA XOsoft High Availability &


CA XOsoft Replication Release 12

T his solution for maximum business continuity is based on asynchronous real-time


data replication with automatic failover and automatic reset.
It also enables integrated continuous data protection (CDP) against corruption and
automatic disaster recovery tests which ensure that critical company data and applica-
tions can be recovered. CA XOsoft High Availability provides low-priced interrupt-free
availability of servers which run Microsoft Exchange, Microsoft SQL Server or Oracle
databases, Microsoft Internet Information Services (IIS) WebServer, Fileserver, Black-
Berry support and other applications (both on 32 and 64-bit Windows servers).

Contact: cainfo.germany@ca.com, www.ca.com/us


Sun Microsystems / StorageTek

More than 35 years’ experience


in the classic field of data security

A
s a result of the takeover of StorageTek Sun Microsystems has advanced to
become one of the world’s leading providers of tape libraries and drives. Cus-
tomer-oriented solutions with special service offerings are a mark of the com-
pany. Sun Microsystems (JAVA) develops information technology for the global econ-
omy. With its vision “The Network is the Computer” Sun is promoting the distribution of
the Internet and is itself focusing on open innovations, development of communities as
well as market leadership in Open Source. Sun is represented in more than 100 coun-
tries.

Markets

S un Microsystems is a system provider that develops hardware and software. Since


software development concentrates on resolving system-related tasks or setting
strategically important milestones in line with Sun’s vision, Sun is not in competition
with the developers of application programs. On the contrary, firm partnerships exist
with numerous renowned software manufacturers in order to develop offerings
together. In this way, customers retain their freedom, because they can decide in favor
of the best solution on the market. To ensure that systems are integrated at an early
stage Sun Microsystems has set up a number of partner programs, which have pio-
neered both branch-related and task-related methods. There is a solution portfolio of
almost 13,000 commercial and technical applications for Sun Systems with the plat-
form SPARC/Solaris.
In addition to the partnerships with independent software manufacturers, Sun is
very much committed to long-term sales partnerships with innovative distributors and
resellers. These partnerships have enabled fast and competent solutions for end cus-
tomers on a wide-spread basis.
108 Sun Microsystems

The partnership between Fujitsu Siemens Computers and


Sun Microsystems

A partnership of many years exists between Fujitsu Siemens Computers and Sun
Microsystems/StorageTek, which has been extended to include the Nearline (tape)
sector and, since 2008, SPARC-based systems. As a result of OEM contracts as well as
reseller agreements with Sun for Nearline, drive types and Sparc-based products Fujitsu
Siemens Computers has won over a strong partner for disaster recovery and backup.
Sun Nearline products as quasi-standards are convincing with their high performance
levels and above-average error tolerance. On account of their open design and varied
connection options they can be optimally integrated in backup environments with the
systems sold by Fujitsu Siemens Computers, e.g. CentricStor Virtual Tape Appliance.
Cooperation between Sun Microsystems and Fujitsu Siemens Computers has already
proved to be extremely effective in large-scale, complex projects. And this applies to
companies of the most different kind, particularly large enterprises and small to
medium-sized customers. The numerous joint projects, which you can read about in the
case studies, are proof of this.

Sun StorageTek Open Storage solutions

E lectronically stored data are currently one of the most important assets for many
companies. Storing these data securely, managing them intelligently and having
them available in a controlled way at all times is of vital interest for every company.
Unrelenting data growth and the increasing number of statutory regulations with
regard to archiving periods and access controls call for new data storage and adminis-
tration methods. Sun StorageTek Open Storage solutions allow you to face these chal-
lenges successfully. They combine our more than 35 years of experience in classic fields,
such as data security, data availability and failure concepts, with new technologies like
access protection through identity management and storage virtualization. This in turn
enables trend-setting information lifecycle management infrastructures, which not
only store their data securely and in compliance with regulations, but also make them
available to everyone who needs them.
More than 35 years’ experience in the classic field of data security 109

Tape Library StreamLine® SL500

T he Sun StorageTek StreamLine SL500 tape library is the entry-level solution for small
to medium-sized businesses. Tried-and-trusted LTO technology ensures trouble-free
operation. The modular expandability of a standard rack permits customers to adapt
capacities to meet their operating needs at any time.

Tape Library StorageTek SL3000

T he Sun StorageTek SL3000 modular library system is an innovative, environmentally


aware storage concept for the midrange market. The “Any Cartridge Any Slot” tech-
nology enables you to use the best possible drives – depending on access and storage
requirements.

Highlights
– Flexible scaling, for which only the capacity actually used is charged.
– Cost savings through less space and lower power consumption.
– Simpler storage management through partitioning, sharing.
– Innovative technologies, risk reduction and development of new opportunities.
– Support and services for successful installation, optimization and maintenance.

Tape Library StreamLine® SL8500

T he modular Sun StorageTek StreamLine SL8500 tape library enables organizations to


set up their operational processes more efficiently, to maximize availability and cre-
ate optimal prerequisites for auditing reliability with the burden of costs and impair-
ment to their processes remaining absolutely within scope.

Highlights
– Data consolidation: Excellent scaling as well as support for mixed media.
– Joint resources: Designed for use in mainframes, Unix, Linux and Windows.
– Higher availability: Upgrade without downtimes. Redundancy of operation.
– Higher throughput: High-performance throughput and capacity.
– Simple scaling: Without any downtimes, growth according to your requirements.
– Low space requirements: High density of media slots, optimal utilization.

For more information on Sun solutions, visit www.sun.com


Symantec:
Confidence in a connected world

S
ymantec is a global leader in providing security, storage and systems management
solutions to help businesses and consumers secure and manage their information.
Headquartered in Cupertino, Calif., Symantec has operations in more than 40
countries.

Market Categories
Consumer Products; Security and Compliance; Storage and Availability Management;
Symantec Global Services.
Symantec’s leaders bring decades of diverse experience and a history of success.
Combining business acumen with technical savvy, these executives guide more than
17,500 talented employees to create innovative products and solutions that enable
customers around the world to have confidence in their infrastructure, information and
interactions.

Partnership between Fujitsu Siemens Computers


and Symantec

N early all Symantec products (and those formerly of VERITAS) are certified for Fujitsu
Siemens Computers products. Above all, Fujitsu Siemens Computers is offering the
VERITAS Storage Foundation as the strategic file systems and volume management
solution for its PRIMEPOWER servers.
NetBackup and Enterprise Vault are strategic products and integral components of
the Fujitsu Siemens Computers storage software portfolio. The Fujitsu Siemens Com-
puters CentricStor Virtual Tape Appliance is already being optimized through additional
functions to Symantec products in its operations.
The cooperation between Symantec’s software-solutions alongside the platforms of
Fujitsu Siemens Computers has already proved to be very effective in large and complex
projects. This applies for all varieties of businesses, especially for large enterprises and
SME customers.
Fujitsu Siemens Computers is using Altiris technology for building its DeskView and
ServerView solutions.
112 Symantec

Veritas NetBackup – High performance data protection

V eritas NetBackup Enterprise Server delivers high performance data protection that
scales to protect the largest Unix, Windows, Linux and NetWare environments.
Offering complete protection from desktop to data center to vault, NetBackup software
offers a single management tool to consolidate all backup and recovery operations.
NetBackup helps organizations take advantage of both tape and disk storage with its
advances in disk and snapshot-based protection, off-site media management, and
automated disaster recover. For the ultimate in data protection, Net-Backup offers data
encryption that transmits and stores data using the latest encryption technologies on
the market today. To reduce the impact on business critical systems, NetBackup soft-
ware provides online database and application aware backup and recovery solutions for
all leading databases and applications to deliver data availability for Utility Computing.
Veritas NetBackup offers granular file-level and image-level recovery from a single
backup, automatic discovery of virtual machines (winner of 2007 “Best of VMWorld” for
Data Protection).

Symantec Enterprise Vault – E-Mail Archiving

S ymantec Enterprise Vault™ provides a software-based intelligent archiving platform


that stores, manages, and enables discovery of corporate data from e-mail systems,
file server environments, instant messaging platforms, and content management and
collaboration systems. Because not all data is created equally, Enterprise Vault utilizes
intelligent classification and retention technologies to capture, categorize, index, and
store target data in order to enforce policies and protect corporate assets while reduc-
ing storage costs and simplifying management. Enterprise Vault also provides special-
ized applications, such as Discovery Accelerator and Compliance Accelerator, that mine
archived data to support legal discovery, content compliance, knowledge management,
and information security initiatives. It also eliminates PST headaches and removes the
problems of backup, security, stability and storage consumption inherent to PST files.
Archiving for legal and compliance enables you to retrieve e-mail content and corpo-
rate documents to meet legal and regulatory compliance requirements.

Further information may be found on our website at: www.symantec.com


Contact: Harald Derbsch: harald_derbsch@symantec.com
Rene Köhnen-Wiesemes: rene_wiesemes@symantec.com
Storage Basics Storage Basics
An introduction to the fundamentals
of storage technology An introduction to the fundamentals
of storage technology
Contents
Section 1 –
The information society – saving data and knowledge at new levels
Section 2 –
Tiered storage: intelligent information management in the company
Section 3 –
Online storage: disks and reliability
Section 4 –

Storage Basics – An introduction to the fundamentals of storage technology


Storage networks – spoilt for choice
Section 5 –
Backup & Restore: an unloved compulsory exercise
Section 6 –
Storage management – making complex storage networks manageable
Section 7 –
Virtualization – some catching up is necessary regarding storage topics
Section 8 –
The storage strategy of Fujitsu Siemens Computers and its partners
Forecast: Future storage trends
Glossary
As well as articles about storage from our partners
Brocade, CA, EMC, NetApp, Sun and Symantec

Price: 14.99 Euros

Anda mungkin juga menyukai