storage system

Results 1 - 25 of 357Sort Results By: Published Date | Title | Company Name
By: NetApp     Published Date: Dec 13, 2013
Interested in running a Hadoop proof of concept on enterprise-class storage? Download this solutions guide to get a technical overview on building Hadoop on NetApp E-series storage. NetApp Open Solution for Hadoop delivers big analytics with preengineered, compatible, and supported solutions based on high-quality storage platforms so you reduce the cost, schedule, and risk of do-it-yourself systems and relieving the skills gap most organizations have with Hadoop. See how on going operational and maintenance costs can be reduced with a high available and scalable Hadoop solution.
Tags : open solutions, hadoop solutions guide
     NetApp
By: IBM     Published Date: Sep 02, 2014
This book examines data storage and management challenges and explains software-defined storage, an innovative solution for high-performance, cost-effective storage using the IBM General Parallel File System (GPFS).
Tags : ibm, software storage for dummies
     IBM
By: IBM     Published Date: Sep 02, 2014
Life Sciences organizations need to be able to build IT infrastructures that are dynamic, scalable, easy to deploy and manage, with simplified provisioning, high performance, high utilization and able to exploit both data intensive and server intensive workloads, including Hadop MapReduce. Solutions must scale, both in terms of processing and storage, in order to better serve the institution long-term. There is a life cycle management of data, and making it useable for mainstream analyses and applications is an important aspect in system design. This presentation will describe IT requirements and how Technical Computing solutions from IBM and Platform Computing will address these challenges and deliver greater ROI and accelerated time to results for Life Sciences.
Tags : 
     IBM
By: IBM     Published Date: Sep 02, 2014
Learn how to manage storage growth, cost and complexity, while increasing storage performance and data availability with IBM Software Defined Storage solutions including the IBM General parallel File System (GPFS).
Tags : ibm, storage for dummies
     IBM
By: IBM     Published Date: Sep 16, 2015
The IBM Spectrum Scale solution provided for up to 11x better throughput results than EMC Isilon for Spectrum Protect (TSM) workloads. Using published data, Edison compared a solution comprised of EMC® Isilon® against an IBM® Spectrum Scale™ solution. (IBM Spectrum Scale was formerly IBM® General Parallel File System™ or IBM® GPFS™, also known as code name Elastic Storage). For both solutions, IBM® Spectrum Protect™ (formerly IBM Tivoli® Storage Manager or IBM® TSM®) is used as a common workload performing the backups to target storage systems evaluated.
Tags : 
     IBM
By: Storiant     Published Date: Mar 16, 2015
Read this new IDC Report about how today's enterprise datacenters are dealing with new challenges that are far more demanding than ever before. Foremost is the exponential growth of data, most of it unstructured data. Big data and analytics implementations are also quickly becoming a strategic priority in many enterprises, demanding online access to more data, which is retained for longer periods of time. Legacy storage solutions with fixed design characteristics and a cost structure that doesn't scale are proving to be ill-suited for these new needs. This Technology Spotlight examines the issues that are driving organizations to replace older archive and backup-and-restore systems with business continuity and always-available solutions that can scale to handle extreme data growth while leveraging a cloudbased pricing model. The report also looks at the role of Storiant and its long-term storage services solution in the strategically important long-term storage market.
Tags : storiant, big data, analytics implementations, cloudbased pricing model, long-term storage services solution, long-term storage market
     Storiant
By: Data Direct Networks     Published Date: Apr 08, 2014
DataDirect Networks (DDN), the largest privately-held provider of high-performance storage, has a large and growing presence in HPC markets. HPC users identify DDN as their storage provider more than any other storage-focused company, with twice the mentions of EMC, and more the twice the mentions of NetApp, Hitachi Data Systems, or Panasas.(5) DDN’s strength in HPC is anchored by its Storage Fusion Architecture (SFA), winner of the HPCwire Editor’s Choice Award for “Best HPC Storage Product or Technology” in each of the past three years. The DDN SFA12KX combines SATA, SAS, and solid-state disks (SSDs) for an environment that can be tailored to a balance of throughput and capacity
Tags : 
     Data Direct Networks
By: Intel     Published Date: Aug 06, 2014
Around the world and across all industries, high-performance computing is being used to solve today’s most important and demanding problems. More than ever, storage solutions that deliver high sustained throughput are vital for powering HPC and Big Data workloads. Intel® Enterprise Edition for Lustre* software unleashes the performance and scalability of the Lustre parallel file system for enterprises and organizations, both large and small. It allows users and workloads that need large scale, high- bandwidth storage to tap into the power and scalability of Lustre, but with the simplified installation, configuration, and monitoring features of Intel® Manager for Lustre* software, a management solution purpose-built for the Lustre file system.Intel ® Enterprise Edition for Lustre* software includes proven support from the Lustre experts at Intel, including worldwide 24x7 technical support. *Other names and brands may be claimed as the property of others.
Tags : 
     Intel
By: Intel     Published Date: Aug 06, 2014
Designing a large-scale, high-performance data storage system presents significant challenges. This paper describes a step-by-step approach to designing such a system and presents an iterative methodology that applies at both the component level and the system level. A detailed case study using the methodology described to design a Lustre* storage system is presented. *Other names and brands may be claimed as the property of others.
Tags : 
     Intel
By: Intel     Published Date: Aug 06, 2014
Powering Big Data Workloads with Intel® Enterprise Edition for Lustre* software The Intel® portfolio for high-performance computing provides the following technology solutions: • Compute - The Intel® Xeon processor E7 family provides a leap forward for every discipline that depends on HPC, with industry-leading performance and improved performance per watt. Add Intel® Xeon Phi coprocessors to your clusters and workstations to increase performance for highly parallel applications and code segments. Each coprocessor can add over a teraflops of performance and is compatible with software written for the Intel® Xeon processor E7 family. You don’t need to rewrite code or master new development tools. • Storage - High performance, highly scalable storage solutions with Intel® Lustre and Intel® Xeon Processor E7 based storage systems for centralized storage. Reliable and responsive local storage with Intel® Solid State Drives. • Networking - Intel® True Scale Fabric and Networking technologies – Built for HPC to deliver fast message rates and low latency. • Software and Tools: A broad range of software and tools to optimize and parallelize your software and clusters. Further, Intel Enterprise Edition for Lustre software is backed by Intel, the recognized technical support providers for Lustre, and includes 24/7 service level agreement (SLA) coverage.
Tags : 
     Intel
By: Dell and Intel®     Published Date: Aug 24, 2015
To extract value from an ever-growing onslaught of data, your organization needs next-generation data management, integration, storage and processing systems that allow you to collect, manage, store and analyze data quickly, efficiently and cost-effectively. That’s the case with Dell| Cloudera® Apache™ Hadoop® solutions for big data. These solutions provide end-to-end scalable infrastructure, leveraging open source technologies, to allow you to simultaneously store and process large datasets in a distributed environment for data mining and analysis, on both structured and unstructured data, and to do it all in an affordable manner.
Tags : 
     Dell and Intel®
By: Infinidat EMEA     Published Date: May 14, 2019
Big Data and analytics workloads represent a new frontier for organizations. Data is being collected from sources that did not exist 10 years ago. Mobile phone data, machine-generated data, and website interaction data are all being collected and analyzed. In addition, as IT budgets are already under pressure, Big Data footprints are getting larger and posing a huge storage challenge. This paper provides information on the issues that Big Data applications pose for storage systems and how choosing the correct storage infrastructure can streamline and consolidate Big Data and analytics applications without breaking the bank.
Tags : 
     Infinidat EMEA
By: Infinidat EMEA     Published Date: May 14, 2019
Continuous data availability is a key business continuity requirement for storage systems. It ensures protection against downtime in case of serious incidents or disasters and enables recovery to an operational state within a reasonably short period. To ensure continuous availability, storage solutions need to meet resiliency, recovery, and contingency requirements outlined by the organization.
Tags : 
     Infinidat EMEA
By: Infinidat EMEA     Published Date: May 14, 2019
Even after decades of industry and technology advancements, there still is no universal, integrated storage solution that can reduce risk, enable profitability, eliminate complexity and seamlessly integrate into the way businesses operate and manage data at scale? To reach these goals, there are capabilities that are required to achieve the optimum results at the lowest cost. These capabilities include availability, reliability, performance, density, manageability and application ecosystem integration? This paper outlines a better way to think about storing data at scale—solving these problems not only today, but well into the future?
Tags : 
     Infinidat EMEA
By: Infinidat EMEA     Published Date: Oct 10, 2019
Big Data- und Analytik-Workloads bringen für Unternehmen neue Herausforderungen mit sich. Die erfassten Daten stammen aus Quellen, die vor zehn Jahren noch gar nicht existierten. Es werden Daten von Mobiltelefonen, maschinengenerierte Daten und Daten aus Webseiten-Interaktionen erfasst und analysiert. In Zeiten knapper IT-Budgets wird die Lage zusätzlich dadurch verschärft, dass die Big Data-Volumen immer größer werden und zu enormen Speicherproblemen führen. Das vorliegende White Paper informiert über die Probleme, die Big Data-Anwendungen für Storage-Systeme mit sich bringen, sowie darüber, wie die Auswahl der richtigen Storage-Infrastruktur Big Data- und Analytik-Anwendungen optimieren kann, ohne das Budget zu sprengen.
Tags : 
     Infinidat EMEA
By: Infinidat EMEA     Published Date: Oct 10, 2019
Auch nach jahrzehntelangen Fortschritten in Industrie und Technik existiert immer noch keine universelle integrierte Storage-Lösung, die das Risiko verringern, die Profi tabilität sichern, die Komplexität beseitigen und sich nahtlos in die Betriebs- und Verwaltungsprozesse der Unternehmen für Daten in großem Maßstab einfügen können. Zur Erreichung dieser Ziele gibt es Fähigkeiten, um optimale Ergebnisse zu möglichst niedrigen Kosten zu erreichen. Diese Fähigkeiten umfassen Verfügbarkeit, Zuverlässigkeit, Performance, Datendichte, Managementfähigkeit und Integration in Anwendungs-Ökosysteme. Das vorliegende White Paper skizziert ein besseres Konzept für die Speicherung von Daten in großem Maßstab, so dass diese Probleme nicht nur heute, sondern auch in Zukunft gelöst werden.
Tags : 
     Infinidat EMEA
By: SDL     Published Date: Sep 20, 2019
Component Content Management: A New Paradigm in Intelligent Content Services While technology has changed the world, the way that companies manage information has inherently stayed the same. The advent of near-ubiquitous connectivity among applications and machines has resulted in a data deluge that will fundamentally alter the landscape of content management. From mobile devices to intelligent machines, the volume and sophistication of data have surpassed the ability of humans to manage it with outdated methods of collection, processing, storage, and analysis. The opportunity afforded by the advent of artificial intelligence (AI) has stimulated the market to search for a better way to capture, classify, and analyze this data in its journey to digital transformation (DX). The paradigm of document-based information management has proven to be a challenge in finding, reusing, protecting, and extracting value from data in real time. Legacy systems may struggle with fragmented information
Tags : 
     SDL
By: Bell Micro     Published Date: Jun 14, 2010
This paper describes the benefits, features and best practices using a XIV in a SAP environment.
Tags : bell micro, sap application, best practices, ibm xiv, storage system
     Bell Micro
By: Bell Micro     Published Date: Jun 14, 2010
In this Customer Solution Profile a Fortune five organization put the IBM XIV storage system through its paces in their Microsoft Exchange 2003 environment.
Tags : bell micro, microsoft exchange, ibm xiv storage system, resilience, performance, email system
     Bell Micro
By: Bell Micro     Published Date: Jun 14, 2010
This paper discusses how the IBM XIV Storage System's revolutionary built-in virtualization architecture provides a way to drastically reduce the costs of managing storage systems.
Tags : bell micro, ibm xiv, storage system management, virtualization architecture
     Bell Micro
By: Bell Micro     Published Date: Jun 14, 2010
In this white paper, we describe the XIV snapshot architecture and explain its underlying advantages in terms of performance, ease of use, flexibility and reliability.
Tags : bell micro, hardware, storage management, vmware, ibm xiv
     Bell Micro
By: Hewlett Packard Enterprise     Published Date: Jul 25, 2019
"Flash has permeated enterprise storage in small, medium-sized, and large enterprises as well as among webscale customers like cloud and service providers. It is available in a variety of different system architectures, including internal storage, hyperconverged and converged platforms, and shared storage arrays. Revenue driven by shared storage arrays, of both the hybrid flash array (HFA) and the all-flash array (AFA) type, is much larger than from other segments today and will continue to dominate enterprise storage spend through at least 2020."
Tags : 
     Hewlett Packard Enterprise
By: Hewlett Packard Enterprise     Published Date: May 10, 2019
Nimble Secondary Flash array represents a new type of data storage, designed to maximize both capacity and performance. By adding high-performance flash storage to a capacity-optimized architecture, it provides a unique backup platform that lets you put your backup data to work. Nimble Secondary Flash array uses flash performance to provide both near-instant backup and recovery from any primary storage system. It is a single device for backup, disaster recovery, and even local archiving. By using flash, you can accomplish real work such as dev/test, QA, and analytics. Deep integration with Veeam’s leading backup software simplifies data lifecycle management and provides a path to cloud archiving.
Tags : 
     Hewlett Packard Enterprise
By: HPE     Published Date: Jan 04, 2016
EMC to 3PAR Online Import Utility leverages storage federation and Peer Motion to migrate data from EMC Clariion CX4 and VNX systems to HP 3PAR StoreServ. In this ChalkTalk, HPStorageGuy Calvin Zito gives an overview.
Tags : 
     HPE
By: HPE Intel     Published Date: Jan 11, 2016
Want to know where flash storage technology is heading? Watch Part V of our "Mainstreaming of Flash" video series to hear what's next with this exciting technology! HPE 3PAR StoreServ was built to meet the extreme requirements of massively consolidated cloud service providers. Its remarkable speed—3M+ IOPS—and proven system architecture has been extended to transform mainstream midrange and enterprise deployments, with solutions from a few TBs up to 15PB scale.
Tags : 
     HPE Intel
Start   Previous   1 2 3 4 5 6 7 8 9 10 11 12 13 14 15    Next    End
Search White Papers      

Add White Papers

Get your white papers featured in the insideBIGDATA White Paper Library contact: Kevin@insideHPC.com