systems

Results 1 - 25 of 3483Sort Results By: Published Date | Title | Company Name
By: NetApp     Published Date: Dec 13, 2013
Interested in running a Hadoop proof of concept on enterprise-class storage? Download this solutions guide to get a technical overview on building Hadoop on NetApp E-series storage. NetApp Open Solution for Hadoop delivers big analytics with preengineered, compatible, and supported solutions based on high-quality storage platforms so you reduce the cost, schedule, and risk of do-it-yourself systems and relieving the skills gap most organizations have with Hadoop. See how on going operational and maintenance costs can be reduced with a high available and scalable Hadoop solution.
Tags : open solutions, hadoop solutions guide
     NetApp
By: IBM     Published Date: Sep 02, 2014
Research teams using next-generation sequencing (NGS) technologies face the daunting challenge of supporting compute-intensive analysis methods against petabytes of data while simultaneously keeping pace with rapidly evolving algorithmic best practices. NGS users can now solve these challenges by deploying the Accelrys Enterprise Platform (AEP) and the NGS Collection on optimized systems from IBM. Learn how you can benefit from the turnkey IBM Application Ready Solution for Accelrys with supporting benchmark data.
Tags : ibm, accelrys, turnkey ngs solution
     IBM
By: IBM     Published Date: May 20, 2015
Software defined storage is enterprise class storage that uses standard hardware with all the important storage and management functions performed in intelligent software. Software defined storage delivers
Tags : 
     IBM
By: IBM     Published Date: Sep 16, 2015
The IBM Spectrum Scale solution provided for up to 11x better throughput results than EMC Isilon for Spectrum Protect (TSM) workloads. Using published data, Edison compared a solution comprised of EMC® Isilon® against an IBM® Spectrum Scale™ solution. (IBM Spectrum Scale was formerly IBM® General Parallel File System™ or IBM® GPFS™, also known as code name Elastic Storage). For both solutions, IBM® Spectrum Protect™ (formerly IBM Tivoli® Storage Manager or IBM® TSM®) is used as a common workload performing the backups to target storage systems evaluated.
Tags : 
     IBM
By: Storiant     Published Date: Mar 16, 2015
Read this new IDC Report about how today's enterprise datacenters are dealing with new challenges that are far more demanding than ever before. Foremost is the exponential growth of data, most of it unstructured data. Big data and analytics implementations are also quickly becoming a strategic priority in many enterprises, demanding online access to more data, which is retained for longer periods of time. Legacy storage solutions with fixed design characteristics and a cost structure that doesn't scale are proving to be ill-suited for these new needs. This Technology Spotlight examines the issues that are driving organizations to replace older archive and backup-and-restore systems with business continuity and always-available solutions that can scale to handle extreme data growth while leveraging a cloudbased pricing model. The report also looks at the role of Storiant and its long-term storage services solution in the strategically important long-term storage market.
Tags : storiant, big data, analytics implementations, cloudbased pricing model, long-term storage services solution, long-term storage market
     Storiant
By: RYFT     Published Date: Apr 03, 2015
The new Ryft ONE platform is a scalable 1U device that addresses a major need in the fast-growing market for advanced analytics — avoiding I/O bottlenecks that can seriously impede analytics performance on today's hyperscale cluster systems. The Ryft ONE platform is designed for easy integration into existing cluster and other server environments, where it functions as a dedicated, high-performance analytics engine. IDC believes that the new Ryft ONE platform is well positioned to exploit the rapid growth we predict for the high-performance data analysis market.
Tags : ryft, ryft one platform, 1u deivce, advanced analytics, avoiding i/o bottlenecks, idc
     RYFT
By: MEMSQL     Published Date: Apr 12, 2016
The pace of data is not slowing. Applications of today are built with infinite data sets in mind. As these real-time applications become the norm, and batch processing becomes a relic of the past, digital enterprises will implement memory-optimized, distributed data systems to simplify Lambda Architectures for real-time data processing and exploration.
Tags : 
     MEMSQL
By: Data Direct Networks     Published Date: Apr 08, 2014
DataDirect Networks (DDN), the largest privately-held provider of high-performance storage, has a large and growing presence in HPC markets. HPC users identify DDN as their storage provider more than any other storage-focused company, with twice the mentions of EMC, and more the twice the mentions of NetApp, Hitachi Data Systems, or Panasas.(5) DDN’s strength in HPC is anchored by its Storage Fusion Architecture (SFA), winner of the HPCwire Editor’s Choice Award for “Best HPC Storage Product or Technology” in each of the past three years. The DDN SFA12KX combines SATA, SAS, and solid-state disks (SSDs) for an environment that can be tailored to a balance of throughput and capacity
Tags : 
     Data Direct Networks
By: IBM     Published Date: Feb 13, 2015
Software defined storage is enterprise class storage that uses standard hardware with all the important storage and management functions performed in intelligent software. Software defined storage delivers automated, policy-driven, application-aware storage services through orchestration of the underlining storage infrastructure in support of an overall software defined environment.
Tags : 
     IBM
By: Intel     Published Date: Aug 06, 2014
Powering Big Data Workloads with Intel® Enterprise Edition for Lustre* software The Intel® portfolio for high-performance computing provides the following technology solutions: • Compute - The Intel® Xeon processor E7 family provides a leap forward for every discipline that depends on HPC, with industry-leading performance and improved performance per watt. Add Intel® Xeon Phi coprocessors to your clusters and workstations to increase performance for highly parallel applications and code segments. Each coprocessor can add over a teraflops of performance and is compatible with software written for the Intel® Xeon processor E7 family. You don’t need to rewrite code or master new development tools. • Storage - High performance, highly scalable storage solutions with Intel® Lustre and Intel® Xeon Processor E7 based storage systems for centralized storage. Reliable and responsive local storage with Intel® Solid State Drives. • Networking - Intel® True Scale Fabric and Networking technologies – Built for HPC to deliver fast message rates and low latency. • Software and Tools: A broad range of software and tools to optimize and parallelize your software and clusters. Further, Intel Enterprise Edition for Lustre software is backed by Intel, the recognized technical support providers for Lustre, and includes 24/7 service level agreement (SLA) coverage.
Tags : 
     Intel
By: Dell and Intel®     Published Date: Aug 24, 2015
New technologies help decision makers gain insights from all types of data - from traditional databases to high-visibility social media sources. Big data initiatives must ensure data is cost-effectively managed, shared by systems across the enterprise, and quickly and securely made available for analysis and action by line-of-business teams. In this article, learn how Dell working with Intel® helps IT leaders overcome the challenges of IT and business alignment, resource constraints and siloed environments through a comprehensive big data portfolio based on choice and flexibility, redefined economics and connected intelligence.
Tags : 
     Dell and Intel®
By: Dell and Intel®     Published Date: Aug 24, 2015
Organizations working at gathering insights from vast volumes of varied data types understand that they need more than traditional, structured systems, and tools. This paper discusses how the many Dell | Cloudera Hadoop solutions help organizations of all sizes, and with a variety of needs and use cases, tackle their big data requirements.
Tags : 
     Dell and Intel®
By: Dell and Intel®     Published Date: Aug 24, 2015
To extract value from an ever-growing onslaught of data, your organization needs next-generation data management, integration, storage and processing systems that allow you to collect, manage, store and analyze data quickly, efficiently and cost-effectively. That’s the case with Dell| Cloudera® Apache™ Hadoop® solutions for big data. These solutions provide end-to-end scalable infrastructure, leveraging open source technologies, to allow you to simultaneously store and process large datasets in a distributed environment for data mining and analysis, on both structured and unstructured data, and to do it all in an affordable manner.
Tags : 
     Dell and Intel®
By: Kx Systems     Published Date: Jan 16, 2015
?Kdb+ is a column-based relational database with extensive in-memory capabilities, developed and marketed by Kx Systems. Like all such products, it is especially powerful when it comes to supporting queries and analytics. However, unlike other products in this domain, kdb+ is particularly good (both in terms of performance and functionality) at processing, manipulating and analysing data (especially numeric data) in real-time, alongside the analysis of historical data. Moreover, it has extensive capabilities for supporting timeseries data. For these reasons Kx Systems has historically targeted the financial industry for trading analytics and black box trading based on real-time and historic data, as well as realtime risk assessment; applications which are particularly demanding in their performance requirements. The company has had significant success in this market with over a hundred major financial institutions and hedge funds deploying its technology. In this paper, however, we wa
Tags : kx systems, kdb+, relational database
     Kx Systems
By: snowflake     Published Date: Jun 09, 2016
Today’s data, and how that data is used, have changed dramatically in the past few years. Data now comes from everywhere—not just enterprise applications, but also websites, log files, social media, sensors, web services, and more. Organizations want to make that data available to all of their analysts as quickly as possible, not limit access to only a few highly-skilled data scientists. However, these efforts are quickly frustrated by the limitations of current data warehouse technologies. These systems simply were not built to handle the diversity of today’s data and analytics. They are based on decades-old architectures designed for a different world, a world where data was limited, users of data were few, and all processing was done in on-premises datacenters.
Tags : 
     snowflake
By: snowflake     Published Date: Jun 09, 2016
THE CHALLENGE: DATA SOLUTIONS CAN’T KEEP PACE WITH DATA NEEDS Organizations are increasingly dependent on diff erent types of data to make successful business decisions. But as the volume, rate, and types of data expand and become less predictable, conventional data warehouses cannot consume all this data eff ectively. Big data solutions like Hadoop increase the complexity of the environment and generally lack the performance of traditional data warehouses. This makes it difficult, expensive, and time-consuming to manage all the systems and the data.
Tags : 
     snowflake
By: BitStew     Published Date: May 26, 2016
The heaviest lift for an industrial enterprise is data integration, the Achilles’ heel of the Industrial Internet of Things (IIoT). Companies are now recognizing the enormous challenge involved in supporting Big Data strategies that can handle the data that is generated by information systems, operational systems and the extensive networks of old and new sensors. To compound these issues, business leaders are expecting data to be captured, analyzed and used in a near real-time to optimize business processes, drive efficiency and improve profitability. However, integrating this vast amount of dissimilar data into a unified data strategy can be overwhelming for even the largest organizations. Download this white paper, by Bit Stew’s Mike Varney, to learn why a big data solution will not get the job done. Learn how to leverage machine intelligence with a purpose-built IIoT platform to solve the data integration problem.
Tags : 
     BitStew
By: Spectrum Enterprise     Published Date: Oct 10, 2018
To thrive in today’s global business economy, you need voice systems that keep your clients connected and your employees productive. Spectrum Enterprise provides next generation communications solutions that deliver superior quality and cost advantages with modern collaboration features to keep your workforce and clients connected.
Tags : 
     Spectrum Enterprise
By: Ricoh     Published Date: Oct 02, 2018
Your business is changing. As a finance leader, you know that accounting is a labour-intensive, costly process where systems often don’t allow for expedient exception handling and many days are fraught with difficulty in matching invoices to other databases for reconciliation. Like most companies, you know where you want to go but may not have infrastructure or internal expertise to handle electronic fund transfers, credit card payments or cheque processing— all the pieces required to make your vision for an efficient, integrated operation a reality.
Tags : 
     Ricoh
By: BigCommerce     Published Date: Oct 16, 2018
Businesses who have lived through the evolution of the digital age are well aware that we’ve experienced a generational shift in technology. The rise of software as a service (SaaS), cloud, mobile, big data, the Internet of Things (IoT), social media, and other technologies have disrupted industries and changed customers’ expectations. In our always-on, buy anything anywhere world, customers want their shopping experiences to be personalized, dynamic, and convenient. As a result, many businesses are trying to reinvent themselves. Success in a fast-paced economy depends on continually adapting and innovating. Companies have to move quickly to keep up; there’s no time for disjointed technologies and old systems that don’t serve the customer-obsessed mentality needed to thrive in the digital age.
Tags : 
     BigCommerce
By: BigCommerce     Published Date: Oct 16, 2018
Whether your company has been selling online for 20 minutes or 20 years, you are undoubtedly familiar with the PCI DSS (Payment Card Industry Data Security Standard). It requires merchants to create security management policies and procedures for safeguarding customers’ payment data. Originally created by Visa, MasterCard, Discover, and American Express in 2004, the PCI DSS has evolved over the years to ensure online sellers have the systems and processes in place to prevent a data breach.
Tags : 
     BigCommerce
By: Akamai Technologies     Published Date: Nov 02, 2018
What to expect and how to prepare your best defense DDoS attackers launch hundreds of DDoS attacks every day. When a DDoS attack strikes an organization, panic is common. No one knows why systems and applications are failing, who to call, or what questions to ask. Meanwhile, senior leaders want answers. Planning ahead and being prepared is your best defense against DDoS attacks. This white paper explores best practices for building and maintaining a DDoS protection plan. You’ll learn what to expect from a DDoS attack, how to prepare, what to put in a DDoS runbook, and how to test your plan. In viewing this Akamai content, we would like to share your data with Akamai. Click here for more info or to opt out.
Tags : 
     Akamai Technologies
By: Aberdeen     Published Date: Jun 17, 2011
Download this paper to learn the top strategies leading executives are using to take full advantage of the insight they receive from their business intelligence (BI) systems - and turn that insight into a competitive weapon.
Tags : aberdeen, michael lock, data-driven decisions, business intelligence, public sector, analytics, federal, state, governmental, decisions, data management
     Aberdeen
By: Zebra Technologies     Published Date: Oct 01, 2018
The consistent and accurate identification of patients has long presented a challenge for healthcare organizations across the globe. Why? Because the process of patient identification remains largely manual and persists across the entire patient journey. As such, it is highly prone to errors and can lead to serious patient harm. Automated patient identification systems are essential to correcting this long-standing problem and improving the health and well-being of patients as well as protecting the integrity of the healthcare system.
Tags : 
     Zebra Technologies
By: Dell     Published Date: Nov 12, 2018
Today, IT leaders address the PC lifecycle across a continuum from control to transformation. Control is geared to optimization, while transformation focuses on the business impact of technology. Though the two approaches differ, they are not in opposition. They strive for the same goals and face similar challenges. As IT leaders provide their workforce with the tools to carry out the corporate mission, they should develop a PC lifecycle strategy that encompasses the key organizational needs of systems management, end-user productivity, business innovation and data-centric security. Read this Dell whitepaper to learn more about the findings of a recent Forrester Consulting study, “Digital Transformers Innovate, Digital Controllers Optimize”. This paper will help clarify the PC lifecycle continuum, from the basics of control to the advanced levels of transformation, so you will be better equipped to determine the needs of your organization on that spectrum.
Tags : 
     Dell
Start   Previous   1 2 3 4 5 6 7 8 9 10 11 12 13 14 15    Next    End
Search White Papers      

Add White Papers

Get your white papers featured in the insideBIGDATA White Paper Library contact: Kevin@insideHPC.com