For a long time, the industry’s biggest technical challenge was squeezing as many compute cycles as possible out of silicon chips so they could get on with solving the really important, and often gigantic problems in science and engineering faster than was ever thought possible. Now, by clustering computers to work together on problems, scientists are free to consider even larger and more complex real-world problems to compute, and data to analyze.
IBM (NYSE: IBM) announced today that Ping An Insurance (Group) Company of China, Ltd. (Ping An), the largest private insurer in the country, has implemented an IBM Software Defined Storage solution to help it speed data collection from one month to one hour, dramatically improving its ability to meet new regulatory requirements such as the annual audit and industrial data analysis.
As compute speed advanced towards its theoretical maximum, the HPC community quickly discovered that the speed of storage devices and the underlying the Network File System (NFS) developed decades ago had not kept pace. As CPUs got faster, storage became the main bottleneck in high data-volume environments.
Intel’s White Paper, “Architecting a High-Performance Storage System,” shows you the step-by-step process in the design of a Lustre file system. It is available for download at insideBIGDATA White Paper Library. “Although a good system is well-balanced, designing it is not straight forward. Fitting components together and making the adjustments needed for peak performance is challenging. The process begins with a requirement analysis followed by a design structure (a common structure was selected for the paper) and component choices.”
IBM is announcing a new software defined storage-as-a-service on IBM SoftLayer, code named Elastic Storage on Cloud, that gives organizations access to a fully-supported, ready-to-run storage environment, which includes SoftLayer bare metal resources and high performance data management and allows organizations to move data between their on-premise infrastructure and the cloud.
“Dolphin helps companies manage data volume and optimize processes so they can balance the performance and processing capabilities of SAP systems against the cost of running those systems. We develop a data volume management strategy so our customers can keep business critical data in SAP HANA, to get the fast efficient processing they need, and move static or business complete data on to other storage where it is still accessible. With a data volume management strategy in place, our customers are better prepared to go live on HANA and improve their return on investment.”