Enabling Cloud Computing and Server Virtualization with Improved Power Efficiency

Transcription

1 Enabling Cloud Computing and Server Virtualization with Improved Power Efficiency I. Enabling cloud computing and server virtualization without power penalties Enterprise servers are the workhorses of today s data centers, running enterprise applications such as DBMS, CRM, ERP and mail services; executing database queries; and powering search engine operations. In addition, virtualized servers are key to cloud computing. All these functions, of both physical and virtual servers, require massive amounts of memory, as well as bandwidth. However, enterprise servers are challenged today, more than ever, to meet the growing demands for memory capacity and available bandwidth. Scaling server memory technology to higher capacity and bandwidth requires pushing beyond existing technological capabilities, forcing IT and data-center administrators to make trade-offs that prevent optimization. 1

2 As server systems push to higher speeds, the ability to drive more memory becomes a challenge due to increased memory loading. To minimize or eliminate these issues, Inphi Corp. developed a new memory-buffer architecture called Isolation Memory Buffer (imb ) technology that isolates the memory from the CPU. imb technology can scale the server memory modules capacities 2x to 4x while operating at much higher frequencies. This memory scaling is accomplished by using imb technology in a memory module called as Load-Reduced DIMM (LR-DIMM). Data centers using imb technology to scale memory capacity and bandwidth in their servers can increase the utilization of future server base. Accomplishing existing capacities with smaller number of DIMMs or Servers not only reduces the procurement cost of the data-center infrastructure, but also helps to lower the overall powering and cooling cost thereby reducing the total cost of ownership (TCO). II. Options for expanding memory capacity Few options exist that support scaling to higher memory density and higher bandwidth. The key challenge in scaling is that the memory controller is required to drive higher numbers of modules and memory components and therefore higher loads on the Command/Address bus (C/A) and data (also called DQ) that form the memory interface. Current generations of memory controllers can drive up to three memory slots per channel. There is a trade-off between how fast the channel can run vs. the number of modules that are installed because the memory controller must drive longer trace lengths that increase electrical loading on the memory controller drivers. These requirements make it even more difficult to achieve the timing and signal integrity needed to run high-performance applications. Each of the following options provides a way to mitigate this challenge, resulting in varying degrees of capacity and bandwidth scalability. A. Standard RDIMM speed and density baseline RDIMM, or registered DIMM, is an architectural implementation that was introduced to facilitate higher memory capacities and bandwidths, relative to the un-buffered DIMM (UDIMM). A logic component, the register, is the core logic on an RDIMM. It buffers the C/A signals, or the set of signals that is routed to each DRAM component on the DIMM. Relative to an un-buffered DIMM, RDIMM provides higher capacity and faster speeds for single-, dual- and quad-ranked memory modules. 2

3 For the purpose of this discussion, we consider RDIMM as the baseline for speed and density. Below is an illustration of the inverse capacity and frequency scaling for an Enterprise Server test system (Dual Processor) based on DDR3 RDIMM DR3 Server Memory Analysis using 2Gb, x4 DRAM on Enterprise Server Test System (Dual Processor) MT/s 1066 MT/s 1333 MT/s 1600 MT/s Server using RDIMM (in GB) Although RDIMM improves the capacity and frequency scaling to some extent, its performance is reaching its limits. This is because at higher capacities, the data signals (DQ) driven on the memory controller and DRAM interface experience excessive loading. B. Load-Reduced DIMM (LR-DIMM) LR-DIMM seeks to alleviate many of the problems faced by RDIMM by extending the concept of buffering to include data signals as well. LR-DIMM incorporates an imb component that buffers DQ and command, address and control signals on its pre-buffer interface to the memory controller. The result is reduced load on the memory controller as the memory controller has to drive only a single load to the imb component on all its interface signals. The imb component assumes responsibility for interfacing with the memory elements on its post-buffer interface. In addition to reducing the load, imb technology seeks to increase capacity by accessing more DRAM on the post-buffer interface. Using a new option called Rank Multiplication, imb technology enables more ranks of DRAM to be populated on the memory module and seamlessly accessed by the CPU memory controller, thereby accomplishing higher capacities. The above concept allows LR-DIMM to scale to higher capacities and frequencies using cost-effective mainstream DRAM technology. The graph below indicates that at higher operating frequencies, LR-DIMM allows for the server system capacity to increase2x-4x compared to the capacity of RDIMM, at difference bandwidth points. 3

5 C. Proprietary ASICs on the motherboard On some enterprise servers, a concept similar to imb technology is incorporated on the system motherboard itself. A logic memory element is implemented on the system board, also sometimes known as buffer on board (BoB). BoB essentially performs the buffering of the DQ, C/A and control signals on the interface with the memory controller on the motherboard itself. On the DRAM interface, BoB provides increased memory channels that incorporate memory slots. BoB can be a compelling solution for even higher memory capacity. Yet, this comes at the cost of a fixed cost overhead and reduced flexibility for end customers. III. Architecture of the imb technology imb technology is central to the load-reduced DIMM architecture. It is the logic to which all the memory controller interface signals get routed; from there, they are fanned out on the other side of the DRAM interface. imb Besides the load-reduction feature, an imb component also contains logic that allows the memory controller and BIOS to program the drive strengths and the termination to optimal values. imb component also assumes responsibility for ensuring a proper initialization of the DRAM interface during the initial system power-up. Specifically, imb component performs particular functions such as write-leveling and read-training to ensure that the imb and the DRAM devices operate at high speeds within the margins of timing and signal integrity. Following are some of the high-level features of imb technology. 5

6 A. Key imb features and benefits 1. Reduced memory bus load imb technology allows DDR3 server systems to scale to higher memory capacities and operating frequencies by reducing the loading on the memory controller s DQ, C/A and control signals. This distributes the capacitive loading on the memory sub-system uniformly, which enables the memory controller and the rest of the memory sub-system to operate at higher frequencies, even with higher loading. Low-cost multi-sourced DIMMs In conjunction with its imb development, Inphi has been working with many ecosystem players, in particular DRAM vendors, to ensure that higher-capacity DRAM components such as dual-die and quad-die packages are available to the market at the same time as imb components. This coordination will allow 16GB and 32GB LR-DIMM to be available in the first half of Inphi also has championed the imb technology at JEDEC for more than a year in an effort to standardize this architecture for the benefit of the enterprise server market. In addition to making progress within the ecosystem, this standardization effort will ensure multiple sources for the LR-DIMM and make higher capacities cost-effective while establishing a robust supply chain. 2. Support for extending the use of a given memory technology imb technology increases the ranks of memory components on the post-buffer interface, using a concept commonly known as rank multiplication. This means that the controller addresses x number of logical ranks on the pre-buffer interface, and the imb can increase the number of physical ranks on the post-buffer interface to 2x or 4x. In order to fully exploit this capability, DIMM and DRAM vendors are developing a wide range of memory densities using 1Gb (2Gb dual-die packages and 4Gb quad-die packages), 2Gb (4Gb dual-die packages and 8Gb quaddie packages) and 4Gb (8Gb dual-die packages) DRAM densities. This allows increased memory capacity using mainstream x4 memory components, or similar capacities using cheaper x8 memory components both of which help reduce procurement costs by as much as 60%, depending on the lifecycle and cross-over of the current-generation and next-generation memory technologies. 6

8 IV. Greatest impact to Total Cost of Ownership As mentioned in the introduction, data centers of any scale can use imb-based LR-DIMMs to reduce their TCO. By maximizing the memory capacity and bandwidth on the servers, administrators can avoid purchasing additional servers that would otherwise be needed to achieve equivalent capacities and bandwidths. In addition, data centers can lower their power profile by reducing their powering and cooling requirements. The following chart illustrates how data centers of any size can reduce their TCO using imb technology. $9, $8, $7, $6, $5, $4, $3, $2, $1, $0.00 Average savings in hardware cost (in $, k) Average savings in powering and cooling cost (in $, k) Small data-centers (<100 Servers) Small data-centers (Up to1000 Servers) Small data-centers (Up to 10,000 Servers) V. Conclusion Inphi s imb technology is a compelling approach that solves serious challenges for data center servers. imb technology has immediate application in the current generation of DDR3-based servers, while the scalability of the technology ensures a long life and continued application for future DDR3 and DDR4 memory-based servers. The load reduction and rank multiplication features provide 2x 4x increase in the memory capacity of the servers at different frequencies, while using mainstream memory components. Savings on procurement costs of memory components might be as high as 60%, depending on the lifecycle and cross-over of the current-generation and next-generation memory technologies. Overall, the imb-based LR-DIMM technology is poised to significantly reduce the total cost of ownership of data centers by reducing procurement costs as well as costs related to powering and cooling. Inphi is committed to developing imb technology-based products to serve the enterprise needs of today and tomorrow. Inphi Corporation 2393 Townsgate, Suite 101 Westlake Village, CA Phone: (805) Inphi Corporation 3945 Freedom Circle, Suite 1100 Santa Clara, CA Phone: (408) Web: Inphi Corporation. All rights reserved. Inphi, the Inphi logo, and ExacTik are registered trademarks of Inphi Corporation. imb is a trademark of Inphi Corporation. Updated 09/07/2010

Super Micro Computer, Inc. Memory Configuration Guide X9 Socket R Series DP Motherboards Version 1.1R March 12, 2012 Introduction This document is designed to provide the reader with an easy-to-use guide

WHITE PAPER Sponsored by High Speed Memory Interface Chipsets Let Server Performance Fly SUMMARY The demands on server performance continue to increase at a tremendous pace. New requirements from large

Kingston Technology KingstonConsult WHD Local October 2014 2012 Kingston Technology Corporation. All rights reserved. All trademarks and registered trademarks are the property of their respective owners.

An evaluation of blade server power efficiency for the, Dell PowerEdge M600, and IBM BladeCenter HS21 using the SPECjbb2005 Benchmark The HP Difference The ProLiant BL260c G5 is a new class of server blade

Overview Is finding reliable DRAM memory for your HP ProLiant Server series in your data center a major challenge? When you choose HP SmartMemory you get the same high quality, reliability, and confidence

Achieving a High Performance OLTP Database using SQL Server and Dell PowerEdge R720 with This Dell Technical White Paper discusses the OLTP performance benefit achieved on a SQL Server database using a

KINGSTON.COM Best Practices Server: Performance Benchmark Memory channels, frequency and performance Although most people don t realize it, the world runs on many different types of databases, all of which

Intel Virtualization and Server Technology Update Petar Torre Lead Architect Service Provider Group 29 March 2012 1 Legal Disclaimer Intel may make changes to specifications and product descriptions at

THE CHALLENGE Meeting database performance demands. Increasing data that needs to be read quickly can be bottlenecked by the reluctance to add memory due to cost. THE SOLUTION The SQL Server 214 BPE feature

White Paper EMC XtremSF: Delivering Next Generation Performance for Oracle Database Abstract This white paper addresses the challenges currently facing business executives to store and process the growing

Architecture Enterprise Storage Performance: It s All About The Interface. A DIABLO WHITE PAPER APRIL 214 diablo-technologies.com Diablo_Tech Enterprise Storage Performance: It s All About The Architecture.

How Microsoft Designs its Cloud-Scale Servers How Microsoft Designs its Cloud-Scale Servers Page 1 How Microsoft Designs its Cloud-Scale Servers How is cloud infrastructure server hardware design different

Front cover Quantifying the Power Savings by Upgrading to DDR4 Memory on Lenovo Servers Demonstrates how clients can significantly save on operational expense by using DDR4 memory Compares DIMMs currently

Transition Guide November 2010 2 Introduction Key points Apple will not be developing a future version of Orders for will be accepted through January 31, 2011 Apple will honor all warranties and extended

: Moving Storage to The Memory Bus A Technical Whitepaper By Stephen Foskett April 2014 2 Introduction In the quest to eliminate bottlenecks and improve system performance, the state of the art has continually

Dell Reliable Memory Technology Detecting and isolating memory errors THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL ERRORS AND TECHNICAL INACCURACIES. THE CONTENT IS

High Performance NVMe JBOF The High Capacity, Fastest and Most Modular Direct Attached NVMe Storage Device for VDI and HPC BLAZING 12GB/s SPEED 8 DRIVES [1U ] WORKS WITH ANY SERVER The Magma StorageBox

HP and NX Introduction The purpose of this document is to provide information that will aid in selection of HP Workstations for running Siemens PLMS NX. A performance study was completed by benchmarking

IBM Chipkill Memory Advanced ECC Memory for the IBM Netfinity 7000 M10 IBM Chipkill Memory (Chipkill), part of the IBM Netfinity X-architecture initiative, will be available during the first quarter of

CHAPTER 1 INTRODUCTION 1.1 Background The command over cloud computing infrastructure is increasing with the growing demands of IT infrastructure during the changed business scenario of the 21 st Century.

Cloud computing is a compelling way to deliver web-based and non-web-based applications that better utilize the physical infrastructure, while lowering costs by moving from silos of expensive customized

Motherboard- based Servers versus ATCA- based Servers Summary: A comparison of costs, features and applicability for telecom application hosting After many years of struggling for market acceptance, it

Building High-Performance iscsi SAN Configurations An Alacritech and McDATA Technical Note Building High-Performance iscsi SAN Configurations An Alacritech and McDATA Technical Note Internet SCSI (iscsi)

Juniper Networks QFabric: Scaling for the Modern Data Center Executive Summary The modern data center has undergone a series of changes that have significantly impacted business operations. Applications

TECHNOLOGY CONNECTED Advances with System Area Network Speeds Data Transfer between Servers with A new network switch technology is targeted to answer the phenomenal demands on intercommunication transfer