Extending RDMA for Persistent Memory over Fabrics

For datacenter applications requiring low-latency access to persistent storage, byte-addressable persistent memory (PM) technologies like 3D XPoint and MRAM are attractive solutions. Network-based access to PM, labeled here PM over Fabrics (PMoF), is driven by data scalability and/or availability requirements. Remote Direct Memory Access (RDMA) network protocols are a good match for PMoF, allowing direct RDMA Write of data to remote PM. However, the completion of an RDMA Write at the sending node offers no guarantee that data has reached persistence at the target. This webcast will outline extensions to RDMA protocols that confirm such persistence and additionally can order successive writes to different memories within the target system.

The primary target audience is developers of low-latency and/or high-availability datacenter storage applications. The presentation will also be of broader interest to datacenter developers, administrators and users.

After you watch, check-out our Q&A blog from the webcast: http://bit.ly/2DFE7SL

For a long time, the architecture and best practices of storage networks have been relatively well-understood. Recently, however, advanced capabilities have been added to storage that could have broader impacts on networks than we think.

The three main storage network transports - Fibre Channel, Ethernet, and InfiniBand – all have mechanisms to handle increased traffic, but they are not all affected or implemented the same way. For instance, placing a protocol such as NVMe over Fabrics can mean very different things when looking at one networking method in comparison to another.

Unfortunately, many network administrators may not understand how different storage solutions place burdens upon their networks. As more storage traffic traverses the network, customers face the risk of congestion leading to higher-than-expected latencies and lower-than expected throughput. Watch this webinar to learn:

•Typical storage traffic patterns
•What is Incast, what is head of line blocking, what is congestion, what is a slow drain, and when do these become problems on a network?
•How Ethernet, Fibre Channel, InfiniBand handle these effects
•The proper role of buffers in handling storage network traffic
•Potential new ways to handle increasing storage traffic loads on the network

In the short period from 2014-2018, Ethernet equipment vendors have announced big increases in line speeds, shipping 25, 50, and 100 Gigabits-per -second (Gb/s) speeds and announcing 200/400 Gb/s. At the same time Fibre Channel vendors have launched 32GFC, 64GFC and 128GFC technology while InfiniBand has reached 200Gb/s (called HDR) speeds.

But who exactly is asking for these faster new networking speeds, and how will they use them? Are there servers, storage, and applications that can make good use of them? How are these new speeds achieved? Are new types of signaling, cables and transceivers required? How will changes in PCIe standards keep up? And do the faster speeds come with different distance limitations?

Watch this SNIA Networking Storage Forum (NSF) webcast to learn how these new speeds are achieved, where they are likely to be deployed for storage, and what infrastructure changes are needed to support them.

Traditionally, much of the IT infrastructure that we’ve built over the years can be divided fairly simply into storage (the place we save our persistent data), network (how we get access to the storage and get at our data) and compute (memory and CPU that crunches on the data). In fact, so successful has this model been that a trip to any cloud services provider allows you to order (and be billed for) exactly these three components.

We build effective systems in a cost-optimal way by using appropriate quantities of expensive and fast memory (DRAM for instance) to cache our cheaper and slower storage. But currently fast memory has no persistence at all; it’s only storage that provides the application the guarantee that storing, modifying or deleting data does exactly that.

Memory and storage differ in other ways. For example, we load from memory to registers on the CPU, perform operations there, and then store the results back to memory by using byte addresses. This load/store technology is different from storage, where we tend to move data back and fore between memory and storage in large blocks, by using an API (application programming interface).

New memory technologies are challenging these assumptions. They look like storage in that they’re persistent, if a lot faster than traditional disks or even Flash based SSDs, but we address them in bytes, as we do memory like DRAM, if more slowly. Persistent memory (PM) lies between storage and memory in latency, bandwidth and cost, while providing memory semantics and storage persistence. In this webcast, SNIA experts will discuss:

•Traditional uses of storage and memory as a cache
•How can we build and use systems based on PM?
•What would a system with storage, persistent memory and DRAM look like?
•Do we need a new programming model to take advantage of PM?
•Interesting use cases for systems equipped with PM
•How we might take better advantage of this new technology

Kubernetes (k8s) is an open-source system for automating the deployment, scaling, and management of containerized applications. Kubernetes promises simplified management of cloud workloads at scale, whether on-premises, hybrid, or in a public cloud infrastructure, allowing effortless movement of workloads from cloud to cloud. By some reckonings, it is being deployed at a rate several times faster than virtualization.

In this presentation, we’ll introduce Kubernetes and present use cases that make clear where and why you would want to use it in your IT environment. We’ll also focus on the enterprise requirements of orchestration and containerization, and specifically on the storage aspects and best practices.

•What is Kubernetes? Why would you want to use it?
•How does Kubernetes help in a multi-cloud/private cloud environment?
•How does Kubernetes orchestrate & manage storage? Can Kubernetes use Docker?
•How do we provide persistence and data protection?
•Example use cases

Has hybrid cloud reached a tipping point? According to research from the Enterprise Strategy Group (ESG), IT organizations today are struggling to strike the right balance between public cloud and their on-premises infrastructure. In this SNIA webcast, ESG senior analyst, Scott Sinclair, will share research on current cloud trends, covering:

We’re all accustomed to transferring money from one bank account to another; a credit to the payer becomes a debit to the payee. But that model uses a specific set of sophisticated techniques to accomplish what appears to be a simple transaction. We’re also aware of how today we can order goods online, or reserve an airline seat over the Internet. Or even simpler, we can update a photograph on Facebook. Can these applications use the same models, or are new techniques required?

One of the more important concepts in storage is the notion of transactions, which are used in databases, financials, and other mission critical workloads. However, in the age of cloud and distributed systems, we need to update our thinking about what constitutes a transaction. We need to understand how new theories and techniques allow us to undertake transactional work in the face of unreliable and physically dispersed systems. It’s a topic full of interesting concepts (and lots of acronyms!). In this webcast, we’ll provide a brief tour of traditional transactional systems and their use of storage, we’ll explain new application techniques and transaction models, and we’ll discuss what storage systems need to look like to support these new advances.

And yes, we’ll explain all the acronyms and nomenclature too.

You will learn:

• A brief history of transactional systems from banking to Facebook
• How the Internet and distributed systems have changed and how we view transactions
• An explanation of the terminology, from ACID to CAP and beyond
• How applications, networks & particularly storage have changed to meet these demands

Join SNIA Solid State Storage Initiative (SSSI) Education Chair and leading analyst Tom Coughlin and SSSI member Motti Beck of Mellanox Technologies for a journey into the requirements and trends in worldwide data storage for entertainment content acquisition, editing, archiving, and digital preservation. This webcast will cover capacity and performance trends and media projections for direct attached storage, cloud, and near-line network storage. It will also include results from a long-running digital storage survey of media and entertainment professionals. Learn what is needed for digital cinema, broadcast, cable, and internet applications and more.

This webcast will present an overview of scale-out file system architectures. To meet the increasingly higher demand on both capacity and performance in large cluster computing environments, the storage subsystem has evolved toward a modular and scalable design. The scale-out file system is one implementation of the trend, in addition to scale-out object and block storage solutions. This presentation will provide an introduction to scale-out-file systems and cover:

The SNIA Swordfish™ specification for the management of storage systems and data services is an extension of the DMTF Redfish® specification. Together, these specifications provide a unified approach for the management of servers and storage in converged, hyper-converged, hyperscale and cloud infrastructure environments.

To help speed your Swordfish development efforts, SNIA has produced open source storage management tools available now on GitHub for your use. Join this session for an overview of these open source tools, which include a Swordfish API Emulator, a Swordfish Basic Web Client, an example Swordfish plugin for the Microsoft Power BI business analytics service, and an example Swordfish plugin for the Datadog monitoring service.

Containers are a big trend in application deployment. The landscape of containers is moving fast and constantly changing, with new standards emerging every few months. Learn what’s new, what to pay attention to, and how to make sense of the ever-shifting container landscape.

This live webcast will cover:
•Container storage types and Container Frameworks
•An overview of the various storage APIs for the container landscape
•How to identify the most important projects to follow in the container world
•The Container Storage Interface spec and Kubernetes 1.13
•How to get involved in the container community

After you watch the webcast, check out the Q&A blog at http://bit.ly/2GPkFET

Cloud data centers are by definition very dynamic. The need for infrastructure availability in the right place at the right time for the right use case is not as predictable, nor as static, as it has been in traditional data centers. These cloud data centers need to rapidly construct virtual pools of compute, network and storage based on the needs of particular customers or applications, then have those resources dynamically and automatically flex as needs change. To accomplish this, many in the industry espouse composable infrastructure capabilities, which rely on heterogeneous resources with specific capabilities which can be discovered, managed, and automatically provisioned and re-provisioned through data center orchestration tools. The primary benefit of composable infrastructure results in a smaller grained sets of resources that are independently scalable and can be brought together as required. In this webcast, SNIA experts will discuss:

•What prompted the development of composable infrastructure?
•What are the solutions?
•What is composable infrastructure?
•Enabling technologies (not just what’s here, but what’s needed…)
•Status of composable infrastructure standards/products
•What’s on the horizon – 2 years? 5 Years
•What it all means

One of the great advantages of Hyperconvergence infrastructures (also known as “HCI”) is that, relatively speaking, they are extremely easy to set up and manage. In many ways, they’re the “Happy Meals” of infrastructure, because you have compute and storage in the same box. All you need to do is add networking.

In practice, though, many consumers of HCI have found that the “add networking” part isn’t quite as much of a no-brainer as they thought it would be. Because HCI hides a great deal of the “back end” communication, it’s possible to severely underestimate the requirements necessary to run a seamless environment. At some point, “just add more nodes” becomes a more difficult proposition.

In this webinar, we’re going to take a look behind the scenes, peek behind the GUI, so to speak. We’ll be talking about what goes on back there, and shine the light behind the bezels to see:

•The impact of metadata on the network
•What happens as we add additional nodes
•How to right-size the network for growth
•Tricks of the trade from the networking perspective to make your HCI work better
•And more…

Now, not all HCI environments are created equal, so we’ll say in advance that your mileage will necessarily vary. However, understanding some basic concepts of how storage networking impacts HCI performance may be particularly useful when planning your HCI environment, or contemplating whether or not it is appropriate for your situation in the first place.

After you watch the webcast, check out the Q&A blog at http://bit.ly/2Va4wwH

When it comes to storage, a byte is a byte is a byte, isn’t it? One of the enduring truths about simplicity is that scale makes everything hard, and with that comes complexity. And when we’re not processing the data, how do we store it and access it?

In this webcast, we will compare three types of data access: file, block and object storage, and the access methods that support them. Each has its own set of use cases, and advantages and disadvantages. Each provides simple to sophisticated management of the data, and each makes different demands on storage devices and programming technologies.

Perhaps you’re comfortable with block and file, but are interested in investigating the more recent class of object storage and access. Perhaps you’re happy with your understanding of objects, but would really like to understand files a bit better, and what advantages or disadvantages they have compared to each other. Or perhaps you want to understand how file, block and object are implemented on the underlying storage systems – and how one can be made to look like the other, depending on how the storage is accessed. Join us as we discuss and debate:

In the storage world, NVMe™ is arguably the hottest thing going right now. Go to any storage conference – either vendor- or vendor-neutral, and you’ll see NVMe as the latest and greatest innovation. It stands to reason, then, that when you want to run NVMe over a network, you need to understand NVMe over Fabrics (NVMe-oF).

TCP – the long-standing mainstay of networking – is the newest transport technology to be approved by the NVM Express organization. This can mean really good things for storage and storage networking – but what are the tradeoffs?

In this webinar, the lead author of the NVMe/TCP specification, Sagi Grimberg, and J Metz, member of the SNIA and NVMe Boards of Directors, will discuss:
•What is NVMe/TCP
•How NVMe/TCP works
•What are the trade-offs?
•What should network administrators know?
•What kind of expectations are realistic?
•What technologies can make NVMe/TCP work better?
•And more…

After the webcast, check out the Q&A blog http://sniaesfblog.org/author-of-nvme-tcp-spec-answers-your-questions/

With all the different storage arrays and connectivity protocols available today, knowing the best practices can help improve operational efficiency and ensure resilient operations. VMware’s storage global service has reported many of the common service calls they receive. In this webcast, we will share those insights and lessons learned by discussing:
- Common mistakes when setting up storage arrays
- Why iSCSI is the number one storage configuration problem
- Configuring adapters for iSCSI or iSER
- How to verify your PSP matches your array requirements
- NFS best practices
- How to maximize the value of your array and virtualization
- Troubleshooting recommendations

After you watch the webcast, check out the Q&A blog at http://bit.ly/2WjmFJW

Kick off the new year with a new SNIA Persistent Memory and NVDIMM Special Interest Group webcast on how applications can take advantage of Persistent Memory today with NVDIMM - the go-to Persistent Memory technology for boosting performance for next generation storage platforms. NVDIMM standards have paved the way to simple, plug-n-play solutions. If you're a developer or integrator who hasn't yet realized the benefits of NVDIMMs in your products, you will want to attend to learn about NVDIMM functionality, applications, and benefits. You'll come away with an understanding of how NVDIMMs fit into the persistent memory landscape.

Join SSSI members and respected analysts Tom Coughlin and Jim Handy for a look into their new Emerging Memory and Storage Technologies Report. Tom and Jim will examine emerging memory technologies and their interaction with standard memories, how a new memory layer improves computer performance, and the technical advantages and economies of scale that contribute to the enthusiasm for emerging memories. They will provide an outlook on market projections and enabling and driving applications. The webcast is the perfect preparation for the 2019 SNIA Persistent Memory Summit January 24, 2019.

If you’re a storage equipment vendor, management software vendor or end-user of the ISO approved SNIA Storage Management Initiative Specification (SMI-S), you won’t want to miss this presentation. Enterprise storage industry expert Mike Walker will provide an overview of new indications, methods, properties and profiles of SMI-S 1.7 and the newly introduced version, SMI-S 1.8. If you haven’t yet made the jump to SMI-S 1.7, Walker will explain why it’s important to go directly to SMI-S 1.8.

The SNIA Swordfish™ specification helps to provide a unified approach for the management of storage and servers in hyperscale and cloud infrastructure environments, making it easier for IT administrators to integrate scalable solutions into their data centers. Swordfish builds on the Distributed Management Task Force’s (DMTF’s) Redfish® specification using the same easy-to-use RESTful methods and lightweight JavaScript Object Notation (JSON) formatting. Join this session to receive an overview of Swordfish including the new functionality added in version 1.0.6 released in March, 2018.

Scale-out storage is increasingly popular for cloud, high-performance computing, machine learning, and certain enterprise applications. It offers the ability to grow both capacity and performance at the same time and to distribute I/O workloads across multiple machines.

But unlike traditional local or scale-up storage, scale-out storage imposes different and more intense workloads on the network. Clients often access multiple storage servers simultaneously; data typically replicates or migrates from one storage node to another; and metadata or management servers must stay in sync with each other as well as communicating with clients. Due to these demands, traditional network architectures and speeds may not work well for scale-out storage, especially when it’s based on flash. Join this webinar to learn:

•Scale-out storage solutions and what workloads they can address
•How your network may need to evolve to support scale-out storage
•Network considerations to ensure performance for demanding workloads
•Key considerations for all flash

After you watch the webcast, check out the Q&A blog: http://bit.ly/scale-out-q-a

The Storage Networking Industry Association (SNIA) is a non-profit organization made up of member companies spanning information technology. A globally recognized and trusted authority, SNIA’s mission is to lead the storage industry in developing and promoting vendor-neutral architectures, standards and educational services that facilitate the efficient management, movement and security of information.