Home Blog Page 416

Veaam triumvirate takes charge – again

With the dust settling after Peter McKay’s October 2018 departure from being Veeam’s CEO how has the company changed and where is it going?

A conversation with Ratmir Timashev, co-founder of the Veeam rocketship-like data protection business, and now its EVP for sales and marketing, reveals that, in a way, Veeam has swung back to its roots while talking advantage of the business extension that McKay built.

Building the enterprise business

When McKay became co-CEO and President in May 2017 his brief included building up Veeam’s enterprise business. It already had a successful core small, medium and mid-level enterprise transactional business which had broken records with its growth rate capitalising on the VMware virtual server wave washing over the industry. The deal sizes were $25,000 to $150,000 and business was consistently booming, but the feeling was there that they were leaving money on the table, so to speak, by not selling to larger enterprises.

That was for McKay to do, and he did so, establishing Veeam as an enterprise supplier and building up its North American operation. He moved quickly and spent money and organisational effort reshaping Veeam. A little too quickly as it happened.

Ratmir Timashev

Timashev explains that the changes took investment away from the main Veeam business with marketing to the core transactional business customers suffering as did inside sales.

So, eighteen months after joining Veeam, McKay left, and three main executives, the tight triumvirate of ; Timashev, co-founder Andrei Baranov who was previously a CTO and then co-CEO with McKay, and Bill Largent, a senior board member from May 2017 after being CEO, resumed operational control, running Veeam in the same kind of way as they had before Mckay’s time.

Enterprise – SME balance

Baranov is now the CEO with Largent the EVP for operations. The three have set up a 70:30 balance scheme for spending and investment.

Seventy per cent goes to the core transactional business, Veeam’s high-velocity business, focussing on TDMs (Technical Decision Makers) in the small, medium and mid-level enterprises.

The remaining thirty per cent is focussed towards EDMs or Executive Decision Makers in the larger enterprises.

Timashev says very few businesses successfully sell to the whole range of small-to-medium-to-mid-to-large-enterprise customers. It’s hard to do with products and its also hard to do with sales and marketing. He cites Amazon and Microsoft as having managed to get over that high bar.

It’s clear he wants Veeam to be in that category.

Hybrid cloud moves

All customers are aware of and generally pursuing a hybrid cloud strategy, meaning cloud-style on-premises IT infrastructure combined with using the public cloud, and generally more than one public cloud. Their board directors and executives have charged the IT department with finding ways to leverage the cost and agility advantages of the public cloud.

For Veeam, with it being a predominantly on-premises data protection modernisation business, this hybrid cloud move represents an opportunity but also a threat. Get it wrong and its customers could buy data protection services elsewhere.

Timashev said Veeam is pursuing its own hybrid cloud strategy. All new products have a subscription license model. The main data protection product still has a perpetual license model but is also available on subscription.

Subscription licenses can be moved from on-premises to the public cloud at no cost with Veeam Instance Licensing. VIL is portable and can be used to protect various workloads across multiple Veeam products and can be used on premises, in the public cloud and anywhere in between. Timashev believes Veeam is the only data protection vendor to offer this capability.

Veeam has also added tiering (aka retiring) of older data to the AWS and Azure clouds, or to on-premises object storage. It has its N2SW offering for backing up AWS EC2 instances and RDS data.

It wants to do more, helping customers migrate to the public cloud if they wish, and provide data mobility between the on-premises and public cloud environments.

Veeam’s triple play

Can Veeam, with what we might call the old guard back in control, pull off this triple play; reinvigorating core transactional business growth, building up its enterprise business, and migrating its product, sales and marketing to a hybrid cloud business model paralleling its customer’s movements?

The enterprise data protection business features long established stalwarts, like Commvault, IBM and Veritas, and newer fast-moving and fast-growing upstarts such as Cohesity and Rubrik who are moving into data management. 

We have to add SaaS data protection companies like Druva to the mix as well. Veeam is going to need eyes in the back and sides of its head as well as the front to chart its growth course through this crowded forest of competing suppliers. 

It used to say it was a hyper-availability company and will need to be hyper-agile as it moves ahead. It has the momentum and we’ll see if it can make the right moves in what is going to be a mighty marketing, sales and product development battle.

Lenovo gets into fast virtual SAN array business with Excelero

Lenovo is buddying up with Excelero to provide NVMe over Fabrics access speed to data for its ThinkSystem servers.

The Data Centre Group at Lenovo has done the global reselling deal with Excelero and it gets Lenovo into NVMe-oF block data access without having to build a storage array of its own. Its own servers with their local storage provides the data store for Excelero’s NVMesh 2 software. 

Excelero NVMesh diagram

NVMesh involves intelligent client block device drivers – agent software – in data accessing servers. These use a Remote Direct Drive Access (RDDA) protocol to access data on target servers with their local (SSD) storage.

The system can provide a shared external array or operate in hyper-converged mode with a virtual SAN. This initially used RoCE (RDMA over Converged Ethernet) and Infiniband.

The second generation NVMesh 2 added support for standard gigabit Ethernet and TCP/IP and Fibre Channel. V2.0 also added erasure coding N+M type data redundancy. It is a parity-based data redundancy scheme running as a distributed service, meaning that the parity calculations are running on the clients, in a decentralised fashion. 

This was designed for large scale and adding more clients increases the overall CPU power available for parity calculations. Striping, mirroring and striping and mirroring are also available in what’s called a MeshProtect offering.

V2 also has a MeshInspect facility to provide cluster-wide and per-object statistics on throughput, IOPS, and latency.

Lenovo joins NVMe-oF mainstream

The Excelero partnership gets Lenovo into the NVMe-oF game, alongside all the mainstream storage vendors. Just this week both IBM with Storwize  V5000E and V5100 products, and Quantum with its F2000 array added end to end NVMe fabric access to these products.

Interestingly Excelero partnered Quantum with an NVMe-oF and StorNext combination at an industry event last September.

Lenovo and Excelero have Big Data and web-scale deployments in mind. They think that, as NVMe flash nears price parity with traditional flash, the NVMe over Fibre Channel and ordinary Ethernet means customers can get more fast storage capacity, flexibility and scale-out support without changing networking protocols or storage fabrics. Expensive RoCE or InfiniBand networking gear is no longer mandatory.

For Excelero, Lenovo gives it a useful additional channel into the enterprise storage market as it fights to grow its startup business with fellow fast data access startups Apeiron, E8 and Pavilion Data Systems, as well as competing with existing suppliers such as Dell EMC, HPE, IBM, NetApp, Quantum and Pure Storage.

Patrick Guay, Excelero’s VP for strategic accounts, issued a prepared quote: “We’re extremely proud to be the first shared NVMe storage software chosen by Lenovo, and look forward to expanding our business together.”

Is Optane DIMM endurance good enough? Quick answer…Yes, Intel has delivered

Update; April 9, 2019 – Intel statement added as final section.

After the dust has settled from Intel’s monster gen 2 Xeon, Optane, Ethernet and FPGA launch this week the enduring question about Optane DC’s Persistent Memory endurance is closer to being settled, with a view emerging that Intel has finally delivered in Optane’s promise.

A  Intel Optane DC Persistent Memory Module (PMM) | StorageReview.com – Storage Reviews gives Optane DIMM endurance at about 360 PBW (petabytes written) for the 256 GB drive.

I calculate that (360 PBW divided by 5 and then again by 365 and then by 2,546 to get to write cycles per day) to mean 1,406,520 write cycles, which agrees with MemVerge co-founder and CEO Charles Fan’s thinking.

He says 3DXP’s theoretical endurance is in the range of 1 million to 100 million writes (ten to the 6 to ten to the 8), with Intel’s Optane being towards the lower end of that.

If the Optane DIMM endures for 1,406,250 write cycles, let’s write the full drive contents every minute. How long will it last?

With the 256 GB drive we do 60 writes an hour, 1,440 a day, 525,600 a year and the drive dies after 2.675 years.

Let’s write the full drive contents every second. How long will it last then?

With the 256 GB drive we would do 60 writes a minute, 3,600 a minute, 86,400 a day, and the drive dies after 16.27 days. That looks very bad. But is it a real possibility?

Blocks & Files asked industry figures what they thought of this (simplistic) calculation.

Jim Handy

Jim Handy, the General Director at Objective Analysis, said: “My first impulse was to double-check the figures, but then I decided not to, and to instead consider whether it’s at all likely that Optane will be used in such a write-intensive environment.  I would think not.

“In Memory Mode the Optane is only written to when a dirty line is evicted from the DRAM cache.  It would take a LOT of cache thrashing to get even close to 60 full drive writes per hour.

“In App Direct Mode the software is choosing which items need to persist and which can be satisfactorily handled by the DRAM.  I would assume that Optane’s slow writes would further motivate the application developer to limit the number of persistent writes.  The only things that really need to persist are already separated out by most transactional databases because persistent stores are so slow on today’s systems.

“Charge card processing is an application that probably has one of the highest-volume requirements for persistent transactions.  VISA says (as of 2017) that its VisaNet can handle 65,000 transactions per second.  Let’s assume each transaction consumes 16 bytes (put your own guess here).  That would take 17 minutes to fill a 128 GB DIMM, assuming that all of VisaNet used a single Optane DIMM.  (A 256-byte transaction in a single-DIMM Optane system would meet your one-fill-per-minute number.)

“You may be right  about the ability to wear the  module out really fast, but it might not be important.  Remember that we had something similar with SSDs some time back when SSD makers competed to get to 20 or 25 DWPD, perhaps even higher, before users figured out that  they didn’t need that kind of endurance, especially if it cost more.  Now there are many 1 DWPD SSDs on the market, even for the data centre.”

An industry expert on background

Blocks & Files was told on background by an industry expert that 360 PB written (1,406,250 complete device writes) into a 256 GB device is quite impressive, especially compared to NAND. 

For comparison, a high-end NAND-based 256 GB SSD with 10 DWPD (drive writes per day) for 5 years, is only 18,250 complete device writes; 4.672 PBW.  So, the Optane DIMM has approximately 77 times the endurance of an enterprise, high endurance NAND-based device.  

Most typical SSDs have only 1 DWPD, in typical use, so the Optane DIMM compared to that device has ~770 times the endurance.

If Intel decides to issue DIMMs with more over-provisioning, the PBW will rise, of course.

Mark Webb

Consultant Mark Webb of MKW Ventures says the Optane DIMM is over-provisioned by 20 per cent and each die is over-provisioned by 20 per cent. He says: ‘Intel manages the endurance heavily to prevent  wearout from impacting the user. This is part of the controller firmware.”

He also says Intel doesn’t specify endurance other than saying there are no limitations for the warranty period; “So you can run it full blast.”

Chris Evans

Architecting IT consultant and analyst Chris Evans said of the Optane DIMMS: “They essentially have ‘unlimited’ endurance, or at least unlimited in the normal 3-5 year replacement cycle.”

Twitter Optane endurance debate

Amnon Izhar, a Distinguished Engineer at Dell EMC,  points out that it’s physically impossible to write 256 GB/sec to the DIMM. The DDR4 interface doesn’t support it.

Pure Storage Software Engineer Roland Dreier says “A DDR4 channel is 25 GB/sec, give or take,” which backs up this point. Anandtech’s Ian Cutress, supplying an Intel slide (below), says Optane is fixed at DDR4-2666, but the official numbers are around 8.3 GB/s per module for Optane (so, 50 GB/s ish with 6 per socket).” 

F Steven Chalmers, an independent consultant and advisor, takes a different tack; “If we are using this as if memory, a hot kernel data structure or file directory can be rewritten to failure in minutes/hours.”

For example, an HPC; “ping-pong benchmark between a CPU and a non-coherent accelerator using memory locations could easily cycle in a microsecond, meaning a specific DRAM location could be written once per microsecond indefinitely.  Such an algorithm could deliver 1.4e6 writes to a specific location in a few seconds. 

“My point is not that the new tech is useless, it’s that the new persistent memory parts have to be used in a very very controlled way consistent with their write endurance, much as flash has to be used in a very controlled way.  Useful tool in toolbox.  Not best and highest use of persistent memory by a long shot.”

Izhar said: “if wear levelling is done right, it shouldn’t matter if user overwrites the same file/address or sprays data all across. Host/used does not access a specific physical cell – it accesses LBA or logical address.” The address that you target on the Optane DIMM will change internally over time as a result of wear levelling.

Although wear levelling is needed, the DIMM is byte-addressable and so no block erase takes place. That means garbage collection is not needed.

His verdict is that: ‘Intel has very good engineering team. It took them time to get 3DXP to market, but they delivered.”

Last word – Intel

An Intel spokesperson said: “The endurance of Intel Optane DC Persistent memory (DCPMM) for any given time frame is a function of both media write endurance AND bandwidth.

“Published PBW is estimated assuming maximum DCPMM bandwidth at the target power of 15W for 5 years across various access patterns including 100 per cent write.

“In the example chosen (256GB PMM writing full drive contents every minute or second), we are bandwidth limited, not write cycle limited.  Our estimation for 5 years was 360PBW.  Media write cycles would extend well past 5 years.”

Datrium out-guns VMAX, XtremIO and FlashArray all-flash boxes

Hyper- or hybrid converged system supplier Datrium’s DVX storage and compute system out-performs Dell EMC’s VMAX and XtremIO all-flash arrays as well as Pure Storage’s FlashArrays.

Datrium wants to be seen as a converged enterprise storage and compute systems supplier, and says it converges five use cases onto its DVX system; primary storage, backup, encryption, mobility and policies. These, it asserts, are traditionally stored and processed in different silos, with different suppliers’ products, be they on-premises or in the cloud.

The DVX platform can converge these. Datrium has a distinctly enterprise focus, saying it can support remote and branch offices and small departments; the traditional hyper-converged infrastructure (HCI) market, but its focus is on the enterprise.

In that case it has to do well at the basic enterprise primary storage game, where capacity, performance and low latency count. Has it got the chops for this?

According to its internal tests using IOMark it does.

Datrium compared its results with publicly-available information about Dell EMC and Pure Storage arrays and came up with these comparisons;

It beat Dell EMC’s VMAX 950F and XtremIO X2 arrays and the Pure Storage Flash Array//m70 and //x70 in a range of mixed random read and write workloads and bandwidth tests. Pretty convincing.

The business’ products are aimed at hybrid multi-clouds with, Datrium says, the same customer experience whether it be on-premises or in the cloud.

Expansion into Europe

Datrium CEO Tim Page says his firm comes out top in 100 per cent of the customer POCs (Proof of Concepts) its entered. Its last two quarters have been knockouts in terms of revenue and the firm is expand outside its North America heartland with Europe its first port of call.

Sean Mullins, ex-Dell EMC, has been appointed as VP of Sales in Europe.

 It will focus on building out its UK presence this year—hiring and training new sales staff and developing relationships with organisations and partners throughout the region—with more ambitious expansion plans beginning in January 2020.

Quantum jumps onto all-flash NVM Express drive and fabric train

File manager Quantum has jumped aboard the NVM Express drive and fabric access train with a 2U, 24-slotter box and new software to accelerate users’ access to its StorNext files.

StorNext is a scale out, multi-tiered file data management product, including software and hardware such as the Xcellis arrays and Lattus object storage. It’s popular with Media and Entertainment (M&E) customers. NVMe drive support came with was V6.2 of StorNext in September last year.

The new F-Series NVMe arrays uses new software; Quantum’s Cloud Storage Platform. Jamie Lerner, Quantum’s President and CEO, said: “This platform is a stepping stone for us, and for our customers, to move to a more software-defined, hyper-converged architecture, and is at the core of additional products we will be introducing later this year.” 

He declared: “This is the most significant product launch we’ve done in years.”

F2000 diagram

Quantum’s F2000 array is the first product in an F-Series and, as the diagram above indicates, can hold up to 184TB of data in its active:active, dual-controller configuration using off-the-shelf hardware. It uses dual-ported drives for added reliability.

The software stack is  Quantum’s Cloud Storage Platform, offering block access and tuned for video and video-like data to maximise streaming performance. It supports NVMe-oF access from host application servers as well as Fibre Channel, iSCSI, iSER and RDMA.

Quantum says it’s tightly integrated with StorNext and designed for M and E activities such as;

  • Post-production: real-time editing of 4K and 8K content.
  • Sports video: coping with tens to hundreds of cameras generating multiple concurrent ingest streams and playing the content in real-time.
  • Rendering and simulation: deployable in animation and visual effects studios needing high IOPs and low-latencies between large-scale render farms and the storage subsystem.

StorNext customers should see an immediate uplift in data access speed and lower latencies, along with reduced rack space needs compared to disk-based arrays.

Tom Coughlin, the President of, Coughlin Associates, provided a supportive quote: “Combined with StorNext and its out-of-band metadata management, the F-Series provides storage tiering, content protection and diverse software defined management capabilities that should be well received by the media and entertainment industry.”

Blocks & Files thinks we can expect hyper-converged Quantum products to arrive later this year. This NVMe drive and fabric all-flash array catapults Quantum into the modern storage array age and should be a building block for even more development. Storage-class memory anyone?

IBM refreshes Storwize V5000 array line-up

IBM has replaced its five Storwize V5000 arrays with four faster, bigger and cheaper models.

V5000E

The V5000 and V7000 products are classic dual-controller storage arrays which use IBMs Spectrum Virtualise software to provide a pool of SAN (block access) storage. They come in a 2U x 24 2.5-inch slot base chassis.

The old and new model ranges look like this;

As you can see the V5000E and V5100 replaces the original V5000 line. The table below shows the feature spec of old and new.

Bold text highlights changes and yellow columns indicate new products

The V5010E, with E indicating Extended, replaces the V5010 and V5020. The V5030E replaces the V5030 and all-flash V5030F.

V5100

A new V5100 line which pushes the V5000E range closer to the high-end V7000s. It looks like this:

FCMs are IBM’s proprietary FlashCore Modules, its own design SSDs. The V5100F is the all-flash mode.

These have considerably more cache memory than the V5000Es and can support NVMe drives and NVMe over Fibre Channel to provide low-latency and fast access to data. IBM said the V5100s are ready for storage-class memory and 32Gbit/s Fibre Channel.

The V5100s can have two nodes in a cluster while the V7000s can cluster up to four nodes together. If you can bear looking at another table here is one showing some of their features:

Updated Storwize V7000 table.

Update: Blocks & Files was told by Eric Herzog that; ” The Storwize V7000 Gen3 introduced in October ’18 was also updated, along with the new Storwize 5000/5100 family announced 2 April, with the same 30.72TB SSDs, the 14TB HDDs, the 25Gbit/s Ethernet and the 32Gbit/s FC.”

Storwize management

All the Storwize arrays are managed through IBM’s Storage Insights, a cloud-based management and support tool, which provides analytics and automation.

Bullet points from IBM’s Storwize update briefing deck:

  • V5010E has 4x more cache than the V5010, 2x its IOPS, and scales to 12PB maximum flash capacity.
  • V5010E is 30 per cent cheaper than the V5010 and can be upgraded to the V5030E. 
  • V5030E has compression and deduplication. It scales to 23PB of all-flash storage in a single system; 32PB with 2-way clustering.
  • V5030E has a 30 per cent lower street price than the V5030 and 20 per cent more IOPS.
  • V5100s can have 2PB of flash in a 2U chassis, and scale out like the V5030F; 23PB in a single system or 32PB with 2-way clustering.
  • V5100s have 9x more cache than the V5030, and pump out 2.4x more IOPS than the V5030F (with data reduction), while costing 10 per cent more.

All-in-all, the refreshed V5000E and V5100s, and updated V7000s go faster than the previous models, store more data and enable users to get at it faster for less cost.

Intel announces Optane DIMM support with Gen 2 Xeon SP processors and QLC ruler

As expected Intel has updated its Xeon processor line and shone a bright light on Optane DIMM (DC Persistent Memory) general availability.

It’s also announced a dual-port Optane SSD and a QLC (4bits/cell) ruler format drive.

Gen 2 Xeons

Intel has updated its Xeon CPU line to the Gen 2 Xeon Scalable Processors (SP) which were called the Cascade Lake line. The top end is the Platinum 9200 with up to 56 cores; 112 in a 2-socket system. There can be 12 channels of memory support with up to 4.5TB of Optane PMEM per socket.

Intel says there can be up to 36TB of system-level memory capacity in an 8-socket system when combined with traditional DRAM. That’s twice the normal (DRAM-only) system memory.

The Platinum 8200 has up to 28 cores and comes in 2, 4 and 8+ socket configurations. There are also Gold 6200s (to 24 cores), Gold 5200s (to 18 cores), Silver 4200s (12 cores) and Bronze 3200 series CPUs (to 8 cores).

Customers have to buy the Optane DIMM and then pay again for Optane support in the Xeon CPU. For example;

A Xeon Platinum 8280L can support 3TB of Optane and 1.5TB of DDR4 DRAM; that’s 4.5TB memory support, and it costs $17,906 whereas an 8280 with 1.5TB DRAM support only costs $10,009. The Optane costs an additional $7,897; eye-watering. Anandtech has more price and configuration details.

Intel’s also announced Xeon D-1600 SoC processors for network edge use. They have up to 8 cores.

It’s new Agilex FPGAs, built with 10nm process technology, also support Optane DC Persistent Memory and the Compute Express Link (CXL), Intel’s latest cache and memory coherent interconnect.

Optane DIMM endurance

Intel is still not releasing detailed Optane DIMM performance numbers. A Storage Review article says the 256GB module has an over 350PBW rating for 5 years at a 15W power level. A chart illustrates this and indicates a 360-370 PBW value.

Assuming a 360PBW value, that’s 72PB/year, 197.26TB/day or 770.5 write cycles a day, 281,250/year, and 1,406,250 over 5 years.

Dual-port Optane SSD and QLC ruler

Joining the single port P4800X is the dual-port Optane DC SSD D4800X (NVMe).  Intel says it delivers 9x faster read latency compared to a NAND dual port SSD, under write pressure. No detailed performance numbers were provided.

Intel has also announced a QLC (4 bits/cell) 64-layer 3D NAND SSD, the D5-P4326  which uses the roughly 12-inch EDSFF ruler format to offer 15.36 and 30.72TB capacities. The 15.36TB capacity is also available in a U.2 form factor product. The 30.72TB ruler can enable 1PB of storage in a 1U chassis.

The random read/write IOPS are 580,000/11,000 and the sequential bandwidth is 3.2/1.6 GB/sec. Average read latency is 137μs. Its endurance is 0.18 drive writes per day (DWPD) for random I/O, and 0.9 DWPD for sequential I/O.

The claim is it enables HDD and TLC SSD replacement in warm storage. On eBay a 15.56TB ruler costs $3,920.39; $251.95/TB and $0.25/GB.

Intel QLC ruler

They join Intel’s existing 7.68TB QLC P4320 drive.

Gen 2 Xeon and Optane DIMM ecosystem

A raft of server and other suppliers is supporting the Gen 2 Xeons and Optane DIMMs; Cisco, Fujitsu, Supermicro and VMware to name a few.

VMware is supporting Optane DIMMs in “App-Direct” mode as well as “Memory” mode. vSphere is certifying the following maximum capacity for vSphere 6.7 release as follows:

  • Up to 6TB for 2-socket and 12TB for 4-socket of Optane  DC Persistent Memory in memory mode
  • A combination of DRAM and Optane DC Persistent Memory in App-Direct mode with a combined limit of 6TB for 2-socket and 12TB for 4-socket

VMware vSAN will not have support for “App-Direct” mode at this time as cache or capacity tier devices of vSAN. Expect more developments to come.

Blocks & Files welcomes the dawning of the persistent memory era and notes that system and application software developments are needed before IT users will see the speed benefits. Bring it on.

MemVerge virtualizes DRAM and Optane DIMMs into single clustered memory pool

Startup MemVerge is developing software to combine DRAM and Optane DIMM persistent memory into a single clustered storage pool for existing applications to use with no code changes.

It says it is defeating the memory bottleneck by breaking the barriers between memory and storage, enabling applications to run much faster. They can sidestep storage IO by taking advantage of Optane DIMMs (3D XPoint persistent memory). MemVerge software provides a distributed storage system that accelerates latency and bandwidth.

Data-intensive workloads such as AI, machine learning (ML), big data analytics, IoT and data warehousing can use this to run at memory speed. According to MemVerge, random access is as fast as sequential access with its tech – lots of small files can be accessed as fast as accessing data from a few large files.

The name ‘MemVerge’ indicates converged memory and its patent-pending MCI (Memory Converged Infrastructure) software combines two dissimilar memory types across a scale-out cluster of server nodes. Its system takes advantage of Intel’s Optane DC Persistent memory – Optane DIMMs – and the support from Gen 2 Xeon SP processors to provide servers with increased memory footprints.

MemVerge discussed its developing technology at the Persistent Memory summit in January 2019 and is now in a position to reveal more details.

Technology

The technology clusters local DRAM and Optane DIMM persistent memory in a set of clustered servers presenting a single pool of virtualised memory in which it stores Distributed Memory Objects (DMOs). Data is replicated between the nodes using a patent-pending algorithm. 

MemVerge software has what the company calls a MemVerge Distributed File System (MVFS.)

The servers will be interconnected with RDMA over Converged Ethernet (RoCE) links. It is not a distributed storage memory system with cache coherence.

DMO has a global namespace and provides memory and storage services that do not require application programming model changes. It supports memory and storage APIs with simultaneous access.

MemVerge concept

Its software can run in hyperconverged infrastructure (HCI) servers or in a group of servers representing an eternal, shared, distributed memory pool; memory constituting the DRAM+ Optane DIMM combination. However, in both cases applications up the stack “see” a single pool of memory/storage.

Charles Fan, CEO, says MemVerge is essentially taking the VSAN abstraction layer lesson and applying it to 3D XPoint DIMMs.

The Tencent partnership, described here, involved 2,000 Spark compute nodes and MemVerge running as their external store. However MemVerge anticipates that HCI will be the usual deployment option.

It’s important to note that Optane DIMMs can be accessed in three ways:

  1. Volatile Memory Mode with up to 3TB/socket this year and 6TB/socket in 2020. (See note 1  below.) In this mode the Optane is twinned with DRAM and data in the two is not treated as persistent but volatile. (See note 2.)
  2. Block storage mode in which it is accessed, but not at byte-level, via storage IOs and has a lower latency that Optane SSDs.
  3. App Direct Mode in which it is byte-accessible using memory semantics by applications using the correct code. This is the fastest access persistent memory mode.

MemVerge’s DMO software uses App Direct Mode but the application software up the stack does not have to change.

Product delivery

The product can be delivered as software-only or inside an off-the-shelf, hyperconverged server appliance. The latter is based on 2-socket Cascade Lake AP server nodes with up to 512 GB DRAM and 6TB of Optane memory and 360TB of physical storage capacity, provided by 24 x QLC (or TLC) SSDS. There can be 128 such appliances in a cluster, linked with two Mellanox 100GbitE cards/node. The full cluster can have up to 768TB of Optane memory and 50PB of storage.

The nodes are interlinked with RoCE but they can also be connected via UDP or DPDK (a UDP alternative). Remote memory mode is not supported for the latter options.

DRAM can be used as memory or as an Optane cache. A client library provides HDFS-compatible access. The roadmap includes adding NFS and S3 access.

MemVerge appliance.

The hyperconverged application software can run in the MemVerge HCI appliance cluster and supports Docker containers and Kubernetes for app deployment. Fan envisages an App Store. and likens this to an iPhone for the data centre.

The software version can run in the public cloud.

Speed features

MemVerge says its MCI technology “offers 10X more memory size and 10X data I/O speed compared with current state-of-the-art compute and storage solutions in the market.”

According to Fan, the 10X memory size idea is based on a typical server today having 512GB DRAM. The 10X data I/O speed claim comes from Optane DIMMs having 100 – 250ns latency while the fastest NVME drive (Optane SSD) has a 10 micro seconds latency.

He says 3D XPoint’s theoretical endurance is in the range of 1 million to 100 million writes (ten to the 6 to ten to the 8), with Intel’s Optane being towards the lower end of that. It is dramatically better than NAND flash with its 100 to 10,000 write cycle range (ten to the 2 and ten to the 4) – this embraces SLC, MLC and TLC flash.

MemVerge background

MemVerge was founded in 2017 by VMware’s former head of storage, CEO Charles Fan, whose team developed VSAN, VMware’s HCI product; chairman Shuki Bruck,  the co-founder of XtremIO, and CTO Yue Li. There are some 25 employees.

Fan and Bruck were also involved in founding file virtualization business Rainfinity, later bought by EMC.

The company has announced a $24.5 million Series A funding round from Gaorong Capital, Jerusalem Venture Partners, LDV Partners, Lightspeed Venture Partners and Northern Light Venture Capital.

The MemVerge Beta program is available in June 2019. Visit www.memverge.com to sign up.

Note 1. The jump from 3TB/socket this year to 6TB/socket next year implies either that Optane DIMM density will double or that more DIMMs will be added to a socket.

Note 2. See Jim Handy’s great article explaining this.

Google endorses Elastifile Cloud File Service for GCP

Google is endorsing Elastifile’s Cloud File Service service for its cloud, despite having its own Cloud Filestore service.

Cloud Filestore supports NFS v3 and comes in standard (up to 180MB/sec and 5,000 IOPS) ) and premium (700MB/sec and 30,000 IOPS) editions.

Elastifile’s CFS has far wider file protocol support; NFS v3/4, SMB, AWS S3 and the Hadoop File System. It provides varying levels of service with differing cost and performance, and integrates tiering between file and object storage. Pricing for provisioned capacity starts at 10¢/GB/month.

CFS can scale out to support 1,000s of nodes and PBs of capacity, according to Elastifile, and deliver millions of IOPS. It features automated snapshots and asynchronous replication.

Elastifile CFS graphic.

A big appeal of Elastifile is that it runs both on-premises and in the cloud and so can function as a cloud on-ramp for Google. The two say It helps bridge the gap between traditional and cloud-native workflows, making cloud integration easier than ever before.

Dominic Preuss, Google’s director of product management, is cited as the co-writer of an Elastifile blog, along with Elastifile CEO Erwan Menard, and he says: “We’ve been working closely with Elastifile on this effort to bring you scale-out file services that complement our Cloud Filestore offering and help you meet high-performance storage needs.”

Preuss says there is a deep engineering partnership with Elastifile.

CFS is said to be well-suited for horizontal use cases such as persistent storage for Kubernetes, data resilience for pre-emptible cloud VMs and scalable Network Attached Storage (NAS) for cloud-native services. An example given is CFS used to run SAP on the Google cloud with NetWeaver and HANA workflows.

The intent to add CFS to the Google Cloud Platform was announced in December. Now the Elastifile Cloud File Service is actually available in the Google Cloud Marketplace.

Pure Storage buys Compuverde (IBM’s Spectrum NAS software supplier)

Pure Storage has bought NAS software supplier Compuverde, best known for making the software behind IBM’s Spectrum NAS product.

Compuverde was founded in 2008 by Stefan Bernbo, CEO, system architect Christian Melander and Roger Persson who is named in several software patent filings. The funding history is unknown though chairman Michael Blomqvist has been described as its financier.

The Compuverde team joins Pure Storage, which also gets to work with Compuverde’s partners.

The Swedish-based company sells vNAS and hyperconverged storage products, with vNAS scalable to more than 1000 nodes and capacity ranging from a 1TB to several EBs. The hyperconverged system is suitable for VMware vSphere, HyperV, XEN and KVM hypervisors.

At the time of the IBM deal in 2018 Compuverde software scaled out to 100s of nodes and featured self-healing capability, and erasure coding with data striped across nodes and locations, as well as disks. It supported NFS v3, v4, v4.1, as well as SMB v1. v2. v3, Amazon S3 and OpenStack Swift.

Compuverde vNAS features in 2018

A virtual IP mechanism ensured all nodes in a cluster appear available at all times, even when a particular node is taken down for upgrade or has failed, and the software supports intelligent locking and snapshots.

IBM Spectrum NAS was targeted at home directories, general and virtual machine file serving, and to provide NAS storage for Microsoft applications.

Pure flying into filers

Pure says the acquisition will expand file capabilities. Charles Giancarlo, Pure’s Chairman and CEO, stated: “We’re excited about the opportunities that the Compuverde team and technology bring to Pure’s existing portfolio. As IT strategies evolve, enterprises look to leverage the innovations of the public cloud in conjunction with their on-prem solutions.”

Blocks & Files expects Pure to use Compuverde software in the public cloud as well as offer all-flash filers. Watch out,Elastifile, Isilon, NetApp, Panasas, Qumulo and WekaIO.

Acquisition details are not revealed and the deal is expected to close this month.

NetApp adds MAX Data supports for new Intel Xeons and Optane DIMMs

NetApp’s updated v1.3 MAX Data supports new Intel Xeon CPUs and Optane DC Persistent Memory in the host servers to which it feeds data from ONTAP arrays. 

Optane DC Persistent Memory is 3D XPoint media installed on DIMMs. It is also known as storage-class memory, and MAX Data combines it with the host server’s DRAM to provide a unified data store. NetApp says it helps customers use their data without having to redesign their critical applications.

NetApp MAX Data scheme.

NetApp said last year that MAX Data supported Optane DIMMs, but they were not available when that was stated. Now they are becoming available, along with new Xeon processors.

Jennifer Huffstetler, VP and GM of Intel’s Data Center Product Management and Storage, said in a canned quote: “With the second generation of Intel Xeon Scalable processors and Intel Optane DC persistent memory, customers can discover the value of their stored data. By working together with innovative companies such as NetApp, we can move, store and process more data.”

Optane DC Persistent Memory

Up until now Intel has talked about its Cascade Lake Xeon CPUs with the 2-socket AP version having explicit Optane DIMM support and the standard version thought to have the same support, albeit without a unified memory controller. Now Intel is using the term ‘second generation’ to describe its Optane DIMM-supporting Xeon CPUs.

NetApp says MAX Data is the first enterprise storage solution in the market that uses Intel Optane DC persistent memory in its servers for storing persistent data. Blocks & Files thinks it won’t be the last.

Note: See NetApp press release here.

AMD gets Western Digital Memory Extension tech for EPYC Optane battle

AMD has announced it’s using Western Digital’s Memory Extension technology to provide servers with greater effective memory capacity and so take on Optane DRAM extension technology.

The Ultrastar DC ME200 Memory Extension Drive is a NAND device available in 1TiB, 2TiB and 4TiB capacities. Its use requires no modifications to the host server’s operating system, system hardware, firmware or application stacks. The ME200 has an NVMe/PCIe interface and comes in U.2 and AIC (add-in-card) HH-HL form factors.

The ME200 is a tweaked version of WD’s Ultrastar SN200 SSD, built with planar (single layer) 15nm MLC (2bits/cell) NAND. It uses vSMP software from ScaleMP to provide replacement memory management unit (MMU) functionality and virtualises the drive to form a virtual memory pool along with the host system’s DRAM.

ScaleMP’s vSMP ServerONE software can, it says: “aggregate up to 128 EPYC-based servers into a single virtual machine. This translates to 256 EPYC processors, up to 16,384 CPUs, and with up to 512 TB of main memory.”

Western Digital says the ME200 drive improves the EPYC-based server memory-to-core ratio compared to conventional scale-out DRAM compute clusters using only DIMMs. The drive also enables lower TCO of in-memory infrastructure through consolidation. 

An example; a 30-node cluster, holding 30TiB of data in memory, can be reduced to a 8-node/32TiB one with 4TiB system memory each, and with increased per-node CPU utilisation.

A 1U server can support up to 24TiB of system memory using the Ultrastar memory drive for in-memory compute clusters.

WD suggests its Ultrastar memory drive is good for in-memory database engines like SAP HANA, Oracle, IBM, and Microsoft, and scale-out memory-centric architectures like Redis, Memcached, Apache Spark and large-scale databases. 

Regarding Optane

Optane persistent is storage-class memory use is restricted to Intel server processors, notably the Cascade Lake CPUs, now known as Gen 2 Xeon SPs.

The likelihood of Intel extending Optane support to AMD processors is as likely as the Moon reversing its orbital direction. Hence AMD’s working with Western Digital and ScaleMP.

Compared to the use of Optane DIMMs to expand effective memory, the ME200 costs less money, is probably simpler to implement and is available for AMD EPYC as well as X86 processors. Optane-enhanced memory servers may well go faster though. 

Back in 2016 ScaleMP said its software can pool Optane SSDs and DRAM as well as NAND SSDs and DRAM. We don’t hear so much about this now.

Get an ME200 datasheet here. Get a ScaleMP vSMP white paper here.