Home Blog Page 341

Hitachi Vantara packs Optane into HCI systems

Hitachi Vantara has updated its hyperconverged infrastructure systems with faster processors.

Loosely speaking, the company’s Unified Compute Platform (UCP) HC and RS lines are similar to Dell EMC’s VxRail and VxRack.

Speed is a key aspect of the upgraded HCI set. All the new HC boxes except one support Optane in Persistent Memory (DIMM form) to expand memory, and in SSD form for use as a cache tier. The GPU-supporting V225G supports Optane in SSD form only.

The UCP HC line features clusters of up to 64 x 1U and 2U nodes – Hitachi Vantara servers with Cascade Lake Xeon Refresh processors. Different models support all-flash, all-NVMe flash, and hybrid flash/disk storage media. All the models support vSphere and vSAN.

The RS line, positioned as a software-defined data centre (SDDC) product, supports 4,256 HC nodes in one to eight racks. It includes built-in Cisco Nexus switches plus public cloud support, and works with VMware Cloud Foundation and Nutanix.

Hitachi has also updated its HCI management software. The upgrade includes faster provisioning, certified SAP HANA support and consumption-pricing options.

Western Digital carves flash and disk drives into separate divisions

Western Digital’s new boss CEO David Goeckeler is making his mark by reorganising the flash and disk operations into two business units.

WD previously segmented its business between Data Center Devices & Solutions, Client Devices, & Client Solutions, while also providing total revenue results for HDD and Flash. 

WD said the reorg “will align the company’s diversified portfolio to an acceleration strategy designed for growth, profitability and development agility.”

The company aims to deliver “both Flash and HDD market-leading innovation through highly focused product development organisations … and uniquely address broad customer storage needs from edge devices to cloud infrastructure.”

Goeckeler, who joined WD from Cisco in March this year, said in the press statement: “With Flash and HDD product development businesses each focused on driving the storage innovation our customers need to support a broader set of use cases, the company is uniquely positioned in the age of cloud-based digital business.”

WD’s Flash revenues have overtaken its disk drive revenues.

Goeckeler has hired ex-Cisco guy Robert Soderbery to run the flash business.

Robert Soderbery

Soderbery said: ”What captured my imagination about Western Digital’s Flash business is that it has the potential to power the growth of every interesting end market in technology, including cloud, mobile, gaming, video and intelligent devices.”

Soderbery ran the $20bn+ enterprise networking business at Cisco. He left the company in 2016 to become an advisor to Rubrik, a senior advisor at McKinsey, and President at Uplift, a travel financing firm.

Goeckeler has hired headhunters to find a boss for the disk drive business unit. The reorg has prompted the departure of Mike Cordano, President and COO. He has teamed up with Mark Long, former EVP for strategy and corporate strategy, who left WD last year, as co-founders and partners of Prime Impact Capital.

Fungible Inc. scoops up Cloudistics assets

Fungible, the data processing unit (DPU) unit chip startup, has bought the assets of a data centre composability business called Cloudistics for an undisclosed sum.

The company said the “Cloudistics team [has] core competencies in the software needed to compose disaggregated data centre resources.”

Pradeep Sindhu, CEO of Fungible, said in a statement: “We are doubling down on our vision of revolutionising data centre infrastructure at all scales. We see progress towards this grand vision in two parts: The first, which we’ve already achieved, is to develop a new class of microprocessors called the Fungible DPU that promises to provide significant improvements in infrastructure performance, economics, security and reliability.

“The second, equally important part, is to build software that dynamically composes disaggregated DPU-powered servers into one or more computational clusters, each designed to run a specific workload. With these two parts working together synergistically we are taking a leap forward toward the holy grail of providing high performance and high agility in data centres.”

We can expect Fungible to announce Cloudistics software support for its DPU hardware in pretty short order. Sindhu said: “We call a data centre powered by the Fungible DPU and the Fungible Composer a Fungible Data Center.”

Cloud Platform

Cloudistics was established in 2013 and has taken in $15.m in an A-round. Twenty-three employees are identified on LinkedIn.

The Cloudistics Cloud Platform OS runs on host servers and aggregates compute, storage, network resource into pools along with virtualization and management. Independent virtual data centres can be provisioned in a peer-to-peer geo-distributed private cloud from these pools.

Cloudistics Ignite

In 2018, Cloudistics launched Ignite, a set of software services to handle IaaS deployments on qualified hardware, with a cloud-hosted multi-tenant management portal. The company said it saw a 400 per cent increase in customer adoption in 2018 compared with the prior year.

Lenovo announced the ThinkAgile CP Series composable cloud platform, a ‘cloud-in-a-box’, using Cloudistics software, in May 2019. 

Ceph scales to 10 billion objects

Ceph, the open source integrated file, block and object storage software, can support one billion objects. But can it scale to 10 billion objects and deliver good and predictable performance? 

Yes, according to Russ Fellows and Mohammad Rabin of the Evaluator Group who set up a Ceph cluster lab and, by using a huge metadata cache, scaled from zero to 10 billion 64KB objects.

In their soon-to-be published white paper commissioned by Red Hat, “Massively Scalable Cloud Storage for Cloud Native Applications”, they report that setting up Ceph was complex – without actually using that word. “We found that, because of the many Ceph configuration and deployment options, it is important to consult with an experienced Ceph architect prior to deployment.” 

The authors suggest smaller organisations with smaller needs can use Ceph reference architectures. Larger organisations with larger needs better “work with Red Hat or other companies with extensive experience in architecting and administering Ceph.”

The scalability issue

Analysis of unstructured data, files and objects is required to discern patterns and gain actionable insights in a business’s operations and sales.

These patterns can be discovered through analytics and by developing and applying machine learning learning models. Very simply, the more data points in an analysis run, the better the resulting analysis or machine learning model.

It is a truism that object data scales more easily than file storage because it has a single flat address space whereas files exist in a file-folder structure. As the number of files and folders grow, the file access metadata also grows in size and complexity – and more so than object access metadata.

File storage is generally used for applications that need faster data access than object storage. Red Hat wants to demonstrate both the scalability of object storage in Ceph and its speed. The company has shown Ceph can scale to a billion objects and perform well at that level via metadata caching on NVMe SSDs.

However, Red Hat wants to go further and has commissioned the Evaluator Group to scale Ceph tenfold, to 10 billion objects, and see how it performed.

The benchmark setup

The Evaluator test set-up had six workload generating clients driving six object servers. Each pair of these accessed, in a split/shared-nothing configuration, a Seagate JBOD containing 106 x 16TB Exos nearline disk drives; 5PB of raw capacity in total spread across three storage JBODS.

Evaluator Group 10 billion Ceph object cluster lab

Each object server had dual Xeon 18-core CL-6154 processors, 384GB of DRAM, six Intel DC P4610 NVMe 7.6TB write-optimised NAND SSDs for metadata caching, and 384GB of DRAM.

Ceph best practice recommends not exceeding 80 per cent capacity and so the system was sized to provide 4.5PB of usable Ceph capacity. Each 64KB object required about 10KB of metadata, meaning around 95TB of metadata for the total of 10 billion objects.

Test cycles

The Evaluator Group testers ran multiple test cycles, each performing PUTS to add to the object count, then GETS and, thirdly, a mixed workload test. The performance of each successive workload was measured to show the trends as object counts and capacity both increased.

The measurements of GETs (reads) and PUTs (writes) performance showed a fairly linear pattern as the object count increased. PUT operations showed linear performance up to 8.8 billion objects; 80 per cent of the system’s usable Ceph capacity, and then dropped off slightly. GET operations showed a dip to a lower level around 5 million objects and a more pronounced decline after the 8.8 million objects level.

Test chart for 10 billion small (64KB) objects

GET performance declined once the metadata cache capacity was exceeded (yellow line on chart) and the cluster’s usable capacity surpassed 80 pr cent of actual capacity. Once the cache’s capacity was surpassed the excess metadata had to be stored on disk drives, and accesses were consequently much slower. 

Performance linearity at this level would requite a larger metadata cache.

The “deep scrubbing” dip on the chart occurred because a Ceph parameter set for deep scrubbing, to help with data consistency, came into operation at 500 million objects. Ceph was reconfigured to stop this.

The system exhibited nearly 2GB/s of sustained read throughput and more than 1GB/sec of sustained write throughput.

Larger objects

The Evaluator Group also tested how Ceph performed with up to 20 million 128MB objects. In this test the metadata cache capacity was not exceeded and performance was linear for reads and near-linear for writes as the object count increased;

There is less metadata with the smaller number of objects, meaning no spill over of metadata to disk. The GET and PUT performance lines are both linearish – ‘deterministic’ is the Evaluator Group’s term, with performance of 10GB/sec for both operation types.

Comment

Suppliers like Iguazio talk about operating at the trillion-plus file level. That’s extreme but today’s extremity is tomorrow’s normality in this time of massive data growth That suggests Red Hat will have keep going further to establish and then re-establish Ceph’s scalability credentials.

Next year we might see a 100 billion object test and, who knows, a trillion object test could follow some day.

Red Hat shifts automated data pipeline into OpenShift

Red Hat today released OpenShift Container Storage 4.5 to deliver Kubernetes services for cloud-native applications via an automated data pipeline. 

Mike Piech

Mike Piech, Red Hat cloud storage and data services GM, Piech, said in his launch statement: “As organizations continue to modernise cloud-native applications, an essential part of their transformation is understanding how to unlock the power of data these applications are generating.

“With Red Hat OpenShift Container Storage 4.5 … we’ve taken a significant step forward in our mission to make enterprise data accessible, resilient and actionable to applications across the open hybrid cloud.”

OpenShift is RedHat’s container orchestrator, built atop Kubernetes. Ceph open source storage provides a data plane for the OpenShift environment.

Red Hat OpenShift with Container Storage. (Red Hat diagram.)

The automated data pipeline is based on notification-driven architectures, and integrated access to Red Hat AMQ Streams and OpenShift Serverless. AMQ Streams is a massively scalable, distributed, and high-performance data streaming platform based on the Apache Kafka project.

OpenShift Serverless enables users build and run applications so that, when an event-trigger occurs, the application automatically scales up based on incoming demand, or scales to zero after use.

Red Hat says that, with the recent release of OpenShift Virtualization, users can host virtual machines and containers on a single, integrated platform which includes OpenShift Container Storage. This is what VMware is doing with its Tanzu project

New features in OpenShift Container Storage 4.5 include:

  • Single, integrated, storage facility for containers and virtual machines,
  • Shared read-write-many (RWX) block access for enhanced performance,
  • External mode deployment option with Red Hat Ceph Storage, which can deliver enhanced scalability of more than 10 billion objects, without compromising performance.
  • Integrated Amazon S3 bucket notifications, enabling users to create an automated data pipeline to ingest, catalog, route and process data in real time. 

Druva pushes Kubernetes container protection into beta

Druva is adding backup for Kubernetes-orchestrated containers, as it aims to be a one-stop shop for data protection needs

The service, currently in beta, will provide application-consistent data protection, including resources inside and outside Kubernetes clusters. That includes dependencies on external storage and databases. There will be secure SaaS management, with a platform built on AWS for global control. Altogether there will be centralised protection, management, and compliance with self-service recovery for DevOps and application owners.

Stephen Manley.

Stephen Manley, Druva’s CTO, said in a statement: “With Kubernetes usage on the rise, too many production workloads lack adequate data protection. Now, central teams can securely and seamlessly incorporate Kubernetes protection into their existing policies, while enabling application owners to recover, migrate, and clone applications when necessary.”

Druva says its software identifies and protects resources inside and outside Kubernetes deployments. As application owners build and update their environment, Druva automatically stores snapshots of the entire environment, This enables customers to recover applications quickly and easily. These snapshots can be restored to a new location for migration, cloning, compliance and troubleshooting of production workloads.

Druva customers can protect production application deployments by choosing the “Kubernetes” option in the CloudRanger offering. Early access will be available in Q4, with wider availability expected in the first half of 2021.

Container storage is on the move

Kubernetes is the dominant container orchestrator. A Sysdig report says it has a 77 per cent share of the market by use. This expands to 89 per cent when Red Hat OpenShift and Rancher – both built with Kubernetes – are included.

Sysdig chart of orchestrator use.

When new workloads come along new data protection products and suppliers come along in their wake; think VMware and Veeam, HYCU and Nutanix, and also SaaS apps and in-cloud protection with Clumio and Druva. Now organisations are adopting containers and Kubernetes orchestration and storage provisioning with data protection added by new suppliers like Kasten and Portworx.

Veeam is partnering with Kasten and its K10 product. Pure Storage bought Portworx last week. Veritas protects containers with NetBackup. So too do Actifio, Cohesity, Commvault, and HYCU  as a wave of containerised app protection moves across the data protection supplier landscape. Rubrik has yet to announce its move into this space.

Ionir: let us be your Kubernetes data mover

Ionir, a software-defined storage startup, came out of stealth today with the news that it has bagged $11m in A-series funding.

The company has developed Data Teleport, a container storage platform that manages Kubernetes workloads in the cloud. This software is based on Magellan, which was developed by Reduxio, a startup that appears to have died (see our 2019 write-up of the company).

Jacob Cherian

Reduxio’s website is 404ing today and at least four execs are now working at Ionir, including CEO Jacob Cherian and co-founder and CTO Nir Peleg.

Reduxio took in $50.4m in VC funding since its inception in 2012. Ionir’s investors include Jerusalem Venture Partners (JVP) alongside C5 Capital, a UK VC, which both previously invested in Reduxio. Under the “once bitten, twice shy” rubric they must really believe in the company. Also, it does no harm that enterprise container app storage is a hot technology right now, as Storage’s Portworx acquisition last week, and NetApp’s Project Astra developments, show.

Ionir thinks storage for containers needs to be containerised itself and not provided by links to external silos. Container mobility is limited without accelerated data migration.

Ionir CEO Jacob Cherian today said that with Data Teleport, “persistent volumes can be moved or copied between Kubernetes clusters and clouds in under 40 seconds independent of the size of the volume or the amount of data involved.”

The idea is that containers are mobile across systems in the on-premises and public cloud worlds. This is fine so long as the containers are stateless and have no data associated with them. However, most enterprise apps are stateful and rely on reading and writing persistent data. When the containers move the data associated with them also move. This can be time-consuming especially when the network is slow and the data large.

DataTeleport makes a clone of the data using metadata references, and ships the clone description to the target site. That takes about 40 seconds. The migrated container can then be instantiated and access the data as if it were all present in the target system (which it is not). But desired bits of data on the source system are moved over first.

Cherian blogs: “For read operations, a combination of global deduplication, compression  and smart prefetch is used to minimise the read-on-demand and the overall amount of data transferred between the clusters.”

 “All write operations (which tend to be latency sensitive) are executed on the target, dismissing the need for invalidation on the source.”

This means that the time to access data on the target system varies between fast – meaning it’s present – and slow, meaning it’s still on the source system waiting to be migrated. This is similar to data being present or not in a cache.

Ionir said Data Teleport will support future interfaces and technologies without the need for forklift upgrades.

The ransomware attack happened on your watch. Why shouldn’t I fire you?

Firing Squad painting by Goya.
Firing Squad painting by Goya.

A ransomware attack happens. An organisation has to pay a ransom to get its encrypted files back. Why? Because there were no clean backup copies. At first sight, this appears negligent. So who’s to blame? Suppose the CEO said to the CIO: “Why should you keep your job?”

What reasonable response could the CIO make?

  1. It was a supplier’s fault; the ransomware protection technology failed.
  2. It was a staff member’s fault; they didn’t do what I directed.
  3. It was very clever ransomware; no-one could have predicted that type of attack.
  4. It was the business’ fault; it didn’t do what I recommended as I described in these mails and documents, which you saw.

Apart from number 4, the CEO might say: “But it happened on your watch and you are responsible.”

Heads should roll?

I asked my part of the Twittersphere what it thought. The answers are personal views.

Supercomputing engineer OxKruzr replied: “”Offline backup of critical data” is something that is just foundational to the [CIO] position.”

Andre Buss, a UK technology analyst, agreed that the CIO should go: “Absolutely – there are enough mitigations that any competent CIO should have in place to recover from this. If not, they are not on top of their brief.”

Yev Pusin, director of marketing at Backblaze, tweeted: “We think about ransomware a lot, and while I hate to put that much weight on it, if the company is unable to withstand the fallout, meaning if it is company ending, then I think so. It may even be career-ending.”

He added: “I always try to give the benefit of the doubt, so that’s why I’m thinking about it in terms of total exposure. If you lose some of your design assets, unless you’re a design firm it may not be company fatal but if you’re losing client data financial information, it’s bad.”

Flashdba, a staffer at an all-flash array software supplier, thought the affected organisation would “need to take into account tenure in the role. Some organisations take a long time for a new CIO to affect change, even for something so obviously critical as this.” That seems reasonable.

Andrew Dodd worldwide marcoms manager at HPE’s storage media devision, thought the CIO’s head should roll “only if there’s evidence of extreme negligence as to the risks. Otherwise, ransomware is just too prevalent and cybercriminals too well resourced and focused for businesses to be immune, however well prepared. Never say never – and always backup to an OFFLINE target.” This relates to the number 3 and 4 defences above.

‘Joshuad_’, a UK based techie, said: “I definitely think [the CIO should go], unless they’ve highlighted the risk before and been shot down.” That’s the number 4 defence in a nutshell.

Not just the CIO

Other respondents thought the responsibility was greater than that of CIO’s North Carolina-based Chad Richardson pointe to the Chief Technology Officer (CTO) as well: “I feel like that would fall more to the CTO.  If it was negligence to prioritise spending for that scenario it would be the CIO.”

VMware marketeer John Nicholson looks higher up the business hierarchy for the responsibility, to the CFO’s office: “Allow shareholders to sue the external auditors who didn’t identify this as a business risk. If the CFO signed off on a quarterly filing that didn’t detail their risk exposure, he should be barred from being an officer of a public company for 5 years.”

He added: “Expecting a silo inside a company to police itself at scale is madness. We have an audit and reporting framework. They should be leveraged.”

If not now, when?

Ransomware is vast in its scale and impact but in essence it is just a new form of data loss.

The best defence is to to get prepared. Get yourself guaranteed clean copies of files (and objects and databases) that can be restored quickly – and document this.

Nearly all data protection suppliers now offer ransomware protection. For example; Acronis, ArcServe, Actifio, Asigra, Clumio, Cohesity, Commvault, Dell EMC, Druva, Exagrid, HPE, HYCU, Rubrik, StorageCraft, TMS, Veeam, Veritas and many more.

And let’s not forget that the IT department has used RAID, backup technology and DR for decades to guard against drive, system failures and user errors such as inadvertent deletion. In essence, these are ways to create data copies so that data on crashed drives can be recovered through a rebuild process.

Cohesity’s marketing VP Chris Wiborg said: “For most of our customers, this conversation has shifted from ‘if’ to ‘when’ they will be subject to such an attack. 

“Our advice to CIOs would be to ensure their current backup and recovery provider provides at a minimum: immutability, anomaly detection, and the ability to instantly restore data and applications en masse so that recovery is ensured and this never becomes a question they have to face.” 

Your occasional storage digest with NetApp, VxRail and more

A mixed bag in this this week’s roundup, with NetApp changing its logo – and trying to redefine HCI to suit itself, a new report on backup, and Dell EMC VxRail getting close to Tanzu so it can run containerised apps.

HCI = ‘Hybrid Cloud Infrastructure‘?

Brad Anderson, NetApp GM, claims in this blog that analysts are ‘seeing the benefits of disaggregated architectures, and other vendors now agree that HCI should stand for hybrid cloud infrastructure.”

What do the competitors say?

Dell Technologies products webpage defines HCI as Hyperconverged Infrastructure. And Nutanix? Its products webpage shows Hyperconverged Infrastructure.  Cisco? Its HyperFlex product is defined as Hyperconverged Infrastructure. HPE? Its integrated systems webpage says HCI is Hyperconverged Infrastructure.

Dell, Nutanix, Cisco and HPE account for about 80 per cent of the HCI market. So who is Anderson referring to when he says “other vendors now agree that HCI should stand for hybrid cloud infrastructure”?

Incidentally, Tim Doering, a NetApp senior product manager, has an HCI demo where HCI means “Hyper Converged Infrastructure.”

VxRail gets Tanzu’d

Dell EMC has integrated VxRail HCI systems with the latest VMware HCI software and Tanzu offerings. Tanzu enables vSphere to run virtual machines and containers in the same server. The company says VxRail delivers “Kubernetes at Cloud Speed” with multiple, fully-integrated HCI infrastructure options to run vSphere with Tanzu. It claims to be the only HCI supplier doing this.

Customers can develop on a validated platform-as-a-Service or container-as-a-service with Tanzu Architecture for VxRail. Alternatively, they can get started with a Kubernetes deployment of vSphere with Tanzu on VxRail. They can also deploy Kubernetes at scale on a turnkey private cloud deployment, using VMware Cloud Foundation with Tanzu on VxRail.

NetApp: Black is the new Blue

NetApp is changing its logo colour. Emily Miller, VP for Brand and Influence Marketing at NetApp, blogs: “As part of that process, we are launching a new look, tone, and feel, including moving away from NetApp blue – Pantone 300 – and making our logo all one colour, black. 

Old NetApp blue logo – now not incredibly cool

“What does this change give us?  A cleaner, more contemporary look, and a freer, more flexible way to bring in colour, imagery, and animation. See for yourself in this identity reel. I hope you’ll agree: it’s incredibly cool.”

The identity reel video declares NetApp is now the Cloud Storage Specialist; “In a world full of generalists, NetApp is a specialist. We are relentlessly focused on helping you get more out of cloud than you ever thought possible. Whether on premises, in the cloud or anywhere in between, no one integrates, secures and connects your storage like NetApp does.”

Data Protection report

Storage architect Chris Evans has published “Essential Capabilities for Modern Data Protection”. The 22-page buy-and-download report is aimedT professionals, CTOs and CIOs, and suggests they should have four key questions in mind when evaluating suppliers.

  • Can my provider offer inclusive protection for on-premises, cloud (IaaS) and SaaS-based data?
  • Can my provider offer a holistic view of all my data management needs?
  • Can my provider offer flexible licensing that matches the data type or usage profile?
  • Can my provider offer data mobility and ensure applications can be protected/recovered wherever the data moves to? 

Evans notes that “where data and applications were once intrinsically commingled, the use of containers and serverless frameworks has decoupled data from the application code.”

The report looks at the data protection background, drivers and challenges and presents a brief history of backup. Evans discusses essential capabilities for modern data protection and looks at deployment models. Finally it looks at seven vendors: Clumio, Cohesity, Commvault, Druva, HYCU, Rubrik and Veeam.

Short items

Sumo Logic raised $325.6m in an IPO this week, valuing the data analytics company at $2.17bn.

Backblaze explains here how it built and authenticated its S3-compatible APIs for its cloud storage service.

Wells Fargo analyst Aaron Rakers notes Yangtze Memory Technologies Corp (YMTC) China’s national NAND flash provider, gets 80%+ of its chip manufacturing equipment from US & Japan. YMTC plans to invest $22 billion in its Wuhan fab for the production of 3D NAND. Rakers highlight the shared view of AMAT & Lam Research that domestic China memory players could not successfully build a fab without US tools. He points out that increasing restrictions on the supply of US equipment would devastate the Chinese semiconductor industry.

Kingston Digital Europe has announced the availability of the 128GB DataTraveler 2000 encrypted USB flash drive. This has an alphanumeric keypad so users can lock the drive with a word or number combination. An auto-lock is activated when the drive is removed from the host device without being properly shut down. 

Redstor, a UK cloud data management supplier, has launched Redstor Marketplace, the industry’s first self-service portal for MSPs. It makes it easier for MSP organisations to deploy backup, recovery and archiving offerings to manage their customers’ on-premises and cloud data, including Microsoft 365 and G Suite.

VMware and Nutanix grab lion’s share of spoils, in ‘difficult quarter’ for HCI vendors

IDC’s converged systems tracker for Q2 2020 shows a decline in a “difficult quarter”. But HCI sales did grow well in China and Japan, and HPE did well with its HCI systems.

As is now traditional, VMware led Nutanix and these two were a long way ahead of trailing pair HPE and Cisco.

IDC splits the market three ways;

  • HCI – $1.9bn revenues, 1.1 per cent growth, 47.1 per cent revenue share
  • Certified Reference Systems & Integrated Infrastructure – $1.5bn -7.6 per cent decline, 39.1 per cent share
  • Integrated Platforms – $544m -13.1per cent decline, 13.8 per cent share

IDC senior research analyst Paul Maguranis said in a statement: “The certified reference systems & integrated infrastructure and integrated platforms segments both declined this quarter while the hyperconverged systems segment was able to witness modest growth despite headwinds in the market.”

Charting revenue share trends since 2017 shows that certified reference systems like FlexPod are closing the gap with HCI (hyperconverged infrastructure), as HCI revenues decline from a peak at the end of 2019; 

B&F Chart using IDC numbers.

IDC looks at the top three vendors in the HCI market, checking the branded product sales and, separately, the HCI software owner’s sales. Both VMware and Nutanix HCI software is sold by other vendors such as HPE.

Branded HC sales numbers show a growth spurt by HPE;

HPE’s Y/Y 53.5 per cent revenue growth, from $84.7m to $130m, stands out from declines by Dell, Nutanix and the rest of the market. The picture revealed by slicing revenues by SW owner is quite different.

HPE’s 2Q20 revenues were $83.3m, meaning it sold ($130m – $83.3m = ) $46.7m of someone else’s HCI SW running on its servers. We understand this to be a mix of VMware and Nutanix. Without this software, HPE’s own SimpliVity HCI sales – (IDC only counts SimpliVity) – fell 4.9 per cent. HPE notes Nimble dHCI grew 112 per cent year-over-year in the period.

This HCI software owners table also shows Cisco suffering a 32.8 per cent fall in HCI revenues Y/Y. Both VMware and Nutanix outgrew the market.

We checked to see if HCI is taking more revenue from the overall external storage market. The Q2 IDC numbers for both storage categories are:

  • HCI – $1.9bn revenue; 1.1 per cent change Y/Y.
  • External storage – $6.26bn revenue, -1.1 per cent change Y/Y.

Charting the trends to see the longer term picture shows a growing gap between the two; 

B&F Chart using IDC numbers.

The HCI-external storage sales gap (blue line on chart) has declined for the most recent two quarters in what could be a seasonal pattern. The blue line shows fourth quarter HCI sales in any year on the chart peak higher than fourth quarter external storage sales. 

NetApp is morphing into cloud beast. But can it change its spots quickly enough?

Reports emerging from this week’s NetApp’s Financial Analysts’ Day show the storage hardware veteran is changing its identity to present itself as a hybrid cloud data services company.

But according to William Blair analyst Jason Ader, “NetApp remains a ‘show-me’ story —in particular with respect to meeting its cloud targets and growing share in a crowded on- premises storage market.” 

Wells Fargo senior analyst Aaron Rakers said NetApp emphasised the importance of its Cloud Data Services (CDS) unit. He quoted this claim from Anthony Lye, who runs CDS: “NetApp’s Cloud Volumes is the #1 shared storage platform (NFS, SMB), which is sold as a first party solution by Microsoft and available on GCP and AWS.

“The Spot acquisition brings compute and storage optimisation with a portfolio of services that provision, run, and scale applications in multiple clouds. With CloudJumper, NetApp can provide a virtual desktop service (VDS) and a managed virtual desktop service in the cloud. The company’s Cloud Insights solution then provides real time analytics across these services.” 

CEO George Kurian said: ‘NetApp’s strategy is to bring the simplicity and flexibility of cloud to the enterprise data center and to bring enterprise data services to the public cloud.” He noted that 93 per cent of enterprises now have a multicloud strategy and 87 per cent have a hybrid cloud strategy.

Cloud first

NetApp expects to drive revenue growth via all-flash array share gains and CDS expansion, Rakers reported. “The company now expects CDS ARR (Annual Recurring Revenue) to reach $250-$300m exiting F2021 (+141 per cent y/y at midpoint), $400-$500m exiting F2022 (+64 per cent y/y at midpoint), and >$1B exiting F2025. This compares to NetApp exiting F1Q21 with ARR at $178m.” 

NetApp pointed out 80 per cent of its engineers are software engineers. All-in-all, Rakers thinks “NetApp will benefit from its continued work to establish a software-first narrative.” 

Ader also highlighted NetApp’s CDS-first story. “To hammer home its bullish outlook on the cloud opportunity – and its current momentum – the company initiated a cloud ARR target of $1bn in fiscal 2025.”

“Management sees cloud services as complementary versus cannibalistic to its business,” he wrote, “with the firm already seeing its cloud technology bringing new logos and workloads into the fold.”

He said NetApp provided “incremental detail on the two strategic priorities outlined in recent quarters: 1) expanding the scale and scope of its cloud business, aided by the recent acquisitions of Spot, CloudJumper, and Talon and its deep partnerships with CSPs; and 2) returning to enterprise storage market share gains, primarily through differentiation in all-flash arrays (AFAs) and object storage.”

All-flash

Ader reported NetApp will have “a mid-range storage refresh (from hybrid systems to AFAs)”. Timings were not mentioned. Also, “Management noted that AFAs still only account for about 25 per cent of NetApp’s installed base systems, leaving significant runway for AFA refreshes over the next several years.”

Ader said NetApp’s Brad Anderson “commented that he expects virtually all performance workloads will move to all-flash in the near term while an increasing percentage of capacity-based systems will move to all-flash in the future as the cost of NAND flash comes down.”

NetApp estimates nine per cent CAGR for the AFA market, and 13 per cent CAGR for object storage. It sees object storage growing as a cold data tier for all-flash array data and as archiving data for verticals such as media and entertainment, and oil and gas.

“Management noted that NetApp has grown its object storage sales a striking 50 per cent annually over the last five years, largely under the radar of investors,” Ader wrote.

Seagate ships 18TB Exos disk drive

Seagate has begun shipping the 18TB nearline Exos X18 disk drive and is quoting an MSRP of $561.75.

The nearline disk drive capacity race is hotting up – Western Digital has shipped 18TB drives since July and Seagate aims to ship 20TB drives by the end of the year. We expect Seagate to ship 18TB IronWolf NAS drives later this month. 

The company today also released upgraded Exos AP controllers that enable its AP-2U12 storage arrays to handle Exos X18 drives.

Ken Claffey, GM of enterprise data solutions at Seagate, said the Exos X18 drives and upgraded controllers means the company “offers the industry’s leading density and configurability with ease of deployment for data lakes and private storage clouds.”

Exos X18

The helium-filled Exos X18 spins at 7,200rpm and has either single 6Gbits/SATA or dual 12Gbit/s SAS interfaces. The drive has a 256MB cache, features Instant Secure Erase, a 2,.5m hours MTBF rating, and comes with a five-year warranty.

We think it has nine platters, like the 16TB Exos, and understand it is the highest capacity conventionally recorded drive Seagate will produce – before moving to heat-assisted magnetic recording (HAMR) technology at the 20TB capacity level.

AP arrays

Seagate today announced the Exos AP 2U12 array, using the 18TB spinners, and said the AP 4U100 is getting a new controller. It also has a Nytro AP 2U24 system coming soon. 

With the latest announcement there are four AP systems:

  • Exos AP 2U12 –  12 x 3.5-inch 18TB Exos drives (216TB)
  • Nytro AP 2U24 – 24 x 2.5-inch SSDs (coming soon)
  • AP 4U100 – up to 1.6PB 
  • AP 5U84 – up to 1.344PB

Seagate got into storage arrays in 2015 via the $694m acquisition of Dot Hill, an OEM supplier of drive array chassis. This business is now part of Seagate’s Cloud Systems division. The portfolio includes the Exos drives and Application Platform (AP) arrays, which Seagate describes as compute and convergence platforms. They are sold to OEM customers such as Cloudian, in a continuation of the Dot Hill business  model.