Home Blog Page 48

MongoDB says it’s winning with targeted commercial AI projects

Database provider MongoDB has launched the latest version of its document-based system, claiming it can be at the center of moves toward commercial AI deployments.

Competing against traditional relational database heavyweights like Oracle and IBM, MongoDB sells both an on-premises version of its database and a cloud version, called Atlas, which is sold through the AWS, Azure, and Google Cloud.

MongoDB 8.0 is now generally available and comes with various data management and throughput improvements.

Architectural optimizations in 8.0 have significantly reduced memory usage and query times, and it has more efficient batch processing capabilities than previous versions. Specifically, 8.0 promises 32 percent better throughput, 56 percent faster bulk writes, and 20 percent faster concurrent writes during data replication.

In addition, 8.0 can handle higher volumes of time series data, and can perform complex aggregations more than 200 percent faster, with lower resource usage and costs, according to MongoDB.

Horizontal scaling is also “faster and easier than ever.” Horizontal scaling allows applications to scale beyond the limits of traditional databases by splitting data across multiple servers (sharding), without pre-provisioning additional compute resources for a single server. Sharding improvements in MongoDB 8.0 distribute data across shards “up to 50 times faster,” without the need for additional configuration or setup.

As part of the global launch of the improved database, a reveal in London took place at one of the company’s regular local events for customers, partners, and developers, which Blocks & Files attended. At the moment, MongoDB is believed to only hold about 2 percent of the total global database market by sales, although most analysts put it in the top five providers when it comes to developer use.

We wanted to know how the company intended to scale up through wider cloud use and, of course, as a result of wider AI-driven workloads. The business was notably upbeat. In fact, it has established a whole unit around commercial AI deployments in industry verticals, and claims it’s winning.

Greg Maxson, MongoDB
Greg Maxson

Greg Maxson, senior director of AI GTM (go-to-market), said businesses were being tested by the rapid marketing cycle of AI, uncertainty about which AI language models to use and which technology stacks to implement, and a lack of skills and resources to do it.

“Two months ago we established our MongoDB AI Application Program (MAAP), and have listed the seven general stack technologies that customers must start with, when it comes to AI projects. It’s foundational stuff, but we are already winning key customers around it on the services side.”

Maxson said a “large automotive company in France” wanted to better use its mechanical data, including video clips and manuals, to find out what the faults were in its cars when they were returned to dealers. “We came up with 20 AI models they could test to find the right solution, and now one of them is being used with MongoDB at dealerships across France,” said Maxson.

He claimed another firm – a “global household appliance manufacturer” – came to MongoDB because it wanted to integrate digital sound analysis from its products, including vacuum cleaners, into its manufacturing and quality control systems using AI. The chosen MongoDB system is now being used.

“We’ve brought all three main cloud providers into this process, and global system integrator Accenture is also involved, and we are aiming to set industry vertical standards to make AI projects work.”

The AI hype cycle is huge when it comes to data processing, data management, and data storage. But to make AI work across these areas, there has to be focus to enable delivery. Establishing an AI GTM unit at product and service providers is likely a solid first step in guiding potential AI customers through confusion.

Bootnote

MongoDB was initially called 10gen but wanted to indicate the product created in 2007, now called MongoDB, could scale to handle massive amounts of data – hence the name change.

Storage news ticker – October 3

Adaptive announced a control plane that brokers access across all surfaces and safeguards sensitive information, protecting data closer to its source, minimizing sensitive data exposure. It uses an agentless architecture that doesn’t require any change to existing workflows and tools. By understanding an organization’s data context, Adaptive simplifies protecting data at scale, allowing organizations to manage privileges and mask sensitive information efficiently. It says it’s at the forefront of an integrated, new approach that ends longstanding silos between data security and IAM (Identity and Access Management).

Adaptive argues that traditional data protection tools focus on protecting data at rest. However, in modern organizations, a large number of identities consume data for reporting, analysis, internal tools, ETL pipelines, and more. Protecting data at rest alone is ineffective, and the only way to safeguard data is to manage and control access across channels. Data protection in modern organizations is actually an access problem. Since data protection tools and access management systems have operated in silos, they leave blind spots in an organization’s security and don’t solve for escalating insider threats and cyber attacks.

Asianometry has produced a video on “The Wobbly Future of the Hard Disk Drive Industry.” It’s 18.5 minutes on the situation of the disk drive industry, threatened by SSDs, and facing production issues with new tech, such as HAMR and MAMR. It’s a tad simplistic, but it could fill up a coffee break.

Video screenshot on hard drive storage

Ceramic-based data storage startup Cerabyte announced that its president, Steffen Hellmold, will deliver a main stage presentation titled “The New Storage Tier to Enable the Yottabyte Era” at Yotta 2024, October 7–9, in Las Vegas, Nevada.

Data lakehouse supplier Databricks announced a new regional hub in London for the EMEA region by opening a new office in the Fitzrovia neighborhood. It says the Databricks Data Intelligence Platform provides a unified foundation for all data and governance, combined with AI models tuned to an organization’s unique characteristics. Databricks has over 400 employees in the UK and its local business has experienced over 60 percent annualized growth over the past three years. Customers include Gousto, Heathrow, Rolls-Royce, SEGA Europe, Shell, and Virgin Atlantic. Databricks has also formed partnerships with more than 35 universities across the country.

HPC parallel file system and enterprise storage supplier DDN is partnering with YTexas, a business network and community organization supporting businesses relocating to or expanding within Texas. DDN says it’s expanding its AI research footprint in Texas. Through its partnership with YTexas, DDN hopes to contribute to the state’s burgeoning business ecosystem, driving technological breakthroughs, job creation, and robust economic growth.

Enterprise cloud collaboration and file services supplier Egnyte has added five more patents to its existing portfolio of 45 or so:

  • System and Method for Enhancing Content Collaboration by Conflict Detection and Resolution in a Hybrid Cloud Cache – This solution addresses challenges in hybrid cloud environments by managing file versions and preventing data loss due to conflicting uploads. 
  • Event-Based User State Synchronization in a Local Cloud of a Cloud Storage System – This patent enables synchronization of user definitions between remote and local cloud systems, allowing centralized control over user access and near real-time updates. It maintains a global user directory on a remote cloud server, which can be synchronized with local user directories on multiple local cloud servers.
  • System and Method for Sensitive Content Analysis Prioritization Based on File Metadata – This technology estimates the likelihood of sensitivity for file objects. 
  • Storage Agnostic Large Scale Permissions and Access Analytics – This solution simplifies and consolidates permission sets from multiple heterogeneous file storage systems.
  • System and Method for Serving Subject Access Requests – This patent details a method for efficiently handling SARs in compliance with data privacy regulations, like GDPR.

Information management and governance supplier EncompaaS announced a strategic partnership with Dutch records retention pioneer Filerskeepers. EncompaaS enables organizations to discover, understand, govern, and use their data to promote automated governance at scale. Filerskeepers identifies country-specific data retention obligations and assists its clients with implementing those obligations to their data, no matter where in the world they operate. 

A mass storage roadmap is available here at the IEEE Xplore website. It covers NAND, SSDs, HDDs, tape, optical disks, and DNA storage, about which it says: “DNA data storage has been demonstrated in the lab, but the costs of reading and writing data on synthetic DNA are currently too expensive for practical applications.”

Mass storage roadmap

NetApp has expanded its Google Cloud partnership to integrate its unified data storage and intelligent services into the Google Distributed Cloud architecture. Google Distributed Cloud extends customers’ cloud infrastructure and services to on-premises sites and it now supports NetApp ONTAP and StorageGRID to support its own services including databases, AI, and analytics.

Fabless semiconductor company Primemas, which has a system-on-chip (SoC) hub chiplet platform, is partnering with Samsung to collaborate on the development of advanced Compute Express Link (CXL) memory products. Primemas will conduct joint R&D with Samsung using its CXL 3.0-enabled SoC hub chiplet (hublet) and FPGA chiplet to provide optimized products for next-generation datacenters and AI systems.

Primemas graphic

Quantum announced that its object storage software has extended its Veeam Ready qualifications to include the new ActiveScale 7.0 software, which support all-flash systems for fast ransomware recovery, and ActiveScale Cold Storage support for Veeam Archive Tier for low-cost, air-gapped retention of Veeam backups. With Veeam’s new Direct to Archive feature, backups can be sent directly from Performance Tier storage to the Archive Tier.

Red Hat OpenStack 2024.2 Dalmatian is the latest release of the open ource and modular cloud infrastructure software stack with its Nova (compute), Swift (object storage), Cinder (block storage), Neutron (networking), and Horizon (dashboard) components. Skyline and its modernized web UI are now fully supported as part of the official OpenStack release. Blazar introduced support for reserving compute (including GPU) instances based on existing Nova flavors. There are security updates as well. In Nova, with the libvirt driver and libvirt version 7.3.0 or newer, mediated devices for vGPUs are now persisted across reboots of a compute host. Download OpenStack Dalmatian here.

The OpenStack market is estimated to be worth $22.81 billion in 2024 and is expected to reach $91.44 billion by 2029, growing at a CAGR of 32 percent. OpenStack is currently experiencing a significant surge in adoption, attributed largely to the software’s popularity as a VMware alternative, and also to its AI workload support. 

Rowan Trollope, CEO of database supplier Redis, argues that AI is only as good as the quality and availability of data fed into it, and the exponential growth of GenAI technologies could take a quick downward turn if not supported by the right databases – like, say, Redis, which powers the most innovative GenAI tools available today, including those developed by OpenAI. Redis 8, the latest release, has 16x more throughput and 4x faster response time. Trollope reckons it’s important not to let hyperscalers – such as Google and AWS – monopolize the game. Redis recently changed its licensing as the previous setup was allowing Google and Amazon to take something Redis put out as open source and sell it to their customers.

High-availability supplier SIOS announced LifeKeeper for Linux version 9.9.0 with advanced DR features, including synchronous and asynchronous data mirroring, intelligent failover mechanisms, enhanced security management, expanded language support, and additional Linux operating system compatibility. More information can be found here.

Stratio BD’s Generative AI Data Fabric product helps businesses manage and use vast quantities of data and enables all users (data managers, data scientists, as well as business users) to complete complex data queries in their natural language using its Talk to Your Data feature. The Generative AI Data Fabric 4.4 release allows users to complete complex data transformation and ML processes within seconds using natural language prompts. Users can have entire conversations with Stratio GenAI in almost any language, providing the world’s fastest and most comprehensive natural language data management product. Stratio BD’s ““”Productivity, Performance, and Potential: Generative AI In Action” White Paper can be read here.

A recent study by TRG Datacenters aimed to identify how much data volume grows over time by collecting apps in four main categories: communication, navigation, work, and social media. The download sizes of these apps were traced over time using the Wayback Machine. The study also gathered data on the most sold phones from 2014 to 2024, focusing on their photo, video, and non-expandable memory characteristics. By calculating the average non-expandable memory and the sizes of photos and videos, the study compared these sizes over the years to determine the file size change both in megabytes (MB) or gigabytes (GB) and as a percentage. No surprise: file sizes have grown a lot, with video games growing the most.

Data storage growth

Following AT&T’s claim that Broadcom proposed a VMware price increase of 1,050 percent, Dave Russell, SVP and head of strategy at Veeam, stated: “We have seen a 300 percent price increase in the VMware products we’re using. This is in line with what I’ve heard from customers, some of whom have reported even higher price increases … Most large enterprises have renewed their VMware products, giving them time to decide whether they still plan to use them going forward. Meanwhile, smaller customers are more sensitive to pricing changes … The vast majority of large enterprises put themselves in a position to push any concerns out. Some customers are looking at whether an alternative is ‘feasible.'”

William Blair analysts write: “We have confidence that the AI investment cycle has several years to run. Qualitatively, the executives at Microsoft/OpenAI, Alphabet, Oracle, Amazon, and Meta have all indicated they see a multi-year investment runway related to building out AI infrastructure, training next-gen models, and building new AI-native applications. Alphabet’s CEO put it most succinctly, indicating that the real risk in this AI platform shift is under-investment rather than over-investment.

“While training-based demand for infrastructure has been highly concentrated in a handful of the largest technology companies (40 percent of Nvidia revenue comes from four companies), inference demand promises to be more diffuse. The general rule of thumb for inference costs are that they are the square root of training costs. Importantly, while inferencing ASPs are lower, the potential market opportunity is much broader, applicable eventually to the entire global population as AI becomes integrated into applications across industries and geographies.”

Western Digital opens separate websites for HDD and flash businesses

Western Digital has advanced the previously announced split between its HDD and flash businesses by launching separate customer websites. The separation aims to improve operational focus and market agility in pursuit of improved sales and margins.

“We are now operating as two specialized websites: WesternDigital.com for HDDs and platforms, and SanDisk.com for flash technology, including SSDs, memory cards, USB flash drives, and more,” a canned statement from the storage titan read.

As announced in October last year, Western Digital plans to separate its HDD and flash businesses, creating two independent, public companies with market-specific, strategic focus. The separation will “better position each franchise to execute innovative technology and product development, capitalize on unique growth opportunities, extend respective market leadership positions, and operate more efficiently with distinct capital structures,” Western Digital explained. “The creation of two specialized websites is a necessary step in the company separation process.”

Although the websites are separate, there’s still some crossover between brands. On WesternDigital.com, you can shop for all HDD and platform products from the following brands: Western Digital, WD, WD_BLACK, and SanDisk Professional. On SanDisk.com, you can shop for all flash products, such as SSDs, memory cards, and USB flash drives from the Western Digital, WD, WD_BLACK, SanDisk, and SanDisk Professional brands.

For support, customers go through the relevant website, with G-Technology customers going through WesternDigital.com.

All warranty claims for HDDs and platform products from Western Digital, WD, WD_BLACK, SanDisk Professional, and G-Technology should be submitted through the Western Digital Support account. After signing in, select your registered product and “Request Warranty Replacement.” If you have not registered your product yet, select “Register a New Product.”

All warranty claims for flash products such as SSDs, memory cards, and USB flash drives from Western Digital, WD, WD_BLACK, SanDisk, and SanDisk Professional should be submitted through the SanDisk Support account.

The formal business split is expected to be completed in the “second half of 2024” and since last year the firm has established legal entities across around 20 countries. Once complete, both divisions will operate as publicly traded companies.

David Goeckeler, Western Digital’s CEO, will lead the SanDisk business, and Irving Tan, currently executive vice president of global operations, will become the CEO of WD.

Discover what it takes to build your AI-ready infrastructure strategy

SPONSORED POST: Organisations across the UK are rushing to find new ways of using artificial intelligence (AI) to streamline their operations and build new products and services for their customers and stakeholders.

A report published by the UK Office of National Statistics (ONS) last year suggests that the majority of UK organisations are yet to implement any form of AI within their business, though there has been a considerable expansion in awareness of the technology and willingness to explore its capabilities.

Certainly, the potential to extract intelligence and insight from the vast amount of data now at their disposal is huge. But identifying and implementing machine learning, deep learning, generative AI (GenAI) and other forms of the technology to fulfil that ambition still poses a significant challenge for many.

It’s not just about the tech stack – there are hurdles with data quality, supply chains, legacy systems, costs and operational complexity to navigate too. Building a hybrid IT infrastructure capable of handling AI is an important step. The larger the workload and the more data it consumes, the more likely that a data centre will be need to host and process all of that information – an environment that offers the scalability and agility to quickly expand further in support of additional and ever larger AI workloads and datasets as the business requires.

Few organisations will have that knowledge or infrastructure capability in-house, so choosing a partner with the expertise to guide them through the implementation journey will be critical.

Digital Realty has put together a white paper specifically designed to provide guidance on the importance of having a robust enterprise infrastructure to support an organisation’s AI needs. Titled “AI for IT Leaders: Deploying a Future-Proof IT Infrastructure”, it offers advice on the strategic goals of enterprise AI adoption together with the common AI workloads, challenges and solutions needed to deploy AI-ready infrastructure. Digital Realty also provides a list of evaluation criteria which will help you choose the right partners to build an AI infrastructure stack within your organisation that will yield the best performance.

You can register download a copy of the Digital Realty whitepaper by clicking this link.

Sponsored by Digital Realty.

Analyst warns of memory oversupply from AI server demand

Analyst Jim Handy thinks the memory market is being artificially buoyed by AI server demand, with oversupply and a price correction coming.

Jim Handy

Handy is a senior analyst at Objective Analysis. In a post titled How Long Will Memory Growth Continue? he writes: “The memory business is currently faring pretty well [but] the business is definitely off-trend.”

He says that the memory market is characterized by capacity-driven cycles. As market demand rises, suppliers build more fabs, but the market can’t often absorb their output, prices fall due to over-supply, and the market slumps.

The current memory market is demand-driven with a lot “coming from massive AI purchases in hyperscale datacenters.” Forecasting the duration of this cycle is tricky as “demand-driven cycles tend to be caused by factors that are very hard to predict.”

At the FMS event in Santa Clara in August, Handy “presented a chart showing the history of hyperscaler capital expenditures … It’s unusually high at the moment, but it’s not clear how long these companies will continue their high expenditures. They haven’t had a matching revenue surge, so they can’t fund accelerated spending forever.” 

Objective Analysis chart

He superimposes the current memory market revenue history on a second chart showing two previous demand cycles, “normalized to the market’s underlying trend, so they are all expressed in a percentage relating to how far off-trend they are, rather than in absolute revenues.”

Objective Analysis chart

Also, all three curves start from the same point, month 1, and then rise and fall over time. A 2017 cycle is shown by the red line and a 2021 cycle by a black line. Their duration, rise, and fall are identical within a two-month window.

The green line shows the current cycle, with the dashed extension being “a projection of where it might head if it performs as did the prior two cycles. Today we appear to be in Month 18.” If the projection is correct, this suggests the current cycle will peak in month 21, by the end of 2024, with the demand declining as 2025 progresses through to the fall.

Handy says this isn’t as scientific a forecast as those Objective Analysis usually produces for its clients, but adds: ”Today’s heavy AI spending can’t last forever, and when it does end, there will undoubtedly be an oversupply with a subsequent price correction, if not a collapse like those seen in 2018 and 2022.”

This line of analysis suggests that the high-bandwidth memory (HBM) boom being enjoyed by SK hynix and Micron, and pursued by Samsung, could be relatively short-lived. 

Pliops XDP LightningAI is a memory tier for GPU compute 

The Pliops LightningAI product functions as a memory tier for GPU servers and can provide a more than 2x speed up for large language model (LLM) responses.

Ido Bukspan, Pliops
Ido Bukspan

Pliops is an Israeli server CPU offload startup that has developed XDP (Extreme Data Processor) key-value store technology with its AccelKV software running in an FPGA to accelerate low-level storage stack processing, such as RocksDB. It has now developed a LightningAI product, using ASIC hardware inside a 1 or 2RU server, applicable to both training and inference LLM workloads.

CEO Ido Bukspan said: “We saw how we can leverage our technology to something even more that changed the needle significantly in the world. And the potential is huge.”

He said that Pliops developed the core product then “took it all the way and developed further than our product in order to show the end-to-end, amazing value of XDP to performance … It’s not just a specific area. It can be expanded. We did it all the way, developed all the stack and pieces of software needed in order to prove the value that our new AI tool can help AI developers get out much more from their existing GPUs.”

XDP LightningAI’s best fit is with inference workloads, where it enables an LLM, running a multi-tier inference process, to “remember” cached data but then replaces intermediate responses and data – the attention state – needed for a subsequent response, speeding up the end-to-end LLM processing time.

Pliops slide

The LLM, running in a GPU server with, for example, high-bandwidth memory, and accessing NoSQL and vector databases, runs out of memory capacity during a multi-tier response. This requires old data, previously evicted from the HBM prefill cache, to be reloaded. LightningAI serves as a persistent memory tier for such data, enabling the GPU to avoid the HBM reload time penalty. 

Pliops slide

It runs in a networked x86 server networked by NVMe-oF to a GPU, and enables the GPU to sidestep a memory wall, more than doubling its speed, and also be around 50 percent more power-efficient. Pliops sees it as a great benefit to inference workloads using retrieval-augmented generation (RAG) and vectors, where the GPU servers will have limited memory capacity and operate in power-constrained environments.

Pliops slide

A GPU will run the Pliops LLM KV-Cache Inference Plug-in software. It will use a Pliops API to issue standard GPU-initiated IO requesting Pliops CUDA key-value activity. The GPU servers’ BlueField DPUs send the request across a 400 GbE RDMA Ethernet fabric to ConnectX-7 NICs in the nearby (in-rack) XDP LightningAI server. There, it’s sent to the XDP-PRO ASIC, which wrangles the data operations using direct-attached SSDs. 

Pliops slide

The Pliops stack includes application (vLLM) modifications, a GPU CUDA library for NVMe key-value commands, and a NVMe-oF initial target for GPU and Lightning servers. The system can be deployed on standard 1 or 2RU ARM or x86-based servers, and is fully compliant with the vLLM framework. A single unit can serve multiple GPUs.

Pliops is working with potential customers, OEMs and ODMs. They can inspect demonstration and proof-of-concept XDP LightningAI units now, and the company will be at SC24 in Atlanta, November 17-22. We can expect additional GenAI applications to be supported beyond LLMs in the future, as well as even more LLM acceleration, between 2.5x and 3.0x.

NetApp responds to VAST with ONTAP Data Platform for AI

Analysis NetApp is one of the first of the major incumbent storage vendors to respond to VAST Data’s parallel NFS-based data access for AI work, with its internal ONTAP Data Platform for AI development, after HPE rolled out its AI-focused Nvidia partnership in March.

VAST Data has penetrated the enterprise data storage market with its DASE (Disaggregated Shared Everything) architecture, which provides a single tier of storage with stateless controllers driving low-latency, high-bandwidth, all-flash storage across an internal RDMA-type fabric with metadata stored in storage-class memory type drives. The company pitches its AI-focused software stack built on this base as providing costs that are close to that of disk, parallel access, and a single namespace used by a data catalog and unstructured data store plus a structured database, IO-event triggered data engine, and now an InsightEngine using Nvidia GPUs as compute node units and embedded NIM microservices.

VAST’s platform and products are being presented at its Cosmos marketing event. Until now, none of the enterprise storage incumbents – with the exception of HPE, which announced the AI-focused Nvidia partnership earlier this year – has responded to the tech other than to adopt lower-cost QLC (4 bits/cell) flash technology. HPE developed its Alletra MP hardware architecture and runs VAST’s file software on that with its own block storage offering separately available. Quantum’s Myriad OS development shares many of these concepts as well.

Now NetApp has just announced its own disaggregated compute/storage architecture development at its Insight event and a white paper, ONTAP – pioneering data management in the era of Deep Learning, fleshes out some details of this ONTAP Data Platform for AI project.

Currently, NetApp has three ONTAP storage hardware/software architectures:

  • FAS – clustered, dual-controller base Fabric-Attached Storage for unified files and blocks on hybrid disk and SSD drives
  • AFF – all-flash FAS with SSD drives only
  • ASA – all-flash SAN Array, AFF with block storage optimizations

Now a fourth ONTAP Data Platform for AI architecture is being developed, with NetApp saying it’s “a new design center in NetApp ONTAP built on the tenets of disaggregation and composable architecture.”

It’s a ground-up concept, starting with separate compute controllers, running ONTAP instances, “embellished with additional metadata and data services,” and storage nodes filled with NVMe SSDs, forming a single storage pool, accessed across a high-speed, low-latency, Ethernet-based RDMA fabric. Both compute units and storage nodes can be scaled out with dynamic workload balancing.

Blocks & Files diagram comparing VAST and NetApp storage stack approaches
Blocks & Files diagram comparing VAST and NetApp storage stack approaches

The system supports file, block, and object storage with underlying Write Anywhere File Layout (WAFL) storage and a single namespace. “Physical block space is now distributed across multiple [drive] enclosures, thus creating a single extensible namespace” and “each compute unit or node running the ONTAP OS has full view of and can directly communicate with the storage units providing the capacity.”

The ONTAP instances provide data protection (snaps, clones, replication, anti-ransomware), storage management (speeds, feeds, protocols, resiliency, scale), and intelligent data functions (exploration, insights, getting data AI-ready).

File locking can disrupt parallel access. NetApp is developing “the concept of independently consistent micro file system instances. Each micro file system instance operates as a fully functional file system and provides consistency across data and metadata operations … Since each micro file system instance has exclusive ownership of its resources at a given point in time, they can operate safely on file system internal data structures in parallel to other instances.”

NetApp white paper diagram
NetApp white paper diagram

NetApp says “these micro file system instances are decoupled from the front end or application facing constructs. As an example, a file system client mounting file shares and performing data and metadata operations has no visibility to which micro file system instance is processing the request. The client will communicate with the file server as per semantics prescribed during mount.”

The design achieves parallelism at three levels:

  • Client and server-side protocol stack 
  • File system namespace and object management subsystem 
  • File system block layer managing on-disk layout 

The white paper says the “WAFL on-disk layout will ensure that each individual file or a collection of files within a file share will have their data blocks distributed across multiple disk enclosures to drive massive parallelism and concurrency of access. Each instance of the ONTAP OS will have high bandwidth connectivity across the backend disk enclosures and can leverage RDMA constructs to maximize performance as well as ensure quality of service end to end.”

Metadata engine

A structured metadata engine “extracts the data attributes (or metadata) inline. Once the attributes are extracted, the metadata engine indexes and stores this metadata to enable fast lookups. A query interface allows applications to query for this metadata. The query  interface is extensible to enable semantic searches on the data if exact key words are not known.” 

It provides “a fast index and search capability through the metadata set. AI software ecosystems deployed for data labeling, classification, feature extraction or even a RAG framework deployed for generative AI inferencing use cases can significantly speed up time-to-value of their data by leveraging the structured view of unstructured data presented by the metadata engine.” 

The data in the system is made ready for AI as ”NetApp’s powerful SnapDiff API will track incremental changes to data in the most efficient manner. The metadata engine in ONTAP will record these changes and leverage its trigger functionality to initiate downstream operations for data classification, chunking and embedding creation. Specialized algorithms within ONTAP will generate highly compressible vector embeddings that significantly reduces both the on-disk and in-memory footprint of the vector database (significantly shrinking infrastructure cost). A novel in-memory re-ranking algorithm during retrieval ensures high precision semantic searches.”

The generated embeddings are stored in an integrated vector database backed by ONTAP volumes.

Looking ahead

NetApp’s ONTAP Data Platform for AI project validates VAST’s architectural approach and raises a question for the other enterprise storage incumbent suppliers. If NetApp sees a need to spend deeply on a new ONTAP data architecture, what does that mean for Dell, Hitachi Vantara, IBM, and Pure Storage? Do they have their product design engineers poring over VAST and working out how they could develop competing technology on PowerStore or Power Scale, VSP One, FlashSystem, and FlashArray/FlashBlade base architecture systems?

Secondly, with VAST, HPE, and NetApp providing, or soon to provide, parallel NFS-based data access for AI work, where does that leave previously HPC-focused parallel file system suppliers looking to sell their storage into enterprises for AI workloads? We’re thinking DDN (Lustre), IBM (StorageScale), Quobyte, and VDURA (PanFS). Is there some kind of middle ground where a parallel file system meets disaggregated architecture?

Answers to these questions will likely emerge in 2025, when we might also expect a VAST IPO.

Veeam survey uncovers apathy toward EU’s NIS2 security directive

A Veeam survey reveals that only 43 percent of EMEA IT decision-makers believe NIS2 will significantly enhance EU cybersecurity, yet 90 percent of respondents reported at least one security incident that the directive could have prevented in the past 12 months.

Some 44 percent of respondents experienced more than three cyber incidents, with 65 percent categorized as “highly critical.”

Andre Troskie, Veeam
Andre Troskie

Andre Troskie, EMEA Field CISO at Veeam, stated: “NIS2 brings responsibility for cybersecurity beyond IT teams into the boardroom. While many businesses recognize the importance of this directive, the struggle to comply found in the survey highlights significant systemic issues.”

The EU cybersecurity NIS2 (Network and Information Security 2) directive updates the 2016 NIS directive, which aimed to improve the cyber-resilience of critical infrastructure and services across the European Union. Operators of such services had to set up risk management practices and report significant incidents. EU member states had to set up national cybersecurity strategies and Computer Security Incident Response Teams (CSIRTs), and there was an EU-wide Cooperation Group to encourage cybersecurity information  sharing.

NIS2, which takes effect on October 18, broadens the scope of NIS, has stricter security requirements, faster incident reporting, a focus on supply chain security, harsher penalties for non-compliant organizations, harmonized rules across the EU, and better member state information sharing.

It represents an additional cost for businesses and other organizations. French digital services company Wallix states: ”Compliance with Directive NIS2 is non-negotiable and has significant financial implications for companies. According to the impact assessment associated with the directive, it is expected that companies will increase their spending on computer security by up to 22 percent in the first years following its implementation.”

Economics consultancy Frontier Economics assessed the NIS2 costs and its report states: “The direct costs of implementing the regulation on firms across the EU is €31.2 billion per year representing 0.31 percent of total turnover across all of the sectors that are affected by the NIS2 Directive … This represents a large increase in costs given that the EC estimated that average ICT security spending as a percentage of turnover was 0.52 percent in 2020.” 

It provided a chart of likely costs per affected economic sector:

Wallix suggests: “Although this increase in spending may seem substantial, it is expected to be offset by a significant reduction in costs associated with cybersecurity incidents.”

Veeam Software commissioned the survey from Censuswide, which gathered the views of more than 500 IT decision-makers from Belgium, France, Germany, the Netherlands, and the UK. The UK was included due to its significant business ties with EU countries. Nearly 80 percent of businesses are confident in their ability to eventually comply with NIS2 guidelines, but up to two-thirds state they will miss this deadline.

The main reasons cited were technical debt (24 percent), lack of leadership understanding (23 percent), and insufficient budget/investments (21 percent). The survey found that 42 percent of respondents who consider NIS2 insignificant for EU cybersecurity improvements attribute this to inadequate consequences of non-compliance. That’s led “to widespread apathy towards the directive.”

Troskie said: “The combined pressures of other business priorities and IT challenges can explain the delays, but this does not lessen the urgency. Given the rising frequency and severity of cyberthreats, the potential benefits of NIS2 in preventing critical incidents and bolstering data resilience can’t be overstated. Leadership teams must act swiftly to bridge these gaps and ensure compliance, not just for regulatory sake but to genuinely enhance organizational robustness and safeguard critical data.”

Veeam provides an NIS2 compliance checklist, assessment, and white paper.

VAST links up with Equinix and Cisco

VAST data is putting its AI storage and processing kit in Equinix colos and running its software on Cisco’s UCS servers as it broadens its routes to market. 

The company has an existing deal with Cisco concerning Cisco’s Nexus 9000 Ethernet switches used in HyperFabric AI clusters. VAST has certified Cisco Nexus Ethernet-based switches for validated designs with its storage. Cisco customers can monitor and correlate storage performance and latency using VAST’s APIs, feeding network and storage telemetry into the Nexus HyperFabric.

Now Cisco plans to offer the VAST Data Platform software natively on select UCS servers as an integrated system via its global sales team and channel partners. Cisco and VAST say “this full-stack enterprise AI solution simplifies the design, deployment, and management of AI infrastructure for Generative AI, RAG-based inferencing, and fine-tuning AI workloads.”

John Mao

John Mao, VP, Technology Alliances at VAST, stated: “This tight integration and joint selling motion between VAST and Cisco will help to accelerate enterprise AI adoption by providing end-to-end visibility of compute, networking, storage and data management – allowing organizations to seamlessly build and scale their AI operations.

Jeremy Foster, SVP and GM of Cisco Compute, claimed the VAST-Cisco partnership would “massively simplify the overall operation of AI-ready data centers, enabling customers to reduce time, resources and costs required by delivering an integrated stack of the next-generation of storage, compute and networking.”

Cisco UCS servers with VAST Data software and Cisco’s Nexus HyperFabric AI will be available in the first half of 2025.

Equinix and VAST

Equinix has an existing set of 236 International Business Exchange (IBX) globally-distributed co-location centres providing compute, storage and networking gear from suppliers such as Dell, NetApp, PureStorage and Seagate (Lyve Cloud). Some of these are made available to customers at 26 IBX locations through its Equinix Metal as-a-service business model.

It’s now going to provide VAST’s Data Platform for Nvidia DGX systems, including SuperPOD, and the Nvidia AI Enterprise platform in IBX colos as well. VAST and Equinix says this “leverages and supports Nvidia accelerated computing with the VAST Data Platform to deliver a parallel file and object storage system that is ideal for model training and distribution – speeding AI adoption and time to market.”

Renen Hallak

Renen Hallak, Founder and CEO of VAST Data, stated: “By combining supercomputing services with VAST’s AI-native offering for scalable and secure data services deployed in Equinix data centers, we’re setting a new standard in delivering high-performance, scalable, simple and sustainable accelerated computing infrastructure.”

Jon Lin, EVP and GM, Data Center Services at Equinix, said: “Equinix is helping customers access the benefits of AI by providing a fully managed global platform in close proximity to their data through private, high-bandwidth interconnections to cloud providers.”

VAST integrates Nvidia GPUs and NIM for AI insights

VAST Data has brought Nvidia GPU hardware and NIM microservices software into its AI storage and data processing to create an InsightEngine product providing real-time and automatically triggered AI model data access and analytical insights.

It has announced partnerships with Cisco and Equinix to widen its product’s route to market and set up a Cosmos AI user community centered on building an ecosystem of partners and users exchanging ideas around building AI deployments and use-cases using its products. We’ll cover the Cisco, Equinix and Cosmos news in separate stories and focus on the Nvidia GPU and NIM news here.

VAST slide deck diagram
VAST slide deck diagram

This VAST and Nvidia announcement builds on VAST’s storage+server data platform, which has a base of all-flash storage and a software stack comprising a Data Catalog, global namespace (DataSpace), unstructured DataStore, structured DataBase, and AI process-triggering Data Engine. Its all-flash storage has a DASE (Disaggregated and Shared-Everything Architecture) with scale-out x86-based controller nodes (C-nodes) linking to data-storing, all-flash, D-nodes across InfiniBand or RoCE links with a 200/400 Gbps networking fabric. The C-node and D-node software can run in shared industry-standard servers and the D-node software can also run in Nvidia BlueField3 DPUs.

The existing Nvidia partnership has VAST’s system certified for the DGX SuperPOD. This has been extended so that Nvidia GPUs can now be VAST controller nodes:

VAST slide deck diagram
VAST slide deck diagram

That means that the GPUs can work directly on data stored in the VAST array without it having to be first moved to the GPU server. Secondly, Nvidia’s NIM microservices software now runs natively inside the VAST software environment. NIM provides Gen AI Large language models (LLMs) as optimized containers. These simplify and accelerate the deployment of custom and pre-trained AI models across clouds, datacenters and workstations.

Justin Boitano, Nvidia
Justin Boitano

Justin Boitano, VP Enterprise AI at Nvidia, stated: “Integrating Nvidia NIM into VAST InsightEngine with Nvidia helps enterprises more securely and efficiently access data at any scale to quickly convert it into actionable insights.”

VAST says its software with NIM embeds “the semantic meaning of incoming data using advanced models powered by Nvidia GPUs. The vector and graph embeddings are then stored in the VAST DataBase within milliseconds after the data is captured to ensure that any new file, object, table or streaming data* is instantly ready for advanced AI retrieval and inference operations.”

InsightEngine uses VAST’s DataEngine to trigger the Nvidia NIM embedding agent as soon as new data is written to the system, allowing for real-time creation of vector embeddings or graph relationships from unstructured data. Such vectors and graphs are used in RAG (Retrieval-Augmented Generation) whereby a customer’s proprietary data is used to inform LLM query responses to make them more accurate and less prone to fabricate data relationships (hallucinations).

The VAST DataBase can store “exabytes of both structured and unstructured enterprise datasets” and “trillions of embeddings,” and run “real-time similarity search across massive vector spaces and knowledge graphs.” 

VAST slide deck diagram
VAST slide deck diagram

Data indexing occurs at the point of data ingestion and “this architecture eliminates the need for separate data lakes and external SaaS platforms.” The data is held and processed securely. VAST says “any file system or object storage data update is atomically synced with the vector database and its indices, offering comprehensive, secure data access management and global data provenance to ensure data consistency across multi-tenant environments.”

VAST diagram
Jeff Denworth, VAST Data
Jeff Denworth

Jeff Denworth, co-founder at VAST Data, stated: “With the VAST Data Platform’s unique architecture, embedded with Nvidia NIM, we’re making it simple for organizations to extract insights from their data in real-time. By unifying all elements of the AI retrieval pipeline into an enterprise data foundation, VAST Data InsightEngine with Nvidia is the industry’s first solution to provide a universal view into all of an enterprise’s structured and unstructured data to achieve advanced AI-enabled decision-making.”

VAST’s InsightEngine with Nvidia will be generally available in early 2025. Learn more here.

Bootnote

*We understand block data support is coming to VAST Data.

VAST sets up Cosmos partner-user group to spread platform

VAST Data is setting up a Cosmos community of customers, partners and practitioners to exchange VAST Data use case information.

This announcement comes as VAST also announces Nvidia GPU and NIM adoption in its Data Platform, plus bundling its SW with Cisco’s UCS servers and making it available in Equinix IBX co-location sites.

VAST says Cosmos aims to streamline AI adoption for its members by offering a comprehensive, interconnected ecosystem that facilitates conversation, shares use cases, and provides learning opportunities through labs, vendor showcases, and general AI research news. As AI usage is still in its early days, VAST says, Cosmos will help members stay informed and be supported.

Renen Hallak

Renen Hallak, Founder and CEO of VAST Data, stated: “With the VAST Data Platform at the center of this comprehensive, interconnected AI ecosystem of technology leaders and AI practitioners, Cosmos will help accelerate discovery, empowering innovation, and enabling the transformation of entire industries.”

There are three main claimed membership benefits, with the first being faster AI development and deployment helped by AI Labs with pre-sales-type demos of working AI reference architecture systems. AI stack component providers can partner in infrastructure building exercises and, thirdly, Cosmos will facilitate knowledge sharing by hosting interactive events with industry insiders and AI experts.

These, VAST says, will “allow participants to ask in-depth questions, receive tailored advice, and gain clarity on complex topics.” It sounds very much like a classic legacy mainframe and enterprise app supplier user group.

Early Cosmos participants include VAST Data, of course, Nvidia, Elon Musk’s xAI human-centered AI business, server vendor Supermicro, consultancy Deloitte, technology services provider WWT, Cisco, GPU-as-a-service supplier CoreWeave, WWT’s data center platform Core42, healthcare and tech venture capital business NEA, tech systems house Impetus, AI infrastructure supplier Run:AI, and datalake supplier Dremio.

Cosmos is a selling opportunity for product and service-supplying members, as Jeetu Patel, EVP and Chief Product Officer at Cisco, indicated: “We’re in a new era. With the promise and the complexity of AI, data centers, both public and private, must be reimagined to meet the needs of these new AI workloads. The scale of this change will only be possible if we collaborate across the technology stack. Cisco is working with VAST, NVIDIA and others to build modular infrastructure that allows organizations to quickly deploy these AI workloads, including the networks to support them.”

Ditto Stephen Brown, AI Factory Leader at Deloitte, commented: “A strong data foundation is critical for successfully scaling AI and we look forward to collaborating with members of the Cosmos community to help clients extract tangible value from their GenAI initiatives.”

Mitch Ashley, Chief Technology Advisor for The Futurum Group possibly went a tad over the top in his statement: “Cosmos is a once-in-a-generation opportunity for industry and technology leaders to garner the once unimaginable benefits from AI, which would be unachievable if we go it alone. It’s incumbent upon us to take bold steps like Cosmos that can reshape our future solutions possible with AI.”

Interested potential Cosmos participants can join the community here, register for a World Tour with sessions in US, European and Asia-Pacific region cities, and read a VAST blog. There is also a Cosmos web event on October 2.

Eon secures $127 million to turn cloud backup headaches into searchable snapshots

Stealth-emerging Israeli startup Eon says it transforms traditional, hard-to-use cloud backups into easy-to-manage assets with granular restores and searchable database snapshots.

Eon, previously known as PolarKeep, was formally founded in January this year by CEO Ofir Ehrlich, CTO Ron Kimchi, and CRO Gonen Stein. It says it monitors cloud resource sprawl and brings cloud backup posture management (CBPM) to enterprises. Eon replaces legacy backup tools and generic snapshots, transforming backups into useful, easy-to-manage assets.

The three founders have an AWS background. Ehrlich co-founded Israeli DR startup CloudEndure in 2014 as VP for R&D. When it was bought by AWS for around $200 million in January 2019, he became AWS’s head of engineering, app migration services, and elastic DR – effectively the CloudEndure business unit inside AWS until March 2023. He is an angel investor with a substantial startup portfolio.

From left, Gonen Stein, Ofir Ehrlich and Ron Kimchi

Stein was a CTERA tech sales director then a co-founder of Cloud Endure. After the acquisition, he became AWS’s product owner for migration and DR services. Kimchi was AWS’s general manager for DR and cloud migration, joining AWS from Algotech in September 2019.

Since January 2024, Eon says it has secured a $20 million seed round led by Sequoia Capital, with participation from Vine Ventures, Meron Capital, and Eight Roads. Then a $30 million A-round was led by Lightspeed Venture Partners with participation from Sheva, and a $77 million B-round led by Greenoaks with participation from Quiet Ventures followed.

That makes a total of $127 million raised, from seed funding to B-round, in just nine months – surely some kind of startup funding record, and indicative of Eon hitting its product development milestones rapidly.

Ehrlich said: “We are fortunate to have supportive funding partners who deeply understand the value of unlocking cloud backups to be truly automated, globally searchable, portable, and useful.”

Greenoaks partner Patrick Backhouse said: “Storage and backup are among the largest parts of the IT budget. Yet customers are stuck with frustrating, outdated options, leaving them with poorly optimized costs, incomplete data inventories, and shallow classification. Eon has the team, the expertise, and the ambition to develop an entirely new product that we believe will become the cognitive referent for cloud-native backup.”

Sequoia partner Shaun Maguire said: “In an industry where file restoration can take weeks, Eon’s novel backup solution pinpoints data instantly, saving time, money, and compliance headaches for customers.”

Eon notes that the global cloud infrastructure market is growing at an aggressive pace, expected to reach $838 billion by 2034, with enterprises estimating that 10 to 30 percent of their total cloud bill will be spent on backup storage and management. Current backup management methods, Eon claims, require time-consuming, manual data classification and tagging processes, agents and appliances, with mounting prohibitive costs, and ultimately produce backups that are not accessible.

Ehrlich states: “Eon has reimagined what backups can be for enterprises by introducing a new era of cloud backup storage and management.”

Eon claims its software, which is fully managed and portable, autonomously scans, maps, and classifies cloud resources continuously, providing backup recommendations based on business and compliance needs, and ensuring the appropriate backup policy is in use. Existing backup offerings, it claims, rely on snapshots, which are non-searchable black boxes that require full restores and are vendor-locked. 

In contrast, Eon’s backup storage provides global search capabilities, enabling customers to find and restore individual files. They can even run SQL queries on Eon’s backed-up database snapshots, which are searchable, without any resource provisioning. This suggests that Eon produces and stores backup file and database record metadata to provide the search repository.

We asked Gonen Stein questions about Eon’s technology.

Blocks & Files: Which cloud resources does Eon automatically scan, map and classify? 

Gonen Stein: Eon supports scanning and backing up cloud resources including; block, file, and object storage, as well as managed and unmanaged databases. Eon will continue to roll out support for additional cloud infrastructure resources.

Blocks & Files: How does Eon provide continuous backup recommendations based on business and compliance needs, and what do you mean by ‘the appropriate backup policy’?

    Gonen Stein: Eon continuously scans cloud resources and then maps and classifies them based on environment type (prod/dev/staging / QA…), and data classes (PII, PHI, financial info…), all with no tagging required.

    After mapping and classifying resources, Eon applies the appropriate backup policies on the mapped resources, based on the customer backup requirement (i.e.: production workloads containing PII data need to be backed up for 90 days, across cloud regions). This helps customers set backup retention to the right period, reducing storage costs associated with over-backing-up data, while preventing unnecessary business exposure.

    This automated approach is in contrast to today’s completely manual process, where customers need to constantly tag resources, and then manually associate backup policies based on resource tags.

    Blocks & Files: You say Eon’s next generation of backup storage is fully managed and portable- where is it portable to? 

    Gonen Stein: Eon creates a new tier of backup storage (we call them Eon snapshots), which does not rely on traditional vendor-locked cloud snapshots. Eon snapshots can be backed up from one cloud provider to another and also support restoration to a different cloud provider.

      Blocks & Files: How does it provide global search capabilities? 

        Gonen Stein: Eon’s new tier of storage (Eon snapshots), is automatically indexed, and unlike traditional black-boxed snapshots, is globally searchable. This means that a user can search for any backed-up files, or DB tables, without having to know what snapshot it was stored in.

        Blocks & Files: You have said that snapshots are non-searchable, so how do your customers find and restore individual files and run SQL queries on backed-up database snapshots? 

          Gonen Stein: To clarify, traditional snapshots (such as EBS snapshots and RDS snapshots) are not searchable. Eon snapshots are searchable. In addition to Eon’s global search capabilities, Eon also provides a database explorer, which allows customers to run a SQL query right on top of database backups, without requiring the customer to restore and provision full databases, before attempting to retrieve DB records.

          Blocks & Files: How do you restore individual files without any resource provisioning? 

            Gonen Stein: Eon Snapshots allows restoring files directly from the Eon console by searching for files in the backups, selecting specific files (from any backed-up version), and restoring them. This is in contrast to files stored in traditional snapshots, which require the customer to first figure out where the files are stored, then restore the whole snapshot (all or nothing), and only then locate the specific file in the restored volume.

            ****

            There is no connection, as we understand it, between Eon and Infortrend’s EonCloud products, nor between Eon and the similarly named German multinational electric utility company.