Home Blog Page 11

MinIO gives object storage a GenAI upgrade with Iceberg

Interview: MinIO is building a linking layer between its object store and GenAI and that involves embracing structure in the form of Iceberg tables. It also entails building some linkage between them and, up until now, the vector-focused GenAI models and agents.

We talked to MinIO co-founder and co-CEO AB Periasamy and Erik Frieberg, his CMO, to explore this topic. It’s fairly complex and we’ll cover it in two parts, with this being part one. The interview has been edited for brevity.

Blocks & Files: What’s the main thing affecting MinIO presently?

AB Periasamy

AB Periasamy: The best thing that happened to us is GenAI and in every enterprise the budget is now towards GenAI and the race is starting to shift towards data. Who has more data and how to put data to use? That is directly contributing to our company’s growth. Because once you start scaling your data and AI data infrastructure, it points to object store and customers are also looking beyond cloud. Cloud is already behind GenAI and the scale of data is pushing the customers to good private cloud and the private cloud object store, everything is in our backyard. We created that market outside of AWS and that’s contributing to our growth. And you can see the company has been aggressively growing hiring in on all sites, engineering, like marketing, sales, even. We launched a government BU and are launching a partner program. Overall, the company is going through an upgrade and aggressive growth.

Blocks & Files: How many people work at MinIO now?

AB Periasamy: I stop counting, Erik. How much? It is like 150-plus.

Erik Frieberg: I would say 160 something now. It’ll be 170-ish next week.

Blocks & Files: What’s happening with MinIO marketing?

Erik Frieberg

Erik Frieberg: I joined a little while ago to run marketing. The go-to-market organization a month before I joined in November was dramatically different than it is today. I think, that’s no joke, eight or nine times of what it was a year ago.

Blocks & Files: Why the strategy to have that built up so quickly?

AB Periasamy: I made the decisions. I’ll just give the high level view that, even though we are hiring aggressively relative to others, we are very prudent about hiring because, in this AI world, we have to think about hiring very differently. All the junior-level jobs today, machines can do so much better and most other companies hire aggressively and then fire aggressively. 

For us, we treat our team like family and it’s so important to take your time, be prudent about every one of those who you bring on board. Two years from now, five years from now, are they going to be with the family and are they going to be more productive because they can combine themselves with AI? That’s the only hope. Human plus machine is going to be better than human and machine in isolation. 

So be prudent about bringing the right people in who will fill the holes we have inside the organization. That’s how we are looking at it. So, in many ways, I actually look at this as this is not fast enough, but I would rather hire less and not hire anybody than hire aggressively.

Erik Frieberg: I’m going to say three things. One is I would actually say AB probably started hiring too late, so this should have been on the rise already. Second is the difference of having AIStor. I come from an open source background. I was at MongoDB where there was no commercial differentiation and it makes a big difference. There’s always only so much support and services you can sell to customers. 

You’ve got to have differentiated production software that has features that are a catalyst for growth. And that [AIStor] came out in the October-ish time frame and I think people are now really understanding it, seeing the difference. You’re going to see some announcements coming in the future on this, so that’s creating more opportunity. And then like AB said, it’s not a target. It’s not that he’s saying: “Go hire five people in marketing.” He’s saying: “Hey, we need more product marketing now because we have more to talk about.”

Blocks & Files: AB, MinIO is a storage company and for a customer of yours who is using generative AI, large language models and agents, there’s a software stack between the storage layer and their agents and large language models. So where in that stack do you see MinIO’s activities ending? Where’s the border and is it moving up the stack?

AB Periasamy: Yes. In fact, the story we have been telling all along resonates more with the customers now than the past. We always said MinIO object store is a key value store. What is a key value store? It’s a data store. We are closer to Oracle than EMC in the past. The data store is the dumbest of all databases. It is a key value store with get and put objects, but it needs to process data at massive scale without losing a single object. It has to be transaction ACID-compliant, all of that and scale, while it has the simplicity at its roots. So we always thought of ourselves as a database company, except that we are talking about unstructured data. Things are starting to evolve there because of GenAI.

We brought Erik in because he came from MongoDB. Our culture is closer to MongoDB than Pure or NetApp or MCA software company and in modern times [with] GenAI, the scale is much larger. The only way you can go is scale-out and then go software-defined. There are some interesting things that are happening [with] the object store and the layer above Gen AI, where the integration is going on.

The AI is directly talking to the object store through the MCP server. It’s the agents that are interacting with the object store, both in terms of administration as well as the data discovery and dealing with the data itself. There are some interesting challenges that are emerging here. If it’s a spreadsheet, if it’s a CSV file, single object? Say I have a database file, can I analyze this? We are talking about enterprises having 10 petabyte tables; data sitting in Iceberg tables. In the past it was proprietary data sitting in proprietary databases. They are increasingly now moving towards Iceberg open table format.

How can you have large language models understand that scale of data? That’s a bridging factor that we need to work on and those are the areas that we are actually actively investing in as well. The model may still be running on OpenAI or Anthropic, but the enterprise data never leaves their facility. How do they discover the data? Only the metadata or an anonymized form of the intelligence is shared. We are actively working on that part of the stack.

There’s also another interesting twist here that, if it’s unstructured data like blobs, photos, videos, documents, then everybody knows how LLM take advantage of them. The new emerging area is structured data, structured data like Iceberg tables, and at large scale. How can LLMs understand structured data?

Comment 

The point AB is making is that Iceberg tables are structured and it makes no sense to vectorize them as, unlike words or images, they don’t stand alone with dimensional aspects that you can vectorize. There needs to be some intervening logic/abstraction between the tables and the GenAI LLMs and agents that bridges their unstructured data vector focus and the searching of structured data. We’ll explore this more in part two.

The Iceberg age: StarTree latest to adopt popular table format

Iceberg is becoming the lingua franca of datalake table formats with StarTree the latest supplier to embrace it as a real-time backend.

Open source Iceberg is an open table format for large-scale analytics which doesn’t store or execute queries itself, as a traditional database would. It operates as a software layer atop storage systems like Parquet, ORC, and Avro, and cloud object stores such as AWS S3, Azure Blob, and the Google Cloud Store, which handle handles metadata, partitioning, and schema evolution.  Iceberg provides ACID transactions, schema versioning, and time travel, with data querying and processing handled by separate SW such Apache Flink, Presto, Spark, Trino, and other analytics engines. 

The StarTree Cloud is a fully managed, cloud-native platform built on Apache Pinot, a real-time distributed OLAP (Online Analytical Processing) datastore. StarTree Cloud is designed for OLAP and enables low-latency querying (milliseconds) and high-throughput processing (10,000+ queries/second) of large-scale data from streaming sources (e.g., Apache Kafka, Amazon Kinesis) and batch sources (e.g., AWS S3, Snowflake). Now it can be both the analytic and serving layer on top of Iceberg.

StarTree claims Iceberg support can transform it from a passive storage format into a real-time backend capable of powering customer-facing applications and AI agents with high concurrency serving thousands of simultaneous users with consistent speed and reliability. 

Kishore Gopalakrishna on the summit of Half Dome, Yosemite.

Kishore Gopalakrishna, StarTree co-founder and CEO, stated: “We’re seeing explosive growth in customer-facing, and increasingly agent-facing, data products that demand sub-second responsiveness and fresh insights. At the same time, Iceberg is emerging as the industry standard for managing historical data at scale.”

“As these two trends converge, StarTree is delivering unique value by acting as a real-time serving layer for Iceberg empowering companies to serve millions of external users and AI agents securely, scalably, and without moving data.”

Recent Iceberg adoptees include Snowflake, Confluent, AWS with S3, SingleStore, and Databricks

Paul Nashawaty, principal analyst at theCUBE Research, said: “Apache Iceberg is rapidly becoming the de facto standard for managing large-scale analytical data in the open data lakehouse—adoption has surged by over 60 percent year-over-year, according to theCUBE Research.”

StarTree asserts that most existing query engines built around Iceberg and Parquet struggle to meet the performance SLAs required for external-facing, high-concurrency analytical applications, and companies have historically avoided serving data directly from their lakehouse. It claims that by combining Iceberg and Parquet open table formats Pinot’s indexing and high-performance serving capabilities, StarTree offers real-time query acceleration directly on native Iceberg tables.

StarTree Cloud graphic.

Unlike Presto, Trino and similar engines, StarTree says it’s built for low-latency, high-concurrency access, integrating directly with Iceberg, boosting performance with features such as:

  • Native support for Apache Iceberg and Parquet in StarTree Cloud
  • Real-time indexing and aggregations, including support for numerical, text, JSON, and geo indexes
  • Intelligent materialized views via the StarTree Index
  • Local caching and pruning for low-latency, high-concurrency queries
  • No data movement required—serve directly from Iceberg
  • Intelligent prefetching from Iceberg, minimizing irrelevant data scans 

Nashawaty reckons: “StarTree’s ability to serve Iceberg data with sub-second latency and without data duplication is a unique and timely advancement. It addresses a critical performance need for accessing historical data in modern data products.”

Support for Apache Iceberg in StarTree Cloud is available today in private preview. For more information, visit www.startree.ai.

Kioxia unveils highest capacity SSD at 245.76 TB

Online storage capacity constraints are melting away with Kioxia announcing the largest capacity SSD at 245.76 TB, opening the door to a 200-plus petabyte rack and a five-rack exabyte configuration.

It complements the already announced 122.88 TB model and is intended for use in data lakes – where massive data ingestion and rapid processing are required – and generative AI, which requires storing large datasets for training large language models (LLMs), as well as the embeddings and vector databases that support inference via retrieval-augmented generation (RAG).

These drives are aimed at very large organizations and hyperscalers that need this level of capacity muscle. 

Neville Ichhaporia, Kioxia
Neville Ichhaporia

Neville Ichhaporia, SVP and GM of the SSD business unit at Kioxia America, said the LC9 series sets “a new benchmark for capacity and innovation in the SSD space. As generative AI reshapes datacenter architecture, we’re delivering the kind of breakthrough technology that enables our customers to optimize their IT infrastructure investments – pushing boundaries in performance, efficiency, and scale.”

Kioxia now has twice as much drive capacity as competitors Phison (122.88 TB) and Solidigm (122 TB), which announced their drives in November last year. Since Sandisk is a NAND fab joint-venture partner of Kioxia, it could also produce a similar-sized SSD to the LC9.

The LC9 drives use Kioxia’s BiCS8 218-layer 3D NAND organized in QLC (4bits/cell) format with 2 Tb dies. These 32 dies are stacked to produce 8 TB of capacity in a compact 154-ball BGA (Ball Grid Array) package. Kioxia says this is an industry-first. 

We imagine that Phison might find scope for parallelism with around 30 of these BGAs needed for a 245.76 TB drive, were it to be minded and able to buy the component chips and produce one.

A Kioxia slide provides LC9 speeds-and-feeds information:

The drives are available in U.2 (2.5-inch), E1.S, and E1.L formats, though maximum capacity is limited to 122.88 TB in the 2.5-inch and E1.S variants.

These SSDs have a PCIe 5×4 interface, or PCIe 5×2 in a dual-port config, along with NVMe. Performance is heavily skewed to reads, with an up to 12 GBps bandwidth versus just 3 GBps for writes. This is not that fast for PCIe Gen 5 SSDs. Kioxia’s own CD9P drive offers up to 14.8 GBps reads and 7 GBps writes but it does use faster TLC NAND.

Phison’s 128 TB Pascari D205V uses QLC NAND, like the LC9, and is faster, delivering up to 14.6 GBps read bandwidth and 3.2 GBps write speeds. The SK hynix PS1012 drive is also a QLC SSD and is slightly faster as well. It pumps out 13 GBps read bandwidth at the 61 TB capacity level. A 122 TB version is due later this year.

Kioxia’s LC9 SSDs

The LC9 has Flexible Data placement support to improve endurance, at 0.3 drive writes per day, and performance. It also has various security features, such as self-encryption, and meets the Open Compute Project (OCP) Datacenter NVMe SSD v2.5 specification, but not all requirements.

We envisage other SSD suppliers announcing their own similarly capacious SSDs and pitching them at the same market soon enough. It’s noteworthy that Pure Storage is introducing 300 TB versions of its proprietary NAND drive, the Direct Flash Module (DFM), and Meta has licensed DFM IP. Both Micron and Kioxia are NAND chip suppliers to Pure.

The pace of SSD capacity expansion is rapid, far outpacing disk drive capacity growth, which is currently in the 32-36 TB area with 40 TB drives forecast. There is no pricing information available for the LC9 so we can’t make a $/TB comparison with HDDs. A rack full of LC9s could potentially significantly outperform disk drives in TCO terms because of the reduced floor and rack space, cooling, and most likely the power they need.

Kioxia’s coming BiCS 10 generation of NAND will feature 332 layers, 52 percent more than BiCS 9. That would enable, in theory, a 3 Tb QLC die and, by extrapolation, a 368 TB SSD. This would have seemed crazy three or four years ago. Such is the way our expectations of what’s normal in the NAND world have changed.

Kioxia LC9 Series SSDs are now sampling with select customers and will be featured at the Future of Memory and Storage 2025 conference, taking place August 5-7 in Santa Clara.

LTO tape market rebounds in 2024 with 176.5 EB shipped

LTO tape
LTO tape

There was 15.4 percent more LTO tape capacity shipped in 2024 compared to 2023 as growth resumed following three years of more or less level pegging after a pandemic dip.

The LTO Program Technology Provider Companies (TPCs) – HPE, IBM, and Quantum – released their annual tape media shipment report, saying a record 176.5 exabytes (EB) of total compressed tape capacity shipped in 2024, assuming a 2.5:1 compression ratio. Similar numbers are available for 2019 to 2023, but not for earlier years. However, the LTO org has previously produced a bar chart showing the 2008 to 2019 history. We have combined the old and new charts to provide a full 2008 to 2024 tape capacity shipped history: 

There was a dip in 2018 due to a patent dispute between the two tape ribbon manufacturers, Fujifilm and Sony, which restricted the global supply of LTO-8 tape media until a cross-licensing deal was agreed between the two.

Capacity shipped then grew in 2019 to 114 EB, but was hit hard by the COVID-19 pandemic in 2020 with a fall to 105 EB. There was a ramp back up in 2021 to 148 EB, but dismal 0.5 percent growth in 2022 to 148.3 EB, and a laggardly 3.14 percent rise to 152.9 EB in 2023. Happily, the need to archive unstructured data has now resumed growing and we have a healthier 15.4 percent growth from 2023 to 2024.

Bruno Hald

Looking ahead to the hoped-for AI archiving demand, Bruno Hald, General Manager, Secondary Storage, at Quantum, said: “Setting a new growth record for the fourth year in a row, LTO tape technology continues to prove its longevity as a leading enterprise storage solution. Organizations navigating their way through the AI/ML era need to reconfigure their storage architectures to keep up, and LTO tape technology is an essential piece of the puzzle for those seeking a cost-friendly, sustainable, and secure solution to support modern technology implementation and the resulting data growth. We look forward to introducing the next iteration of LTO tape technology this year to bring enhanced storage capabilities to the enterprise.” Hald is referring to the LTO-10 format.

The LTO org says in its announcement that tape complements SSD and disk as an archival storage medium “to provide a low-cost, last line of defense against ransomware and other security threats.” And it mentions air-gapping, of course: “Further, offline copies of LTO tapes separate data archives from the connected environment, providing greater assurances for data recovery.”

It does not mention that LTO-10, unlike previous generational LTO changes, does not provide a speed increase over LTO-9, only saying: “LTO-10 … will provide the same blazing speed as LTO-9.” Nor does it mention LTO-10’s loss of backwards-compatibility or initial 30 TB raw capacity rather than the expected 36 TB raw capacity. Setting these relatively minor points aside, the future for LTO tape shipments looks good for the rest of 2025 and 2026. But if Microsoft’s Project Silica or Cerabyte or HoloMem get their various glass tablet technologies productized by 2027, 2028, or 2029, then we might see some real competition for tape developing.

MASV Express aims to automate complex file workflows without code

Massive file transfer supplier MASV has an upcoming MASV Express product that can automate complex media ingest and delivery processes without coding or any complicated setup.

It can move files to the cloud or connected on-prem storage, MAMs (Media Asset Managers), and individual recipients with no coding, no desktop client, drag-and-drop workflows, two-way encryption, secure authentication, and privacy controls. The files can be up to 20 GB or more in size.

Majed Alhajry, MASV
Majed Alhajry

Majed Alhajry, CTO and interim CEO of MASV, stated: “File transfer is no longer just a background process; it’s a critical function in the global media supply chain. When files move faster, more reliably, and with less friction, content gets to market sooner and that means more opportunities to monetize across broadcast, digital, and OTT (Over the Top) platforms.”

OTT refers to the delivery of audio, video, and other media content directly to viewers over the internet, bypassing traditional distribution channels like cable or satellite TV. MASV’s file transfer moves files at up to 10 Gbps across an AWS server and content delivery network. It uses TCP-based file delivery, with chunking – splitting files into smaller parts for parallel transfer – and AWS’s private network infrastructure to accelerate data movement.

A customer such as Troveo, which helps creators license unused video footage for AI training, uses MASV for fast, no-code ingestion of over 6,000 TB of video each month. Sarah Barrick, Head of Growth at Troveo, said: “If something like MASV did not exist, it likely would have precluded a lot of creators from monetizing their video assets with Troveo. The ultimate factor was the usability. It was just so easy for users of all kinds, from small creators to larger enterprises – they didn’t even have to download a desktop client. They didn’t need any technical expertise, and could use a private, secure portal to begin uploading right away.”

MASV partners with Qumulo, and has integrations with AWS, Backblaze, Frame.io, Iconik, Mystika Workflows, Wasabi, and other suppliers. Think of MASV as a dedicated large file transfer service, overlapping in that regard with Arcitecta, LucidLink, and PeerGFS, which all offer more functionality alongside basic file moving. The MASV Express offering will be showcased at IBC2025 in The Netherlands, Stand 1.B01, RAI Amsterdam, September 12-15, 2025.

Bootnote

MASV was a passion project within a small Canadian networking company before spinning off in late 2019. That company was LiveQoS where Majed Alhajry was a software engineer between 2014 and 2017. LiveQoS acquired him and his SuperBeam peer-to-peer file sharing app in June 2014. We understand that MASV started life as a LiveQoS VPN project to improve the reliability and performance of Office 365. MASV then evolved to move massive media files, such as rushes, across the internet, and became MASV Rush. It was in a preliminary development phase from late 2017 to late 2019 with Alhajry the technical team lead. Effectively, he was a MASV co-founder.

MASV was officially founded in December 2019 as a pay-as-you-go large file transfer service, with co-founder Gregg Wood as CEO. He is an ex-Corel Senior Director of Product Management who left in January 2017 and became a C-Level advisor at the Positive Venture Group in Ottawa. Early funding for MASV included investment from Peter Lalonde, a Positive Venture Group partner. Wood stepped down from the C-Level Advisor role in September 2018, initially becoming MASV’s VP Product Management and then, when it officially launched, its CEO.

MASV CFO Jeff Stoss is another co-founder, having been in at the start and CFO at LiveQoS from 2016. Stoss was also the co-founder, CFO, and C-Level Advisor at the Positive Venture Group. This was sold to Consero Global in July 2021. LiveQoS, with its residual SD-WAN IP, was sold to IP licensing business Adaptiv Networks in 2019.

We asked MASV some quesions about its recent history;

Blocks & Files: Is Majed Alhahry a co-founder of MASV? Same for CFO Jeff Stoss?

MASV: Yes, Majed and Jeff are co-founders along with Greg Wood who remains active in MASV via his vice-chairman role. 

Blocks & Files: Why did CEO Greg Wood become vice-chairman at MASV?

MASV: Co-founder Greg has made the transition from CEO to Vice Chairman after his successful 5 year stint in leading MASV from its founding. This is a strategic succession to cofounding CTO Majed Alhajry, reflecting the company’s evolution into a deeply technical platform for secure enterprise data exchange. Greg did a fantastic job building the business and is now working on new initiatives to take MASV into new markets. Greg remains actively engaged supporting Majed and Team MASV.

Blocks & Files: Why is Majed Alhajry an interim CEO? I understand he’s been in that post since January when Greg became vice-chairman.

MASV: Majed brings a distinguished track record of technical and business leadership as the founder of SuperBeam and co-founder of MASV. His recent appointment as interim CEO, in addition to CTO, reflects a natural progression in his journey with MASV, leveraging his deep expertise to drive the company’s strategic vision. In this expanded role, Majed is spearheading MASV’s push into the enterprise market, bolstered by a growing team, strategic partnerships, and purpose-built solutions like MASV Express. These initiatives—combining cutting-edge features with enterprise-grade capabilities—position MASV to solidify its foothold and accelerate growth in the corporate sector.

Blocks & Files: What is cut-through delivery technology?

MASV: Cut-through delivery is no longer a term MASV will use going forward to refer to Express. Instead, they have explained how Express works:

MASV Express moves files from origin to the destination without first needing the complete file in MASV’s secure container. This way, Express dramatically accelerates turnaround time, on top of MASV’s already super-fast file transfer speed, preserving MASV’s operational agility. MASV Express keeps content flowing smoothly through any network environment, ensuring the fastest, most reliable, and most secure delivery experience available. Express builds on MASV’s current capability to overcome any point-to-point delivery issues—like propagating outages and network congestion—by preserving reliability of the file transfer while decreasing turnaround times.

Storage news ticker – July 22

Adeptia Connect is a business data exchange software moving data between applications, partners, and customers, supporting data integration, application integration, and B2B/EDI processes. It can be single or multi-tenant and deployed on-premises or in the cloud or both. It has AI-driven data mapping, pre-built connectors, and reusable templates, allowing non-technical users to create and manage data workflows. Adeptia has just announced self-managed and hybrid deployment options for Connect. “Hosted on Microsoft Azure and AWS and managed by Adeptia’s specialized Cloud Infrastructure Engineering team, Adeptia’s self-managed option combines the simplicity of SaaS with the enterprise security standards that customers demand. Adeptia handles complex operational tasks – such as deployment optimization, monitoring, backups, disaster recovery, and version upgrades – while maintaining rigorous security compliance, including SOC 2 certification.” More info here.

Data source connector supplier Airbyte has added new features to its Enterprise product: the ability to facilitate data sovereignty and compliance across different global regions; synchronize data with its metadata to assist AI models’ ability to improve reasoning and accuracy; and directly load data volumes to BigQuery and Snowflake, which we’re told increases data sync speeds and reduces compute costs. Direct loading can reduce compute costs by 50 to 70 percent, and increases speed by up to 33 percent depending on the use case. 

Airbyte now enables synchronized unstructured files and structured records to be transferred in the same data pipeline, which is the most efficient way to preserve metadata and data relationships – enabling richer data context and significantly improving the performance of AI Large Language Models (LLMs) using that data. There are more details in the blog post and more about Airbyte Enterprise here.

Assured Data Protection (ADP) is partnering with technology services distributor Avant in North America to offer Rubrik and Nutanix-based managed backup, disaster recovery (DR), and cyber resiliency services. It will mean that Avant can offer Rubrik’s backup and DR technology to customers for the first time with Nutanix technology also being available, depending on needs.

Ataccama ONE v16.2 introduces AI-driven data lineage features designed to help business users audit, trust, and act on their data, without writing a single line of SQL. It says: “Business users can now trace a data point’s origin and understand how it was profiled or flagged without relying on IT. Ataccama shows how data flows through systems and provides plain-language descriptions of the steps behind every number. For example, in a financial services setting, a data steward can immediately see how a risk score was derived or how a flagged transaction passed through a series of enrichment and quality checks. That kind of visibility shortens reviews, streamlines audits, and gives business teams the confidence to act on the data in front of them.”

Broadcom’s Tomahawk Ultra is an Ethernet switch for HPC and AI workloads with 250 ns switch latency at full 51.2 Tbps throughput. It delivers line-rate switching performance even at minimum packet sizes of 64 bytes, supporting up to 77 billion packets per second. It reduces header overhead from 44 bytes down to as low as 6 bytes, while maintaining full Ethernet compliance – boosting network efficiency and enabling flexible, application-specific optimizations. The switch incorporates lossless fabric technology that eliminates packet drops during high-volume data transfer, and implements Link Layer Retry (LLR) and Credit-Based Flow Control (CBFC) to eliminate packet loss and ensure reliability. It’s designed with topology-aware routing to support advanced HPC topologies including Dragonfly, Mesh, and Torus.

Cadence announced the tapeout of the industry’s first LPDDR6/5X memory IP system solution optimized to operate at 14.4 Gbps, up to 50 percent faster than the previous generation of LPDDR DRAM. The Cadence IP for the JEDEC LPDDR6/5X standard consists of an advanced PHY architecture and a high-performance controller designed to maximize power, performance and area (PPA) while supporting both LPDDR6 and LPDDR5X DRAM protocols for optimal flexibility. It says its new LPDDR6/5X memory IP system offering enables scaling up AI infrastructure to accommodate the memory bandwidth and capacity demands of next-generation AI LLMs, agentic AI and other compute-heavy workloads for various verticals. Multiple engagements are currently underway with leading AI, high-performance computing (HPC) and datacenter customers.

Research from data streamer Confluent says more than two-thirds (67 percent) of UK IT leaders see disconnected, siloed data as preventing them from fully adopting AI and machine learning (ML) – making fragmented data systems the largest obstacle to AI success. There was:

  • Ambiguity surrounding data lineage, timeliness, and quality assurance – 63 percent
  • Insufficient skills and expertise in managing AI projects and workflows – 61 percent
  • Limited ability to seamlessly integrate new data sources – 57 percent
  • Insufficient infrastructure for real-time data processing – 50 percent

Download Confluent’s 2025 Data Streaming Report here.

Confluent announced that Confluent Cloud is now available in the new AI Agents and Tools category on AWS Marketplace. It says this enables AWS customers to easily discover, buy, and deploy AI agent solutions, including its fully managed data streaming platform, Confluent Cloud, using their AWS accounts, for accelerating AI agent and agentic workflow development. Confluent Cloud streams, connects, processes, and governs data as it flows throughout the business in real time, enabling customers to build and deploy intelligent, adaptable, and scalable AI applications faster. By replacing fragmented, batch-based systems with a continuously updating stream of business context, Confluent creates a shared, real-time data backbone that AI agents can subscribe to, interact with, and reason over. Confluent Cloud empowers customers to confidently scale their businesses for the future with AI at the forefront. 


China’s CXMT, the parent of DRAM fabber ChangXin Memory, is preparing for an IPO in China. This is another sign, together with YMTC’s NAND plans below, of China’s growing self-sufficiency in memory technology and production. China could emerge as a DRAM and NAND supplier on the world stage.

Data management supplier Denodo’s cloud-based Agora offering is now available in the AWS Marketplace’s new AI Agents and Tools category. Denodo is offering both a SaaS version (Agora) and an Amazon Machine Image (Denodo Enterprise Plus), giving AWS customers a flexible path to build Agentic AI ecosystems using open standards. With support for the Model Context Protocol (MCP) built into the latest Denodo AI SDK, organizations can now unify, govern, and deliver real-time data to GenAI apps without having to replicate or re-engineer their entire data architecture.

Denodo has launched DeepQuery, a new capability that is bringing reasoning-based GenAI to enterprise systems. While most GenAI tools stop at retrieving facts, DeepQuery connects to live, governed enterprise data and delivers answers that explain the why behind the what. These answers are fully reasoned, cited, and ready for decision-making. Rather than pulling data into a silo, DeepQuery applies structured reasoning directly on real-time systems to support sales teams, finance, HR, and more. It’s the first enterprise-grade GenAI to offer this depth of synthesis with full explainability and traceability marking an evolution toward cognitive GenAI, where models don’t just answer – they investigate. Denodo is offering early access to select partners through its AI Accelerator Program. Denodo also announced the availability of Model Context Protocol (MCP) support as part of the Denodo AI SDK.

China’s Exascend offers products like PCIe NVMe SSDs, SATA SSDs, CFast, and CFexpress cards, tailored for demanding applications in aerospace, AI, IoT, and video surveillance. It has two new PE4 SSDs, 7 mm U.2 models: 23.04 TB and 30.72 TB. They “enable system integrators to maximize capacity in edge servers, rugged industrial computers, on-premises AI nodes, and space-constrained datacenters where every watt and every millimeter count.” The PE4 30.72 TB doubles the density of competing 7 mm U.2 SSDs and enables up to 737 TB of flash in a single 1U server, cutting rack space requirements by 50 percent compared to 15 mm drives and delivers a two-fold jump in TB-per-U efficiency.

The PE4 series have a PCIe Gen 4 x 4 interface and use 3D TLC NAND. They are rated for up to 1 drive write per day (DWPD) and offer a 2-million-hour mean time between failures (MTBF). Engineering samples of the PE4 Series U.2 SSD in 23.04 TB and 30.72 TB models are shipping now to select customers, with volume production scheduled for Q3 2025. The PE4 7 mm family is already available in 7.68 TB and 15.36 TB capacities.

Frontgrade Technologies, which provides high-reliability microelectronics for space and national security, has unveiled its embedded MultiMediaCard (eMMC) – a space-grade NAND memory product offering 32 GB density, the highest available in today’s market. Designed to deliver maximum capacity and resilience in space environments, this radiation-tolerant, non-volatile flash data storage device is optimized for low Earth orbit (LEO) missions and beyond. It has an integrated flash controller with built-in FTL for wear leveling, bad block handling, and ECC, and has radiation-tolerant construction suitable for many mission orbit requirements.There is a standard eMMC 5.1 interface for streamlined system integration. Frontgrade’s eMMC is currently available for sampling. Flight units are expected to be available in Q4 2025. Download a datasheet from this webpage

IBM says integrating Storage Ceph with its Storage Deep Archive “presents a compelling solution for organizations looking to optimize their data archival strategies. The combination delivers substantial cost efficiency, operational simplicity, air gap resiliency, and policy-driven flexibility.” Ceph now supports policy-based transitions of data to external S3-compatible storage tiers, including those that present Glacier Retrieval classes. IBM’s Storage Deep Archive has “ultra-low-cost archival capabilities” and “IBM Storage Deep Archive uniquely complements Ceph’s capabilities by offering an S3 Glacier-compatible interface without the traditional overhead associated with tape systems,” although it is a tape library system. An IBM blog has more information.


Note: IBM Storage Deep Archive is an on-premises archival data storage system integrated within an IBM Diamondback server and a 5U top rack, utilizing tape storage with an S3 Glacier command interface. It offers up to 27 petabytes of capacity in a 19-inch rack.

William Blair analyst Jason Ader says of Nutanix: “With respect to the competitive fallout from Broadcom’s acquisition of VMware, Nutanix sees this as a multi-year displacement opportunity, less characterized by a specific tipping point and more by a continuum that will build on itself over time (putting more pieces in place right now to execute on the land grab, including channel enablement, OEM and external storage partnerships, and incremental sales reps). While Nutanix has already seen some benefit from VMware displacements (as evidenced by the recent uptick in new logos and the strong and growing seven-figure pipeline), inertia with VMware customers is real (especially with larger, complex deployments) and software/hardware refresh cycles often need to align for customers to shift off VMware.”

VergeIO is offering replacement storage with the launch of The AFA Replacement Kit. VergeIO says it’s “an offering designed to replace traditional all-flash arrays with a simpler, more cost-effective infrastructure solution. VergeIO customers have reported reducing storage costs by a factor of ten, in addition to the added savings from eliminating expensive VMware licensing and support agreements. This VergeIO kit brings together three Solidigm 4 TB enterprise SSDs and a VergeOS server license combined into one streamlined platform. Along with your servers, it’s a complete, ready-to-run infrastructure solution. All IT needs to do is insert the included flash drives into empty drive bays in existing servers, and they’re ready to deploy VergeOS.” VergeIO marketing head George Crump said: “It is 12 TB per node, typical customer has at least four nodes. If a customer needs more, we have flexibility.” 

We don’t see customers choosing this as a general external AFA replacement, e.g. rejecting NeApp AFF A150, which is both a scale up (2RU x 24 slots) and scale-out external storage system (to 12 x A pairs) whereas the VergeIO AFF Replacement is a scale-out system with three-drive (slot) nodes. We see it specifically as a VMware vSAN replacement.

VergeIO was been asked about this point and a spokesperson said: “The VergeIO AFA Replacement Kit is not limited to three drives or three nodes—those were illustrative starting points for the press release. The solution can scale to any number of drives per server and any number of servers. It’s designed to give IT teams flexibility to size configurations based on workload requirements, not fixed appliance specs.” 

Also the product is both a vSAN and general external AFA replacement, with VergeIO saying: “The AFA Replacement Kit serves both purposes. It functions as a general AFA replacement, suitable for customers considering alternatives to traditional SANs and filers from vendors like Dell, NetApp, Pure, HPE, and others.[And] It is a strong alternative for VMware customers who are moving off vSAN and want server-side flash with integrated compute and storage.”

VergeIO wants us to know “that research from ESG indicates that over 53% of SANs are used exclusively to support virtualized workloads, and over 90% support at least some virtualized workloads. VergeIO is designed to meet that need directly. While VergeOS is not designed to run non-virtualized workloads, many customers find that they can migrate bare-metal applications they previously believed too resource-intensive for virtualization. This is common in VergeOS environments thanks to its efficiency, where storage, compute, AI, and networking are integrated into a single software codebase.”

DigiTimes reports China’s NAND fabbing YMTC aims to pilot production of NAND chips using China-only tech this year. It was estimated to hold 8 percent of the global NAND market (a monthly capacity of 130,000 wafer starts per month) by the end of 2024, and wants to reach a 15 percent share by the end of 2026. This is based on ramping up WSPM to 150,000 and its latest gen 5 NAND memory with 294 layers built from bonded 150L and 144L components (Xtacking 4). 

YMTC is on the US tech export sanctions list and cannot buy NAND fab tools using the latest technology from suppliers such as ASML. Tom’s Hardware has YMTC NAND die info:

  • X4-9070 – 1 TB TLC 3600 MT/s – in production
  • X4-6080 – ? TB QLC – H2 2025
  • X5-9080 – 2 TB TLC with 4800 MT/s – 2026
  • X5-6080 – ? TB QLC with 4800 MT/s – 2026

YMTC’s gen 6 die will likely feature 300+ layers built from 3 component stacks (Xtacking 5).

Zadara software is being used by East Africa services provider TouchNet at an iXAfrica co-location datacenter in Nairobi, Kenya, to provide a sovereign cloud, offering computer, networking, storage, and GPUaaS. The Zadara AI Cloud enables businesses, managed service providers (MSPs), and cloud service providers (CSPs) to optimise operations without the need to rely on offshore services. Zadara provides its own take on the public cloud with fully managed compute (EC2 compatible), storage (block, file, object), and network systems hosted in a global point-of-presence (POP) network, called edge clouds. It can also be run on AWS, Azure, GCP, and Oracle public clouds, or in a customer’s own datacenters – as in this iXAfrica case. This deployment enables sovereign cloud capabilities hosted locally in Kenya, and a public cloud-like experience (compute, storage, networking) using Zadara’s EC2- and S3-compatible services. There is full regulatory and data residency compliance for sectors requiring localized data control (e.g. government, finance, healthcare).

Where to point your RAG: Sourcing proprietary data for LLMs and AI agents

Digital brain

Analysis: There are many places within IT infrastructure that organizations can use to get a single proprietary data source for their large language models and AI agents doing RAG (retrieval-augmented generation) – backup stores, data lakes, data management catalogs, and storage arrays. Let’s sketch the background then take a look at how they compare to each other.

LLMs and agents can scan all kinds of unstructured information in text, numbers, audio, image, and video formats in response to natural language inputs and generate natural language outputs. However, they are initially trained on general data sources and will make better responses to an organization’s users if they have access to the organization’s own data. But that data is spread across many disparate systems and formats.

For example, datacenters (mainframes and x86 servers), edge sites, public cloud instances, SaaS app stores, databases, data warehouses, data lakes, backups on disk and tape on-premises and in the public cloud, and archives. The data can be accessed via structured (block) or unstructured (file and object) protocols with many subsidiary formats: Word, Excel, PDF, MPEG, etc.

There are four theoretical approaches to getting all these different data sources available for RAG. First, you could go to individual specific sources each time, which would require you to know what they are, where they are and what they contain. This approach operates at the level of individual storage arrays or cloud storage instances.

Secondly, you could go to existing agglomerated data sources, meaning databases, data warehouses, and data lakes. Third, you could go to the backup stores and, fourthly, use a data management application that knows about your data.

These second, third, and fourth options already have metadata catalogs describing their content types and locations which makes life a lot easier for the RAG data source hunter. All else being equal, accessing one location to find, filter, select, extract, and move RAG data is better than going to myriad places. It makes AI data pipeline construction much simpler.

A couple of other observations. The more of your proprietary data the single place contains the better as you have to make fewer exception arrangements. Secondly, the LLMs and agents need vectorized unstructured data for their semantic searches, and this needs producing and storing. Any central RAG data source facility needs to support vectorization and vector storage.

That’s the background. Let’s take a gander at the four main RAG data sources.

Storage arrays

Virtually every storage array vendor we know, both hardware and software, software-defined, file and object, is building in some kind of Gen AI support. Cloudian, DDN, Dell, HPE, NetApp, Scality, StorONE, VAST Data, WEKA, and others are all piling in.

Arrays (or SW-defined storage) which have a fabric connecting on-premises arrays, and also public cloud instances with a global namespace  will have an advantage; their reach is obviously greater than non-fabric arrays. This is true for cloud file services suppliers, such as Box, CTERA, Egnyte, Nasuni, and Panzura as well.

However, such vendors can only supply data for RAG stored on their systems and nowhere else, unless they have connectors giving them wider data access. An agent granted access to a Dell, NetApp, DDN, HPE, whatever array won’t be able to see data on another supplier’s array in that organization though the NetApp, DDN, HPE, whatever array RAG lens.

Database/warehouse/lake/lakehouses

A database, data warehouse, data lake and lakehouse store more information respectively and, generally, in more formats as we head from databases to data warehouses, then data lakes and lakehouses. At the database end of this spectrum, specialized vector databases are appearing, from suppliers such as Pinecone and Zilliz (Milvus). They say they offer the best of breed vector storage, filtering, extract, and support for AI pipelines.

Other databases aim to be multi-modal, like SingleStore. Its SingleStoreDB is a performant, distributed, relational, SQL database with operational, analytical, and vector data support, integration with Apache Iceberg, Snowflake, BigQuery, Databricks, and Redshift, and a Flow ingest feature.

Data warehouses are rapidly becoming AI data feed warehouses, witness Snowflake and its AI Data Cloud. Data lakes and lakehouses are also very much GenAI-aware and rapidly developing features to support it. As an example consider Databricks which recently added a Lakebase Postgres database layer to its lakehouse, enabling AI apps and agents to run analytics on operational data within the Databricks environment. It also introduced Agent Bricks, a tool for automated AI agent development.

These are all excellent products, and connectors exist to bring them data from multiple different sources or allow them to access external table software data.

Their operational and admin environment becomes more and more complicated as they extend their reach to external tables or use connectors to bring in data from distant sources

Backup stores

A backup vault stores data copied from multiple places within your organization. It can function as a single source for LLMs and agents needing access to your proprietary information. It’s not real-time information but it can be very close to that and it, obviously, has troves of historical information.

The backup vault storage media is an important consideration. Tape is obviously a no-no due to lengthy data access times. Disk is better while an all-flash vault is probably the fastest. 

Cohesity, Commvault, Rubrik, and Veeam are well aware of this strength they possess, as a vast potential RAG data store, and building out features to capitalize on it. Cohesity has its Gaia initiative. Rubrik recently acquired Predibase for agentic AI development functionality.

An obvious caveat is that the backup vendors can only serve data to LLMs and agents that they have backed up. Anything else is invisible to them. This could encourage their customers to standardize on one backup supplier across their organization.

Data Managers

Data managers such as Arcitecta, Datadobi, Data Dynamics, Hammerspace, and Komprise almost never act as data sources. We say almost never because Hammerspace is edging towards that functionality with the Open Flash Platform initiative. The data managers manage data stored on multiple storage arrays and in public cloud storage instances. This partially reflects a hierarchical life cycle management background common to several of them.

The data managers don’t store data, but they do index and catalog it; they know what data their customers have, and they often have data moving capability, coming from a migration background in some cases. That means they can construct interface processes to AI data pipelines and feed data up them. Komprise is highly active here. Datadobi with its StorageMap, policy-driven workflows and data moving engine has a great base on which to build AI model-specific functionality.

Arcitecta’s Mediaflux product can be used to curate high-quality datasets from vast volumes of unstructured data, making it easier to feed data into large language models (LLMs) and other AI systems. We expect more functionality specific to AI to emerge here.

To expand their reach and give them more relevance here the data managers could usefully partner with the database/warehouse, lake and lakehouse suppliers. Hammerspace and Snowflake is an example of this.

Other data silos

SaaS apps such as Salesforce are another data source for RAG. They cannot see outside their own environment and, absent explicit connectors, other data sources, with one exception, cannot see into theirs. Their own environments – Microsoft’s, for example – can of course be large.

The exception is backup. If a SaaS app has its customer data backed up, by a Cohesity, Commvault, Druva, HYCI, etc., then that data is inside the backup vendor’s purview.

Summing up

There is no silver bullet here, no instant fix, no one all-embracing Gen AI RAG data source. You can only search out the best available Gen AI data source or sources for your organization considering your on-prem and public cloud IT data estate. All the suppliers mentioned are developing features, building partnerships and even acquiring companies that own relevant technology. They all have their merits and will have more next month, next year, and so on.

Wherever you are starting from you can perhaps place candidate suppliers into one of the four categories: storage, database/warehouse/lake/lakehouse, backup store and data management. That will give you a quick overall view of their scope. You can then, hopefully, decide which route or routes you might take to arrive at the most efficient way to give the LLMs and agents you use access to the proprietary data they need

Competitive Corner breaks down key shifts in 2025 Gartner backup MQ

The Competitive Corner provides the best and most detailed analysis of vendor ratings in the 2025 Gartner Magic Quadrant (MQ) for Backup and Data Protection Platforms that we have encountered.

We covered this MQ in a storage ticker on June 30, noting Druva got promoted from Visionary to Leader, Huawei entered as a Challenger, the separate Cohesity and Veritas entries have been combined, and Microsoft exits as a Niche Player. Otherwise it was pretty stable with small vendor position changes. Competitive Corner analyst Matt Tyrer applied his microscope to the MQ, plotting the changes in detail, looking at why they took place, and comparing the positioned vendors to a depth we have not seen before. 

Matt Tyrer

Tyrer is an ex-head of competitive intelligence at Druva and Commvault.

He notes: “We did see some significant changes in the Leaders quadrant. First off, Druva makes their debut as Leader this year – making them the only 100 percent SaaS platform in the Leaders quadrant and first new entrant to that space since 2020. Veritas is no longer separately listed now that their acquisition by Cohesity is complete, and Dell Technologies continues their inexorable slide out of the Leaders space – something I’m expecting to see in the 2026 MQ.”

Tyrer comments: “Technically, Veeam barely holds onto their position as ‘Highest in Ability to Execute’ (by a sliver) and Rubrik stays just ahead of Cohesity to be a repeat ‘Furthest in Vision.’ We’ll explore why in the next section when we delve into each Leader in more detail.”

“Outside of the Leaders there were a few things of note. As previously mentioned, Microsoft dropped out of the Magic Quadrant this year after a brief appearance in last year’s report. Huawei is the only new vendor this year, serving as the only representative in the Challenger’s space. HYCU lost some ground in terms of Vision, while Arcserve gained position on that axis. Lastly, IBM slightly moved closer to that Leader space.”

Tyrer looks at the leader vendors in some detail and issues verdicts:

  • Cohesity: With the addition of NetBackup, Cohesity’s workload support across the board is unmatched. Only Commvault can compare to the sheer diversity of workloads and data sources that Cohesity can protect. While there is indeed some expected chaos associated with the Cohesity + Veritas post-merger initiatives, those won’t be a distraction for long and Cohesity has the very real potential to be THE platform to beat in the enterprise with their combined technology stack – if executed properly.
  • Commvault has expanded their portfolio faster than ever over the past 12 months, which has introduced the previously mentioned challenges around complexity and usability, but those are issues Commvault has successfully addressed in the past and can likely correct moving forward. Despite these concerns, they remain one of the most feature-rich technology stacks in the industry and can address many use cases and data sources others cannot.
  • For customers that are invested in the broader Dell technology portfolio, the Dell data protection suite (DPS) should be adequate for their needs. The solutions are proven, but do lag behind the market leaders in terms of innovation and features. Customers with mixed heterogeneous vendor environments, particularly those with investments in cloud, will need to properly evaluate which of the many Dell solutions they will need to deploy – Dell PowerProtect Backup Service (OEM’d from Druva) is by far the best option for those types of Dell customers needing on-prem, hybrid, cloud, and SaaS workload coverage.
  • Druva’s SaaS solution delivers an easy to deploy, simple to manage solution that truly hits the “secure by design” cyber security goals by not having any infrastructure for bad actors to directly attack – something customers would otherwise have to design/architect themselves if using competing solutions in the MQ. This is a great fit for security-minded customers and those investing into hybrid and cloud.
  • Rubrik’s portfolio continued to grow, and with a pivot back to more of a data protection focus they appear to be finding balance between expanding their core backup capabilities and enhancing their cyber security products. Rubrik doesn’t yet have the broad workload coverage available from Commvault or Cohesity, but it is quickly catching up. Rubrik remains simple to deploy and operate, but that simplicity also translates into Rubrik’s platform being more limited in terms of more advanced features and functionality many enterprise customers need such as in depth reporting and robust cross-platform disaster recovery.
  • Veeam’s well established market presence and customer base are quick to adopt new Veeam features and products, but Veeam very often shares their footprint with other backup solutions – seeing multiple backup products on the customer floor to address different needs. The Veeam roadmap is ambitious, striving to deliver on 34 new products and enhancements, including significant expansion of their VDC offering and a new release of their core platform. So, look for Veeam to continue to impact the data protection market with a host of new capabilities over the next year.

He also has a look at Arcserve, Huawei, HYCU, and IBM, saying:

  • Arcserve’s simple, turnkey UDP appliance offering is well proven in the market and a great fit for midmarket customers or enterprise customers with multiple sites they need to easily manage for protection. Their pricing is flexible and with these latest investments into their product, the Arcserve solution supported by these new resilience features is one to watch.
  • Huawei is an emerging player in the enterprise market, and while their support for multi-cloud and availability is mainly limited to their own cloud they are slowly expanding. For now, it’s likely that customers that are invested in the Huawei cloud ecosystem and those with many sites in APAC will be better suited for their solution, but as workloads and geographic availability/support grow so too will their suitability for a broader market.
  • HYCU’s focus on Nutanix and SaaS applications gives them a solid advantage for customers heavily invested in either. There is no other vendor close to providing the SaaS coverage available from HYCU R-Cloud. Customers with larger multi-cloud and on-prem environments may need to look at other solutions outside of HYCU to complete their backup coverage, but the differentiated approach HYCU is taking to the backup market continues to see some interesting innovations. If they continue to expand their cyber resilience features and workload support they will be moving in the right direction again on the MQ.
  • IBM’s technology is proven and steadily catching up with today’s business needs for data protection. Their innovations in cyber security integrations and broad channel support make them a reliable partner for enterprises. With some cloud expansion and continued development of their first party IBM Storage Defender solution, we could see them eventually punch into the Leaders quadrant.

Tyrer’s analysis is available as a blog at his Competitive Corner website, with a 14-minute read time. We’d recommend it as the single best Gartner enterprise backup MQ analysis we have ever seen.

ADP plots next move beyond Rubrik and Nutanix

Storage array customers, like those using Rubrik and Nutanix, increasingly require as-a-service backup and disaster recovery. ADP is looking to expand into this area, and is also exploring how it might use AI with the customer data it collects.

Assured Data Protection (ADP) is a global managed services provider, offering Backup-as-a-Service and DR-as-a-Service. It is Rubrik’s largest MSP and, as of this year, also supports Nutanix. It has main offices in both the UK and US, and operates 24/7 in more than 40 countries where it has deployments, with datacenter infrastructure in six worldwide locations. The UK operation is helmed by co-founder and CEO Simon Chappell and his US counterpart is co-founder Stacy Hayes in Washington, DC. 

Simon Chappell, ADP
Simon Chappell

We met Simon Chappell in London and discussed how ADP is doing and where it might be going in the future.

ADP launched operations in the Middle East in October last year via a strategic partnership with local value-added distributor Mindware. Mindware will establish local datacenters to help clients manage data sovereignty issues and minimize latency in data transfer. In February, ADP partnered with Wavenet, the UK’s largest independent IT MSP, to provide enterprise-grade backup and disaster recovery solutions to UK customers.

ADP has set up an Innovation Team aimed at expanding the company’s DR, backup, and cyber resiliency services with the addition of new technologies that complement current data protection services.

Rubrik recently acquired Predibase to help with AI agent adoption. The agents would be helped to generate responses by accessing Rubrik-stored data, a Rubrik data lake. You can use AI to clean the data but you can also use AI to look into it and analyze it.

What would Chappell think about providing AI support services for Rubrik customers who are using Rubrik as a data source for their AI? Chappell said: “There you go. That’s the $64 million question. The thing we are really good at is disaster recovery and cyber resilience. If we start to think, oh we’ve got all this data, are we as Assured going to do some service analytics around that? That is a big leap because we’ve gone from being pre-eminent at cyber resilience and disaster recovery to something else.”

Our understanding is that ADP is the custodian of the customer data it has backed up using Rubrik, and it could provide the data to large language models. In fact, it could go so far as to build its own LLM. Such an LLM could be used to clean recovered data in a clean room environment. We’re familiar with Opswat and Predatar malware presence scanning capabilities and wonder if ADP might be thinking along these lines.

This is all speculation on our part, based on the conversation with ADP and how it could potentially evolve. We think that there would need to be some kind of metadata filtering as the dataset size could be enormous, in the 500 TB+ area, leading to prolonged full scan time. You would use metadata to reduce the scan set size.

Another potential direction for ADP’s evolution could be to look for adjacent services possibilities, such as a strategic storage array partner. It could provide BaaS and DRaaS capabilities using the Rubrik and Nutanix-focused infrastructure it has built up in North and South America, the Middle East and Europe. Our understanding is that such a vendor would not be a legacy incumbent but rather a storage array supplier similar in its field to Rubrik and Nutanix in theirs. That would mean a relatively new-ish supplier with a strong and growing upper-mid-market-to-lower-enterprise customer base and a long runway ahead of it; an incumbent-to-be so to speak.

We would not be surprised to find out that, in 6 to 12 months time, ADP has a third strategic partner alongside Rubrik and Nutanix, and that it would be a storage supplier. Similarly, we wouldn’t be surprised to discover that ADP is offering malware cleansing services for recovered data.

SNIA tech review maps road to commercial DNA storage

The SNIA’s DNA Storage Alliance has published a 52-page technology review looking at the data encode/decode tech, commercial readiness metrics, and the challenges ahead.

DNA data storage relies on the encoding of digital information using sequences of the four nucleotides in strands of DNA. The four nucleotides are adenine (A), guanine (G), cytosine (C), and thymine (T), and they are found in the double helix formation of the DNA biopolymer molecule, located in the cells of all living organisms. Synthetic DNA can store data in a form orders of magnitude smaller than other storage media and can endure for centuries. It relies on chemical reactions at the molecular level and these are slower than electrical operations in semiconductors. Therein lies its core challenge.

DNA storage diagram

DNA data storage has been demonstrated in research projects, but the data write/read speeds, as well as equipment size, complexity, and cost, are all far from any successful commercial product. This technical document reviews the field, looking at the DNA codec, synthesis, storage and retrieval, sequencing, and commercialization challenges.

DNA storage diagram
DNA Storage Alliance tech review paper diagram showing an example of a DNA Channel: In step 1, the source bitstream is randomly scrambled, mitigating problematic sequences, packetized into large blocks which are then coded with ECC (outer code), and encoded from bits-to-bases (inner code), which divides the large blocks into small DNA sequences that are compatible with the properties of the Physical Layer chemistry. Also, object tags (primers) may be added that can be used to retrieve all DNA segments in a pool associated with a particular digital object. Next, the now fully “line coded” DNA sequences are passed to the DNA Physical Layer for writing (synthesis), storing, retrieval, and reading (sequencing). Lastly, the recovered DNA sequences are passed back to the codec where they are converted back to bits and decoded, reversing all the transformations, error correction, packetization etc. done on the encoding side. 

A look at the challenges reveals five: data throughput, total ownership costs, media endurance and data retention metrics, bio-security and data security, and standardization. On the throughput topic, the paper reveals: ”The requirements for moving data into and out of traditional storage far exceed the current capabilities of writing and reading DNA in biotechnology use cases.” 

DNA storage latency

It says: “The most fundamental challenge for DDS (DNA data storage) systems is to increase the throughput of DNA write and read operations. The underlying write and read operations for DNA are relatively slow chemical reactions (high latency), so the emphasis for increasing throughput involves enabling parallelism.”

It recommends: “DDS technology providers must increase the throughput of the underlying write and read operations, as well as reducing the time required to move molecules between operations, all while maintaining a competitive TCO for the use case at hand.” 

The paper ends on an optimistic note: “While DNA data storage is still quite nascent and there remain significant challenges to commercialization, the foundations of writing, storing, retrieving, and reading data using DNA have been shown to work on scalable technology platforms. Moreover, the ongoing investment in DNA technology, driven by biological and scientific applications, will continue to drive innovations that enhance DNA data storage capabilities.”

DNA data storage will augment, not replace, existing archival storage technologies, “resolving the ‘save/discard’ dilemma with a viable TCO for zettabyte scale and data preservation.”

Use cases, it says, will emerge over the next three to five years for DNA archival data storage.

The paper has lots of terminology that will be unfamiliar to people working with electricity-based digital storage, such as homopolymer, oligonucleotide, ligation, and polymerases, but that’s because it’s molecular organic chemistry. The document is freely downloadable and an excellent introduction to DNA data storage.

Bootnote

The DNA Storage Alliance is an SNIA community, with around 35 members and a six-member board:

  • Esther Singer, Director, DNA Data Storage. Twist Bioscience
  • Stephane Lemaire, Co-founder and Senior Innovation Officer, Biomemory
  • David Landsman, Director Industry Standards, Western Digital
  • David Turek, CTO, Catalog
  • Marthe Volette, Director of Technology, Imagene (an AI biotech company in Israel)
  • Julien Muzar, Technologist, Life Science, Entegris (8,000 employee supplier of advanced materials and process solutions for the semiconductor and other high-tech industries)

Twist Bioscience is a member and has board representation and recently changed its stance towards DNA data storage. It spun off its DNA business as Atlas Data Storage, a commercializing startup led by Varun Mehta, co-founder and CEO of HPE-acquired Nimble Storage. Twist retains an ownership stake and Atlas raised a $155 million seed funding round in May. We expect Atlas will take over the Bioscience membership and possibly its board position.

DNA Storage alliance

Esther Singer is still a Twist employee, being Director of Product and Market Development. In our opinion, the three most important DNA storage technology companies are Biomemory, Catalog and Atlas Data Storage.

DDN touts Infinia storage as key to faster, cheaper AI inference

DDN has released performance benchmarks showing it can can speed up AI processing time by 27x because of the way it handles intermediate KV caching.

An AI LLM or agent, when being trained on GPUs or doing inference work on GPUs and possibly CPUs, stores existing and freshly computed vectors as key-value items in a memory cache, the KV cache. This can have two memory tiers in a GPU server; the GPUs’ HBM and the CPUs’ DRAM. If more data enters the KVCache, existing data is evicted. If needed later, it has to be recomputed or, if moved out to external storage, such as locally attached SSDs or network-attached storage, retrieved, which can be faster than recomputing the vector. Avoiding KV cache eviction and recomputation of vectors is becoming table stakes for AI training storage vendors, with DDN, Hammerspace, VAST, and WEKA as examples.

Sven Oehme, DDN
Sven Oehme

Sven Oehme, CTO at DDN, states: “Every time your AI system recomputes context instead of caching it, you’re paying a GPU tax – wasting cycles that could be accelerating outcomes or serving more users. With DDN Infinia, we’re turning that cost center into a performance advantage.”

Infinia is DDN’s multi-year, ground-up redesigned object storage. It provides sub-millisecond latency, supports more than 100,000 AI calls per second, and is purpose-built for Nvidia’s H100s, GB200s, and Bluefield DPUs. DDN reminds us that Nvidia has said that agentic AI workloads require 100x more compute than traditional models. As context windows expand from 128,000 tokens to over 1 million, the burden on GPU infrastructure skyrockets – unless KV cache strategies are deployed effectively.

The company says that the traditional recompute approach with a 112,000-token task takes 57 seconds of processing time. Tokens are vector precursors, and their counts indicate the scope of an AI processing job. When the same job was run with DDN’s Infinia storage, the processing time dropped to 2.1 seconds, a 27-fold speedup. It says Infinia can cut “input token costs by up to 75 percent. For enterprises running 1,000 concurrent AI inference pipelines, this translates to as much as $80,000 in daily GPU savings – a staggering amount when multiplied across thousands of interactions and 24/7 operations.”

Alex Bouzari, CEO and co-founder of DDN, says: “In AI, speed isn’t just about performance – it’s about economics. DDN enables organizations to operate faster, smarter, and more cost-effectively at every step of the AI pipeline.”

It is unclear how DDN’s implementation compares to those from Hammerspace, VAST Data, and WEKA, as comparative benchmarks have not been made public. We would suppose that, as KV caching is becoming table stakes, suppliers such as Cloudian, Dell, IBM, HPE, Hitachi Vantara, NetApp, PEAK:AIO, and Pure Storage will add KV cache support using Nvidia’s Dynamo offload engine.

Bootnote

The open source LMCache software also provides KV cache functionality, as does the Infinigen framework.

Storage news ticker – July 18

AWS said customers can boot Amazon Elastic Compute Cloud (Amazon EC2) instances on AWS Outposts using boot volumes backed by NetApp on-premises enterprise storage arrays and Pure Storage FlashArray, including authenticated and encrypted volumes. This enhancement supports both iSCSI SAN boot and LocalBoot options, with LocalBoot supporting both iSCSI and NVMe-over-TCP protocols. Complementing fully managed Amazon EBS and Local Instance Store volumes, this capability extends existing support for external data volumes to now include boot volumes from third-party storage arrays, providing customers with greater flexibility in how they leverage their storage investments with Outposts.

… 

Rai Way, part of the RAI group (Italy’s state TV broadcaster), and a specialist in digital infrastructure and media, has signed a Business Alliance Partnership with decentralized storage and compute provider Cubbit, adopting its DS3 Composer technology to integrate cloud storage services into its portfolio. Rai Way now offers a fully Italian edge-to-cloud service for storing customer data, enabled by Cubbit. The joint system ensures maximum resilience, data sovereignty, cost-effectiveness, and performance – even in low-connectivity areas – thanks to Rai Way’s distributed datacenter network, proprietary fibre infrastructure, and Cubbit’s geo-distributed technology. The initial capacity is 5 petabytes, with data hosted across Rai Way’s first five newly built edge datacenters in Italy.

DapuStor announced the launch of its ultra-high capacity J5060 QLC SSD, delivering 22.88 TB of capacity. Delivering up to 7,300 MBps read speeds at just 13 W, and maxing out under 25 W, it offers a best-in-class data-per-watt ratio. The J5060 is engineered for large, sequential data workloads. It uses coarse-grained (large-granularity) mapping and a dual-PCB hardware design to minimize DRAM usage and overcome capacity limitations.

– Sequential Read up to 7.3 GBps
– Sequential Write up to 2.8 GBps
– 4K Random Read 1.5M IOPS
– 32 KB Random Write 15K IOPS
– 4K Random Read Latency as low as 105 μs 

Research house DCIG is researching the unstructured data management space. It says that, in the last several years, DCIG has published more than ten TOP 5 reports covering various aspects of unstructured data management. However, DCIG discovered a significant gap in the marketplace. “There’s really no good mental model for people to understand the overall unstructured data management space,” Ken Clipperton explains. After meeting with nearly 20 solution providers, he found widespread confusion about what constitutes unstructured data management. DCIG has devised a “7 Pillars of Unstructured Data Management” framework which provides a cohesive, vendor-neutral, actionable model for understanding the data management challenges faced by organizations today. It is creating a technology report on the seven pillars framework. This report will be available for licensing by solution providers in the second half of 2025.

Fast file and object storage array supplier DDN announced a strategic partnership with Polarise to deliver high-performance, resource-efficient AI infrastructure designed for the next generation of sovereign European AI workloads – at scale, with speed, and with sustainability in mind. Polarise specializes in building turnkey AI factories – end-to-end, AI-centric datacenters based on the Nvidia reference architecture. With locations in Germany and Norway, Polarise offers customers a sovereign alternative for consuming AI computing power through colocation, dedicated/private cloud, or direct API access via its own cloud platform. The companies will initially focus on joint deployments across Germany and Norway, with further European expansion planned in 2025 and beyond.

Deduping and fast restore backup target supplier ExaGrid announced its Tiered Backup Storage appliances can now be used as a target for Rubrik backup software. With advanced data deduplication, it will lower the cost of storing backups using the Rubrik Archive Tier or Rubrik Archive Tier with Instant Archive enabled, as compared to storing the data in the cloud or to an on-premises traditional storage. ExaGrid can achieve an additional reduction of 3:1 to 10:1 in addition to Rubrik’s compression and encryption, further reducing storage by as much as 90 percent. The combined deduplication is between 6:1 and 20:1 depending on length of retention and data types.

Streaming log data company Hydrolix announced support for AWS Elemental MediaLive, MediaPackage, and MediaTailor, as well as client-side analytics from Datazoom. The new integrations provide media and entertainment companies with real-time and historical insights into video streaming performance and advertising delivery, helping optimize viewer experience and ad revenue while significantly reducing the cost and complexity of data storage and analysis. The AWS Elemental and Datazoom integrations complement existing integrations with AWS CloudFront and AWS WAF, as well as other data sources.

HighPoint Technologies’ Rocket 1628A and 1528D NVMe Switch Adapters empower enterprise IT and solution providers to deploy up to 32 NVMe drives and 8 PCIe devices in a single x16 slot. Read more here.

Lexar NM990

Chinese memory supplier Longsys now owns the Lexar brand. Lexar has launched the NM990 PCIe 5.0 SSD with up to 14,000 MBps read and up to 11,000 MBps write (4 TB model), engineered for high-end gamers, professional creators and AI developers. It features:

  • 1-4 TB capacity range
  • Thermal Defender Technology: Ensures 23 percent more efficient power consumption and a smoother experience
  • HMB and SLC Dynamic Cache: Delivers random read/write speeds up to 2000K/1500K IOPS for faster load times and reduced latency
  • 3,000 TBW (4 TB model): Built for long-term durability and heavy workloads
  • Microsoft DirectStorage Compatible: Optimized for next-gen gaming performance

F5 and MinIO are partnering. The use cases addressed by F5’s Application Delivery and Security Platform (ADSP) combined with MinIO AIStor are:

  • Traffic Management for AI Data Ingestion (Batch and Real-Time): Enabling high-throughput, secure ingestion pipelines for data training and inference while eliminating pipeline bottlenecks and hotspots.
  • Data Repatriation from Public Clouds: Supporting organizations that are bringing data back to on-prem or hybrid environments to manage costs and comply with regulations. 
  • Data Replication, Secure Multicloud Backup, and Disaster Recovery: Ensuring data availability, reliability, and security across geographies.
  • AI Model Training and Retrieval-Augmented Generation (RAG): Accelerating model development through real-time access to distributed data.

Cloud file services supplier Nasuni says it achieved top honors in the NorthFace ScoreBoard awards for the fifth year running. The NorthFace ScoreBoard Award is one of the industry’s most respected benchmarks for customer satisfaction. Nasuni achieved a Net Promoter Score (NPS) of 87, a Customer Satisfaction (CSAT) score of 98 percent and an overall ScoreBoard Index rating of 4.8 out of 5 – all above the industry average. Nasuni has also been awarded seven badges in G2’s Summer 2025 Reports, which are awarded based on real-user reviews and market presence. 

Memory and storage semiconductor developer Netlist has expanded recently filed actions against Samsung and Micron in the U.S. District Court for the Eastern District of Texas. The two actions were originally filed on May 20, 2025, to assert Netlist’s new U.S. Patent No. 12,308,087 (“the ‘087 patent”) against Samsung’s and Micron’s High-Bandwidth Memory (HBM) products. The amended complaints add another Netlist patent, U.S. Patent No. 10,025,731 (“the ‘731 Patent”), against the two defendants’ DDR5 DIMM products, as well as their distributor Avnet, Inc. to these actions.

Netlist’s ‘087 patent entitled “Memory Package Having Stacked Array Dies and Reduced Driver Load” covers HBM for current and future AI applications. Analysts currently project U.S. HBM revenues to exceed $25 billion annually for each Samsung and Micron by 2031. The ‘087 patent will expire in November 2031. Netlist’s ‘731 patent reads on DDR5 DIMM with DFE and ODT/RTT circuits. Analysts currently project U.S. DDR5 DIMM revenue to exceed $65 billion annually in 2029. The ‘731 patent will expire in July 2029.

During the past two years, Netlist has obtained jury verdicts awarding combined total damages of $866 million for the willful infringement of its patents by Samsung and Micron.

Patriot Memory announced the launch of its latest flagship PCIe Gen5 SSD – the PV593 PCIe Gen 5 x4 M.2 2280 SSD, delivering industry-leading speeds of up to 14,000 MBps read and 13,000 MBps write. Powered by the advanced SMI SM2508 controller and built on TSMC’s 6nm process, the PV593 redefines next-gen storage for demanding workloads such as AI training, 4K/8K video editing, AAA gaming, and high-throughput multitasking. It features DRAM cache and dynamic SLC caching technology to reduce latency, accelerate system startup, and improve application launch times. With 4K random read/write speeds reaching up to 2000K/1650K IOPS, users can experience lightning-fast responsiveness across all tasks.

Vector search company Qdrant announced the launch of Qdrant Cloud Inference. This fully managed service allows developers to generate text and image embeddings using integrated models directly within its managed vector search engine offering Qdrant Cloud. Users can generate, store and index embeddings in a single API call, turning unstructured text and images into search-ready vectors in a single environment. Directly integrating model inference into Qdrant Cloud removes the need for separate inference infrastructure, manual pipelines, and redundant data transfers. This simplifies workflows, accelerates development cycles, and eliminates unnecessary network hops for developers. 

Qdrant Cloud Inference is the only managed vector database offering multimodal inference (using separate image and text embedding models), natively integrated in its cloud. Supported models include MiniLM, SPLADE, BM25, Mixedbread Embed-Large, and CLIP for both text and image. The new offering includes up to 5 million free tokens per model each month, with unlimited tokens for BM25. This enables teams to build and iterate on real AI features from day one.

We’re told the CEO of Redis halted all product development for a week, just to ensure the entire team knew how to code with AI. Rowan Trollope, a developer-turned-CEO, was initially sceptical of vibe coding – a new phenomenon where developers use AI to help code. Now he’s pushing it across Redis after seeing first-hand how much vibe coding can accelerate delivery, improve code reviews, and reshape workflows. He vibe codes in his spare time, mostly for side projects, and while he knows AI won’t give him perfect code, he sees it as an accelerant. He’s one of the few CEOs with first-hand experience of vibe coding who has both the technical chops and the authority to enforce AI adoption from the top down.

Cyber-resilience supplier Rubrik unveiled upcoming support for Amazon DynamoDB, AWS’s flagship serverless, distributed NoSQL database service, and launched a proprietary cyber resilience offering for relational databases, beginning with Amazon RDS for PostgreSQL. Rubrik enables storage-efficient, incremental-forever backups and provides the flexibility to choose from a full range of Amazon Simple Storage Service (Amazon S3) storage classes, including Amazon S3 Standard, S3 Standard-Infrequent Access, S3 One Zone-Infrequent Access, S3 Glacier Instant Retrieval, S3 Glacier Flexible Retrieval, and S3 Glacier Deep Archive. Read a blog to find out more.

AI Data Cloud company Snowflake announced the appointment of Chris Niederman as Senior Vice President of Alliances & Channels. He joins Snowflake with more than 30 years of technology experience and previously spent 11 years at Amazon Web Services (AWS), most recently serving as Managing Director of the AWS Industries and Solutions team, leading the organization responsible for AWS’s worldwide partner strategy and industry transformation initiatives. Niederman will be responsible for leading the Snowflake global channel and partner ecosystem, and driving growth and collaboration through the Snowflake AI Data Cloud to empower every enterprise to achieve its full potential through data and AI. 

DDN subsidiary Tintri announced “a game-changing and award-winning refresh to its global Partner Programme – introducing powerful new tools, exclusive benefits, and a reinvigorated partner experience designed to ignite growth, accelerate sales, and fuel innovation across the channel.” It includes technical certifications, pre-sales experts, marketing development funds, rewarding incentives and no barriers to entry to join the partner program. Click here to find out more.

Glenn Lockwood

VAST Data has hired Glenn Lockwood as Principal Technical Strategist. He joins from Microsoft, where he helped design and operate the Azure supercomputers used to train leading LLMs. Prior to that, he led the development of several large-scale storage systems at NERSC – including the world’s first 30+ PB all-NVMe Lustre file system for the Perlmutter supercomputer. Glenn holds a PhD in Materials Science. On AI inference, he says: “The hardest data challenges at scale went from ‘read a big file really fast’ to ‘search a vector index, rank and filter documents, load cached KV activations, cache new KV activations, and repeat.’ Not only are access sizes and patterns different across memory and storage, but application developers are expecting data access modalities that are much richer than the simplistic bit streams offered by files.” Read more in a VAST blog.

Veeam ran a survey into six months after the EU’s Digital Operational Resilience Act (DORA) came into effect. It found organizations are lagging: 96 percent of EMEA FS organizations believe they need to improve their resilience to meet DORA requirements. 

  • DORA is a priority: 94 percent of organizations surveyed now rank DORA higher in their organizational priorities than they did the month before the deadline (Dec 2024), with 40 percent calling it a current “top digital resilience priority.”
  • The mental load of mounting regulations: 41 percent report increased stress and pressure on IT and security teams. Furthermore, 22 percent believe the volume of digital regulation is becoming a barrier to innovation or competition.
  • Barriers to compliance: Third-party risk oversight is cited by 34 percent as the hardest requirement to implement. Meanwhile, 37 percent are dealing with higher costs passed on by ICT vendors, and 20 percent of respondents have yet to secure the necessary budget to meet requirements.

Wedbush analysis about the AI phenomenon: “Our bullish view is that investors are still not fully appreciating the tidal wave of growth on the horizon from the $2 trillion of spending over the next 3 years coming from enterprise and government spending around AI technology and use cases. We have barely scratched the surface of this 4th Industrial Revolution now playing out around the world led by the Big Tech stalwarts such as Nvidia, Microsoft, Palantir, Meta, Alphabet, and Amazon.”

“Now the time has come for the broader software space to get in on the AI Revolution as we believe the use cases are exploding, enterprise consumption phase is ahead of us for 2H25, launch of LLM models across the board, and the true adoption of generative AI will be a major catalyst for the software sector and key players to benefit from this once in a generation 4th Industrial Revolution set to benefit the tech space. 2025 so far has been an inflection year within enterprise generative AI as true adoption has begun by going from idea to scale as more companies are looking to invest into AI to decrease costs/increase productivity. Looking forward its all about the use cases exploding which is driving this tech transformation being led by software and chips into the rest of 2025 and beyond and thus speaks to our tech bull and AI Revolution thesis further playing out over the next 12 to 18 months.”

It has “identified the 30 tech companies in our ‘IVES AI 30’ … that define the future of the AI theme over the coming years as we believe these tech stocks are closely tied to the AI Revolution ranging from hyperscalers, cybersecurity, software, semis, internet, and autonomous/robotics.” 

WEKA co-founder and sometime COO Omri Palmon is departing for something, possibly a startup, that’s not public yet. A LinkedIn post and comments discuss this.