Home Blog Page 14

Cerabyte gets In-Q-Tel cash for data archiving

Ceramic nano-dot archiving startup Cerabyte has received an investment from In-Q-Tel as part of a strategic initiative to speed up its product development.

Cerabyte’s technology uses femto-second laser-created nano dots in a ceramic-coated glass storage tablet that is stored in a robotic library. The medium is immutable, can endure for 1,000 years or more and promises to be more performant and cost-effective than tape. It is based in Germany and has opened offices in Silicon Valley and Boulder, Colorado. In-Q-Tel (IQT) is effectively a VC for the US CIA and other intelligence agencies and was set up to bring new commercial information technology to these agencies. 

IQT Munich managing director Greg Shipley stated: “Cerabyte’s innovative technology can significantly enhance storage longevity and reliability while also reducing long-term costs and complexity. This strategic partnership aligns with our mission to deliver  advanced technologies that meet the needs of the national security community.”

Christian Pflaum, co-founder and CEO of Cerabyte, matched this with his statement: “The strategic partnership  with IQT validates our mission and fuels our ability to deliver accessible permanent data storage solutions.”

Cerabyte says that Commercial and government organizations anticipate managing data volumes comparable to those of major corporations such as Meta or Amazon. Most of this data will remain in cold storage for extended periods, aligning with typical declassification timelines of 25 to 50 years. The US National Academies of Sciences, Engineering, and Medicine (NASEM), conducted a Rapid Expert Consultation (REC) on technologies for archival data storage upon request of the Office of the Director of National Intelligence (ODNI); ODNI is the head of the US intelligence community.

We understand a a US National Academies REC is a report by subject matter experts looking at urgent issues where decision makers need expert guidance quickly.

Other recent In-Q-Tel investments include:

  • Participating in $12 million A-round for VulnCheck,  a CVE vulnerability and exploit intelligence monitoring startup.
  • Participating in Israeli defense tech startup Kela’a $11 million seed round.
  • Contributing to US Kubernetes and AI workload security startup Edera’s $15 million A-round.
  • Participating in cloud GPU server farm business Lambda’s $480 million D-round.

It has famously invested in big data analytics supplier Palantir and also in Databricks.

Cerabyte has previously received an investment from Pure Storage and been awarded a portion of a €411 million ($426 million) European Innovation Council (EIC) Accelerator grant.

It’s possible we may have Cerabyte product development news coming later this year.

MariaDB CEO on database company regaining its mojo

Interview: B&F caught up Rohit de Souza, CEO of MariaDB, following his visit to Finland where he met founder Michael “Monty” Widenius.

MariaDB was founded in 2009 by some of the original MySQL developers, spurred into action by Oracle buying MySQL owner Sun Microsystems. They developed their SkySQL open source databases as a MySQL fork. It was an alternative to proprietary RDBMSes and became one of the most popular database servers in the world. It went through some ten funding events, amassing around $230 million before going public through a SPAC transaction in 2023. In February 2022, it was valued at $672 million but started trading at the much lower $368 million. Michael Howard was CEO at the time, but Paul O’Brien took the post in March 2023. Tom Siegel was appointed CRO and Jonah Harris CTO.

Things did not go well. MariaDB posted poor quarterly results, received NYSE warnings about its low market capitalization, laid off 28 percent of its staff, closed down its Xpand and SkySQL offerings, and announced Azure Database for MariaDB would shut down in September 2025. A lender threatened to claw back cash from its bank account. Harris left in April 2024. O’Brien left in May that year.

MariaDB was acquired by private equity house K1 Investment Management in September 2024, at a suggested $37 million price, about a tenth of its initial public trading capitalization.

Rohit de Souza

At that time, MariaDB had 700 customers, including major clients such as Deutsche Bank, Nokia, RedHat, Samsung, ServiceNow, the US Department of Defense, and multiple US intelligence and federal civilian agencies. This gave K1 something to build on. It appointed Rohit de Souza as CEO, with previous executive experience from his roles at Actian and MicroFocus. O’Brien was transitioned out to advisor status. 

De Souza’s job is to pick up the pieces and rebuild the company after its near-disastrous public ownership experience. How has he set about doing that?

He claimed MariaDB had been “completely mismanaged,” but there is a strong community and it’s “still the number three open source database in the world.” It has had a billion downloads of its open source software and it’s used by up to 1.3 million enterprises. “And if that’s not a customer base who wants to continue using its product, and see it develop further,” then he doesn’t know what is, De Souza said.

Some executive musical chairs took place. Siegel left the CRO job in January this year, Widenius was rehired as the CTO, and Mike Mooney became MariaDB’s CRO.

There is a free community edition of MariaDB and a paid-for Enterprise edition that has support. De Souza said there is “a very small sliver of those who are paying for it, but that’s because we’ve not managed that whole process right.” MariaDB has been leaving money on the table.

He said: “For instance, take a big bank, a prominent bank in the UK, they’re using both the community version and the enterprise one. They’re paying us support on the enterprise piece, but there’s been no diligence with them around saying, ‘Hey, if you’re going to get support from us, you need to get support on the entire estate.'”

“People can pay for the support on the enterprise and then use the skills they’ve got to fix the community edition. They use the support on the community stuff.”

This will be one strategy De Souza will use to increase revenue from the customer base. He said “there’s scope for the smooth scheme increase in the discussions” with customers.

How have customers responded to that idea? “So far I’ve touched a lot of customers as we’ve gone forward. The reception has been fabulous. The reception has been, ‘Oh really? We didn’t know you were going down that path. We thought MariaDB was sort of on its way out’.”

He is also updating the product. “We talked to them about what we’re doing around vector, how it’s an obvious development. It’s out with the community edition already and the GA piece for the enterprise edition will be in there soon.”

MariaDB added vector support alongside its existing transactional, analytical, and semi-structured data support in January.

De Souza has been giving MariaDB employees and partners “the message that there’s more value we can provide to our customer base. There’s more revenue that we can get from our customer base for this.”

He told us he had visited Widenius in Finland, and met his daughter Maria, after whom MariaDB is named, and other family members.

He said of Monty: “This guy, this is his life’s work. This is part of why I joined, because I see so much value in this and I see so much innate capability. Until this point engineering at MariaDB was every engineer did what he thought was right. There was no hope. Monty is a fabulous technologist, but you can’t ask him to drive the roadmap. He doesn’t want to do that.”

So, De Souza said: “You need a chief product officer. I’ve got a good chief product officer. I’ve got a good chief technology officer. I brought in a top-notch engineering leader to make sure that the trains run on time, and make sure our releases are right, make sure our roadmaps are correct. I brought in the right AI talent to make sure that we’re working on AI to put next to SQL in the database to do checkpointing.”

“I want to make it really, really simple to use for all of the fellas trying to mine the gold out of the AI system.”

The situation is that a MariaDB customer is looking to use AI agents and have them access structured and unstructured information in the database. That information will need vectorizing and the vectors stored in the database. Is MariaDB developing its own AI agents?

“We’re using traditional LLMs, Langchain or whatever. Whatever you want to use. We’d like to be able to use the typical externally available agents.”

Think of the LLM as an arrow and MariaDB as a data target.”“Use whatever arrows you want and we’ll make our target big and bold for those arrows and we’ll make it usable. We’ll make whatever information you’ve got stored in MariaDB accessible to all your AI, to the rest of your AI estate.”

MariaDB may well add an AI pipeline capability, including data masking to screen out sensitive data, “filtered to remove sensitive information. Then maybe transformed a little and then fed upstream and add to that on the front end the right security models. This is thanks to K1 and their relationships in the security area.”

De Souza is also concerned that his customers don’t get information leakage through AI agent use. “So for example, if you did a query through your large language model of your financial database to, say, give me what’s the revenue of the company, and think, ‘Hey, I’ll email this to my friend downstairs.’ You won’t be able to. You won’t have access to that on the front end.”

He sees AI agents as “extraordinarily powerful microscopes,” and will be looking at partnerships and/or acquisitions to gain technology access.

He said the company is getting enthused and growing. “I’m adding 20 percent to the staff in the company and I’m having no trouble finding the right talent. Because you say to prospective talent, check out MariaDB’s customer base and exposure and heritage. Check out what we can do with it. Do you want to be part of this or would you prefer to go risk yourself elsewhere?”

De Souza is adding senior advisors to MariaDB, such as Linux Foundation exec director Jim Zemlin and ServiceNow president Amit Zavery. He said ServiceNow has been trying to develop its own database. “We’ve just renewed a relationship with them and I’m convinced that at some point [CEO Jim] McDermott is going to see the light” – and use MariaDB.

He said there is a substantial large enterprise appeal in moving off Oracle and saving lots of cash. “We now have lots of interest. We’re in the process of migrating a large Middle Eastern airline off Oracle. They’ve said, ‘Show me that you can move a critical system,’ which is a little unnerving for us, and I’ve had to have all hands on that. But if that’s successful, we’ll have proved – we’ve already proved it in financial services, in retail – we’ll be able to prove it in the airlines and the transportation business.”

Another direction for development is automating the migration off Oracle. De Souza said: “The Oracle mode will only be available for the enterprise, of course, because these are going to be customers who are already paying millions of dollars.”

Storage news ticker – March 24

Airbyte announced new capabilities for moving data at scale for AI and analytics workloads:

  • Support for the Iceberg open standard for moving data into lakehouses
  • Enterprise Connector Bundle that includes connectors for NetSuite, Oracle Database with Change Data Capture (CDC), SAP HANA, ServiceNow, and Workday
  • Support for AWS PrivateLink with secure, private cloud-to-cloud data transfers to ensure sensitive information is controlled

Lenovo announced new full-stack systems using Nvidia components at GTC 2025. The Lenovo Hybrid AI Advantage with Nvidia framework includes the Lenovo AI Library. A new AI Knowledge Assistant is a “digital human assistant” that engages in real-time conversation with attendees to transform their event experiences and navigation. It enables users to customize Lenovo AI Library use cases and operationalize new AI agents and assistants within weeks. Lenovo is helping businesses build, scale, and operate their own AI factories. Lenovo’s agentic AI and hybrid AI factory platforms include Nvidia Blackwell Ultra support.

Phison has updated its aiDAPTIV+ “affordable AI training and inferencing solution for on-premises environments.” Customers will be able to fine-tune large language models (LLMs) up to 8 billion parameters using their own data. The new aiDAPTIVLink 3.0 middleware provides faster time to first token (TTFT) recall and extends the token length for greater context, improving inference performance and accuracy. A Maingear AI laptop PC concept was demonstrated at GTC 2025 using aiDAPTIVCache SSDs, aiDAPTIVLink middleware, and a Pro Suite GUI-based all-in-one feature set for LLMOps, including ingest, fine-tuning, RAG, monitoring, validation, and inference.

It says aiDAPTIV+ is a budget-friendly, GPU memory extension capability allowing users to train an LLM with their data within an on-premises “closed-loop” secure network, while providing a simple user interface to interact with and ask questions about the data. AI Inference for IoT with aiDAPTIV+ now supports edge computing and robotics use cases with Phison’s validation of Nvidia Jetson-based devices.

Predatar, which scans backups for malware, is releasing third-gen CleanRoom tech that supports more storage environments than the previous version, is more cost effective, and significantly quicker and easier to deploy. It has removed the need for customers to purchase hypervisor licences and third-party EDR software to reduce cost and streamline procurement, and can run on commodity hardware. There is a new ISO-based deployment method that enables a CleanRoom to be created as a Virtual Machine, or on bare metal. CleanRoom 3 can be built in just a few hours without the need for highly specialized skills or extensive training.

CleanRoom 3 will be available from March 31 directly from Predatar, Predatar’s Apex Partners, or from any IBM Storage Reseller. 

Cyber-resilience and data protector Rubrik announced a strategic alliance with Big Four accounting firm Deloitte to deliver advanced data security and management to help organizations safeguard their data assets, business continuity, and cyber resilience. The two will:

  • Combine Rubrik’s Zero Trust Data Security software with Deloitte’s technical knowledge in cybersecurity, risk management, and digital transformation
  • Deliver integrated offerings that help clients streamline data management processes, providing efficient data backup and recovery
  • Use cloud technologies and analytics to assist organizations in improving their IT systems

Deloitte employs over 460,000 professionals worldwide and reported revenues of $67 billion in its fiscal year 2024. It will also serve as a sales and integration channel for Rubrik’s offerings.

SK hynix is speeding up high bandwidth memory (HBM) manufacturing following a customer order surge, including from Broadcom, says Korean media outlet TheElec, planning to bring in equipment into its new M15X fab two months earlier than originally planned. Wafer capacity at M15X is planned to increase from 32,000 wafers/month. Broadcom is expected to account for 30 percent of SK hynix’s total HBM capacity by the end of the year.

We understand Broadcom’s Strata DNX Jericho2 (BCM88690) Ethernet switch-router, launched in 2018, delivers 10 Tbps of bandwidth and integrates HBM for large-scale packet buffering. Jericho2c+ announced in 2020, offers 14.4 Tbps and uses HBM for enhanced buffering and power efficiency. The StrataDNX Qumran3D single-chip router provides 25.6 Tbps of routing capacity and uses HBM to support high-bandwidth, low-power, and secure networking for carrier and cloud operators. 

In December 2024, Broadcom announced availability of its 3.5D eXtreme Dimension System in Package (XDSiP) platform technology, enabling consumer AI customers to develop next-generation custom accelerators (XPUs). The 3.5D XDSiP integrates more than 6,000 mm2 of silicon and up to 12 HBM stacks in one packaged device to enable high-efficiency, low-power computing for AI at scale.

TechRadar reported in January that Broadcom CEO Hock Tan told investors the company has three hyperscale customers who are each planning to deploy one million XPU clusters by 2027, and that it has been approached by two additional hyperscalers who are also in advanced development for their own AI XPUs.

Synology is thought to be launching DS925+, DS425+, DS225+, DS1525+, DS1825+, DS725+, DS625slim, DS1825xs+ and RS2825RP+ NAS storage systems between now and summer 2025, according to the NAScompares website, citing a ChipHell source. The new systems have 2.5 Gbps Ethernet ports.

Content Distribution Network supplier Akamai and VAST Data are partnering to profit from edge AI inference by enabling distributed customers to have better local response time and localization for distributed inference, compared to the centralized inference processing offered by hyperscalers. Akamai has a large Edge presence, with 4,300 points of presence and servers in more than 700 cities; effectively a distributed cloud. The partnership suggests that VAST Data storage and its AI software stack will be deployed in Akamai’s server locations.

Akamai says latency, cost, and scalability are hurdles to deploying AI-powered applications at scale. Unlike training, which is resource-intensive but happens in controlled environments, inference must operate continuously, on-demand, and often with ultra-low latency. Adding VAST Data to the Akamai Cloud aims to lower costs and improve customer experience to help democratize edge AI computing.

Agentic AI developer H2O.ai is partnering with VAST Data to bring agentic GenAI and predictive AI to enterprises’ exabyte-scale datasets. The two will combine “VAST Data’s modern infrastructure for storing and managing all data types (video, audio, documents, sensor data) with H2O.ai’s enterprise h2oGPTe platform to deliver retrieval-augmented answers, powerful model fine-tuning options, and grounded insights for enterprise users across on-premise, airgapped, and private cloud deployments.” 

Enterprises can build agentic AI applications on top of the VAST Data Platform. Sri Ambati, Founder and CEO of H2O.ai, said: “h2oGPTe, the world’s #1 Agent AI, with VAST data platform makes the perfect platform for sovereign AI. Agentic AI is bringing incredible agility to businesses and modern automated workflows are transforming SaaS. With the accuracy of H2O.ai and instant access to any volume and variety of data from VAST, enterprises can deliver real-time decisions and become AI superpowers.”

H2O.ai partners include Dell, Deloitte, Ernst & Young (EY), PricewaterhouseCoopers (PwC), Nvidia, Snowflake, AWS, Google Cloud Platform (GCP), and Microsoft Azure. Founded in 2012, H2O.ai has raised $256 million from investors, including Commonwealth Bank, Nvidia, Goldman Sachs, Wells Fargo, Capital One, Nexus Ventures, and New York Life.

Datacenter virtualizer VergeIO, claiming it’s the leading alternative to VMware, has a strategic partnership with app virtualization and VDI supplier Inuvika to deliver a more cost-effective, high-performance virtual desktop and application delivery offering. The two aim to provide customers with “a seamless, scalable alternative to VMware Horizon and Citrix, addressing the increasing cost and complexity of legacy VDI solutions.”

A 2025 Emissions Blindspots Report commissioned by S3 cloud storage supplier Wasabi says “almost half of Europe’s business leaders (47 percent) are afraid to learn the full extent of their emissions data, even though 61 percent fear public backlash if their emissions are too high.” Wasabi surveyed 1,200 business decision makers across the UK, France and Germany and 51 percent are concerned they are going to lose customers if they fully disclose their emissions data. Only 66 percent of European businesses think that they can paint an accurate picture of their tech stack emissions, which includes Scope 3 emissions. You can read the 2025 Wasabi Emissions Blindspots Report in its entirety here.

In July last year, Wasabi and Zero Circle, a sustainable finance marketplace, teamed up to give businesses the data they need to “reduce their environmental impact.” Wasabi uses Zero Circle’s invoice-based carbon footprint calculator to provide customers with “transparency” and a real-time assessment of their emissions.

Storage news ticker – March 21

Data integration supplier Adeptia announced double-digit annual growth since 2021 by enabling enterprises to manage and optimize their first-mile (raw data created outside of an organization’s internal data operations) unstructured and third-party data with its AI-powered Adeptia Connect software. In the past year, it has hired Charles Nardi as CEO, Geoff Mueller as COO, and Ross Trampler as VP Sales. Deepak Singh, Adeptia founder and CTO, has moved into the new role of Chief Innovation Officer. Adeptia received a $65 million strategic growth investment round led by PSG in March 2023.

Alluxio, supplier of open source data store virtualization software using a virtual distributed file system with multi-tier caching, announced a strategic collaboration with the vLLM Production Stack, an open source implementation of a cluster-wide full-stack vLLM serving system developed by LMCache Lab at the University of Chicago. 

AI inference requires low-latency, high-throughput, and random access to handle large scale read and write workloads. This partnership aims to help large language model (LLM) inference by providing an integrated solution for KV Cache management, utilizing both DRAM and NVMe, enabling efficient KV Cache sharing across GPU, CPU, and a distributed storage layer. By optimizing data placement and access across different storage tiers, the solution delivers low-latency, greater scalability, and improved efficiency for large-scale AI inference workloads.

Italy-based Avaneidi launched “CyberStorage by Avaneidi, a scalable and reliable storage infrastructure that seamlessly combines tailored Secure Hardware Storage, proprietary Data Governance and advanced CyberShield into one unified platform.” Avaneidi has introduced a new proprietary SSD, fully designed and manufactured in Europe, which it claims sets new standards for performance, durability, and reliability. Avaneidi says its “Enterprise Solid State Drives (ESSDs) utilize tailor-made chips and advanced algorithms, providing a bespoke solution optimized for performance and cyber security applications” with “extended drive lifetime, improved security, and significant energy savings.”

Avaneidi Goldrake ESSD.

The ESSD, codenamed Goldrake, comes in an AIC HHHL format with an NVMe PCIe gen 3 x4 interface. Capacities are 2TB, 3.84TB and on up to 16TB. It delivers 550,000/310,000 random read/write IOPS and has 2.5GBps/1.5GBps sequential read/write bandwidth. The endurance is 2 DWPD read-intensive and 5 DWPD write-intensive with a 5-year warranty.

Cloud backup and storage supplier Backblaze is partnering with media workflow designer CHESA to help M&E professionals collaborate anywhere by breaking down geographical barriers and connecting creative teams worldwide, leveraging Backblaze’s cloud infrastructure and CHESA’s workflow design. The two say they are “partnering with leaders in the M&E space, like LucidLink, iconik, and many others to offer a unified ecosystem that caters to every stage of the creative process.” CHESA delivers design and implementation expertise, while Backblaze provides scalable, cost-effective cloud storage for media workflows.

Data protection and management provider Cohesity announced an expanded collaboration with Red Hat to provide enhanced support for Red Hat OpenShift Virtualization, now available on the Cohesity Data Cloud. Customers can secure and protect data from virtual machines and containers in their Red Hat OpenShift Virtualization environments with Cohesity DataProtect and NetBackup. Backup and recovery operations for Red Hat OpenShift Virtualization-based virtual machines, along with containers, can be managed using the same workflows as existing data sources on the Cohesity Data Cloud. Watch a video with Cohesity CEO and president Sanjay Poonen with Red Hat CEO Matt Hicks, discussing the news.

Cohesity announced a partnership with AI supplier Glean to become a fully integrated data source for the Glean platform. It involves extending Glean’s connector capabilities to cover data sources protected by Cohesity, such as files stored on virtual machines, physical hosts, and NAS devices. This will “enable mutual customers to leverage high-quality backup data from Cohesity-protected workloads within Glean’s platform, further expanding enterprise knowledge and AI-driven insights.” 

Users can search across historical data, going beyond the current state of data to access past versions of documents, including deleted files and other previously inaccessible data. Historical snapshots will offer a unique advantage, particularly in compliance, forensic investigations, and data recovery efforts. The integration is expected to be available for customers later this year.

Data streamer Confluent announced new capabilities in Confluent Cloud for Apache Flink to streamline and simplify development of real-time AI applications, including Flink Native Inference, running open source AI models in Confluent Cloud without added infrastructure management, Flink Search using just one interface to access data from multiple vector databases, and built-in ML functions to make data science skills accessible to more teams. 

Confluent also announced further advancements in Confluent Cloud, making it easier for teams to connect and access their real-time data, including Tableflow, Freight Clusters, Confluent for Visual Studio (VS) Code, and the Oracle XStream CDC Source Connector. Learn more about these new features here.

Cloud file services vendor CTERA announced the availability of CTERA Edge-to-Cloud File Services in the Microsoft Azure Marketplace. Powered by Azure, CTERA gives “organizations cloud-native file services that extend from edge to cloud, offering enhanced availability, cyber resiliency, and scalability, while significantly reducing costs” with continuous data protection to Azure Blob Storage and global file access. Learn more about CTERA Edge-to-Cloud File Services at its page on the Azure Marketplace.

Vector databases supplier DataStax announced Astra DB Hybrid Search, a product that will improve AI search relevance by 45 percent. Using Nvidia NeMo Text Retriever reranking microservices, part of Nvidia AI Enterprise, Astra DB Hybrid Search, hosted on Astra DB with GPUs, integrates vector search and lexical search to deliver highly accurate, AI-driven search and recommendation experiences. Hybrid search combines two retrieval methods – vector search (for semantic understanding and contextual relevance) and lexical search (for exact keyword matching) – to help ensure both contextual relevance and precise keyword matching.

Astra DB automatically and intelligently reorders search results using fine-tuned LLMs, providing state-of-the-art ranking for more relevant and meaningful responses. Developers can integrate this functionality using the Astra DB Python client and schemaless Data API. The new capability will be available in Langflow, the open source tool and community for low-code AI application development.

DDN, French AI Model developer Mistral AI, and AI Cloud platform supplier Fluidstack announced a strategic alliance designed to make AI investments more successful, more flexible, easier to deploy, and significantly more cost-effective. The three say they deliver the most powerful AI infrastructure for training and deploying massive LLMs for enterprises, governments, and cloud providers. Timothée Lacroix, co-founder and CTO at Mistral AI, said: “DDN’s AI-optimized platform gives us the power to scale massive AI models with superior efficiency, reliability, and speed. Combined with Fluidstack’s cloud flexibility, this partnership ensures that our AI innovations deliver real-world impact – faster and at scale.”

Global engineering consultant Worley is using the Dell AI Factory with Nvidia to develop secure AI systems for employees and customers. One of their first products will be a GenAI app to assist Worley customer procurement teams in technical evaluations of vendor bid documents, reducing the time spent on data extraction and analysis from weeks to hours. Another will help Worley engineers accelerate design time by using GenAI to leverage insights from past projects for faster and more productive project delivery.

Kit being used includes PowerEdge XE9680 servers with Nvidia H100 GPUs, PowerScale F710 storage, AI Optimized Networking, and Nvidia AI enterprise software.

According to a Dell’Oro Group report, Server and Storage Component revenues grew to a record $244 billion in 2024, fueled by strong AI demand, primarily from GPUs, custom accelerators, HBM, and storage systems. Additionally, memory and storage demand for the general-purpose server market saw a strong rebound, accompanied by improved pricing throughout the year. Additional highlights from the Q4 2024 Data Center IT Semiconductors and Components Quarterly Report:

  • The Server and Storage Systems Component market is forecast to moderate to a healthy growth of over 40 percent in 2025. 
  • Nvidia dominated Data Center Component market revenues, leading all vendors, with Samsung and SK Hynix trailing behind in 2024. Nvidia captured nearly half of the total reported revenues, while hyperscalers deploying custom solutions are rapidly gaining ground. 
  • Smart NIC and DPU revenues increased by 77 percent, driven by strong deployment of Ethernet adapters in AI clusters. 
  • The recovery in the CPU market is overdue, with revenues projected to increase by 45 percent in 2025.

The report is purchasable here.

Backup supplier Druva announced a new integration with Microsoft Sentinel for SecOps teams with a unified view of data insights across their company’s security and backup environments. This integration, “a key component of Druva’s ongoing collaboration with Microsoft,” combines backup telemetry, system behaviors, data anomalies, and threat detections into security operations, enabling centralized monitoring and real-time threat detection. SecOps teams get:

  • Bi-Directional Integration: seamless and real-time connection between the Druva Data Security Cloud and Microsoft Sentinel – SecOps teams can quarantine compromised snapshots directly within their Microsoft Sentinel console.
  • Faster Threat Detection & Recovery: With insight into backup telemetry, SecOps teams can detect cyber threats like ransomware or data corruption in real-time, significantly reducing recovery time.
  • Improved Productivity: Respond to incidents faster and more efficiently.
  • Enhanced Security Visibility & Compliance: Centralized management in Sentinel, now augmented by Druva’s data protection insights, providing deep visibility for threat hunting, compliance audits, and proactive threat mitigation.

The integration is available in the Azure Marketplace.

Cloud backup supplier Eon “launched the first cloud-native package to provide protection and recovery from ransomware attacks.” Engineered for immediate recovery, it restores clean data in minutes, as opposed to hours or even days. Eon’s ransomware package aligns with the NIST Framework. It “delivers industry-first ransomware detection across cloud resources, including managed databases and data warehouses, and offers logically air-gapped backup vaults with immutable snapshots.” 

Eon was founded in 2024 by Ofir Ehrlich and Gonen Stein of the CloudEndure founding team (acquired by Amazon), along with Ron Kimchi, former GM of AWS Migration and Disaster Recovery Services. Eon is backed by VCs including Sequoia, Lightspeed, Greenoaks, and BOND, as well as various industry execs. Find out more here.

HPE reported $7.9 billion revenues for its Q1, up 16 percent year-over-year, with a $627 million GAAP profit, compared to the year-ago $387 million. CEO Antonio Neri said: “HPE achieved our fourth consecutive quarter of year-over-year revenue growth, increasing revenue by double digits in Q1.” Server segment revenues were $4.3 billion, up 29 percent. Intelligent Edge earned $1.1 billion, down 5 percent. Hybrid Cloud reported $1.4 billion, up 10 percent driven by Alletra Storage MP and GreenLake Cloud. Financial Services were flat at $873 million. AI systems orders doubled quarter-over-quarter with backlog increasing 29 percent. Its outlook for Q2 is $7.4 billion ± $2 million, just 2.7 percent up at the midpoint. It is cutting costs, with 2,500 jobs going, as the results were not as good as HPE had hoped and the outlook weak.

MagStor announced “the world’s first Thunderbolt 5 LTO drive, marking a monumental step forward in data storage technology.” It “offers unprecedented speed, reliability, and compatibility for professional data backup and archival needs” and is expected to be available by the end of 2025. The product has pending Intel and Apple Thunderbolt 5 certification and is also patent-pending.

VP Product Tim Gerhard tells us: “Thunderbolt 5 series will be available in LTO-7 thru LTO-10 single or dual configurations. Read/Write speeds will be up to 400 MBps on full height drives. In all honesty, there are no performance gains on the LTO drive itself, but you’re getting the latest controller chipsets and the ability to daisy chain TB5/USB4 v2 devices up to 80 GBps bandwidth. Most Thunderbolt 3 chipsets are going end of life and aren’t being produced anymore.” 

Micron says it’s the world’s first and only memory company shipping both HBM3E and SOCAMM (small outline compression attached memory module) products for AI servers in the datacenter. The SOCAMM, a modular LPDDR5X memory, was developed in collaboration with Nvidia to support the GB300 Grace Blackwell Ultra Superchip. It is “the world’s fastest, smallest, lowest-power, and highest capacity modular memory.” Micron HBM3E (High Bandwidth Memory) 12H 36 GB is designed into the Nvidia HGX B300 NVL16 and GB300 NVL72 GPUs while the HBM3E 8H 24 GB is available for the HGX B200 and GB200 NVL72 GPUs.

MSP backup service and cybersecurity supplier N-able announced disappointing Q4 FY 2025 revenue growth up only 0.1 percent sequentially and 7.5 percent year-over-year to $116.4 million with a $3.3 million GAAP profit. It was accustomed to 10 to 12 percent year-over-year revenue growth. Full-year revenues were $466.1 million, up 10.5 percent year-over-year. CFO Tim O’Brien said: “Our product and go-to-market teams executed critical initiatives, the strategic acquisition of Adlumin expanded the aperture of our business, and we once again operated above the Rule of 40. We firmly believe we have the right pieces in place to win in our markets, and are investing to seize an expanding market opportunity and scale N-able to new heights.”

William Blair analyst Jason Ader commented: “At its 2025 investor day, N-able laid out its plan to reaccelerate growth and margins through sales of its Adlumin XDR/MDR solution (new business and cross-sell), upmarket expansion to new channel partners (VARs and distributors), and new pricing and packaging initiatives (to drive bundles and cross-sell).”

….

Nutanix says its Cloud Platform with Nutanix Unified Storage will integrate with the Nvidia AI Data Platform and enable inferencing and agentic workflows deployed across edge, datacenter, and public cloud. As the only HCI supplier openly supporting AI workflows, we wonder if Nutanix will support GPU servers and external parallel access file and object storage. We’re asking Nutanix about this.

PCI-SIG has announced that the PCI Express 7.0 specification, v0.9, the final draft, is now available for member review. The raw bit data rate doubles from PCIe 6’s 64 GT/s to 128 GT/s, and up to 512 GBps bi-directionally via x16 configuration. The PCIe 7.0 spec is designed to support data-intensive market segments such as hyperscale datacenters, high-performance computing, and military/aerospace, as well as emerging applications like AI/ML, 800G Ethernet, and cloud computing. PCI-SIG says it’s on track to publish the full specification later this year. PCI-SIG members can access the PCIe 7.0 specification in the Review Zone on its website.

Quantum announced Quantum GO Refresh for the DXi T-Series deduping appliances to provide flexible payment options and hardware refreshes. Quantum GO Refresh is a new offering within the Quantum GO portfolio offering a turnkey subscription model paid quarterly or annually. It includes the typical hardware, software, and support and a new dimension of ongoing hardware refreshes as the system ages. For more information, click here.

Quantum celebrated its channel partners’ achievements at the 2025 Quantum EMEA Elevate Partner Summit in Kraków, Poland, recognizing six companies for outstanding performance in 2024: ADT Group (Poland), Errevi System Srl (Italy), Fujitsu (Germany), Econocom (UK), SVA System Vertrieb Alexander GmbH (Germany), and Sithabile Technology Services (South Africa).

Parallel file system supplier Quobyte has announced support for Arm-based servers with Arm support from Rocky Linux 9 and Ubuntu 24.04. It will deliver the same high performance as Quobyte’s x86 offering as well as a complete set of features including RDMA support. The new Arm platforms can be used as they are or added to existing Quobyte x86-based clusters in a mixed environment. Supported processors include Nvidia Grace, Ampere CPUs, and the AWS Graviton.

Scale-out file system supplier Qumulo announced back-to-back quarters of record bookings and growth with profitable net operating income in the most recent quarter. President and CEO Douglas Gourlay said: “In our last two quarters Qumulo delivered record-breaking results measured by bookings, ACV, and gross margin generating over seven percent net operating income.” It has seen new product adoption and significant expansion of its cloud product portfolio and client base. Qumulo has more than 1,100 customers. Cloud Native Qumulo, launched in October 2024 on both AWS and Azure, with 165 customers, has experienced “remarkable customer adoption, delivering 485 percent year-over-year increase in cloud storage consumption.” EVP and COO Michelle Palleschi said: “Achieving both GAAP and non-GAAP net operating income [is] a rare feat for a growth-stage business.”

Cyber-resilience data protector Rubrik announced Q4 FY 2025 revenues of $258.1 million, up 47 percent year-over-year, with a GAAP loss of $114.9 million. Full-year revenues were $886.5 million, up 41 percent, with a loss of $1.15 billion. 

CEO and co-founder Bipul Sinha said: “Fiscal 2025 was a milestone year for Rubrik. Our strong growth at scale demonstrates that we’re winning the cyber resilience market. However, we are still very early in Rubrik’s journey to achieve the company’s full potential and I’m confident that what’s ahead of us is even more important and exciting.” Rubrik and Commvault are reporting similar revenue figures:

Silicon Motion announced sampling of its MonTitan SSD Reference Design Kit (RDK) that supports up to 128 TB with QLC NAND, incorporating Silicon Motion’s PCIe Dual Ported enterprise-grade SM8366 controller, which supports PCIe Gen 5 x4 NVMe 2.0 and OCP 2.5. The 128 TB SSD has QLC NAND built with the latest 2 Tb die and sequential read speeds of over 14 GBps and random read performance exceeding 3.3 million IOPS, “providing over 25 percent random read performance improvement to other Gen 5 high-capacity solutions.” It supports NVMe 2.0 FDP (Flexible Data Placement) that improves SSD write efficiency and endurance. The software also uses Silicon Motion’s proprietary PerformaShape technology, utilizing a multi-stage shaping algorithm to optimize SSD performance based on user-defined QoS sets. The new 128 TB SSD RDK is now available for sampling to select partners and customers. Click here for more info.

Silicon Motion MonTitan graphic

SK hynix is sampling 12-hi 36 GB HBM4 DRAM with mass-production slated for the second half of this year. It says the product “has implemented bandwidth capable of processing more than 2 TB of data per second for the first time” and more than 60 percent faster than HBM3E. Bandwidth is the total data capacity that one HBM package can process per second.

Partially filled table of HBM generations and characteristics

Wedbush analyst Matt Bryson says the production schedule “would seemingly put SK hynix in a position to comfortably support Nvidia’s shift to HBM4 with Rubin (expected to ship 2H 2026) and support SK hynix’s continued leadership in this market.”

StorPool Storage says it’s among the first block storage software platforms to support the Arm-based Ampere Altra processors available in HPE ProLiant RL300 servers, providing fast and reliable cloud solutions at a significant cost savings. This “brings predictable high performance, linear scaling, and extreme energy efficiency for cloud and AI workloads.” Cloud services provider CloudSigma will deliver the first commercial application of this technology – an IaaS offering leveraging StorPool software running on Arm-based HPE ProLiant servers.

Demos of this joint StorPool, Ampere, and CloudSigma system are taking place now through March 20 in the KVM Compute Zone at CloudFest 2025 at the Europa-Park resort in Rust, Germany.

DDN’s Tintri subsidiary with its workload-aware, AI-powered data management offering announced its latest Geek Out! Technology Demo Session “Container Boys: Hello Tintri” taking place on Wednesday, March 26 at 10am PT / 1pm ET. Hosted by Tintri CTO Brock Mowry, the session will explore how containerized workloads can streamline workflows and accelerate time-to-market. Register here.

Ahmed Shihab

Western Digital has hired Ahmed Shihab as chief product officer. He most recently served as Corporate VP of Azure Storage at Microsoft, leading the engineering and operations teams for object and block storage, driving the design and development of Microsoft’s cloud storage offerings. Before that he worked as VP Infrastructure Hardware at AWS and VP and Chief Architect (FSG) at NetApp, with an 8.5-year stint at disk storage supplier Xyratex before that. Irving Tan, CEO of Western Digital, said: “Given his tenure with storage and infrastructure design and development, Ahmed is the right leader to move Western Digital forward as we help our customers unleash the power and value of data. Hyperscalers, cloud service providers, and OEMs are looking to Western Digital to provide them with the most cutting-edge HDD technology to grow their businesses.”

Open source backup supplier Zmanda has introduced its Ask Zmanda AI support assistant, pre-loaded with enterprise backup expertise to help users get immediate answers when troubleshooting a backup issue or exploring advanced configurations. It has product-specific guidance for all Zmanda offerings, including Zmanda Pro-Instant answers to detailed technical questions. Try it out here.

Veeam Backup & Replication software has a security flaw that could lead to remote code execution, tracked as CVE-2025-23120, and carring a CVSS score of 9.9 out of 10.0. It affects 12.3.0.310 and all earlier version 12 builds and was reported by Piotr Bazydlo of watchTowr. It is fixed starting in the following build: Veeam Backup & Replication 12.3.1 (build 12.3.1.1139).

Confluent Tableflow unlocks real-time AI with Delta Lake and Iceberg

Data streamer Confluent has updated its Tableflow software to give AI and analytics software access to real-time operational data in data warehouses and lakes.

The Tableflow updates build on the Confluent-Databricks partnership announced in February to add bi-directional integration between Tableflow, Delta Lake, and the Databricks Unity Catalog.

Specifically, support for Apache Iceberg is now generally available and an Early Access Program for Delta Lake support is open. Production workload teams “can now instantly represent Apache Kafka topics as Iceberg tables to feed any data warehouse, data lake, or analytics engine for real-time or batch processing use cases.”

Tableflow also offers enhanced data storage flexibility and integrations with catalog providers, such as the AWS Glue Data Catalog and the Snowflake Open Catalog (a managed service for Apache Polaris). The Apache Iceberg and Delta Lake support enable real-time and batch processing and help ensure real-time data consistency across applications.

Confluent’s Chief Product Officer, Shaun Clowes, stated: “With Tableflow, we’re bringing our expertise of connecting operational data to the analytical world. Now, data scientists and data engineers have access to a single, real-time source of truth across the enterprise, making it possible to build and scale the next generation of AI-driven applications.”

Confluent Tableflow screenshot graphic
Confluent Tableflow screenshot graphic

Citing the IDC FutureScape: Worldwide Digital Infrastructure 2025 Predictions report, which states: “By 2027, after suffering multiple AI project failures, 70 percent of IT teams will return to basics and focus on AI-ready data infrastructure platforms,” Confluent claims AI projects are failing because old development methods cannot keep pace with new consumer expectations. 

The old development methods, or so says Confluent, are related to an IDC finding: “Many IT organizations rely on scores of data silos and a dozen or more different copies of data. These silos and redundant data stores can be a major impediment to effective AI model development.”

The lesson Confluent draws is that instead of many silos, you need a unified system that knows “the current status of a business and its customers and take action automatically.” Business operational data needs to get to the analytics and AI systems in real-time. It says: “For example, an AI agent for inventory management should be able to identify if a particular item is trending, immediately notify manufacturers of the increased demand, and provide an accurate delivery estimate for customers.” 

Tableflow, it declares, simplifies the integration between operational data and analytical systems, because it continuously updates tables used for analytics and AI with the exact same data from business applications connected to the Confluent Cloud. Confluent says this is important as AI’s power depends on the quality of the data feeding it.

The Delta Lake support is AI-relevant as well as it’s “used alongside many popular AI engines and tools.”

Of course, having both real-time and batch data available though Iceberg and Delta Lake tables in Databricks and other data warehouses and lakes is not enough for AI large language model processing; the data needs to be tokenized and vectorized first.

Confluent is potentially ready for this, with its Create Embeddings action, a no-code feature “to generate vector embeddings in real time, from any model, to any vector database, across any cloud platform.”

Users can bring their own storage to Tableflow, using any storage bucket. Tableflow’s Iceberg tables support access for analytical engines such as Amazon Athena, EMR, and RedShift, and other data lakes and warehouses, including Snowflake, Dremio, Imply, Onehouse, and Starburst. 

Apply for the Tableflow Early Access Program here. Read a Tableflow GA product blog to find out more.

Bootnotes

The Confluent Cloud is a fully managed, cloud-native data streaming platform built around Apache Kafka. Confluent was founded by Kafka’s original creators, Jay Kreps, Jun Rao, and Neha Narkhede. They devised Kafka when working at LinkedIn in 2008.

Delta Lake is open source software built by Databricks that is layered above a data lake and enables batch and real-time streaming data processing. It adds a transaction log, enabling ACID (Atomicity, Consistency, Isolation, Durability) transaction support. Users can ingest real-time data, from Apache Kafka, for example, into a dataset, and run batch jobs on the same dataset without needing separate tools. Databricks coded Delta Lake on top of Apache Spark.

Apache Kafka is open source distributed event streaming software built for large-scale, real-time data feeds used in processing, messaging, and analytics. Data in Kafka is organized into topics. These are categories or channels, such as website clicks or sales orders. A source “producer” writes events to topics, and target “consumers” read from them.

Apache Iceberg is an open source table format for large-scale datasets in data lakes, sitting above storage systems like Parquet, ORC, and Avro, and cloud object stores such as AWS S3, Azure Blob, and the Google Cloud Store. It brings database-like features to data lakes, such as ACID support, partitioning, time travel, and schema evolution. Iceberg organizes data into three layers:

  • Data Files: The actual data, stored in formats like Parquet or ORC
  • Metadata Files: Track which data files belong to a table, their schema, and partition info
  • Snapshot Metadata: Logs every change (commit) as a snapshot, enabling time travel and rollbacks

Apache Spark is an open source, distributed computing framework intended for fast, large-scale data processing and analytics. It’s used for big data workloads involving batch processing, real-time streaming, machine learning, and SQL queries.

The Databricks Unity Catalog provides a centralized location to catalog, secure, and govern data assets like tables, files, machine learning models, and dashboards, across multiple workspaces, clouds, and regions in the Databricks Lakehouse environment. It acts as a single source of truth for metadata and permissions.

Micron warns of inventory drag outside datacenters

Micron saw revenues for its DRAM and NAND grow year-over-year as datacenter servers used more high-bandwidth memory but sales outside the datacenter market turned down with buyers using up inventories.

Revenues in the quarter ended December 31, 2024, were $8.05 billion, up 38 percent year-on-year but down 8 percent quarter-on-quarter after seven consecutive growth quarters. There was a $1.58 billion GAAP profit, virtually doubling the year-ago $793 million profit.

CEO Sanjay Mehrotra stated: “Micron delivered fiscal Q2 EPS above guidance and datacenter revenue tripled from a year ago” with record DRAM sales. He added in the earnings call: “Micron is in the best competitive position in our history, and we are achieving share gains across high-margin product categories in our industry.” 

Micron revenue/profit
Note the quarter-on-quarter downturn in Q2 FY 2025 – but Micron expects a rebound next quarter

Financial summary

  • Gross margin: 36.8 vs 18.5 percent a year ago
  • Operating cash flow: $3.94 billion vs $1.2 billion a year ago
  • Free cash flow: $857 million vs -$29 million last year
  • Cash, marketable investments, and restricted cash: $9.6 billion vs $9.7 billion a year ago
  • Diluted EPS: $1.41 vs 0.71 a year ago

Micron has two main tech segments, DRAM and NAND. DRAM revenues of $6.1 billion were up 47 percent year-on-year but down 4 percent quarter-on-quarter, with high-bandwidth memory (HBM) driving the growth.

Mehrotra said: “We expect a strong ramp of HBM throughout calendar 2025. As noted before, HBM3E consumes three times the amount of silicon compared to D5 to produce the same number of bits. Looking ahead, we expect the trade ratio to increase with HBM4 and then again with HBM4E when we expect it to exceed 4-to-1.” HBM DRAM commands a premium over conventional DRAM.

NAND revenues were $1.9 billion, up 18 percent year-on-year and down 17 percent quarter-on-quarter. Mehrotra said: “In calendar Q4 2024, based on industry analyst reports, Micron achieved yet another record high market share in datacenter SSDs with revenue growth in each category, including performance, mainstream, and capacity SSDs.”

“In NAND, we continue to underutilize our fabs, and our wafer output is down mid-teens percentage from prior levels,” he said. However, a rebound is coming with Micron saying that demand moderated in fiscal Q2 due to short-term customer inventory-driven impacts. It sees a return to bit shipment growth in the months ahead and expects to generate multiple billions of dollars in datacenter NAND revenue and grow its datacenter NAND market share in calendar 2025.

Micron DRAM, NAND revenue
It looks like a cyclical revenue trough might be coming – but Micron disagrees with that perception

There are four business segments:

  • Compute & Networking revenues were $4.6 billion, up 47 percent year-on-year and down 4 percent quarter-on-quarter
  • Storage revenues were $1.4 billion, up 54.7 percent year-on-year and down 20 percent quarter-on-quarter with low datacenter sales
  • Mobile revenues dropped 32 percent year-on-year and 30 percent quarter-on-quarter to $1.1 billion as customers used up inventories
  • Embedded market revenues went down 10 percent year-on-year and 3 percent quarter-on-quarter as automakers used up inventory

Sequentially, only the compute and networking segment revenues grew; the HBM effect in action. AI training needs GPU servers, which require HBM and SSD storage but the AI Inferencing market is still nascent. That is affecting NAND growth in datacenters and workstations, much of Micron’s storage market. The mobile and embedded markets seem to be stalled.

Micron business unit performance

HBM is increasingly important to Micron, with HBM DRAM representing a significant growth opportunity for market leader SK hynix and Samsung. Mehrotra said: “We expect multibillion dollars in HBM revenue in fiscal 2025 … As noted before, HBM3E consumes three times the amount of silicon compared to D5 to produce the same number of bits. Looking ahead, we expect the trade ratio to increase with HBM4 and then again with HBM4E when we expect it to exceed 4-to-1.”

Micron is enthusiastic about its coming HBM4, which will ramp in volume in calendar 2026. It aims to deliver “the best HBM4 products to the market across power efficiency, quality, and performance.” And it expects to gain HBM market share compared to SK hynix and Samsung.

PC demand for DRAM will be driven by PC refreshes for AI-capable systems and the Windows 10 end-of-life in October this year. AI PCs require a minimum of 16 GB of DRAM versus the average 12 GB PC content last year. 

Smartphone unit volume growth looks to be in the low single-digit percentage area, and, with inventories emptying, Micron expects mobile DRAM and NAND bit shipment growth in the current quarter. AI-capable phones will need more DRAM and NAND but demand for such devices is low presently. The embedded market is still affected by automotive OEMs, industrial, and consumer-embedded customers adjusting their inventory levels.

On tariffs, Micron says it “serves as the US importer of record (IOR) for a very limited volume of products that would be subject to newly announced tariffs on Canada, Mexico, and China … Where tariffs do have an impact, we intend to pass those costs along to our customers.”

Concerning next quarter’s outlook, Mehrotra said: “We expect fiscal Q3 revenue to be another record for Micron, driven by shipment growth across both DRAM and NAND.” The actual revenue outlook is a record $8.9 billion ± $100 million, a 30.7 percent year-on-year increase at the mid-point. That’s a rebound from the current quarter’s sequential decline.

GigaOm ranks top UDM and STaaS vendors – but are there too many?

Research house GigaOm has re-evaluated unstructured data management (UDM) and Storage-as-a-Service (STaaS) suppliers in updated Radar and Sonar reports.

It reviewed 22 of the top UDM vendors and the Radar report graphic locates them in a 2D concentric circular space, with an outside Entrant ring, a middle Challenger ring, and an inner Leader ring, with closeness to the central bullseye indicating a higher rating in terms of overall value. This circular space is divided into four quadrants by two orthogonal axes running through the bullseye. A horizontal one runs from Feature Play to Platform Play, and a vertical one spans a Maturity to Innovation range. The 22 suppliers are located in one of the three rings and one of the four quadrants, with colored arrows depicting the speed of their progress.

It sounds complicated, but a look at the Radar graphic provides a clear overview of the vendors’ situations and ratings:

GigaOm radar

The top-rated vendors, with averaged ratings across a key features comparison, are Arcitecta, Cohesity, and IBM, each with 4.4 out of the maximum 5 score. BigID is fourth with 4.3, Varonis fifth with 4.1, and Hammerspace and Pure Storage follow, both scoring 4.0.

In a separate business criteria comparison Arcitecta leads (4.7) followed by Cohesity and Hammerspace (4.5 each), Komprise (4.3), and Atempo, BigID, Hitachi Vantara, IBM, and Panzura (4.2 each).

Our first impression is that there are too many suppliers included, as our own list of UDM vendors includes Arcitecta, Datadobi, Data Dynamics, Hammerspace, and Komprise – just five. We wouldn’t, for example, intuitively include Hammerspace and Pure Storage in one unstructured data management market category, nor Seagate and Datadobi for that matter.

Analyst Whit Walters casts his net wide, including cloud file services suppliers CTERA, Nasuni and Panzura, general unstructured data storage suppliers such as Dell, HPE, Hitachi Vantara, IBM, NetApp, Pure Storage, and Quantum, data protector Cohesity, scale-out filer Qumulo, and HDD manufacturer Seagate, which offers the Lyve product line. Add in Aparavi and Atempo, and you have an eclectic group of 22 suppliers.

What are Walters’ criteria for inclusion? They are pretty general, as he looked at suppliers in two market segments:

  • Infrastructure-focused: Solutions for this segment are designed to target data and metadata management at the infrastructure level, including automatic tiering and basic information lifecycle management, data copy management, analytics, indexing, and search.
  • Business-focused: Solutions for this market are designed to solve business-related problems, including compliance, security, data governance, big data analytics, e-discovery, and so on.

The vast majority of vendors, 17, focus on providing general platforms rather than specific features. Nine are rated as more mature than innovative, with eight being rated the opposite way.

There are two mature feature players, Varonis and BigID, and three innovative feature players, Atempo, Aparavi, and Qumulo.

Walters examined two emerging features, AIOps and Model Training, as well as Generative AI for Data Management. HPE scored highly here, with a 4.5 average score, while BigID, CTERA, Dell, and Nasuni each scored a 4.

A second GigaOm Sonar report by analyst James Brown reviews STaaS suppliers, and looks at a smaller group of ten vendors. They are located in what looks like the bottom half of a Radar diagram:

GigaOm sonar

The analyst tells us: “The GigaOm Sonar provides a forward-looking analysis of vendor solutions in a nascent or emerging technology sector. It assesses each vendor on its architecture approach (Innovation) while determining where each solution sits in terms of enabling rapid time to value (Feature Play) versus delivering a complex and robust solution (Platform Play).”

“The GigaOm Sonar chart plots the current position of each solution against these three criteria across a field of concentric semicircles, and solutions set closer to the center are judged to be of higher overall value. The forward-looking progress of vendors is further depicted by arrows that show the expected direction of movement over a period of 12 to 18 months.”

The vendors are spread fairly evenly across this space, with Infinidat leading overall, followed by NetApp, Pure Storage, and then HPE. Challengers are led by Zadara, Dell, and Hitachi Vantara.

Given that Backblaze and Wasabi are included, it raises the question of why AWS, Azure, and Google Cloud are absent. We have asked GigaOm for clarification ad James Brown told us: “The exclusion of AWS, Azure, and Google Cloud from the STaaS report is due to their foundational role in defining the Storage as a Service market, as STaaS is a core component of their established Infrastructure as a Service offering. Including these market giants would likely skew the report’s findings, potentially overshadowing emerging or niche STaaS providers who offer specialized solutions or target distinct market segments. The report focuses on evaluating providers that offer unique value propositions beyond standard object, block, and file storage, such as specialized storage for specific workloads, enhanced data management features, or on-premises and hybrid solutions. These vendors often serve as the benchmark against which other providers are measured, rather than being evaluated themselves.”

GigaOm subscribers can access both reports through their subscriptions.

IBM boosts Storage Scale with content-aware AI integration

IBM Storage Scale, the longstanding parallel file system, now includes content-awareness to enhance its retrieval-augmented generation (RAG) capability.

Storage Scale has a substantial history in high-performance computing (HPC) and enterprise use cases requiring similar high-speed file data access, including large language model (LLM) training and inference. Storage Scale v5.2.1 added high-performance S3 object storage support. The product also supports Nvidia’s GPUDirect host server memory bypass protocol. In September last year, IBM combined its watsonx data lakehouse with Storage Scale to provide object storage facilities beneath a file access overlay. The aim was to improve AI training and inference workload speed.

Now IBM has expanded on this collaboration with Nvidia, announcing:

  • Content-aware Storage Scale (CAS) capability building on Nvidia’s AI-Q blueprint and NeMo Retriever microservice.
  • Expanded watsonx integrations via support for Nvidia NIM microservices for external AI model integration across multiple cloud environments. IBM watsonx.governance will allow enterprises to monitor and govern NIM microservices across any hosting environment.
  • AI-focused IBM Consulting capabilities, with Nvidia providing AI integration services via Nvidia Blueprints, optimizing compute-intensive AI workloads across hybrid cloud environments using Red Hat OpenShift and Nvidia AI.
  • Nvidia H200 GPU instance availability on IBM Cloud. 
  • Storage Scale support for Nvidia’s AI Data Platform reference design.
Vincent Hsu, IBM
Vincent Hsu

IBM Fellow, VP, and CTO Vincent Hsu wrote about CAS, saying it is based on work by IBM Research and Nvidia. Because IBM now has “embedded compute, data pipelines, and vector database capabilities within the storage system, CAS reduces data movement and latency to increase efficiency.”

Storage Scale can obtain data from many sources. It has “global data abstraction services engineered to provide connectivity from multiple data sources and multiple locations to bring data into your AI factory from IBM and third-party storage environments.”

CAS can, using AI-Q and NeMo Retriever, “extract information from text, charts, graphs and even images.” It is able to “watch folders on other storage systems or in the cloud to identify changes as they occur, automatically run pre-built pipelines, and update just the changes to the vector database, ensuring that data is current for AI applications efficiently using Nvidia accelerated computing.”

IBM Storage Scale will respond to queries using extracted and augmented data, accelerating communications between GPUs and storage via BlueField-3 DPUs and Spectrum-X networking.

CAS can detect changes in multiple unstructured data stores, convert the modified data into vectors, and feed it to AI models to improve RAG responses to user or AI assistant requests. Or as Hsu put it: ”By applying innovative natural language processing techniques, we’ve developed ways to much more efficiently extract the semantic meaning from all kinds of content, making it easier to update AI tools to improve the quality of their answers.”

Nvidia Storage Networking Technology VP Rob Davis said: “AI agents need to rapidly access, fetch and process data at scale, and today, these steps occur in separate silos. The integration of IBM’s content-aware storage with Nvidia AI orchestrates data and compute across an optimized network fabric to overcome silos with an intelligent, scalable system that drives near real-time inference for responsive AI reasoning.”

Storage Scale content-awareness facilities will be embedded in the next update of IBM Fusion generally available in the second quarter. Interested parties can learn more by registering here for the webinar “AI First Storage – Enhancing AI Results with Content-Aware Storage,” scheduled for April 8, 2025, at 1100 Eastern Time.

DDN unveils Inferno and xFusionAI with hybrid file-object tech

Analysis: DDN launched its Inferno fast object appliance and xFusionAI hybrid file+object technology at Nvidia’s GTC 2025. We examine both technologies to see what they offer.

From the AI training and inference point of view, DDN has two storage technologies. Its originally HPC-focused EXAScaler line presents a Lustre-based parallel file system running on NVMe SSDs in scale-out storage nodes with client software executing in the GPU server nodes. The newer Infinia technology provides a base key-value store on which data access protocols are layered, with the S3 object protocol first. File, block, and others will be added over time.

Infinia v2.0, released in February, is designed to eliminate AI bottlenecks, accelerate workflow, and scale for complex model listing. It delivers real-time data services, multi-tenancy, intelligent automation, and has “a very powerful AI-native architecture,” DDN says.

But it does not support file access, and most AI development work to date has used fast file access with slower object as the protocol to access mass unstructured data stores. Loosely speaking, in AI today, file is the past, but participates in the present with growing object access, and object will be increasingly important in the future. It will be a hybrid file+object world for AI training and inference generally for the foreseeable future.

The Inferno product adds Nvidia’s Spectrum-X switch, with RoCE adaptive routing, to Infinia storage. DDN says testing showed Inferno outperforming AWS S3-based inference stacks by 12x with sub-millisecond latency and 99 percent GPU utilization. DDN states that Inferno “is a high density and low power 1RU appliance with 2 or 4 BlueFields.” These will be Nvidia’s Arm-powered smartNICS that link to BlueFields in Nvidia-powered GPU servers.

Inferno uses “high-performance NVMe drives for ultra-low latency … supports seamless expansion,” and “is optimized for AI model training and inference.” It is “fully optimized with DDN’s AI and HPC ecosystem, ensuring streamlined deployment.” 

There is no publicly available Inferno configuration or availability information. If such an appliance used 122 TB QLC SSDs, then we could be looking at a ten-bay chassis with 1.2 PB capacity. A rack containing 30 of them would have 36 PB with an NVMe/GPUDirect for Objects-based network fabric comprised of BlueField-3s talking across Spectrum-X networking to BlueField-3s connected to GPU servers. 

DDN describes xFusionAI technology as different, stating that it is “engineered as an integrated AI storage and data management solution that balances high-speed parallel file storage with cost-efficient object storage for AI workflows … There is a single pool of storage that is logically partitioned between EXAScaler and Infinia, rather than two completely separate systems. xFusionAI can be deployed on a unified hardware infrastructure, with both EXAScaler and Infinia software components running within the same system.”

It “features a single-user interface that provides visibility and control over both EXAScaler and Infinia environments, ensuring streamlined management.” 

Infinia is not merely a backend object store, says the vendor, adding that it serves as an intelligent data management layer to complement EXAScaler’s high-speed file performance. Data can be moved between EXAScaler and Infinia using automated policies or manual tiering, allowing users to optimize storage costs and performance.

In effect, we have high-speed file storage (EXAScaler) being added to Infinia, possibly as a stopgap, until Infinia’s native file system support arrives. This means xFusionAI controllers will be more capable than Inferno’s (Infina object-only) controllers, as they must manage both file and object environments and “move” data between them. We put “move” in quotes because the data might not actually physically move; it could somehow be remapped so that it is transferred from an EXAScaler partition to the Infinia partition, and vice versa. Of course, if the Infinia partition used slower QLC drives with the EXAScaler partition using faster TLC drives, then data would physically move.

It will be interesting to understand the details of this hybrid system as they emerge. One insight is that xFusionAI gives DDN a combined file+object AI training and inference storage system to compete with VAST Data’s hybrid file+object storage, which also has block access, less important in the AI world up to now. DDN says the “product is coming soon. Pricing details are available upon request and depend on configuration, capacity, and deployment scale.”

AI puts value in the data. So how do we get it out?

SPONSORED FEATURE: AI is driving an explosion in infrastructure spending. But while GPU-enabled compute may grab the headlines, data management and storage are also central to determining whether enterprises ultimately realize value from their AI investments and drive broader transformation efforts.

The worldwide AI infrastructure market is expected to hit $100bn by 2027, according to IDC. Servers are expected to account for the lion’s share of this spending, but storage investment is increasing in line with overall growth as tech leaders cater for the massive datasets which AI requires along with the need for training, checkpoint and inference data repositories.

While AI is fueling this spending boom, many of the underlying challenges facing CIOs haven’t changed, explains HPE’s SVP and GM for storage, Jim O’Dorisio. These include driving innovation, streamlining operations, and reducing the total cost of operations, all within the maelstrom of a constantly evolving tech and business landscape.

Data, and therefore storage, all play into this. AI relies on data. But so do the myriad of other, more traditional, operations that companies regularly undertake. But it must be the right data, available to the right systems at the right time, and at the right speed, says O’Dorisio.

“If you go back 15 years ago, 10 years ago, storage was really just where data sat. Increasingly, it’s where we create value now, right,” he explains.

Dealing with the issues of data gravity and location is particularly challenging, a situation aggravated by the broader span and complexity in customer IT environments. The last two decades have seen a rush to the cloud, for example. But many enterprises are now wondering just how much they actually need to be off premises, particularly when it comes to managing all the data they need in order to realize value from AI.

That decision may come down to higher-than-expected costs, or any given cloud provider’s inability to meet strict organizational performance or security requirements, especially for real time and/or AI workloads. IDC notes that even cloud native organizations are beginning to question whether private cloud or on-prem has a role to play for them.

And beyond creating value through AI or other advanced applications, enterprise data still needs to be protected and managed as well. The cyberthreat is more acute than ever – with threat actors themselves enthusiastically leveraging AI. 

The cyber challenge is clearly right up there, says O’Dorisio, but this repatriation of data also creates additional hybrid complexity. There’s sustainability to consider as well, for example. Complex systems require energy to run, and data should be managed efficiently. But the underlying storage should also be as efficient as possible. That includes optimizing energy consumption but also considering the impact of overprovisioning and unduly short life cycles.

This is a legacy problem

The crucial question for an organization’s storage systems then is whether they can keep up with the speed of change. The answer, too often, is they can’t. For multiple reasons.

Traditional architectures that rigidly tie together compute and storage can pose problems when scaling up to meet increasingly complex or large workloads. Expanding storage capacity can mean spending on compute that isn’t really needed, and vice versa. This can lead to silos of systems built out for a particular business unit or workload, or a particular location, for example, core datacenters or edge deployments.

Likewise, legacy architectures are often targeted at specific types of storage: block; file; object. But AI doesn’t distinguish between data formats. It generally wants to chew through all the data it can, wherever it is.

This lack of flexibility can be aggravated by legacy systems that were designed for a particular type of organization or scale, e.g. “enterprise” or a medium sized business. Integrating a raft of standalone systems can present a clear architectural issue as well as management challenges.

Disparate hardware often means disparate management systems and consoles for example, meaning managers are left with a fragmented view of their overall estate. That situation can force team members to specialize in a subset of the organization’s infrastructure, which can often result in inefficiencies and increased operational costs.

These fragmented, siloed, and often hard to scale systems don’t lend themselves well to the hybrid operations that are increasingly becoming the norm. Any organization contemplating repatriating some or all of its data will likely balk at losing the ease of use of managing their data in the cloud.

This can all contribute to a massive bottleneck when it comes to maximizing the value of all the data available. “The architectures are typically complex, and they’re siloed, explains O’Dorisio. “And it makes extracting value from the data very difficult.”

Where is the value?

HPE has sought to address these challenges with its HPE Alletra storage MP platform. The architecture disaggregates storage and compute, meaning each can be scaled separately. So, as the demands of AI increase, infrastructure can be scaled incrementally, sidestepping the likelihood of siloes or wasteful overprovisioning, says HPE. This is bolstered by HPE’s Timeless program, which ensures a free, nondisruptive controller refresh, cutting TCO by 30 percent compared to standard forklift upgrades according to HPE estimates.

The MP stands for multiprotocol, with common underlying hardware optimized for particular applications. HPE Alletra Storage MP B10000 modernizes enterprise block storage with AI-driven cloud management, disaggregated scaling, and 100 percent data availability for all workloads, says HPE. Whereas, the HPE Alletra Storage MP X10000 is purpose built for intelligent high-performance object storage. The AMD EPYC embedded processors at their core are designed to offer a scalable X86 CPU portfolio delivering maximum performance with enterprise-class reliability in a power-optimized profile.

An upcoming release of the X10000 system will give the ability to tag data and add metadata as data is being stored. Users will be able to add vector embeddings and similar functions to support downstream Gen AI RAG pipelines. “Our whole notion is really to add the intelligence and create value as the data is being stored, which really significantly reduces time to value for our customers,” O’Dorisio says. Together with the unified global namespace in HPE Ezmeral Data Fabric, customers can aggregate data from across their enterprise to fuel AI initiatives.

But, even if tech leaders have good reason to situate some or even all their storage infrastructure outside the cloud, giving up the ease of management the cloud offers is a harder sell. Step forward the HPE GreenLake cloud, designed to deliver a single cloud operating model to manage the entire storage estate, across the core, edge and cloud.

Any form of disruption to IT operations, whether due to a disaster or a cyberattack, is now considered an inevitability rather than misfortune. However, by harnessing the Zerto ransomware detection and recovery software, organizations “can really recover in hours and days, versus maybe weeks and months when you’re trying to recover a bunch of data from a cloud,” says O’Dorisio.

Intelligent data savings

This intelligent approach to architecture and ownership also supports a reduction in associated emissions by half, O’Dorisio adds, by reducing overprovisioning and the need for forklift upgrades.

HPE’s own research shows that HPE Alletra Storage MP’s disaggregated architecture can reduce storage costs by up to 40 percent. Better still, intelligent self-service provisioning can deliver up to 99 percent operational time savings, calculates the company.

One major global telecom provider recently deployed HPE Alletra Storage MP B10000 to refresh its legacy storage arrays. In the process, the company dramatically reduced the costs associated with support, energy and cooling, as well as datacenter space, says HPE.

The move helped reduce operating expenses by more than 70 percent while allowing the telco to accommodate a higher volume of traditional databases as well as more modern applications. The increased storage capacity with a smaller footprint means the telco provider also now has space in their datacenter to accommodate future growth.

None of that is to suggest that storage in the AI age is anything less than complex.  After all, as O’Dorisio says, “The data really spans, from private to the edge to the public cloud. Data sits across all those environments. Data is more heterogeneous.”

But deploying block, file or object storage services on a common cloud-managed architecture means both managing and extracting value from that data will be much easier and efficient.

Sponsored by Hewlett Packard Enterprise and AMD.

Pure Storage expands AI capabilities with Nvidia certification

Pure Storage has worked with Nvidia to enable existing and new FlashBlade customers to provide data storage for AI models running on Nvidia’s AI Data Platform.

Despite recently announcing its disaggregated architecture FlashBlade//EXA tech – which scales performance and capacity beyond existing FlashBlade products – Pure has ensured compatibility with Nvidia’s new AI Data Platform.  Through reference design adoption and certifications, FlashBlade customers can integrate with Nvidia’s Blackwell GPUs – transferring data via BlueField-3 NICs, Spectrum-X networking, NIM and NeMo Retriever microservices, and the AIQ blueprint.

Rob Davis, Nvidia VP for Storage Networking Tech, stated: “By integrating Nvidia AI Data Platform capabilities into Pure Storage FlashBlade, enterprises can equip AI agents with near-real-time business data to achieve new levels of personalized customer service, operational efficiency, and unprecedented productivity.”

Pure supports the AI Data Platform reference design with its FlashBlade products and is now certified as high-performance storage (HPS) platform for Nvidia’s Partner Network Cloud Partner Reference Architectures, including HGX systems with B200 or H200 GPUs. It has also earned Nvidia-Certified Storage Partner approval at both the Foundation and Enterprise levels, affirming that Pure FlashBlade can serve as a storage component in Nvidia-style AI factories.

Pure Storage CTO Rob Lee
Rob Lee

The Foundation level is an entry point for Nvidia storage partners, validating baseline performance and compatibility with its AI infrastructure for training smaller large language models (LLMs), inference tasks, and initial retrieval-augmented generation (RAG) workflows. The Enterprise level targets large-scale AI deployments powering AI factories and handling massive datasets for agentic AI and other GenAI applications.

Pure provides storage for the converged FlashStack system, built with Cisco servers and networking. FlashStack customers gain a clear pathway to integrating with Nvidia’s AI Data Platform.

Pure CTO Rob Lee asserts: “The incorporation of the Nvidia AI Data Platform into FlashBlade provides the AI-ready storage” that customers need, adding: “Our recent Nvidia certifications affirm that Pure Storage is supporting the pace and scale that AI models need.”

Solidigm promises fanless GPU servers by extending liquid cooling to SSDs

SK Hynix’s SSD business Solidigm has unwrapped an SSD liquid cooling technology it says will result in smaller, fan-less GPU servers while increasing storage density.

GPU servers, and the datacenters that house them, are becoming hellishly hot. If powering the mega chips and associated storage and networking infrastructure wasn’t enough, the need for liquid cooling is increasing the power draw, reducing the real estate available for servers, and making datacenter design increasingly complex.

At the same time, Solidigm senior director AI and leadership marketing Roger Corell said a typical GPU-based AI server typically carried about 30TB of storage across eight slots, “And we don’t see any reason why that capacity per server growth will not continue at a high level”

That storage element is usually air cooled, added Avi Shetty, senior director AI market enablement and partnerships. But as density increases, this becomes a problem. The SSDs themselves get hot, raising the risk of shutdowns, while traditional cooling technologies, ie heat sinks and fans put a brake on density or reducing server size

Solidigm’s answer is the D7 PS1010 E1.S, which is a 9.5mm form factor SSD and a “cold plate technology kit”, aimed at the direct attached storage element in an AI setup.

Shrinking the drive to 9.5 mm provides space for the cold plate which is connected to and chilled the liquid cooling system already supplying the server. That means the drives themselves are still hot-swappable. And it allows for more density in a 1U rack.

The cold plate only touches one side of the drive. However, as Shetty explained, there are components on both sides of the drive, and chilling one side only isn’t really an option.

“What we have is an innovation which allows for heat to dissipate from the front side as well as well as on the back side,” he said.

This results in “an overall thermal mechanism, which allows us to go at much higher watts at the platform level, up to 30 watts, while maintaining full bandwidth of performance.”

There are some liquid cooled consumer devices available and other storage companies have demonstrated professional devices with integrated cooling. However, Solidigm claims its tech will be the first enterprise device to feature “complete liquid cooling”.

The technology, developed in collaboration with an as yet un-named partner is aimed at “future AI servers” – so no retrofitting. It has not set a precise launch date, beyond the second half of this year.

Solidigm is still working through the exact impact of the cooling technology on overall power consumption.

But Corell said, said there was potential to save power in a number of ways. “One, you don’t need to power the fans. And two, you don’t need potentially as low an ambient air temperature inside the aisles of racks to, you know, pull that cooler air over storage, so lowering both HVAC and cooling power requirements.”

Shetty told us, “The liquid cooling loop for storage can be in parallel to CPU/GPU without affecting the cooling efficiency. With fans removed, we expect overall cooling efficiency to beimproved at the server level along with other TCO improvements.”