Home Blog Page 2

How Pure Storage developed the FlashBlade//EXA system

Interview. What led up to Pure Storage developing a new disaggregated system architecture for its FlashBlade//EXA?

Pure Storage was founded as an all-flash storage array (AFA) vendor, surviving and prospering as a wave of similar startups rose, shone briefly, and then were acquired or failed as more established competitors adopted the technology and closed off the market to most new entrants.

In the 2010-2015 era and beyond, the mainstream enterprise incumbents including Dell, Hitachi Vantara, HPE, IBM, NetApp, and Pure ruled the AFA roost with what was essentially a dual-controller and drives architecture with scale-up features. There was some limited scale-out, with examples such as Isilon, but for true scale-out you had to look to the HPC space with parallel file system-based suppliers such as DDN.

Then, in 2016, a disaggregated canary was born in the AFA coal mine: VAST Data emerged with separately scale-out metadata controllers and scale-out all-flash storage nodes talking to each other across an NVMe RDMA fabric. There was a lot more clever software in VAST’s system yet the base difference was its disaggregated design, aided and abetted by its single storage tier of QLC flash drives. 

Moving on, the AI boom started and gave an enormous fillip to VAST Data and parallel file system AFA vendors such as DDN. Only their systems could keep GPU server clusters running GenAI training workloads without data I/O waits. The incumbent enterprise AFA vendors added Nvidia GPUDirect support to speed their data delivery to GPU servers but could not match the data capacities enabled by the disaggregated scale-out and parallel file system vendors. Consequently, VAST Data and DDN prospered in the Nvidia-dominated GPU Server mass data training space.

This caused a revaluation among other suppliers. HPE adopted a VAST-like architecture with its Alletra storage. Dell assembled Project Lightning to add parallelism to its PowerScale/Isilon storage, while NetApp created the ONTAP for AI project. And Pure? It announced its FlashBlade//EXA a couple of weeks ago.

This has a clever twist in that the existing dual-controller FlashBlade system is used as the metadata controller layer system with separately scaled-out QLC flash storage nodes accessed across an RDMA fabric. We spoke to Pure’s VP for Technology, Chadd Kenney, to find out more.

Blocks & Files: FlashBlade//EXA seems like a radical revision of the FlashBlade architecture, giving it a parallel processing style of operation. Would it be possible to do the same sort of thing with FlashArray, and is that a silly question?

Chadd Kenney: No, it’s an interesting one. So let me give you some background on kind of where EXA came about, which may help you understand the logic behind why we decided to build it. It was a very fun project. I think we spent a lot of time thinking about what architectures existed in the market today and which ones did we want to take additional values from, and then what core competences did we have that we could potentially apply to a new solution. 

And so this was an interesting one because we had a couple different options to play with. The first is we obviously could make bigger nodes and we could break away from the chassis kind of configuration and the bigger nodes would give us more compute. We’d have ability to scale them in a different manner. And that gave us a pretty sizable performance increase, but it wasn’t big enough.

Customers were starting to ask us for, it’s funny, we would get in conversations with customers, they would ask us for 10 terabytes or 20 terabytes per second of performance. So we were kind of scratching our head saying, wow, that’s a lot of performance to deliver there. And so we started to think then later about how decoupling the infrastructure could be an interesting possibility, but how would we do it? 

Then we went back and forth on what the core competence that we were trying to solve for customers was. And the one thing we continued to hear was metadata performance was incredibly problematic and in many cases it was very rigid in the way that it actually scaled. So as an example, there are alternate PNFS or Lustre-like solutions that have tried to solve this before where they disaggregated metadata out. The downside was that it wasn’t scalable somewhat independently in a very granular fashion so that I could say I want to have a massive amount of metadata and a tiny bit of capacity, or I want to have massive capacity or a little tiny bit of metadata.

And so that intrigued us a little bit to understand, OK, so how could we actually deliver this in a different mechanism? 

The second thing we started to think about was, if you think about what FlashArray started with, it’s a key-value store that was highly optimized to flash in order to access data. We then took that same exact key-value store and FlashBlade and somewhat scaled it out across these nodes and we realized, OK, there’s a couple of core competencies to this. 

One is incredibly fast with multi-client connections. And then the second part was that we had an object count that was in the 40 quadrillion level. I mean it was ridiculous level for the amount of objects. So as we started to think about that, we said, well, what if we use the metadata engine that was core to FlashBlade and just kept that intact and then with the amount of objects that we could scale, we could infinitely scale data nodes.

We actually don’t even know what the quantity of data nodes we could get to was. So then we had to test to say, well, could these data nodes scale in a linear fashion? And so when we started adding data nodes in this decoupled way we decided to go with PNFS because that was what customers were asking for initially from us. 

Although I think ObjectStore with S3 over RDMA will probably be the longer term approach, PNFS was kind of like what every customer is asking us about. And so when we started building this, we started to realize, oh my God, as we add nodes, we are seeing exactly a linear performance rise. It was between 85 to a hundred gigabytes per second. And we got really excited about the fact that we could actually build a system that could take very different client profiles of access patterns and linearly scale the overall bandwidth.

And so then we all of a sudden got really excited about trying to build a disaggregated product. And the first logical thing we thought about was, well, we’ll build our own data nodes. And we’ve talked about doing those for quite some time and we were back and forth on it. The hyperscale win kind of gave us some ideas around this as well. And so we then looked at talking to customers – how would you consume this? And what was interesting is, most said, I already have a massive data node infrastructure that I’ve invested in. Could I just use those instead? 

And we said even competitive nodes, comically enough, we could potentially use, we just have to meet the minimum requirements, which are pretty low. And so we decided let’s go to market first with actually off-the-shelf whatever nodes that meet the requirements and Linux distribution and we’ll lay a small little set of packages to optimize the workflows, things like our rapid file toolkit and a bunch of other stuff that were on top of it. 

It got really exciting for us that all of a sudden we had this thing we could bring to market that would effectively leapfrog the competition in performance and be able to achieve some of these larger GPU cloud performance requirements that were almost unattainable by anybody without specific configurations out there.

Blocks & Files: I think I’m asking a dumb question by asking if FlashArray could use this same approach because that’s basically like asking could the compute engine in Flash array use the same approach.

Chadd Kenney: It’s an interesting concept because I think we’re always open to new concepts and ideas and we try to prove these out, especially in the CTO office, we spent a lot of time conceptualising could we change the way that we construct things. I mean the one thing that changed with FlashBlade that was different was we went to this open ecosystem of hardware. And so there is a different mode of operation than what we typically build. We typically build these gorgeous, elegant systems that are in highly simple that anyone can get up and running in near minutes. EXA is obviously a little bit different, but what we realized is the customers wanted that build-it-themselves thing. They kind of liked that, whereas in the enterprise model, they’re not that into that. They really want just the more black box, plug it in and it works.

Blocks & Files: An appliance?

Chadd Kenney:  Yes. I think with FlashArray, the one thing that’s tough about it is it’s just so good at what it does in its own ecosystem and we’ve kind of built now every possible tier. In fact, you’ll see some announcements come out from us around Accelerate timeframe … But we’re going to take somewhat of an EXA-level performance in FlashArray too. And so we really like the way it’s built, so we haven’t really found a mechanism of disaggregating it in any way that made sense for us. And I think you’ll see there are maybe new media types we’ll play with that we will talk about later. But QLC is still very much our motion for this way of doing multi-tiers. So I think FlashArray is somewhat going to stay where it is and we just love it the way it is to be honest. There hasn’t really been demands to shift it.

Blocks & Files: Do you see the ability to have existing FlashArrays occupy the same namespace as EXA so the data on them can be included in EXA’s namespace?

Chadd Kenney: Yeah, really good question. I think what you’re going to start seeing from us, and this will be part of the Accelerate announcement as well, is Fusion actually is going to be taking the APIs across products now. 

So for Object as an example, you may or may not see Object show up on the alternate product in the future here. And so the APIs will actually be going through Fusion for Object and then you can declare where you want it to actually land based upon that. So you’ll start to see more abstraction that sits on top of the systems that will now dictate things like policy-driven provisioning and management where consumers very likely won’t even know if it lands on FlashArray or FlashBlade or not. For certain protocols. Obviously for Block, it’s likely going to land on a FlashArray, but they may not know which FlashArray it lands on. 

So some of the cool stuff I’m working on the platform side of the house is trying to take the telemetry data that we have, build intelligence for recommendations and then allow users to interact with those recommendations both through policy-driven automation but also through natural language processing with some of the Copilot capabilities we’re building.

And then inside the platform we’re starting to build these workflow automations and then recipes that sit outside of that to connect to the broader ecosystem. So we will see our platform thing come to life around Accelerate. I won’t steal too much thunder there, but that’s where we’re really trying to go. It’s like instead of actually trying to bridge namespaces together at the array side, we’re trying to build this all into Fusion to actually have Fusion dictate what actually lands on which system.

Blocks & Files: Let’s imagine it’s six months down the road and I’m a Hammerspace salesperson. I knock on the door of one of your customers who’s aware of all this stuff going on and say, I’ve got this brilliant data orchestration stuff, let me in and I’ll tell you about it. Would your customer then be able to say to the Hammerspace guy, sorry, I don’t need you. I’ve already got that.

Chadd Kenney: Yes. So one thing that we embedded into EXA that we didn’t really talk a lot about is we use FlexFiles as well. So data nodes can reside anywhere and those all can be part of the same namespace. And so we didn’t go too big on this initially because, if you think about the core use case, honestly most people are putting it all in the same datacenter right next to the GPUs. 

So I think over the longer term you’ll start to see more geo-distribution. With FlashBlade as a whole, that is a plan. And as we build pNFS into the core FlashBlade systems, you’ll start to see FlexFiles and multi-chassis be kind of stitched together as one common namespace. So I think we’ll have similar use cases. Of course we’re all-flash and they can use other media types out there. So they may get into use cases maybe that we’re not so tied to; ultra archive and those types of things. But I think as we get into 300 terabytes and then 600 terabytes and the incremental growth of capacity, we’ll start to break into those similar use cases as them.

Scale Computing picks Veeam as its bestie backup supplier

Edge, ROBO and mid-market hyperconverged infrastructure biz Scale Computing, is launching support for its Scale Computing Platform (SC//Platform) within the Veeam Data Platform.

Veeam now describes its backup and recovery software as a ‘platform’, saying it goes beyond backup and recovery by offering data management and cyber-resilience. The resilience relies on things like malware detection, backup immutability and incident response.

Jeff Ready.

Scale’s co-founder and CEO Jeff Ready claimed in a statement “The partnership between Scale Computing and Veeam delivers the best of both worlds: streamlined, autonomous IT infrastructure from Scale Computing and the industry’s most trusted data resilience platform from Veeam.”

Customers “have long asked for deeper integration with Veeam,” he added.

Scale’s HyperCore hypervisor scheme treats its software, servers and storage as an integrated system with self-healing features, and has its own snapshot capability with changed block tracking. Scale’s SW pools storage across tiers with its SCRIBE (Scale Computing Reliable Independent Block Engine). Snapshots use SCRIBE to create near-instant, point-in-time, consistent backups of VMs, capturing virtual disks, VM configuration, and, with Guest Agent Tools installed, the running OS and applications.

We envisage SCRIBE-based snapshots being used by Veeam as its basic unit of Scale backup, the entry point for Scale’s customer data into its data protection system.

Customers can, we’re told, choose from a range of Veeam-compatible backup targets, including object storage, tape, and cloud, and use full VM and granular file recovery from SC//Platform to any supported environment. They can also migrate and restore workloads between SC//Platform, VMware vSphere, Microsoft Hyper-V, and major public cloud environments.

Shiva Pillay.

Shiva Pillay, SVP and GM Americas at Veeam, said in a statement: “This collaboration with Scale Computing further strengthens Veeam’s mission to empower organizations to protect and ensure the availability of their data at all times and from anywhere, delivering cyber recovery and data portability across a purpose-built platform tailored for the unique needs of edge IT.”

Support for SC//Platform in the Veeam Data Platform is expected to be generally available in Q4 2025.

Live demonstrations of the Scale Computing solution will be featured at VeeamON 2025, taking place April 21-23, 2025 at booth #G5 in San Diego, CA. An online brochure about the two supplier’s integration can be accessed here.

Seagate claims spinning disks beat SSDs on carbon footprint

A Seagate Decarbonizing Data report says energy usage is a top concern for more than half of business leaders and better use of disks is a wise datacenter choice.

Jason Feist, Seagate
Jason Feist

The report cites a forecast by Goldman Sachs that global power demand from datacenters will increase by as much as 165 percent by 2030, compared with 2023. With this in mind, it says that rising data volumes, slowing power efficiency gains, and increasing AI adoption are putting pressure on organizations to manage carbon emissions, infrastructure expansion, and total cost of ownership (TCO) at the same time.

Jason Feist, SVP of cloud marketing at Seagate, says in the report: “Datacenters are under intense scrutiny – not only because they support modern AI workloads, but because they are becoming one of the most energy-intensive sectors of the digital economy. This calls for a fundamental shift in how we think about data infrastructure – not as a trade-off between cost and sustainability, but as an opportunity to optimize for both.” 

What does Seagate think a fundamental shift consists of? Getting rid of spinning disks? Not at all. The report includes a table showing the embodied carbon of the three main types of storage media – disk, SSD, and tape:

Seagate figures

It concludes:

  • SSDs have the highest embodied carbon, both in total and per TB, making it the most carbon-intensive option among the three storage media.
  • Hard drives exhibit the least carbon footprint, both in total and on a per-TB basis, offering the most carbon-efficient sustainable storage solution.
  • LTO tape shows moderate embodied carbon, but its annual impact is higher than that of hard drives.

The report suggests three strategic pillars for building a more sustainable data future:

  • Technological Innovation: Advances in computational power, storage areal density, and energy-efficient technologies like liquid/immersion cooling and HVAC systems can significantly lower energy consumption and carbon emissions, effectively managing the growing demand profile. 
  • Commitment to life cycle extension and circularity: Refurbishing, reusing, and maintaining storage equipment extends lifespan and reduces waste. Real-time environmental monitoring and transparent reporting can foster accountability across the datacenter environment. 
  • Share accountability across the ecosystem: Achieving meaningful emissions reduction – across Scopes 1, 2, and 3 as outlined in the report – requires collaboration across the entire value chain, including vendors, suppliers, and cloud service providers.
Seagate graphic
Decarbonizing Datareport graphic

An example of tech innovation, Seagate says, is its HAMR-based Mozaic 3+ disk technology, now in volume production. This enables up to three times more capacity in the same footprint than a 10 TB drive, and reduces a drive’s embodied carbon by more than 70 percent per terabyte. It also lowers cost per terabyte by 25 percent, according to IDC.

Feist says: “Sustainability cannot be solved in isolation. A holistic approach spanning infrastructure, life cycle management, and industry-wide accountability could ensure that the growth of AI and datacenter operations does not come at the expense of the environment.”

Download the Decarbonizing Data report here.

Bootnote

Seagate produces and sells its own Nytro SSD products for datacenter use. Flash drive-based vendor Pure Storage has a different point of view, as you might expect. It takes a system-level view rather than per-drive. Pure suggests that, for a 1 exabyte deployment over ten years with a five-year life cycle for the HDDs and a ten-year life cycle for DirectFlash Modules, it finds that the HDD system emits 107,984 metric tons of carbon whereas a Pure-based system emits 14,779 tons. You can find more details here.

Hammerspace secures $100M to chase AI-driven data growth

Data orchestrator Hammerspace has ingested $100 million in funding to accelerate its global expansion.

Up until two years ago, Hammerspace, founded in 2018, was financed by founder and CEO David Flynn along with cash from a group of high net worth individuals, and long-term investing sources. It raised $56.7 million in an A-round in 2023, saying it planned to expand its sales and marketing business infrastructure. Now it has raised another $100 million as it races to capitalize on anticipated demand for AI, with customers demanding widespread, rapid access to data – a challenge Hammerspace says it can solve through orchestration.

David Flynn, Hammerspace
David Flynn

Flynn stated: ‘AI isn’t waiting. The race isn’t just about raw throughput – it’s about how fast you can deploy, move data, and put your infrastructure to work. Every delay is unrealized potential and wasted investment. We built Hammerspace to eliminate friction, compress time-to-results, and significantly increase GPU utilization. That’s how our customers win.”

In June last year, Flynn was looking forward to Hammerspace becoming cash flow-positive and thinking about a possible IPO in the next 18 to 24 months. He said: “We are building a very financially disciplined company, but we are reaching a point where we have to grow a lot faster to get in front of the opportunities being created by AI.” But he cautioned: “It’s a tough market for an IPO right now, with budgets going down in some areas, and Nvidia sucking the air from everybody else at the moment. It was a tough time to envisage an IPO as Nvidia was so dominant in AI.”

Ten months later, Nvidia’s dominance has only grown, the AI surge continues, and unpredictable tariff shifts have made short-term business planning increasingly difficult. Hammerspace has decided to go for AI-fueled growth and needs fresh funding to achieve it.

This B-round was led by Altimeter Capital with participation from Cathie Wood’s ARK Invest, and a combination of new and existing investors. ARK is an existing investor, having participated in the 2023 A-round.

These investors have bought into Flynn’s view that the AI wave needs a unified data infrastructure that allows AI agents to query and process as much of an organization’s information assets as possible. That means the organization must have complete visibility and control over its data estate – regardless of protocol or location – all accessible through a single pane of glass. This is Hammerspace’s Global Data Environment.

Flynn said: “We didn’t build orchestration for the sake of it. We orchestrate data to the GPU faster regardless of where it is physically stored. We instantly assimilate data from third-party storage so it’s ready to process faster. We deploy and scale easily and quickly so our customers can achieve their business outcomes faster. 

We expect storage vendors like Arcitecta, Dell, DDN, NetApp, Pure Storage, VAST Data, and WEKA to vigorously challenge Hammerspace and fight hard for the enterprise AI data infrastructure business.

Pure debuts RC20 FlashArray//C with reused controllers for smaller deployments

Pure Storage has introduced a more affordable, low-capacity FlashArray//C model, the RC20, built with reconditioned controllers and aimed at edge deployments and smaller workloads.

The FlashArray//C series uses Pure’s proprietary DirectFlash Module (DFM) drives built from QLC flash chips. They are currently at the R4 (release 4) level with the new //RC20 added to the existing C50, C70, and C90 models, as the table indicates:

Pure’s reconditioned controller CPUs and PCIe bus plus the reduced drive drive bays mark out the RC20.

Shawn Hansen, Pure VP and GM of the FlashArray business unit, blogged about the new smaller FlashArray box, saying: “This capacity-optimized all-flash system delivers enterprise-class performance, reliability, and agility at capacities and prices that are more accessible for edge deployments and smaller workloads.”

It has a 148 TB entry-level capacity compared to the prior entry-level //C50’s 187 TB. This could consist of 2 x 75 TB DFMs, whereas the C50 uses 2 x 75 TB and 1 x 36 TB DFMs but Pure can mix and match drive capacities as it sees best.

Hansen says: “We heard loud and clear that SMBs and smaller deployments or scenarios with remote office/branch office (ROBO) were missing out. They were looking for the full capabilities of the Pure Storage platform but with smaller capacities and at a competitive price point.”

The “R” in the RC20 name denotes “the use of factory-renewed controllers, which in conjunction with other new parts, such as new chassis, delivers a product that provides many of the same benefits as other members of the FlashArray//C family, all while reducing e-waste and aligning with our commitment on delivering a sustainable data platform.”

The RC20 can be non-disruptively upgraded to larger products in the FlashArray//C family and/or to next-generation controllers in the future. A Pure chart depicts this, though it lacks numerical values on its axes:

Hansen is keen to show the low electricity needs of the RC20, as expressed in a table:

No pricing details were provided. Check out a FlashArray//RC 20 datasheet here.

Lightbits: record Q1 with rise in sales and deal sizes

Privately-owned Lightbits says it has broken growth records in the first 2025 quarter with a business surge – though we have no way of verifying this claim.

It supplies a disaggregated virtual SAN, block storage accessed by NVMe/TCP that runs either on-prem or in the Azure, AWS, or Oracle clouds, using ephemeral instances to deliver ultra-high performance.

There was, we’re told, a 4.8x increase in software sales, a 2.9x increase in average deal size, a jump in new customers, and a 2x year-over-year license increase. We have nothing to compare this to in terms of actual baseline numbers but it sounds impressive. New customers came from financial services, service providers, and e-commerce organizations with analytics, transactional, and mixed workload environments that need massive scale, high-performance, and low-latency access storage.

Eran Kirzner, Lightbits
Eran Kirzner

CEO and co-founder Eran Kirzner stated: “The quarter close marked significant progress financially and strategically … We now service Fortune 500 financial institutions, as well as some of the world’s largest e-Commerce platforms and AI cloud companies.” 

Lightbits cited AI cloud service providers Crusoe and Elastx and cloud company Nebul as recent customer wins. It says its storage software scales to hundreds of petabytes and delivers performance of up to 75 million IOPS and consistent sub-millisecond tail latency under a heavy load.

CRO Rex Manseau said: “We’re seeing a consistent pattern of engagement with customers finding that other software-defined storage can only accommodate low and middle-tier workloads. They adopt Lightbits for tier 1 workloads, and then we move downstream to their utility tier, as well. And customers seeking VMware alternatives like Lightbits for its seamless integrations with OpenShift and Kubernetes to enable their infrastructure modernization.” 

The intention from Lightbits management is “to expand its global install base, prioritizing key markets across the Americas and Europe, and other high-growth regions.” It has agreements coming this year as well as new products to enable a broader workload coverage area.

Lightbits has raised a total of $105.3 million in funding since being founded in 2016. The most recent C-round in 2022 brought in $42 million.

InfluxData rolls out InfluxDB 3 to power real-time apps at scale

InfluxData has released InfluxDB 3 Core and Enterprise editions in a bid to speed and simplify time series data processing.

InfluxDB 3 Core is an open source, high-speed, recent-data engine for real-time applications. According to the pitch, InfluxDB 3 Enterprise adds high availability with auto failover, multi-region durability, read replicas, enhanced security and scalability for production environments. Both products run in a single-node setup, and have a built-in Python processing engine “elevating InfluxDB from passive storage to an active intelligence engine for real-time data.” The engine brings data transformation, enrichment, and alerting directly into the database.

Paul Dix, InfluxData
Paul Dix

Founder and CTO Paul Dix claimed: “Time series data never stops, and managing it at scale has always come with trade-offs – performance, complexity, or cost. We rebuilt InfluxDB 3 from the ground up to remove those trade-offs. Core is open source, fast, and deploys in seconds, while Enterprise easily scales for production. Whether you’re running at the edge, in the cloud, or somewhere in between, InfluxDB 3 makes working with time series data faster, easier, and far more efficient than ever.”

A time-series database stores data, such as metrics, IoT and other sensor readings, logs, or financial ticks, indexed by time. It typically features high and continuous ingest rates, compression to reduce the space needed, old data expiration to save space as well, and fast, time-based queries looking at averages and sums over time periods. Examples include InfluxDB, Prometheus, and TimescaleDB.

Evan Kaplan, InfluxData
Evan Kaplan

InfluxData was founded in 2012 to build an open source, distributed time-series data platform. This is InfluxDB, which is used to collect, store, and analyze all time-series data at any scale and in real-time. CEO Evan Kaplan joined in 2016. The company raised around $800,000 in a 2013 seed round followed by a 2014 $8.1 million A-round, a 2016 $16 million B-round, a 2018 $35 million C-round, a 2019 $60 million D-round, and then a 2023 $51 million E-round accompanied by $30 million in debt financing.

Kaplan has maintained a regular cadence of product and partner developments:

  • January 2024 – InfluxDB achieved AWS Data and Analytics Competency status in the Data Analytics Platforms and NoSQL/New SQL categories.
  • January 2024 – MAN Energy Solutions integrated InfluxDB Cloud as the core of its MAN CEON cloud platform to help achieve fuel reductions in marine and power engines through the use of real-time data.
  • March 2024 – AWS announced Amazon Timestream for InfluxDB, a managed offering for AWS customers to run InfluxDB within the AWS console but without the overhead that comes with self-managing InfluxDB.
  • September 2024 – New InfluxDB 3.0 product suite features to simplify time series data management at scale, with performance improvements for query concurrency, scaling, and latency. The self-managed InfluxDB Clustered, deployed on Kubernetes, went GA, and featured decoupled, independently scalable ingest and query tiers.
  • February 2025 – InfluxData announced Amazon Timestream for InfluxDB Read Replicas to boost query performance, scalability, and reliability for enterprise-scale time series workloads.

The new InfluxDB 3 engine is written in Rust and built with Apache Arrow, DataFusion, Parquet, and Flight. We’re told it delivers “significant performance gains and architectural flexibility compared to previous open source versions of InfluxDB.” The engine can ingest millions of writes per second and query data in real-time with sub-10 ms lookups.

The Python engine “allows developers to transform, enrich, monitor, and alert on data as it streams in, turning the database into an active intelligence layer that processes data in motion – not just at rest – and in real-time.” This reduces if not eliminates the need for external ETL pipelines.

Both new products fit well with the existing InfluxDB 3 lineup, which is designed for large-scale, distributed workloads in dedicated cloud and Kubernetes environments and has a fully managed, multi-tenant, pay-as-you-go option.

InfluxDB 3 Core is now generally available as a free and open source download. InfluxDB 3 Enterprise is available for production deployments with flexible licensing options. Read more here.

NetApp joins NFL roster as official tech partner

Paul KunertEdit Profile

NetApp is sponsoring the US National Football League (NFL) becoming its Official Intelligent Data Infrastructure partner.

The NFL is a trade association and sports league consisting of 32 professional American football teams organized into four divisions each within the American Football Conference (AFC) and National Football Conference (NFC). NetApp has been raising its sports sponsorship game recently. In January, it became a founding-level partner of the San Francisco 49ers professional football team in a multi-year deal and is now its Official Intelligent Data Infrastructure Partner. The 49ers play in the NFC West division. NetApp also has sponsorship deals with the San Jose Sharks ice hockey team, Porsche Motorsport, TAG Heuer Porsche Formula E team, and the Formula 1 Aston Martin Aramco Racing team.

Cesar Cernuda, NetApp
Cesar Cernuda

Cesar Cernuda, president at NetApp, provided a statement: “The NFL has been a long-standing customer of NetApp, and it is gratifying to see their interest in expanding our relationship. By helping the NFL build an intelligent data infrastructure, we are elevating the sport for fans of football everywhere. Bringing intelligence to the league’s data storage allows them to modernize their infrastructure and gives them the tools they need to agilely adapt to the future.”

The NFL sports league says it has more than 218 million fans in the US with more overseas. We’re told: “NetApp will play a key role globally with the NFL, activating across the entire slate of 2025 International Games.” This will include being the presenting sponsor of the 2025 NFL London Games and the first-ever regular season NFL game in Madrid.

American football is the most-watched sport in the US, “with Super Bowls being among the highest-rated TV broadcasts in US history.” There were 37 league-level sponsors for the 2023-2024 season, including Best Buy, Campbell Soup, FedEx, Frito-Lay, Marriott, Rocket Mortgage, Subway, and Visa.

NetApp will provide storage technology to the NFL, with security mentioned, though few other details were provided. Renie Anderson, EVP and CRO at the NFL, said: “The global partnership between the NFL and NetApp represents a shared commitment to leveraging intelligent data to drive transformative solutions for the league. By combining our expertise with NetApp’s industry-leading intelligent data infrastructure, we can unlock new efficiencies that accelerate innovation within our game.”

Gary Brantley, chief information officer at the NFL, said: “Working with NetApp has helped us serve football fans worldwide by streamlining our technology operations and enabling us to be better custodians of our sport … Over the years we have worked together, NetApp has earned our trust that they can meet those high standards. With an intelligent data infrastructure powered by NetApp technology, we have a secure data storage strategy that can carry us into the future.”

How much is NetApp spending? According to Research and Markets, the average price of an annual NFL sponsorship deal for the 2023-2024 season was roughly $39 million with some deals at the $90 million-plus level. For example, EA Sports, Anheuser-Busch InBev, Nike, Verizon, and Microsoft were high-end sponsors. 

DAOS file system lives on after Optane

The DAOS parallel file system delivers higher IOPS and bandwidth per server than Lustre and WEKA, according to IO500 benchmark data cited by the DAOS Foundation.

DAOS, the Distributed Asynchronous Object Store parallel file system, originated at Intel in 2015 with a research project for the US Department of Energy. Intel evaluated Lustre, StorageScale (GPFS), PanFS, and BeeGFS and concluded it could develop a more advanced solution. DAOS provided an asynchronous I/O model, a distributed key-value store architecture, a user-space implementation to avoid Linux kernel bottlenecks, and was designed to have low latency, high bandwidth, and a high IOPS rating. DAOS was also architected to use Intel’s Optane storage-class memory for faster metadata handling. When Optane was canceled in summer 2022, much of the impetus behind it disappeared.

The software was released under an Apache 2.0 open source license in 2019 and used in the much-delayed Aurora supercomputer, built by Intel and Cray, at the Argonne National Laboratory. Intel and Cray won the deal in 2015 and it was meant to be up and running in 2018 but was actually installed in June 2023, after Optane had been canned, and went live in January 2025.

The DAOS Foundation was formed in November 2023 by Argonne National Lab, Enakta Labs, Google, HPE, and Intel – the main players invested in the software. The UK’s Enakta Labs offers a DAOS-based Enakta Data Platform HPC storage system. HPE bought Cray in May 2019. Google used DAOS in its 2024 Parallelstore HPC file system for AI, ML, and scientific computing, having started a DAOS collaboration with Intel in 2020. VDURA joined in 2024, seeing DAOS as having potential advantages over its long-lived PanFS parallel file system in the AI area.

The Leibniz Supercomputing Centre (LRZ) is also a DAOS user, deploying it in its SuperMUC-NG Phase 2 system in the Munich area in Germany, but LRZ is not a foundation member.

The DAOS Foundation wants to keep DAOS going as a vendor-independent open source project. The Foundation has a technical steering committee, chaired by Johann Lombardi, an HPE Senior Distinguished Technologist. Between 2013 and 2020, he was an HPC Software Architect and then Principal Engineer in Intel’s Extreme Storage Architecture and Development Unit. He is a long-term DAOS developer and advocate.

DAOS history

Lombardi presented a DAOS session to an IT Press Tour event outlining its architecture and how it recovered from Optane’s demise. There is a library of DAOS routines layered above two types of DAOS instance – a DAOS control plane and DAOS engine, which talks to the storage drives via an RDMA link.

DAOS instances

A protocol and middleware layer sits above libdaos and presents file, block, and object interfaces to client applications. This layer includes HPC I/O middleware, and AI and big data frameworks. All these execute as part of client compute instances.

DAOS is inherently multi-tenant and has a dataset concept in which tenants have one or more datasets, their basic unit of storage. Datasets are defined by their capacity, throughput, IOPS rating, and other qualities, such as their type. For example, POSIX-compliance, key-value store or Python. They are spread across drives with I/O handled by the data engines in each node. 

DAOS multi-tenancy

Datasets have access control lists and are a snapshot entity. POSIX datasets can include trillions of files/directories. A tenant dataset is viewed by DAOS as a container with a certain type, which is stored as a set of objects in its multi-level key-value store.

 

DAOS object interface

DAOS metadata and data operations, such as updates and inserts, were previously written to Optane storage-class memory in the particular DAOS node and thus persisted for recovery if a DAOS operation failed mid-transaction. Post-Optane, a Write Ahead Log (WAL) is held in DRAM and persisted to NVMe SSD storage. Data and metadata are written in parallel and can be written to a dedicated SSD or ones shared with normal DAOS storage duties. Checkpoint data goes to the metadata SSD as well. Lombardi said the pre and post-Optane performance was comparable. 

DAOS evolution

The Aurora DAOS system in Optane mode won the top IO500 Production Overall Score slot in 2023. It was an HPE Cray EX Supercomputer with more than 9,000 x86 Sapphire Rapids nodes using >230 PB of DAOS storage across 1,024 nodes with dual Slingshot NICs, delivering some 25 TBps of bandwidth. Each node has two Sapphire Rapids CPUs and six Xe-based Arc GPUs. There is between 220 and 249 PB capacity depending on the redundancy level chosen. 

DAOS won the top two slots in the SC 2024 production list; Argonne first with 300 client nodes and a score of 32,165.09, and LRZ second using 90-client nodes and scoring 2,508.85. An HPE Lustre system (2,080 nodes) was third, scoring 797.04, with a WEKA system (261 nodes) fourth rated at 665.49. A 2,000-node DDN EXAScaler/Lustre system was fifth at 2,000 nodes and scoring 648.96.

Lombardi presented a per-server chart to emphasize DAOS’s superiority:

DAOS is reported to be up to three times faster per server than Lustre or WEKA, based on IO500 benchmarks. What now? The DAOS roadmap looks like this:

DAOS v2.6 in July 2024 was the last Intel release. Version 2.8 will be the first community release and is slated for delivery later this year. A v3.0 release is being worked on for 2026 and there is subsequent version under development as well.

Comment

The DAOS situation is that it has a highly credible flagship win with its Aurora supercomputer user. DAOS is open source HPC/supercomputing software with AI training and inference aspirations needing developer adoption to grow. Existing HPC, supercomputing, and AI training/inference users have adopted other high-performance software, such as Storage Scale, Lustre, VAST Data, WEKA, BeeGFS, or ThinkParQ, or they are using hardware/software combinations like HPE Cray and DDN. They will likely not adopt DAOS, even though it’s open source, because of the conversion cost, unless there is some clear advantage, such as improved performance, lower cost, or freedom from vendor lock-in.

In the AI training and inference area, dominated by Nvidia GPU servers, most suppliers of storage hardware and software support the GPUDirect protocol. DAOS does not. DAOS does use RDMA, however, and a DAOS-backed storage system with RDMA-capable NICs could, in theory, support GPUDirect, enabling direct data paths.

In this area, object storage use is being promoted by Cloudian, DDN (Infinia), MinIO, Scality, and VAST Data, with the encouraging backdrop of Nvidia’s GPUDirect for object storage protocol.

Without GPUDirect support for either file or objects, DAOS faces an obstacle to its ambitions to get into any Nvidia GPU-using AI storage environment, despite having demonstrated IO500 performance advantages over Lustre and WEKA on a per-server basis.

We note that Google is not all-in with DAOS as it has a parallel HPC/AI file system track to its DAOS-based Parallelstore offering. This is a managed Google Cloud Managed Lustre service, powered by DDN. We also note that Intel has its own problems and DAOS promotion will be low on its list of priorities. 

Intel’s DAOS developers did a marvelous job recovering from the Optane fiasco and we have the impressive 2023 Aurora IO500 rating testifying to that. There are no publicly known DAOS adopters beyond the Foundation members and LRZ. DAOS faces an uphill climb to gain wider adoption, and developer adoption is the vital ingredient it needs to grow its market presence.

Developers can find out more at the links below:

Storage suppliers stare down Trump slump after tariff mayhem

Analysis. The latest round of tariff confusion – including a last-minute US Customs exemption for computers, smartphones, and related components, followed by a social media post from President Trump declaring the move temporary – is likely to trigger a slowdown in customer purchases.

The wild movements in tariffs on computer goods, including storage, make it impossible for businesses that rely on them to calculate stable costs. This means they won’t be able to assess how their market will react to the changed prices for the products and services they offer or, indeed, whether particular products and services are worth supplying at all.

The timeline of the recent tariff changes looks like this:

  • April 2 – President Trump announced tariffs on all imported goods, but Executive Order 14257 said that semiconductors were exempt from tariffs, with a link to a 37-page Annex [PDF] listing non-tariffed items.
  • April 7 – A 54 percent tariff on goods from China.
  • April 9 – A 145 percent tariff on goods from China.
  • April 11 – US Customs and Border Protection published “further guidance on the additional duties,” listing 20 Harmonized Tariff Schedule of the United States (HTSUS) codes exempt from tariffs, identifying 8471 and 8471.30.01, which describe computers and laptops, plus 8517.13, which covers smartphones, and 8542, whose subheadings cover CPUs, GPUs, systems on chips, microcontrollers, and memory.
  • April 13 – President Trump declared that the tariff exemptions mentioned on April 11 were temporary as “we are taking a look at Semiconductors and the WHOLE ELECTRONICS SUPPLY CHAIN in the upcoming National Security Tariff Investigations.”

Commerce Secretary Howard Lutnick said on ABC News on April 13 that computers and smartphones would be included in this National Security Tariff Investigation: “They’re going to have a special focused type of tariff to make sure that those products get reshored.”

US buyers of hard disk drives and SSDs, and products like arrays that contain them, are facing cost uncertainty. This makes pricing calculations for HDD and SSD-based storage products and services difficult, if not impossible. It will encourage them to pause sales of such products and services until they have tariff stability and can calculate how to price them and even whether to offer them at all. A net effect could then be a slowdown in orders for HDDs and SSDs, which would depress the revenues of suppliers, such as Kingston, Kioxia, Micron, Phison, Samsung, Seagate, SK hynix, Toshiba and Western Digital, in the current quarter. 

Daniel Ives, Managing Director at Wedbush Securities, commented on X (formerly Twitter) about “the mass confusion created by the constant news flow out from the White House” and asked: “How can companies give guidance in this environment?”

A Trump slump in storage supplier revenues is looking to be a potential outcome of all this tariff turbulence.

Storadera touts low-cost S3 storage hosted entirely in the EU

Estonia-based Storadera’s pitch is simple: low-cost, single-tier S3 storage in the cloud using optimized disk drive storage with higher-capacity shingled drives and AI-improved storage operations on its roadmap.

CEO, founder, and ex-software engineer Tommi Kannisto says he was inspired by US storage biz Backblaze and thought he could offer disk drive-based S3 object storage in Europe. Storadera was incorporated in 2019, went live in Tallinn, Estonia, in 2021 and opened a second datacenter in the Netherlands in 2023.

Storadera offers its single tier S3 storage for €6/TB/month – versus Backblaze’s base price of $6/TB/month – with no additional fees. When comparing starting prices in euros, Backblaze is the lowest-cost supplier as its $6/TB/month converts to €4.75. However, Storadera is based in Europe and its stored data is beyond the direct jurisdiction of non-EU countries. 

Storadera costs marginally less than Wasabi and significantly less than AWS, which also charges data egress fees.

Tommi Kannisto, Storadera
Tommi Kannisto

Kannisto told an IT Press Tour event in the UK that Storadera makes slightly less than €1 million a year in revenue: “We are profitable … we make a very good profit [and] we’re growing 5 percent/month in revenue.”

He said Storadera is “compatible with hundreds of S3-compatible applications. Implementing Storadera with backup tools like Veeam enables [customers] to send the secondary copy to the cloud and increase their backup retention. There is no need to buy new tools to manage cloud storage.”

It has around 100 customers, including the Estonian government and telco Telia. Reseller partners account for about half of its stored data and around 50 percent of its data is coming in from Veeam, so most data is backed up.

Data is stored in standard hard disk drives racked in JBODs, containing 102 drives. Each JBOD is attached to a server with 32 GB of RAM. Services, coming from 100,000 lines of GO code, run in the servers. Kannisto said it’s a hyperconverged setup: “All software runs in all servers and all servers write to all JBODs. There is no load balancer unit.” 

Storadera hardware setup
Storadera hardware setup

The datacenters are autonomous and have bucket geo-replication.

Kannisto buys standard conventional Western Digital drives, with 26 TB ones in the Netherlands datacenter, and said shingled magnetic recording (SMR) media drives are on his roadmap due to their extra capacity with disk write optimizations possible as “most of our load is writing files.” Kannisto reckons “SMR will reduce our capex by 25 percent.”

Asked about using QLC SSDs, he said: “QLC 100-plus TB SSDs are still too expensive – and probably will be for the next ten years.””” He’s blogged on the SSD vs HDD issue, writing: “The price difference can be easily over 10x in favor of hard disks. If we can offer fast enough service on 10x less expensive hardware, then it sounds like magic.”

He also notes: “SSDs have a limited amount of terabytes written (TBW). After that, they just become read-only. Hard disks don’t have such strict limitations.”

Storadera uses variable block sizes in its writes. Small blocks are used at times of low load with bigger blocks used at high load times, made up from batches of small file writes. The system uses 4+2, 6+2, and upcoming 8+2 erasure coding schemes, with chunks striped across servers. All data is encrypted and object locking is used for immutability. Every 60 days, data integrity is checked for bit rot.

The system can handle “close to 300 MBps with 2 MB files,” and it’s easy, with steady streams of data, to write fast to HDD. SSDs are not needed in terms of write speed although metadata is stored on SSDs, accounting for about 0.05 percent of disk capacity.

Kannisto said Storadera is looking to expand its regional coverage, with a German datacenter coming on stream in the middle of this year, followed by further expansion to the UK, then the US or Canada, and APAC. There are no timelines for this.

Possible technology roadmap items include the use of AI to improve storage operations and the concept of smart caching to provide edge access.

Storage news ticker – April 11

Screenshot

Cloud storage and backup provider Backblaze announced a partnership with PureNodal, “a next-generation cloud platform purpose-built for AI and high-performance compute (HPC) workloads,” to deliver an efficient, high-performance, and cost-effective service for enterprise-scale workloads such as AI model training, data analytics, and media rendering. By combining PureNodal’s AI-optimized compute platform—including cloud fabric, low-latency software-defined networking (SDN), and high-performance infrastructure—with Backblaze’s B2 Cloud Storage, the partnership delivers a powerful, scalable foundation for AI and HPC workloads without the complexity or cost of legacy cloud solutions.”

Blackmagic Design’s Cloud Backup 8 is a new rackmount media backup product using HDDs which syncs to Blackmagic Cloud for backing up collaborative projects in DaVinci Resolve. It has 8 independent hard disk slots, 4 x 10G Ethernet ports, and an HDMI monitor output that shows storage status. The storage map shows a graphic representation of the total hard disk capacity and real time read and write access happening for the connected users. There are also graphs that show customers the data transfers on the Ethernet connection. The vendor says it’s “perfect for archiving completed projects to low cost storage” and costs $1,495 from Blackmagic Design resellers.

Connectivity supplier Cloudflare announced the industry’s first remote Model Context Protocol (MCP) server, generally available access to durable Workflows, and a free tier for Durable Objects. “These offerings enable developers to build agents in minutes, rather than months, simply, affordably, and at scale.” MCP servers built on Cloudflare can retain context, providing a persistent, ongoing experience for each user. Read more in a blog.

Cloudflare launched Cloudflare Realtime, a suite of products to help developers build real-time audio, video and AI applications, easily and at scale. Dyte.io, a leading company in  the real-time ecosystem, joined Cloudflare to help create RealtimeKit – a toolkit within Realtime that includes SDKs, server-side services, and UI components to simplify building real-time voice, video, and AI applications. RealtimeKit reduces development time from months to days, offers built-in recording capabilities enables true AI-human conversations with low latency (tens of milliseconds) and scales to millions of concurrent participants.

Cloudian CMO Jon Toor said about the Trump tariffs: “While we can’t address the macro picture, we believe that our software-defined model provides greater hardware sourcing flexibility. And for those who prefer an appliance, we offer the advantage that we’ve stockpiled some kit that will continue to be offered at pre-tariff prices while inventory lasts.  On-prem offers the advantage of offering data sovereignty, as well as being inherently lower cost when demand is continuous (as it usually is in storage).” There’s more in a blog

Cloud monitoring business Datadog has expanded monitoring capabilities  for BigQuery. Now in preview, the >35 extensions help users view BigQuery usage by user and project to identify those incurring the most spend, pinpoint the long-running queries in those segments to optimize, and detect data quality issues. BigQuery is Google Cloud’s fully managed and serverless enterprise data warehouse.

Search company Elastic announced Elasticsearch runs with up to 40 percent higher indexing throughput on C4A VMs, powered by Google Axion, Google’s first custom Arm-based CPU, compared to previous generations of VMs on Google Cloud. Elastic used a macro benchmarking framework for Rally with the elastic/logs track to determine the maximum indexing performance on Google Axion-powered VMs. C4A also powers Elastic Cloud Serverless.

Fivetran introduced its Managed Data Lake Service for Google’s Cloud Storage. It seamlessly ingests data from 700+ connectors into Google Cloud Storage, automates conversion to open table formats like Iceberg and Lake for flexibility and interoperability. The Google Cloud integration enables native support for governance, compliance, and BigQuery. It reduces compute costs while accelerating data pipelines into Google Cloud. More details here.

A Gartner Research paper says businesses and other oganizations should stop buying storage products and buy storage platforms instead. Its summary says: To improve modernization outcomes, I&O leaders must transition from traditional storage infrastructure to infrastructure platforms. This research will help them evaluate their existing infrastructure and explore the requirements, components and selection process for new enterprise storage platforms.”

Google Cloud announced: 

  • Anywhere Cache: the industry’s first consistent zonal cache that works with existing regional buckets to cache data within a specific zone. Store once and access anywhere across a continent, enabling faster data access for AI workloads. Google is seeing up to 2.5TBps per zone and 70 percent lower latency.
  • Rapid Storage: A way to concentrate data in the same zone that GPUs and TPUs run in, with the ability to mount a bucket as a file system to allow AI frameworks to access object storage more seamlessly. Powered by Google’s internal Colossus file system and its stateful protocol – and now directly available to Google Cloud customers. It offers <1ms random read and write latency, 20x faster data access, 6 TBps of throughput, and 5x lower latency for random reads and writes compared to other leading hyperscalers.
  • Storage Intelligence: A feature providing binsights and management capabilities for storage estates at scale, including AI-powered metadata analysis and cost optimization recommendations. It offers metadata analysis at scale and intelligent cost optimization recommendations, making Google Cloud the first hyperscaler to provide such sophisticated, environment-specific storage insights.

Hammerspace has won Buck, “the award-winning creative agency behind 2D/3D animation, branding, and immersive experiences for brands like Airbnb, Apple, FedEx, Microsoft, GitHub, Amazon, and Rivian,” as a client. It’s now operating as one unified studio across facilities in LA, NY, Amsterdam and Sydney for more than 800 artists, “enabling artists and technologists across continents to work in the same live file system – no more wrangling versions, waiting on syncs or duplicating media.” 

HighPoint says its PCIe Gen5 and Gen4 NVMe Switch AICs and adapters can help PCIe expansion and M.2 acceleration, offering bandwidth, flexibility, and scalability. Designed to host multiple PCIe devices, including GPUs and M.2 accelerators, the solutions are aimed at helping businesses to overcome PCIe resource limitations, optimize infrastructure, and deliver next-generation performance for datacenters, edge computing, and high-performance environments. Details here.

Reuters says Huawei’s 2024 revenues rose 22.4 percent, the fastest rate in five years,  to 862.1 billion yuan ($117.9 billion) but its profit declined 28 percent Y/Y to  62.6 billion yuan ($8.63 billion) due to heavy R&D investment.

Malware threat detector Index Engines launched the latest version of CyberSense, v8.10. It’s fully integrated with Dell PowerProtect Cyber Recovery and provides an industry-first raw disk corruption  detection, advanced threat analysis, and seamless integration to fortify cyber resilience. It says that, with >1,500 global installations, “CyberSense’s highly-trained AI ensures data integrity, empowering organizations to detect corruption from cyber threats and recover with confidence.” Custom YARA rules within CyberSense support detection of patterns in files, allowing it to identify even zero-day ransomware that hasn’t been seen before. 

CyberSense has Rapid Threat Detection with Delta Block Analysis. It has ensures seamless integration with leading security and backup solutions, including PowerProtect Data Manager 19.18 & 19.19; Avamar 19.12, NetWorker  19.12; Commvault Backup and Recovery 11.36; Cohesity NetBackup 10.5, including NetBackup OST (Open Storage Technology); and Oracle ASM RMAN. CyberSense is available now through Dell Technologies and its global partner network. 

Cloud data management supplier Informatica announced its CLAIRE Copilot for data integration and CLAIRE Copilot for iPaaS are both in preview. It has GenAI Recipes for application integration for Amazon  Bedrock, Azure OpenAI, Databricks Mosaic AI, Google Cloud Vertex AI and Gemini, Salesforce Pega GenAI, ServiceNow Generative AI, and Oracle Select AI among others. There are new AI-powered  chunking, embedding and PDF-parsing capabilities for unstructured data. 

Informatica MDM SaaS now integrates with CLAIRE GPT, enabling NLP-based search and metadata exploration of business entities and associated attributes within master CLAIRE-generated glossary descriptions. GenAI-Powered Natural Language for Data Marketplace allows users to explore data marketplaces through conversational queries, simplifying data discovery and enhancing accessibility. 

Informatica added support for the Databricks Data Intelligence Platform, API Center and Cloud Data Access Management (CDAM) services running natively on Google Cloud to its Intelligent Data Management Cloud (IDMC) platform. It said the full support for Databricks has300+ connectors for data ingestion, no-code data pipelines running natively within Databricks for data preparation and transformation, data quality and profiling for data within the Databricks Data Intelligence platform, and enterprise-wide data governance with seamless integration with Unity Catalog. 

Kingston has launched a DC3000ME PCIe 5.0 NVMe U.2 SSD using 3D TLC NAND for server applications. It features up to to 2,800,000 random read IOPS, 14 GBps sequential reads, latencies of <10µs read and <70µs write with endurance of 1 DWPD for 5 years and 2 million hrs MTBF. The drive is tuned for strict QoS (99.9999 percent) and has AES-256, TCG OPal 2.0,  onboard power loss protection and 128 namespaces are supported.

Kioxia, AIO Core and Kyocera ave developed an broadband optically connected PCIe gen5 SSD prototype. It combines Kioxia NAND and SSD with AIO Core’s IOCore optical transceiver and Kyocera’s OPTINITY optoelectronic integration module technologies. Kioxia says by replacing the electrical wiring interface with optical and utilizing broadband optical SSD technology it significantly increases the physical distance between the compute and storage devices, while maintaining energy efficiency and high signal quality. 

Pete Paisley, VP Business Development at MagStor, said about IBM’s tape drives and Trump’s tariffs: “I’m highly certain we all get from Japan or China and will be figuring out how to process the additional costs soon. I’d not heard previously that IBM is performing a value added step in AZ prior to your article, regardless as to whether they import at cost or wholesale value it will be a significant impact to costs for all I’m afraid. MagStor is small and nimble and is working on advantageous solutions in the meantime.”

… 

Large file transfer supplier MASV announcing upcoming integration with Adobe’s Frame.io V4. Ti s will enable faster, easier scalable capture-to-cloud workflows, automating repetitive manual tasks and accelerating production timelines for large-scale media enterprise teams by automating the transfer and organization of massive files. The MASV-Frame.io V4 integration will be available to all MASV users later in 2025.

Gene Villeneuve

MASV has hired Gene Villeneuve as its CRO. His experience spans roles at eyko, Tehama, Cognos, OLAP@Work, Business Objects, and IBM. Villeneuve’s deep expertise in building strategic partner ecosystems—including referral networks, resellers, integrators, and value-added resellers (VARs)—will be pivotal in fueling MASV’s momentum across media and entertainment, while spearheading its expansion into new enterprise markets and global territories.

Mainframe app to public cloud migrator Mechanical Orchard launched its Imogen end-to-end mainframe modernization platform, and a partnership with global technology consultancy, Thoughtworks. Imogen rewrites mainframe applications safely and quickly by focusing on the behavior of the system as represented by data flows, rather than on translating code. Using data capture agents, an incremental approach and a patented method of creating deterministic outcomes from generative AI, Imogen helps organizations create a new technology foundation that it says is easy to change and forward-compatible. It uses real-time data flows to rapidly validate code generated by AI so that it’s accurate, well-factored and ready to deploy to the cloud. Find out more here.

A joint offering with Thoughtworks, AI-Accelerated Mainframe Modernization with Mechanical Orchard, “takes a holistic approach to legacy modernization, focusing on system behavior rather than just code, so clients can modernize what matters.” 

Cloud file services supplier Nasuni announced findings from its industry research report, The Era of Hybrid Cloud Storage 2025, based on a survey of 1,000 purchasing decision makers across the US, UK, France, and DACH regions. While AI investment is the top spending priority for nearly half  of businesses, only one-fifth of surveyed companies feel confident their data is AI-ready. Nasuni is pitching that it can get that data AI-ready.

Immutable object storage Veeam backup target supplier Object First has run research finding that 81 percent of IT professionals say immutable backup storage built on Zero Trust principles is the best defense against ransomware, and 54 percent view target backup appliances as more secure than integrated appliances.

According to TechCrunchOpenAI CEO Sam Altman said that OpenAI will add support for Anthropic’s Model Context Protocol, or MCP, across its products, including the desktop app for ChatGPT. MCP is an open source standard that helps AI models produce better, more relevant responses to certain queries.

Perforce has released the 2025 State of Open Source Report  in collaboration with the Eclipse Foundation and the Open Source Initiative  (OSI). The report covers open source investment, culture, compliance risks, and other areas  that hinder further open source usage. It’s available for free download at: The 2025 State of Open Source Report or here.

Proxmox announced Backup Server version 3.4. It has optimized performance for garbage collection, granular backup snapshot selection for sync jobs, static build of Proxmox Backup client, and increased throughput for tape backup. This version is based on Debian 12.10 (“Bookworm”), but uses a newer Linux kernel 6.8.12-9 and includes ZFS 2.2.7 (with compatibility patches for Kernel 6.14). While the kernel 6.8.12-9 is the stable default, Linux kernel 6.14 can optionally be installed for better support of the latest hardware. Proxmox Backup Server seamlessly integrates into Proxmox Virtual Environment—users just need to add a datastore of Proxmox Backup Server as a new storage backup target to Proxmox VE. Proxmox Backup Server 3.4 is now available for download.

CRN reports Pure Storage is planning to reduce special pricing discounts which could mean an overall increase in hardware prices by approximately 10 percent. Services and software prices are not slated to change.

Japan’s Rakuten Symphony announced the addition of object storage to the Rakuten Cloud-Native Storage Platform. “Kubernetes-ready, highly scalable and cost effective, object storage brings greater durability, flexibility and simplification for managing large volumes of data in use cases including data lakes, cloud-native storage, and machine learning. Object storage will be generally available to any Rakuten Cloud-Native Storage customer in May 2025.”

Reltio and Microsoft Fabric have a new, joint product: Reltio Zero Copy Integration with Microsoft Fabric. It allows Azure customers to seamlessly store, manage, and share data on OneLake, using Reltio Data Cloud’s flexible data governance.  Reltio provides Fabric with reliable, real-time customer data, driving business teams with comprehensive  analytics and better enabling AI tools. More information here.

Scality object storage software is being used by document storage business Iron Mountain. The Iron Cloud Object Storage offering “focuses on data archiving and backup, integrating with existing processes to provide a cost-effective cloud-based storage option that protects against accidental deletion and offers rapid recovery, even down to individual file levels. To improve cyber resiliency, Iron Cloud Object Storage offers an offline, air-gapped secure copy of data, ensuring recovery from ransomware attacks and protecting against potential threats that may appear after restoration.” More info here.

HCI Vendor StorMagic has promoted Susan Odle to CEO from Chief Growth Officer. Odle joined StorMagic in August 2024; it’s been a meteoric 8 months. The prior CEO Daniel Beer will remain as a non-executive member of StorMagic’s board of directors. Before joining StorMagic, Odle served as founder and CEO of 8020CS, COO of BDO Lixar, and VP of operations at both GFI Software and Youi.TV. Additionally, she was honoured as one of Top 50 Women in SaaS by The Software Report in 2020. Odle is based in Ottawa, Canada. 

China’s TerraMaster unveiled  its “revolutionary product, the D9-320, a 9-bay intelligent storage enclosure. As a flagship model featuring an innovative “independent drive power control system,” the D9-320 offers a massive 198TB storage capacity, 10Gbps high-speed transmission, and precise nine-channel power management, providing professional-grade storage expansion solutions for small and medium-sized businesses, NAS users, and tech enthusiasts.” The D9-320 supports nine drives in single-disk mode, with a maximum single-drive capacity of 22TB. More details here.

TerraMaster D9-320.

Tessell, a startup offering a multi-cloud database-as-a-service (DBaaS) “that enables enterprises and startups to accelerate database, data, and application modernization journeys at scale, today announced its $60 million Series B funding round, bringing total funding to $94 million. The round was led by WestBridge Capital, with continued strong participation from Lightspeed Venture Partners and new investments from B37 Ventures and Rocketship.vc. This capital will accelerate Tessell’s go-to-market expansion and fuel research and development in AI-powered data management within the evolving enterprise data ecosystem.”

Trump’s tariffs have received a 90 day pause and storage suppliers are affected:

  • China – 145 percent – This affects Sandisk SSDs, Seagate HDDs, Solidigm SSDs, Toshiba HDDs. Recordable Blu-ray/DVD. 
  • India – 10 percent – Recordable Blu-ray/DVD.
  • Japan – 10 percent – Kioxia SSDs, Sandisk SSDs, Sony tape, Toshiba HDDs, Recordable Blu-ray/DVD. Hitachi Vantara storage arrays.
  • Malaysia – 10 percent – Seagate HDDs, Western Digital HDDS.
  • Philippines – 10 percent – Toshiba HDDs.
  • South Korea – 10 percent – Samsung SSDs, SK hynix SSDs
  • Taiwan – 10 percent – Recordable Blu-ray/DVD.
  • Thailand – 10 percent – Seagate HDDs, Western Digital HDDs.

China just raised its import tariffs to 125 percent on US products which will affect US-based storage product suppliers shipping into China. In effect Chinese suppliers will stop shipping into the USA and US suppliers will stop shipping to China until and unless these tariff walls are torn (negotiated) down.

As reported by IT Home, China’s UNIS (Unisplendour – owned by Tsinghua Unigroup) has launched its S5 (1 and 2TB), and the S5 Ultra (2 and 4TB) M.2 2280 SSDs using TLC 3D-NAND and a PCIe gen5 x 4 interface. The S5 is DRAM-less, reads up to 14.9GBps and writes up to 12.9 GBps. Its random read/write IOPS are 1,800,000/1,700,000. The S5 Ultra reads up to 14.2GBps and writes at 13.4GBps and its random read/write IOPS mumbers are 1,700,000/1,600,000. Its endurance is 1200TBW at 2TB and 2400TBW at 4TB.

… 

VMware alternative and virtualized datacenter supplier VergeIO announced the availability of ioMetrics, a Prometheus-compatible monitoring system purpose-built to provide granular observability across VergeOS environments. It gives customers detailed insight into the storage, compute and networking layers – vSAN tier monitoring, cluster insights, node and hardware metrics – “streamlining performance management, capacity planning, and proactive alerting.” It’s designed for integration with observability platforms like Grafana, LogicMonitor, and Dynatrace. 

Open-source distributed SQL database supplier Yugabyte announced the first of its next-generation agentic AI apps, Performance Advisor for YugabyteDB Aeon, its software-as-a-service (SaaS) offering. Yugabyte also announced an extensible indexing framework designed to support the seamless integration of state-of-the-art vector indexing libraries and algorithms, augmenting the capabilities offered by pgvector. The Performance Advisor agentic AI application allows developers to detect potential issues before an application is deployed and offers timely insights to SREs, and platform engineers to help with performance optimization.