Home Blog Page 28

NetApp offloads Spot, Cloudcheckr to Flexera

NetApp is selling its Spot and CloudCheckr CloudOps portfolios to FinOps business Flexera for a reported $100 million.

Spot provides ways to find and use the lowest cost public cloud instances. CloudCheckr is a public cloud cost-optimization facility. Flexera says the combination of Spot’s AI and ML-enabled technology and its own hybrid cloud expertise will provide a comprehensive FinOps offering that enhances financial accountability and efficiency in cloud operations. It also fits in FinOps’ expanding scope, “which now includes datacenters, SaaS applications, and public cloud,” plus software licensing and sustainability.

Jim Ryan, Flexera
Jim Ryan

Flexera CEO and President Jim Ryan stated: “A tsunami of artificial intelligence applications is exponentially increasing organizations’ consumption of cloud resources. Yet we hear from many organizations about the difficulty in answering basic questions like ‘what technology services do we use?’ and ‘why are our cloud bills so high?’ Our acquisition of Spot is the next step in Flexera’s strategic plan to provide organizations with a full panorama of their technology spend and risk across the entire IT ecosystem. We want to make it easy for anyone to find and analyze any data related to spend and risk for any kind of technology, anywhere it lives.”

Haiyan Song, NetApp EVP for Intelligent Operations Services, said: “This decision reflects NetApp’s sharpened focus and underscores our commitment to intelligent data infrastructure and long-term growth opportunities. After a thorough evaluation, it is clear that Flexera’s established expertise and global reach provide the ideal environment for the Spot business to thrive and scale. This move not only allows the Spot team and portfolio to unlock their full potential within Flexera’s ecosystem but also reinforces our dedication to driving value creation and achieving our ambitious growth objectives.”

Haiyan Song, NetApp
Haiyan Song

The Spot portfolio was not storage-focused so fell outside NetApp’s core business. Song said in a blog: “This move will enable us to further our focus on core business areas, aligning the mission of intelligent operations services (also known as CloudOps) to our primary focus on intelligent data infrastructure … Our strategy is to align our CloudOps portfolio with our core storage offerings and focus on delivering the value of intelligent operations services to our customers.”

She revealed: “Many employees currently working on the Spot portfolio are anticipated to join Flexera … We are dedicated to facilitating a smooth transition for all affected employees, providing comprehensive support, transparent communication, and transition assistance.”

Flexera says the acquired software will “provide continuous automation, optimization, and insights on an organization’s cloud infrastructure and applications. The acquisition will add new capabilities such as Kubernetes cost management and commitment management to Flexera’s industry-leading FinOps portfolio. With the acquisition, Flexera will also create a richer ecosystem of FinOps Managed Service Providers (MSPs) to serve customers’ evolving needs and bring new DevOps users into its robust customer community.”

Putting NetApp on the Spot

Financial details of the transaction were not disclosed, but Bloomberg reports that Thoma Bravo-owned Flexera is paying around $100 million for NetApp’s Spot portfolio. 

NetApp bought Israeli startup Spot.io in June 2020 for an undisclosed sum. The price was said to be $450 million, according to the Calcalist, an Israeli business mag. This enabled NetApp to offer containerized app deployment services based on seeking out the lowest cost or spot compute instances.

In October 2021, NetApp acquired CloudCheckr and its cost-optimizing public cloud management CMx platform to expand its Spot by NetApp CloudOps offering. Again, financial details of the transaction were not disclosed. As CloudCheckr’s total funding was $67.4 million, it had been growing fast, and was backed by private equity, thus we expected a $200 million to $300 million acquisition cost range.

This suggests that the total acquisition cost for Spot and CloudCheckr was in the $650 million to $750 million area, far less than the $100 million Flexera is paying.

Spot and CloudCheckr were part of NetApp’s public cloud business, which accounted for 10 percent of its overall revenues in its latest quarter

NetApp cloud revenues

By selling the Spot and CloudCheckr parts of it CloudOps business for $100 million, NetApp will forego their revenues in future and could take a revenue hit in its public cloud business in the next quarter or two.

Wedbush analyst Matt Bryson tells subscribers: “While at one point NetApp had larger ambitions around its cloud overlay offerings with NetApp acquiring numerous companies to form what looked like a full stack of cloud management tools, the past couple of years have seen NetApp reduce its focus on non-storage related cloud businesses. Net, we see the divestment of these non-core businesses as in line with this strategic shift.”

He added: “We believe NTAP’s revised strategy is serving the company well as our conversations suggest it continues to compete effectively vs peers with its core array offerings including making inroads into the block market (in line with results the past few quarters); though we do see some risk of margin compression this quarter and next as favorable eSSD pricing dissipates.”

This Flexera-NetApp transaction is subject to customary closing conditions, including the receipt of required regulatory approvals.

Informatica deepens Databricks dealings

Data integration and management supplier Informatica has strengthened the integration of its Intelligent Data Management Cloud (IDMC) with Databricks’ Data Intelligence Platform, including support for AI Functions.

Databricks supplies an intelligent data warehouse and is growing its business at a furious rate as the generative AI boom pulls in more and more data to be processed. It raised $10 billion in funding late last year and has just augmented that with another $5 billion in debt financing loans, bringing its total funding to $19 billion. Informatica’s Extract, Transform and Load (ETL) and data management and governance offerings help get high-quality data ingested into Databricks for AI training and inference.

Amit Walia, Informatica
Amit Walia

Informatica CEO Amit Walia stated: “We are seeing phenomenal success with our Databricks-related business, with rapid growth and delivering impactful business outcomes for customers such as Takeda, KPMG, and Point72 to name just a few.”

He said: “One of our key priorities while partnering with Databricks is empowering customers to build enterprise-grade GenAI applications. These applications leverage high-quality, trusted enterprise data to provide high-impact GenAI applications with rich business context and deep industry semantic understanding while adhering to enterprise data governance policies.” 

Adam Conway, Databricks SVP of Products, added: “As a leader in cloud-native, AI-powered data management, Informatica is a key partner of ours, supporting everything from data integration and transformation to data quality, governance, and protection.” 

Databricks AI Functions are built-in SQL operations that allow customers to apply AI directly to their data. Informatica’s Native SQL ELT supports Databricks AI Functions through no-code data pipelines, opening Databricks GenAI capabilities to no-code users. Databricks’ AI Functions enable customers to use GenAI capabilities, including sentiment analysis, similarity matching, summary generation, translation and grammar correction on customer data directly from SQL.

The new Informatica Native SQL ELT for Databricks makes it possible to “push down” data pipelines with 50-plus out-of-the-box transformations and support for more than 250 native Databricks SQL functions.

Informatica graphic

In June last year, Informatica integrated its AI-powered IDMC into the Databricks Data Intelligence Platform. Informatica’s GenAI Solution Blueprint for Databricks DBRX provided a roadmap for customers to develop retrieval-augmented generation (RAG) GenAI applications using Databricks DBRX. Native Databricks SQL ELT enables joint customers to perform in-database transformations with full push-down capabilities on Databricks SQL.

Informatica’s CDI-Free offering on Databricks Partner Connect gives customers access to Informatica’s cloud data ingestion and transformation capabilities. Its IDMC platform was validated with the Databricks Unity Catalog.

Altogether, the Informatica IDMC platform includes multiple Databricks-optimized features, such as 300-plus data connectors, the ability to create low-code/no-code data pipelines, data ingestion and replication, and GenAI-driven automation via Informatica’s CLAIRE GPT and CLAIRE copilot offerings.

In Informatica’s third fiscal quarter of 2025, revenues increased 3.4 percent year-over-year to $422.5 million. It surpassed 101 trillion processed cloud transactions per month, with Walia saying: “This accomplishment reflects our commitment to product innovation, customer-centricity, and our goal of being the Switzerland of data and AI. We see great momentum in AI-powered data management use cases.”

VAST enters 2025 with consolidated EBox and GCP support

Analysis: Having erected a substantial AI-focused stack of software components in 2024, announced a partnership with Cisco, and delivered software upgrades, how is VAST Data positioned at the start of 2025?

It’s entering the year with the latest v5.2 Data Platform software release, which features its EBox functionality, first previewed in March 2024. VAST’s basic DASE (Disaggregated Shared Everything) software has a cluster architecture “that eliminates any communication or interdependencies between the machines that run the logic of the system.” It features compute (CNodes) liaising with data box storage enclosures (DNodes) across an internal NVMe fabric. 

The DNodes are just boxes of flash (JBOFs) housing NVMe SSDs. These highly available DNodes store the DASE Cluster system state. VAST says: CNodes “run all the software and DBoxes … hold all the storage media, and system state. This enables the cluster compute resources to be scaled independently from storage capacity across a commodity datacenter network.”

Howard Marks, VAST Data
Howard Marks

The EBox idea colocates a CNode and DNode in one server box, thus preventing the independent scaling of compute and storage. This is justified because, as VAST Technologist blogger Howard Marks says: “The EBox architecture lets us run the VAST Data Platform in environments that, until now, didn’t want, or couldn’t use, highly available DBoxes. These include hyperscalers that have thousands of a very specific server configuration and cloud providers that only offer virtual machine instances. It also allows us to work with companies like Supermicro and Cisco to deliver the VAST Data Platform to customers using servers from those vendors.”

A separate VAST blog states: “The EBox is designed to address the growing needs of hyperscalers and CSPs that require infrastructure capable of handling massive data volumes and complex workloads. By combining the best features of its predecessors into a more compact form factor, the EBox not only saves valuable rack space but also enhances the overall performance and resilience of the datacenter.”

EBox hardware features a single AMD Genoa 48-core processor, 384 GB of DRAM, 3 x storage-class memory (SCM) drives, and 9 x 30 TB NVMe SSDs (270 TB), plus two PCIe slots for front-end cards. There is a minimum cluster size of 11 nodes and metadata triplication “ensuring every read or write operation is replicated across three EBoxes within the cluster.” So the system withstands substantial hardware failure, keeping data secure and ensuring “sustained performance and rapid recovery, even during failures.”

Marks says: “Each x86 EBox runs a CNode container that serves user requests and manages data just like a dedicated CNode would, and DNode containers that connect the EBox’s SSDs to the cluster’s NVMe fabric. Just like in a VAST cluster with CBoxes and DBoxes, every CNode in the cluster mounts every SSD in the cluster.”

v5.2 also includes a global SMB namespace, Write Buffer Spillover, VAST native table support in async replication, S3 event publishing, and S3 Sync Replication, all of which “can streamline complex workloads for enterprise, AI, and high-performance computing environments.” It also has improved write performance, with Marks saying: “We’re taking advantage of the fact that there are many more capacity (QLC) SSDs than SCM SSDs by directing large bursts of writes, like AI models’ dumping checkpoints, to a section of QLC.  Writing to the SCM and QLC in parallel approximately doubles write performance” over the previous v5.1 software release. Since we’re only sending bursts of large writes to a small percentage of the QLC in a cluster, the flash wear impact is insignificant.”

He adds: “We’re also bringing the EBox architecture to the public cloud in 5.2, with fully functional VAST Clusters on the Google Cloud Platform,” which we expect to be announced later this year. 

The S3 event publishing is configured on one or more buckets in the system and provides event-driven workflows triggering functions. When data changes in such a bucket, the VAST cluster will send an entry to a specified Apache Kafka (distributed streaming platform) topic. Specifically, v5.2 VAST software requires the topic to be on an external Kafka cluster and the functions must subscribe to the Kafka topic.

More is coming this year, with Marks writing: “Over the next few quarterly releases, the VAST DataEngine will add a Kafka API-compatible event broker and the functionality to process data,” ending the external Kafka cluster limitation.

Camberly Bates, Futurum
Camberly Bates

Futurum analyst Camberly Bates writes: “VAST’s EBox integration with the Google Cloud Platform is likely to drive further adoption in public cloud environments.” This hints at Azure and AWS support for the EBox concept coming later this year.

We would expect the EBox architecture to support much higher capacity SSDs later this year, with 62 TB drives now available from Micron, Phison, Samsung, and SK hynix, and 122 TB-class SSDs announced by Phison and Solidigm recently.

Bates also suggests, referring to the v5.2 software: “Rivals may introduce similar advancements in replication, namespace management, and performance to remain competitive.” 

Suppliers like DDN, HPE, Hitachi Vantara, IBM, NetApp, Pure Storage, and WEKA are likely going to face continued strong competition from VAST in 2025.

UnifabriX taking CXL external memory mainstream

Israel-based UnifabriX, founded in 2020 by Ronen Hyatt (CEO and chief architect), Danny Volkind (CTO), and Micha Rosling (chief business officer), has taken in $11 million in seed funding to develop Smart Memory Fabric systems based around CXL external memory sharing and pooling technology. The intention is to sidestep the memory capacity limitations of individual CPU and GPU server systems by connecting external memory pools using the CXL (Computer Express Link) scheme, which is based on the PCIe cabling standard.

UnifabriX and Panmnesia are two of the most active CXL-focused startups. We looked at Panmnesia yesterday and now turn our attention to UnifabriX.

It had developed a Smart Memory Node with 32 TB of DDR5 DRAM in a 2RU chassis by April 2023, and now has its MAX (Memory Accelerator) composable memory device based on UnifabriX software and semiconductor IP.

UnifabriX appliance

MAX provides a software-defined memory fabric pool, featuring adaptive memory sharing, and using CXL and UALink cabling and concepts, several of which are mentioned in the slide above. We’ll look at the system-level architecture and then try to make sense of the cabling spaghetti.

UnifabriX diagram

Hyatt talked about this slide: “On top of our FabriX Memory OS, which is a hardened Linux … we have a stream processor that can manipulate the stream of data and the stream of protocols as they come into the memory pool. And this is programmable hardware. You can think of it like the P4 concept that grew in switches and internet switches where you can parse the data as it goes on the fly and edit the protocol messages as they go in and out. 

Ronen Hyatt, UnifabriX
Ronen Hyatt

“So you see here the frontend ports, the six frontend ports go to the host. Today there are CXL 1.1 and 2.0. We have deck and fabric ports and we accelerated the link there to 112G, much faster than CXL supports today. This is NVLink 4-equivalent in terms of speed and we are working on prototyping 224G, which is the equivalent of NVLink 5. Yes, it’s the bandwidth. We wanted to get the highest bandwidth possible on the backend side, on the fabric, when you connect multiple MAX appliances, one to each other.”

CXL cabling situation

The PCIe, CXL, and UALink situation is complex. We should note that there are five CXL standard generations between CXL 1 and CXL 3.1, with also a sixth, CXL 3.2, now available. This adds optimized memory device monitoring and management, extended security, performance monitoring, and is backwards-compatible with prior CXL specifications.

Hyatt tells us: “PCIe was originally built to live inside a platform, serving as a short-distance interconnect superseding PCI, between a CPU and peripheral devices, therefore it does not have a developed ecosystem of cabling. Larger-scale use cases of PCIe emerged only later, with ‘PCIe Fabrics’ that pooled and disaggregated devices such as NVMe storage, NICs, and GPUs.

“Those use cases did not require a lot of bandwidth, and therefore were comfortable with utilizing narrow x4 switch ports and x4 SFF-8644 (mini-SAS) cabling. A few examples here and here.

“The emergence of CXL over PCIe Gen 5 created a new demand for high-performance PCIe cabling that is capable of delivering much higher bandwidth for memory transactions. Since PCIe did not have such solutions ready, the market found interim solutions by utilizing cabling systems from the Ethernet domain, such as:

  • QSFP-DD MSA (x8) – a denser form factor of QSFP, originally created for Ethernet, Fibre Channel, InfiniBand and SONET/SDH. Some people used it (and still use it today) for PCIe x8 connections. See here.
  • CDFP MSA (x16) – originally developed for 400G Ethernet (16 x 25G lanes), but later certified de-facto for PCIe Gen 5. See here and here.

“Today, the PCIe ecosystem is aligning around the OSFP MSA cabling system, with OSFP (x8) and its denser variant OSFP-XD (x16) that both support the latest signaling rate of 224G PAM4 per lane (for example, 8 x 200G = 1.6 Tbps Ethernet), and are therefore also compatible with PCIe Gen 5/CXL 1.1, 2.0 (32G NRZ), PCIe Gen 6/CXL 3.x (64G PAM4), and PCIe Gen 7/CXL 4.x (128G PAM4). i.e. this OSFP cabling system is future-proof for at least two generations ahead in the PCIe domain. It is also ready for UALink that reuses Ethernet IO at the electrical level. One cable to rule them all.”

Nvidia showed a way forward here, with Hyatt explaining: “It takes a lot of market education to bring memory fabrics into the datacenter. Nvidia jumped in to help when it introduced the DGX GH200 system with its NVLink memory fabric, creating a large, disaggregated 144 TB pool of memory. CXL and UALink are the open comparables of NVLink. They all support native load/store memory semantics.

“Nvidia taught the world that memory fabrics (by NVLink) are superior to networks (by InfiniBand). We tend to agree.”

He said: “UnifabriX developed a Fabric Manager (FM) compliant with CXL 3.2 FM APIs including support for DCD (Dynamic Capacity Device), i.e. it is capable of provisioning and de-provisioning memory dynamically, on-demand, using standard, open, CXL APIs. I haven’t seen another DCD Fabric Manager out there, so this may be one of the first FMs that you would encounter that actually does the work.”

There are a couple of other points. Hyatt said: “We are able to mix and match CXL ports and UALink ports, meaning we can provide memory on demand to both CPUs and to GPUs. The UALink connector is based on Ethernet IO, so the same connector, the same OSFP and OSFP XD, is going to be used for both CXL and UALink. You just change the personality of the port.”

Working silicon

The company demonstrated its memory pool dynamically altering in size and composed out to host processors on demand and then returned to the pool. UnifabriX is already earning revenue, with deployments in the data analytics, high-performance computing, public and private cloud areas.

UnifabriX slide
UnifabriX slide

Hyatt said: “We have a few hyperscaler customers [where] the system is there running with the real workloads currently on Emerald Rapids platform and shifting soon towards Granite Rapids and Turin systems with AMD.”

“We have quite a few new customers in different segments of the market, not just the hyperscalers and the national labs. We have drug discovery companies, DNA sequencing. Turns out there are a lot of use cases that sit under the HPC umbrella where people need a lot of memory. Sometimes they need bandwidth, sometimes they need capacity. But having the ability to grow memory on demand and doing it dynamically brings a lot of value, not just on the TCO side.”

He explained: “You see the cloud, the public cloud, national labs. We started with the national labs and animation studios. There’s a lot of digital assets and you need to do rendering and processing, and they’re all working with fast storage systems these days, but they’re not fast enough for what they need. So having a memory pool in between helps to accelerate the whole process.”

Processing in memory

Hyatt talked about MAX being able to do some processing: “It has processing capabilities, which we found very useful for HPC. So we have processing-in-memory or near-memory capabilities. This works great for sparse memory models, for instance, in HPC where you have very large models that fit into petabytes and you need to abstract the memory address space. So you actually expose a huge address space externally. 

“But internally you do the mapping. And this is part of the memory processing that we do here. And this is one example. We have an APU, which is an application processing unit which is exposed to the customer, where the customer can run their own code over containers. So if they want to do something on the memory, like, for instance, checking for malicious code, checking for some abnormal patterns within the memory, this is something that they can run internally. We provide that capability.”

Go to market

How does UnifabriX go to market? Hyatt said: “Currently, we work directly with end customers. And the reason we do it is because this is part of the product definition, like getting the feedback of what customers need. So you don’t want the channel in between because then you lose a lot of the feedback. 

“But we are already engaged with partners. Some of them are platform OEMs that want to have a memory pool as part of their product portfolio. So think about all the big guys that have storage systems and think of a memory pool as a storage server, but it works on memory. So most of the paradigms and the semantics that go with storage would be replicated to the memory world and we are working with them. 

“And on top of that we have several channels, some are specialized for HPC. There are OEM vendors that build unique servers and unique appliances for the HPC market. And HPC is really interested in having the memory bandwidth that CXL provides. There are several system integrators that build the whole racks and ship systems with GPUs and with a lot of compute power. And they actually pack together GPUs, servers, storage, and memory together, and ship it as a rack.”

UnifabriX is planning a new funding round in the second half of 2025. 

The fab process side is developing, with Hyatt saying: “Currently, our silicon is seven nanometer and we plan to have a five nanometer TSMC silicon later, in 2026, early 2027.” This aligns with PCIe Gen 6, as Hyatt pointed out: “CXL itself is moving from PCIe Gen 5 to Gen 6, so we have to upgrade the process. Gen 6 comes with mixed signals … that needs five nanometer to be efficient on power.”

We’ll follow up with an article looking at UnifabriX’s MAX device.

Bootnote

QSFP – Quad Small Form-factor Pluggable standard referring to transceivers for optical fiber or copper cabling, and providing speeds four times their corresponding SFP (Small Form-factor Pluggable) standard. The QSFP28 variant was published in 2014 and allowed speeds up to 100 Gbps while the QSFP56 variant was standardized in 2019, doubling the top speeds to 200 Gbps. A larger variant Octal Small Format Pluggable (OSFP) had products released in 2022 capable of 800 Gbps links between network equipment. 

OSFP MSA – Octal Small Form Factor Pluggable (OSFP) Multi Source Agreement (MSA). The OSFP (x8) and its denser OSFP-XD (x16) variants both support the latest signaling rate of 224G PAM4 per lane (for example 8 x 200G = 1.6 Tbps Ethernet). They are compatible with PCIe Gen5 / CXL 1.1, 2.0 (32G NRZ), PCIe Gen6 / CXL 3.x (64G PAM4) and PCIe Gen7 / CXL 4.x (128G PAM4). This OSFP cabling system is future-proof for 2 generations ahead in the PCIe domain. It is also ready for UALink that reuses Ethernet IO at the electrical level.

CDFP – CDFP is short for 400 (CD in Roman numerals) Form-factor Pluggable, and designed to provide a low cost, high density 400 Gigabit Ethernet connection.

Commvault delivers forest-level Active Directory recovery

The Commvault Cloud Platform (formerly known as Metallic) is now automating protection for Active Directory at forest level.

Active Directory (AD) is a user authentication service in the Microsoft Windows environment. It lists users and resources such as devices, along with the permissions they have or require. AD Domain Services are run on a server and a domain is a central admin unit for management and security. An organization can have more than one domain, and cross-domain access is prohibited unless authorized. Domains are managed in a tree-like structure with a top-level forest managing a set of domains that share a common schema, configuration, and global catalog, and include users, groups, permissions, and domain controllers across the organization.

Pranay Ahlawat, Commvault
Pranay Ahlawat

Commvault CTO and AI Officer Pranay Ahlawat stated: “Recovering Active Directory is foundational to maintaining continuous business after a cyberattack, yet traditional methods are too complex and prone to error. With automated Active Directory forest recovery, we are giving customers game-changing recovery capabilities, and by integrating this into our unique cyber resilience platform with broad workload support, we’re bringing a new era of continuous business to our customers that nobody can match.”

Malware attackers commonly target AD as it is essential in Windows environments, used to authenticate more than 610 million users worldwide. Commvault says: “When disaster strikes, recovering AD is vital, yet traditionally has been very hard to do, requiring intricate, time-consuming, manual processes, as described by Microsoft’s Forest Recovery Guide.” This can take “days or even weeks to complete.”

To use its full name, Commvault Cloud Backup & Recovery for Active Directory Enterprise Edition (CCBRADEE) now enables automated “rapid recovery” of such an AD forest by using automated runbooks. These include “tasks like transferring key roles from an unavailable domain controller to a functioning one, which is essential for a clean recovery.”

Comvault slide.

CCBRADEE has visual topology views of an organization’s Active Directory environment to give admins simple and quick identification of which domain controllers to restore first and how they should be recovered to accelerate availability of AD services. It “integrates AD forest recovery with granular recovery of both Active Directory and Entra ID, the cloud-based identity service, providing comprehensive protection.”

Cohesity and its acquired Veritas business can also safeguard AD forest environments via integrations with Semperis Active Directory Forest Recovery (ADFR) and other tools. More data protection suppliers protect AD at the forest level, including Dell with Recovery Manager for Active Directory Forest Edition, Rubrik with its Security Cloud (RSC), and Veeam. 

Commvault Cloud Backup & Recovery for Active Directory Enterprise Edition is targeted for general availability within the first half of 2025 and priced per user. Explore AD protection on Commvault’s website here.

Can synthetic data help scale AI’s data wall?

COMMISSIONED: As with any emerging technology, implementing generative AI large language models (LLMs) isn’t easy and it’s totally fair to look side-eyed at anyone who suggests otherwise.

From issues identifying use cases that maximize business value to striking the right balance between hard-charging innovation and sound governance, companies face their fair share of GenAI struggles.

Now it seems even those LLMs could use some help. If AI experts have it right, LLMs may be running out of fresh training data, which has the AI sector looking to a possible stopgap: synthetic data.

In the context of LLMs, synthetic data is artificially manufactured using the statistical properties of real-world data without using real information about companies or people and other entities. Using synthetic data helps organizations model outcomes without exposing themselves to security or privacy risks.

Some experts believe that by conjuring new data with which to populate outputs synthetic data can help LLMs clear the so-called data wall. To better understand the value of synthetic data, it helps to grasp the pending limitations posed by real-world data.

The data wall

Academics and AI luminaries alike have noted the probability for LLMs to hit a limit to the amount of human-generated text with which they’re trained – possibly as soon as 2026.

The data shortfall presents a problem because as the volume of training data declines, models can struggle to generalize. This can lead to overfitting, a phenomenon in which a model masters its training data so much that it performs poorly on new data, resulting in less coherent outputs.

And while experts began publicizing the problem shortly after OpenAI’s kicked off the GenAI race by launching ChatGPT two years ago, VCs powerful enough to pull the financial levers of this market have lent their voices to the issue.

“The big models are trained by scraping the internet and pulling in all human-generated training data, all-human generated text and increasingly video and audio and everything else, and there’s just literally only so much of that,” said Marc Andreessen, co-founder of Andreessen Horowitz.

The problem is serious enough that AI companies have gone analog, hiring human domain experts such as doctors and lawyers to handwrite prompts for LLMs.

Barring any breakthroughs in model techniques or other innovations that help GenAI hurdle the coming data wall, synthetic data may be the best available option.

Big brands swear by synthetic data

Synthetic data is particularly useful for helping organizations simulate real-world scenarios, including everything from what merchandise customers may purchase next to modeling financial services scenarios without the risk of exposing protected data.

Walmart, for one, synthesizes user behavior sequences for its sports and electronics categories to predict next purchases. Walmart employees vet the data throughout the process to ensure integrity between the user behavior sequence and the prediction.

The human-in-the-loop factor may be key to harnessing synthetic data to improve outcomes. For example, combining proprietary data owned by enterprises with reasoning from human employees can create a new class of data that corporations can use to create value.

This “hybrid human AI data approach” to creating synthetic data is something that organizations such as JPMorgan are exploring, according to Alex Wang, a senior research associate with the financial services company, who noted that JPMorgan has 150 petabytes of data at its disposal compared to 1 petabyte OpenAI has indexed for GPT 4.

In fact, OpenAI itself has used its Strawberry reasoning model to create data for its Orion LLM. You read that right – OpenAI is using its AI models to train its AI models.

The bottom line

Synthethic data has its limitations. For example, it often fails to capture the complexity and nuances – think sarcasm or turns of phrase – which makes real-world data so rich. This can reduce the relevancy of results, thus limiting the value of scenarios synthetic data is meant to model.

As with real-world data, algorithms used to generate synthetic data can include or amplify existing biases, which can lead to biased outputs. Moreover, ensuring the model trained on synthetic data performs well may require using supplementary real-world data, which can make fine-tuning challenging. Similarly, inaccuracies hallucinations remain an issue in synthetic data.

The challenges that come with using synthetic data require the same sound data governance practices organizations are leveraging with LLMs that train on real-world data. As such, many data engineers view the use of synthetic data to populate models as complementary.

Even so, an existential data crisis isn’t required to capitalize on the benefits of using synthetic data. And your organization needn’t be Walmart of JPMorgan’s scale to take advantage of the opportunities synthetic data has to offer.

Knowing how to effectively leverage synthetic data may be challenging for organizations who haven’t leveraged such techniques to manage and manipulate their data.

Dell Technologies offers access to professional services, as well as a broad open ecosystem of vendors, that can help you embark on your synthetic data creation journey.

Learn more about the Dell AI Factory.

Brought to you by Dell Technologies.

XenData ships hybrid archive appliance

XenData has taken the wraps off its new X10 Media Archive Appliance, which connects to one or two external LTO drives and can manage an “unlimited number” of offline LTO cartridges.

It has both a file-folder interface and a web interface, which provides previews of video and image files. “The X10 is a very cost-effective way to manage a highly scalable media archive,” the company says.

XenData X10 Media Archive Appliance
XenData X10 Media Archive Appliance

The web interface can be securely accessed by on-premises and remote users. They can search and browse for archived files and then play video previews and view low-res versions of image files for all the content held in the archive, including files stored on offline LTO cartridges. The web interface uses HTTPS, and it supports Chrome, Microsoft Edge, and Safari browsers.

The X10 appliance runs a Windows 11 Pro operating system and can be used standalone or connected to a local network. Although optimized for media files, the X10 will archive all file types and file names supported by Windows. It includes a mirrored 4 TB cache volume, “enhancing write and read performance,” said XenData, while being used to store the media file previews.

XenData LTO Archive

The archive file system can be accessed as a standard network share that adheres to the standard Microsoft security model based on Active Directory, and can be easily added to a Windows Domain or Workgroup.

Advanced functionality includes automatic LTO cartridge replication, end-to-end logical block protection, and the creation of LTO cartridge contents reports and the issuance of email alerts. It can also be configured to write to rewritable LTO cartridges using the LTFS interchange format, and can support writing to unalterable LTO WORM cartridges.

Phil Storey, XenData
Phil Storey

“The X10 can manage hundreds of LTO cartridges stored ‘on the shelf’, and allows users to easily find the content they want and bring it back online, even for a very large archive,” said Phil Storey, XenData CEO.

The X10 appliance will be available at the end of this month, priced from $6,950.

XenData introduced its Media Portal viewer to its on-prem and public cloud tape archive library last May so users can see previews of archived image and video files to select content for restoration.

Panmnesia snags award for GPU CXL memory expansion

A Panmnesia scheme to bulk up a GPU’s memory by adding fast CXL-accessed external memory to a unified virtual memory space has won a CES Innovation award.

Panmnesia says large-scale GenAI training jobs can be memory-bound as GPUs, limited to GBs of high-bandwidth memory (HBM), could need TBs instead. The general fix for this problem has been to add more GPUs, which gets you more memory but at the cost of redundant GPUs. Panmnesia has used its CXL (Computer eXpress Link) technology, which adds external memory to a host processor across the PCIe bus mediated by Panmnesia’s CXL 3.1 controller chip, which exhibits controller round-trip times less than 100 ns, more than 3x less than the 250 ns needed by SMT (Simultaneous Multi-Threading) and TPP (Transparent Page Placement) approaches.

A Panmnesia spokesperson stated: “Our GPU Memory Expansion Kit … has drawn significant attention from companies in the AI datacenter sector, thanks to its ability to efficiently reduce AI infrastructure costs.”

The technology was revealed last summer and shown at the OCP Global Summit in October. The company has a downloadable CXL-GPU technology brief, which says its CXL Controller has a two-digit-nanosecond latency, understood to be around 80 ns. A high-level diagram in the document shows the setup with either DRAM or NVMe SSD endpoints (EPs) hooked up to the GPU: 

Panmnesia diagram

In more detail, a second Panmnesia diagram shows the GPU is linked to a CXL Root Complex or host bridge device across the PCIe bus which unifies the GPU’s high-bandwidth memory (host-managed device memory) with CXL endpoint device memory in a unified virtual memory space (UVM). 

Panmnesia diagram

This host bridge device “connects to a system bus port on one side and several CXL root ports on the other. One of the key components of this setup is an HDM decoder, responsible for managing the address ranges of system memory, referred to as host physical address (HPA), for each root port. These root ports are designed to be flexible and capable of supporting either DRAM or SSD EPs via PCIe connections.” The GPU addresses all the memory in this unified and cacheable space with load-store instructions.

A YouTube video presents Panmnesia’s CXL-access GPU memory scheme in simplified form here.

Cohesity adds malware incident response partners to CERT service

Cohesity has added incident response partners to its Cyber Event Response Team (CERT) service to help customers diagnose and set up a recovery response to a malware attack faster.

It set up its CERT service in 2021, stating that it had partnered with the world’s leading cybersecurity incident response (IR) firms. Now it has announced a formal list of IR partners: Palo Alto Networks (Unit 42), Arctic Wolf, Sophos, Fenix24, and Semperis. CERT, available to all Cohesity customers as part of their existing subscription, can share customer-approved operational data – including logs, reports, and inventories – with these IR partners. It has developed a methodology that utilizes native platform capabilities and integrations with its Data Security Alliance to provide greater insight into data breaches.

Sanjay Poonen

Sanjay Poonen, Cohesity CEO, stated: “With ransomware, data breaches, and other cyber threats becoming an unavoidable reality, organizations need the assurance that they can bounce back faster, stronger, and smarter … We’re doubling our commitment to our customers by ensuring they have the expertise and tools to navigate and recover from cyber crises effectively.  Cyber resilience is the cornerstone of modern cybersecurity, and we are committed to helping our customers achieve it.”

An example of such an incident affected the divisional manufacturing plant and R&D center of a global auto parts enterprise. The company used Cohesity backup and its FortKnox immutable isolated data copy vault. The company was hit by ransomware in late 2023, which encrypted all the VM images hosted on ESXi servers. 

The manufacturer contacted CERT the morning after the attack and CERT worked in partnership with Fenix24 to contain and investigate the threat. The threat actor was identified and the IR team found that more than 100,000 files were locked up. Cohesity CERT worked with the division’s IT department and Fenix24 to bring data back from the FortKnox service, validate that the threat had been mitigated, and ensure that remediation steps were successful. 

CERT is available 24×7. “Personnel from Cohesity CERT and its partners are seasoned cybersecurity experts with specialized knowledge in incident response, threat intelligence, and forensics,” we’re told.

Kerri Shafer-Page

Kerri Shafer-Page, Artic Wolf’s VP of Incident Response,  said: ”Cohesity’s quick response toolkit gives us access to all kinds of data that can enable a more comprehensive investigation and quicker recovery. Partnering with Cohesity CERT adds valuable expertise in backup and recovery and helps us ensure our  joint customers are resilient no matter what attackers throw at them.” 

Competitor Commvault also has multi-vendor cyber resilience partnerships, set up in 2023, with:

  • Avira (part of Gen): AI/ML-driven threat intelligence, prediction, analysis and anti-malware technologies. 
  • CyberArk: Identity security platform.
  • Darktrace: Machine learning-based anomaly detection with integration with HEAL and Commvault. 
  • Databricks: Data lake platform for data and AI. 
  • Entrust: Post-quantum cryptography and data encryption.
  • Microsoft Sentinel: SIEM. 
  • Netskope: Zero trust-based Secure Access Service Edge (SASE) web content filtering. 
  • Palo Alto Networks: Threat intelligence repository leveraging Cortex XSOAR to shorten incident response times. 
  • Trellix: Threat detection and response with Intelligent Virtual Execution (IVX) sandbox to analyze and inspect malware in an isolated environment.

Rubrik has a Ransomware Response Team (RRT), a virtual team of experienced people in its global support organization. RRT is available 24x7x365 and composed of critical incident managers and senior support staff. Rubrik’s executive leadership is part of this virtual team and has visibility of every recovery RRT is involved with.

Veeam Software has integrating its data protection reporting with cybersecurity software vendor Palo Alto Networks to enable customers to respond quicker to attacks.

Get more information on Cohesity CERT here.

HighPoint now supports Arm in high-performance NVMe environments

HighPoint Technologies, which provides PCIe NVMe storage and switching solutions, now supports Arm-based computing platforms, increasingly present in datacenters and at the edge.

HighPoint was already an established storage and connectivity provider for Intel-based systems used in high-performance environments. HighPoint’s high-density NVMe technologies support up to 32 devices per PCIe slot, addressing the growing demand for massive storage capacity.

“Integrated hardware sensors, real-time status and health monitoring, and a comprehensive storage management suite now streamline deployment and service workflows for Arm platforms,” said the provider.

The NVMe offerings can be custom-tailored with firmware, driver, and storage management tools for unique applications and easier integration with Arm platforms.

“Our expansion into Arm-based computing platforms reflects HighPoint’s commitment to driving innovation and addressing the evolving needs of datacenter and edge computing environments,” said May Hwang, VP of marketing at HighPoint Technologies. “By leveraging unparalleled NVMe storage performance, scalability, and energy efficiency with our versatile and customizable solutions, we are empowering customers to seamlessly integrate high-performance storage into their IoT servers and AI-driven applications, paving the way for next-generation computing solutions.”

HighPoint’s NVMe product lines span multiple PCIe generations, from Gen 3 to Gen 5 x16, delivering the performance and scalability increasingly required across industrial IoT server environments at the edge.

Last year, HighPoint rolled out a new line of external NVMe RAID enclosures, designed to elevate Gen 4 storage applications to “new heights.” The RocketStor 654x series promised x16 transfer performance and “nearly half a petabyte” of storage capacity, which could be integrated into any x86 platform with a free PCIe x16 slot.

As far as other storage firms integrating with Arm in the datacenter, open source object storage supplier MinIO recently tweaked its code so AI, ML, and other data-intensive apps running on Arm chipsets can achieve higher performance.

BMC modernizes mainframe storage with cloud-first solutions

The new Cloud Data Sets feature in BMC Software’s AMI Cloud Data platform “transforms” mainframe data management by providing direct access to cloud object storage, without requiring modifications to existing JCLs or applications.

The provider claims this enhancement “empowers” ITOps teams to fully replace traditional tape storage with cloud-based solutions, “simplifying” operations and “minimizing” disruption.  BMC bought Model9 in April 2023, and has rebranded its software as the AMI Cloud.

BMC maintains that within the next five years, secondary tape storage and traditional tape software will be “phased out” at most organizations. The future of mainframe data storage lies in cloud-based object storage, it says, with cloud storage “up to 12 times more affordable” than traditional virtual tape library (VTL) solutions, while eliminating the need for costly tape hardware.

South African bank Nedbank is buying into this evolution, and with the help of BMC, the company has transformed its data management. A backup that used to run for 48 hours has been reduced to 36 minutes after switching to a cloud-based solution. The switch has also allowed Nedbank to streamline its disaster recovery and backup processes, reducing complexity while enhancing security and data availability.

“This evolution allows IT operations teams to redirect their tape backups to the cloud without any operational changes, removing a major barrier to cloud adoption in mainframe environments,” says BMC. With Cloud Data Sets, AMI Cloud Data supports all major backup utilities, including EXCP. “We have seen a huge increase in the performance of our backups. One example is that the backup that used to run for an entire weekend for 48 hours, was reduced to 36 minutes when we went to a cloud-based solution,” said Ashwin Naidu, IT manager for enterprise storage and backup at Nedbank. 

BMC AMI Cloud diagram
BMC AMI Cloud diagram

In addition to significantly shorter backup and restore times, the new version of AMI Cloud Data has been optimized for lower CPU consumption.

To deliver on such offerings, BMC says it is combining its software expertise with the data infrastructure of partners, including Hitachi, Mainline, Dell, and AWS.

Cloudian becomes a Hammerspace GDP storage citizen

Object storage supplier Cloudian is partnering unstructured data orchestrator Hammerspace by making its HyperStore object storage a repository used by Hammerspace’s Global Data Platform’s parallel NFS-based file system.

The Global Data Platform (GDP) software provides a global namespace, in which unstructured file and object data is placed, orchestrated, made available for use, and accessed as if it were local. Hammerspace uses its GDE software to manage file and object data in globally distributed locations, across SSDs, disk drives, public cloud services, and tape media. Cloudian’s HyperStore is an object storage system with effectively limitless scalability and supports Nvidia’s GPUDirect for object storage. The Hammerspace software also supports GPUDirect.

Molly Presley

Molly Presley, Hammerspace SVP of Marketing, stated: “Our partnership with Cloudian redefines how enterprises manage and leverage unstructured data. Together, we’re enabling organizations to unify their data ecosystems, eliminate silos, and unlock the full potential of their data for AI, analytics, and other transformative initiatives.”

Cloudian CMO Jon Toor paralleled this comment by saying: “Cloudian is excited to collaborate with Hammerspace to deliver an integrated solution that combines performance, scalability, and simplicity. This partnership gives enterprises the tools they need to conquer the challenges of data growth and drive their digital transformation strategies forward.”

Cloudian Hammerspace diagram

Cloudian claims its storage costs can be as low as 0.5 cents per GB per month, and provides scale-out storage for multiple use cases, such as an AI data lake, with 100x the performance of tape, and costs lower than public cloud. 

There are three main features with this deal:

Jon Toor
  • Seamless data orchestration across tiers and locations as Hammerspace’s data platform automates data movement between Tier 0, Tier 1, and object storage within a unified global namespace, spanning sites, hybrid clouds, and multi-cloud environments.
  • Exabyte-scale and S3-compatible object storage with data protection and ransomware defense.
  • Hammerspace supports standard NFS, SMB, and S3 protocols without proprietary software or networking requirements, enabling seamless integration with Cloudian storage, and both support GPUDirect and its remote direct memory access to storage drives.

Toor told us: “The solution integrates Hammerspace as a storage virtualization layer with Cloudian HyperStore (and other storage devices) as the underlying storage infrastructure, delivering a unified, efficient, and scalable data management ecosystem.

“We see this as an ideal complementary solution. In many of our deployments, Cloudian is implemented alongside other, dissimilar products. With Hammerspace abstracting and virtualizing data across these disparate storage systems – including object storage, NAS, and cloud – the joint solution empowers organizations with high-performance data orchestration, enabling users to access and manage data transparently across on-premises, edge, and cloud environments.

“The power, dependability, and simplicity of this combined solution is proven in multiple Cloudian/Hammerspace deployments today.”

All-in-all, this partnership means, the two suppliers say, that customers can unify their data across edge, core, and cloud environments in a single global namespace. Through the partnership they can extend active production namespaces to provide global unified file access across them all without impacting existing high performance tiers. They get simplified management of and accessibility to unstructured data while supporting performance-intensive applications such as AI training, HPC workloads, and analytics. 

The combined Cloudian-Hammerspace offering is available immediately through Hammerspace and Cloudian’s global partner networks.

Cloudian provides more AI data workflow information here. A solution brief, “End-to-End Media Workflows with Hammerspace & Cloudian,” can be accessed here.