Home Blog Page 83

Dell eyes AI-fueled revival after fiscal 2024 downturn

Strong sequential growth in storage revenues at Dell was not enough to prevent an 11 percent revenue fall for the final fiscal 2024 quarter as the company waits in hope of an AI-driven recovery.

Dell reported $22.3 billion in revenues for the quarter ended February 2, with a $1.2 billion profit, up 91 percent year-on-year. Full year revenues were $88.4 billion, down 13.5 percent from the year before, with a profit of $3.2 billion, 32 percent higher year-on-year. Squeezing more profit from lower revenues is quite the achievement.

Jeff Clarke, Dell
Jeff Clarke

Vice chairman and COO Jeff Clarke said in prepared remarks: “In a year where revenue declined, we maintained our focus on operational excellence delivering solid earnings per share and outstanding cash flow. FY24 was one of those years that didn’t go as planned, but I really like how we navigated it. We showed our grit and determination by quickly adapting to a dynamic market, focusing on what we can control, and extending our model into the high growth AI opportunity.”

CFO Yvonne McGill said: “We generated $8.7 billion in cash flow from operations this fiscal year, returning $7 billion to shareholders since Q1 FY23. We’re optimistic about FY25 and are increasing our annual dividend by 20 percent – a testament to our confidence in the business and ability to generate strong cash flow.”

Dell revenues by quarter/fiscal year
Six successive down quarters with Dell’s Q4 FY24 revenues below the 2019 level

Quarterly financial summary

  • Gross margin: 23.8 percent vs 23 percent a year ago
  • Operating cash flow: $1.5 billion
  • Free cash flow: $1 billion vs $2.3 billion last year; 55 percent lower
  • Cash, cash equivalents, and restricted cash: $7.5 billion vs $8.9 billion last year
  • Diluted earnings per share: $1.59
Dell earnings

Dell has two main business units – Infrastructure Solutions Group (ISG) and Client Solutions Group (CSG). The larger CSG, with its PCs and laptops, had revenues of $11.7 billion, 12 percent down year-on-year, while ISG, with its servers, storage and networking, reported $9.3 billion, six percent lower annually. 

Dell annual revenues, profit
Dell annual revenue history shows FY24 having the lowest revenues since 2018

Servers and networking brought in $4.9 billion, the same as a year ago, while storage was responsible for $4.5 billion, down 10 percent year-on-year but up 16 percent quarter-on-quarter.

Dell server and storage revenues

Clarke said: “Our strong AI-optimized server momentum continues, with orders increasing nearly 40 percent sequentially and backlog nearly doubling, exiting our fiscal year at $2.9 billion,” implying that the next few quarter’s results should be better.

“We’ve just started to touch the AI opportunities ahead of us, and we believe Dell is uniquely positioned with our broad portfolio to help customers build GenAI solutions that meet performance, cost and security requirements.”

Why the quarterly storage revenue rise? Dell said there was demand strength across the portfolio, more than any expected seasonal improvement. Clarke said on the earnings call: “We had year-over-year demand growth in the unstructured space, ECS, as well as PowerScale. They grew quarter-over-quarter and year-over-year on a demand basis. Those are generally good indicators … around AI file and object, which are the data classes that generally feed the AI engines … Our progress in traditional storage was good too. We were ahead of our normal seasonality … It was down year-over-year, but better than we expected across mid-range, across our data protection products and our high-end storage products.”

The outlook is for growth. Clarke said: “We believe the long-term AI action is on-prem where customers can keep their data and intellectual property safe and secure.  PCs will become even more essential as most day-to-day work with AI will be done on the PC. We remain excited about the long-term opportunity in our CSG business.”

He added: “Our storage business will benefit from the exponential growth expected in unstructured data … We think AI moves to the data. More data will be created outside of the data center going forward than inside the data center today. That’s going to happen at the edge of the network. A smart factory, an oil derrick or platform, a deep mine, all variations of this. We believe AI will ultimately get deployed next to where the data is created driven by latency.” 

In his view, enterprises will “quickly find that they want to run AI on-prem because they want to control their data. They want to secure their data. It’s their IP and they want to run domain specific and process specific models to get the outcomes they’re looking for.”

Dell thinks an AI-fueled recovery in storage demand will lag stronger server demand by a couple of quarters.

Revenues for Q1 FY25 are expected to be between $21 billion and $22 billion, 3 percent higher annually at the midpoint. But then growth will accelerate. Clarke sees “modest growth in traditional [servers], stronger growth in AI-enabled servers, and an opportunity with storage as the year progresses.”

Full FY25 revenues should be between $91 billion and $95 billion, up 5 percent year-on-year at the midpoint. McGill said: “We expect ISG to grow in the mid-teens fueled by AI, with a return to growth in traditional servers and storage, and our CSG business to grow in the low single digits for the year.”

HPE storage results vanish in financial reporting revamp

HPE has changed its quarterly financial reporting structure with storage results disappearing into a hybrid cloud category.

Revenues in its first fiscal 2024 quarter ended January 31 were $6.8 billion, 14 percent lower than the year-ago quarter and also below its guided estimate for $6.9 billion to $7.3 billion. There was a $387 million profit, 23 percent less than last year. The move to GreenLake is causing overall revenue declines as subscription revenue is recognized over time whereas straightforward product sales are recognized when the product ships.

Antonio Neri, HPE president and CEO, said in prepared remarks: “HPE exceeded our profitability expectations and drove near-record year-over-year growth in our recurring revenue in the face of market headwinds, demonstrating the relevance of our strategy.” HPE did not exceed its revenue expectations, however. The annual revenue run rate was $1.4 billion, 42 percent more than a year ago, and primarily due to the GreenLake portfolio of services.

HPE revenues
HPE revenues have basically flatlined since 2019

HPE’s new organizational structure consists of the following segments: Server; Hybrid Cloud; Intelligent Edge; Financial Services; and Corporate Investments and Other. It has amalgamated previously separate Compute and HPC & AI reporting lines into a single Server category. The new Hybrid Cloud reporting line includes three elements:

  • The historical Storage segment
  • HPE GreenLake Flex Solutions (which provides flexible as-a-service IT infrastructure through the HPE GreenLake edge-to-cloud platform and was previously reported under the Compute and the High Performance Computing & Artificial Intelligence (“HPC & AI”) segments) 
  • Private Cloud, and Software (previously reported under the Corporate Investments and Other segment)

This means we no longer have a direct insight into the business health of its storage portfolio. HPE says it’s doing this to better reflect the way it operates and measures its business units’ performance. This new hybrid cloud segment is meant to accelerate customer adoption of HPE’s GreenLake hybrid cloud platform.

Financial summary

  • Gross margin: 36.4 percent, up 2.4 percent year-over-year
  • Operating cash flow: $64 million
  • Free cash flow: -$482 million
  • Cash, cash equivalents, and restricted cash: $3.9 billion vs $2.8 billion last year 
  • Diluted earnings per share: $0.29, 24 percent lower than last year
  • Capital returns to shareholders: $172 million in dividends and share buybacks

Server segment revenues were $3.4 billion, down 23 percent year-over-year. Hybrid Cloud pulled in $1.3 billion, down 10 percent, with Intelligent Edge reporting $1.2 billion, up 2 percent. Financial services did $873 million, down 2 percent, while Corporate Investments and Other was responsible for $238 million, up just 1 percent.

CFO Marie Myers said on the earnings call: “Demand in Intelligent Edge did soften due to customer digestion of strong product shipments in fiscal year ’23, which is lasting longer than we initially anticipated and is the primary reason Q1 revenue came in below our expectations.” Specifically, “campus switching and Wi-Fi products eased materially, particularly in Europe and Asia.”

Neri’s top-down view was that “overall, Q1 revenue performance did not meet our expectations.” It was “lower than expected in large part because networking demand softened industry-wide and because the timing of several large GPU acceptances shifted. Additionally, we did not have the GPU supply we wanted, curtailing our revenue upside.”

”This quarter is a moment in time and does not at all dampen our confidence in the future ahead of us.”

Myers commented: ”Demand for our traditional server and storage products has stabilized,” although “our traditional storage business was down year-over-year on difficult compares, given backlog consumption in Q1 ’23. Total Alletra subscription revenue grew over 100 percent year-over-year and is an illustration of our long-term transition to an as-a-service model across our businesses. We are starting to see AI server demand pull through interest in our file storage portfolio. We are also already seeing some cross-selling benefits of integrating the majority of our HPE GreenLake offering into a single business unit.”

Neri said HPE was “capturing the explosion in demand for AI systems” with GPU-enhanced servers, called Accelerator Processing (AP) units, orders rising. AP orders now represent nearly a quarter of HPE’s entire server orders since fiscal 2023’s Q1. Neri said: ”Our pipeline is large and growing across the entire AI life cycle from training to tuning to inferencing.”

Server revenues should grow due to AI demand and better GPU availability, and “hybrid cloud will benefit from continued HPE GreenLake storage demand and the rising productivity of our specialized sales force.”

HPE’s next quarter outlook is $6.8 billion in revenues up or down $0.2 billion, a 2.5 percent annual decrease at the midpoint. Myers said: ”For hybrid cloud, we expect sequential increases through the year as our traditional storage business improves and HPE GreenLake momentum continues. We expect meaningful progress through the year.”

Snowflake replaces CEO as losses widen

After reporting growth in both annual revenues and losses in its latest results, Snowflake chairman Frank Slootman decided to replace Snowflake CEO Frank Slootman, and appointed SVP Sridhar Ramaswamy in his place.

Revenues in cloud data warehouser Snowflake’s final fy2024 quarter, ended Jan 31, were $774.7 million, 31.5 percent more than a year ago, with a $170 million loss, better than the year-ago’s $207.5 million of red ink. The full year revenue number was $2.8 billion, 35.9 percent more than last year. There was a full year loss of $836.1 million, yes, $0.84 billion, contrasting with fy23’s loss of $797 billion.

Frank Slootman.

Chairman Frank Slootman’s prepared statement said: “We are successfully campaigning the largest enterprises globally, as more companies and institutions make Snowflake’s Data Cloud the platform of their AI and data strategy.”

About the resignation of himself as CEO he said: “As the leading cloud data platform, Snowflake is at the epicenter of the AI revolution. There is no better person than Sridhar to lead Snowflake into this next phase of growth and deliver on the opportunity ahead in AI and machine learning. He is a visionary technologist with a proven track record of running and scaling successful businesses. I have the utmost confidence in Sridhar and look forward to working with him as he assumes this new role.”

Presumably Slootman wants Ramaswamy to “Amp it up,” that being the title of Slootman’s recent book about being a CEO and leader. In that book he writes this about a board removing a CEO: “It will treat removing the CEO as a gut-wrenching, high-risk, last-resort move, never to be done casually or for frivolous reasons.” In his view “hiring a new one is a lengthy process.”

Ramaswamy joined Snowflake in May 2023 when it acquired Neeva, the world’s first private AI powered search engine. He was a Neeva co-founder and a Google SVP before that, and led the launch of Snowflake’s Cortex. This is a fully managed service that offers access to AI models, LLMs, and vector search functionality to enable organizations to analyze data and build AI applications. 

Sridhar Ramaswamy.

We note that Ramaswamy’s LinkedIn profile says of his first 8 months at Snowflake: “Learning Snowflake.” Our bet is that he was a CEO-elect from the start.

What Slootman said in the earnings call bears that out: “I was brought to Snowflake five years ago to help the company break out and scale. I wanted to grow the business fast, but not at all costs. It had to be efficient and establish a foundation for long-term growth.”

“I believe the company succeeded in that mission. The board has run a succession process that wasn’t based on arbitrary timeline, but instead, looked for an opportunity to advance the company’s mission, well into the future. The arrival of Sridhar Ramaswamy through the acquisition of Neeva last year represented that opportunity. Since joining us, Sridhar has been leading Snowflake’s AI strategies, bringing new products and features to market at an incredible pace.”

Snowflake hasn’t made a profit since going public, being run for growth. The company now has 9,437 customers, with 461 having trailing 12-month product revenue greater than $1 million, 39 percent more than a year ago, and 691 Forbes Global 2000 customers, representing 39% and 8% year-over-year growth, respectively. It obviously isn’t making a profit from them. 

If 9,437 customers brought in an average of $82,092 each this quarter and Snowflake made a $170 million loss, then it would have needed 2,071 more customers just to break even. Put another way; on these numbers Snowflake needs a 22 percent customer count increase to start making a profit.

Quarterly financial summary

  • Free cash flow: $324.5 million vs $205 million a year ago
  • Cash, cash equivalents and restricted cash: $1.78 billion
  • Gross product margin: 74 percent

AI is the key market driver, with Ramaswamy saying: “You heard the team say many times, “There’s no AI strategy without a data strategy.” And this has opened a massive opportunity for Snowflake to address. To deliver on the opportunity ahead, we must have clear focus and move even faster to bring innovation on the Snowflake platform to our customers and partners. This will be my focus.”

Apart from AI in the USA, Snowflake has a large revenue opportunity outside the USA, where 80 percent of its revenues are based. The EMEA region accounts for 15 percent of them and Asia Pacific a trivial 5 percent.

Snowflake is guiding next quarter’s product revenue to be $747.5 million +/- $2.5 million; a 26.5 percent increase at the midpoint year on year. Full fy2025 revenue guidance is $3.25 billion, with a lower 22 percent Y/Y growth rate. That’s slower growth than Snowflake has been used to enjoying. Perhaps competitors suchas Databricks and Dremio are finally having an impact on its growth.

Pure projects double-digit growth despite revenue dip

Pure Storage has ended its fiscal year as it started – with a drop in revenues. This time it’s due to subscriptions becoming more popular than anticipated, but the company says it expects to resume double-digit growth from the next financial quarter.

Revenues in Pure’s Q4 of fiscal 2024, ended February 4, declined three percent year-on-year to $789.8 million yet beat its own guidance. Net profit dropped 12 percent to $65.4 million. Full year revenues were $2.8 billion, up 3 percent to surpass guidance, and net earnings fell 16 percent annually to $61.3 million profit.

Charles Giancarlo, Chairman and CEO at Pure Storage, said on the earnings call: “We had a solid Q4 performance and ended the year with increasing sales momentum and balanced performance across our theaters and product portfolio. This momentum and growing customer interest in our platform strategy provides us with increased confidence for the coming year.”

Pure noted that subscription revenue of $329 million was up 24 percent while turnover from product sales was down 15 percent to $460.9 million. Giancarlo said: ”We certainly see our way clear to a majority of revenue over time being in the subscription category.”

Quarterly financial summary

  • Gross margin: 72 percent
  • Free cash flow: $200.9 million
  • Operating cash flow: $244.4 million
  • Total cash, cash equivalents and marketable securities: $1.5 billion

What was responsible for the revenue decline? Pure had better than expected sales of Evergreen//One storage-as-a-service (STaaS) and Evergreen//Flex pay-for-what-you-use storage deals at the expense of perpetual license sales. The latter’s revenues are recognized on shipment while the former’s are recognized over the life of the services contract.

During the conference call, Pure’s confirmed its file and object FlashBlade array has brought in more than $2 billion in total sales since launch. Pure has won an eight-figure Evergreen//One FlashBlade deal with one of the largest GPU cloud providers for arrays to be used in AI processing and training stages, it said. Pure, however, is not the only storage supplier to this unnamed customer.

Giancarlo said Pure’s Portworx container storage product “had a record year and accelerated growth based on customers increasingly graduating their container-based development projects to production scale.” The financial sector was highlighted as adopting Portworx in this way.

Answering an analyst’s question, Giancarlo said there were three aspects to AI for Pure: modeling (training) needing lots of GPUs, inference with fewer GPUs, and storage. Storage for AI needs to be in a single fast-access data environment, one that is not spread across silos. He reckons the training market is the smallest one of the three in terms of total market size.

Pure also thinks that the market researchers including Gartner underplay Pure’s market share because they are biased towards perpetual license (capex) sales. Giancarlo said: “They do not incorporate our Evergreen//Forever subscription, which, as you know, means that we don’t resell the same storage when an array becomes obsolete because in our case, arrays don’t become obsolete.”

He said that an analyst’s “basic premise that the market is under accounting, what we believe would be our growth if our Evergreen//One sales were actually in standard product sales, is absolutely correct.”

Pure did not see more sales into Meta’s Research Super Cluster (RSC) this quarter, which meshes with what Hammerspace is claiming, but Pure CTO Rob Lee said: “Our relationship with Meta is stronger than ever. We’re working with them on almost a continuous basis.”

Inevitably Pure was asked about flash replacing disk, and Giancarlo stepped up to that with no hesitation: “I started saying last year that I expected the last disk storage array to be sold in about five years’ time. We’re now four years in front of that. I’m going to stick with that timetable … I am very bullish on this, and I’ll stick with it, that I think the next three to four years, we’re going to see the decline of disk systems.”

Turning to the future, Giancarlo said: “We are beginning to see some encouraging signs of improvement in the macro environment.” CFO Kevan Krysler said in a prepared statement: “We expect double-digit revenue growth and strong growth of RPO, fueled by our highly differentiated data storage platform, and strength of our Evergreen and Portworx consumption and subscription offerings.”

With this in mind, next quarter’s revenues are expected to be $680 million, 15.4 percent up on the year-ago quarter. Pure is looking for 2025 revenues to be around $3.1 billion, meaning a 10.7 percent rise on 2024.

Cohesity pushes out RAG-enhanced Gaia GenAI backup search chatbot

Cohesity has announced a Gen AI chatbot called Gaia that can search through a customer’s backups to find answers to conversational questions.

This builds on its Turing AI/ML gateway initiative which used AWS Bedrock to generate business context-aware answers to questions. Cohesity claims it’s the industry’s first Gen AI-powered conversational search assistant to help businesses transform secondary data into knowledge. That’s because retrieval augmented generation (RAG) adds a company’s backup data in the Cohesity Data Cloud to the Gen AI’s Large Language Model (LLM) inputs.

Sanjay Poonen

Cohesity CEO and President Sanjay Poonen said in an canned statement: “Enterprises are excited to harness the power of generative AI but have faced several challenges gaining insights into secondary data, including backup, archived and vaulted data – because every approach requires re-hydrating the data, and painfully waiting weeks for the data to be available for analytics and insights. Cohesity Gaia dramatically simplifies this process with our patent-pending approach using Retrieval Augmented Generation.”

This means that tasks previously carried out by skilled and expensive data scientists querying data warehouses and the like, with specialized coded programs, can now be done by ordinary managers and employees querying backup data.

Poonen said: “Our approach delivers rapid, insightful results without the drawbacks of more manual and risky approaches. In short, it turns data into knowledge within seconds and minutes.”

Gaia is a Saas offering integrated with Cohesity’s Data Cloud and features:

  • RAG AI conversational search experiences across cloud and hybrid environments. 
  • A fully indexed database of all files, across all workloads, and at all points in time that supports the creation of AI-ready indexes for rapid conversational search and responses. Initially, Cohesity will support Microsoft 365 and OneDrive data and will expand to more workloads over time.
  • All indexed data is immediately available for reading without the need for backups to be reconstructed, allowing the Cohesity Data Cloud to function like a data lake with real-time access to data
  • Granular role-based access controls and zero-trust security principles to protect access to sensitive information and help maintain compliance with various regulatory requirements.

Cohesity says customers can use Gaia:

    • To assess an organization’s level of cyber resilience.
    • To perform financial and compliance audit checks.
    • To answer complex legal questions.
    • To serve as a knowledge base to train new employees.

    It has announced plans with the public cloud big three; Azure, AWS and Google, to bring their LLM services to Gaia.  

    As an example of Gaia use, Cohesity’s Greg Statton from its Office of the CTO, said: “If you notice a rise in costs in a region, typically, you would search for dozens of invoices, review and compare them, and see if you can discover the reason for the cost increases. It could take hours, days, or weeks to resolve. With Cohesity Gaia, you simply ask, ‘Why have costs increased in the region?’, and Cohesity Gaia will pull the relevant data from your stored data, analyze it, and return an answer to your question. It’s that simple.” 

    A spokesperson for Rubrik, which announced its Ruby Gen AI chatbot in November last year, said: “If you don’t have a GenAI product available today, you’re not in the game. We rolled out Rubrik Ruby last year and it’s already available.”

    Gaia will be made generally available on March 15. Get a PDF solution brief here.

    Nutanix makes its first ever profit

    3D rendering of AI brain

    Nutanix has made its first ever profit as its revenues grow 16 percent year-on-year in its ninth successive growth quarter. VMware and CIsco represent clear opportunities while its Gen AI strategy is still emerging.

    Fifteen years after being founded and eight years after its IPO, hyperconverged infrastructure vendor Nutanix has reached business maturity. Revenues in its second fiscal 2024 quarter, ended Jan 31, were $565.2 million compared to the year-ago $486.5 million, and there was a profit of $32.8 million. Quite a turnaround from the $70.8 million loss last year. The company is still growing its customer base, with 440 new customers in the quarter, taking the total to 25,370.

    President and CEO Rajiv Ramaswami was, as usual, measured in his comment, saying: “We’ve delivered a solid second quarter, with results that came in ahead of our guidance. The macro backdrop in our second quarter remain uncertain, but stable relative to the prior quarter.”

    Nutanix is maintaining a quite consistently high growth rate after a slowdown in 2018-2021 period. Ramaswami became the CEO at the end of 2020. He rescued it from a revenue slump.

    He said: ”We achieved quarterly GAAP operating profitability for the first time in Q2, demonstrating the progress we continue to make on driving operating leverage in our subscription model.”

    Financial summary

    • Gross Margin: 85.6 percent vs 82.2 percent a year ago
    • Annual Recurring Revenue: $1.74 billion
    • Free cash flow:  $162.6 million vs year-ago $63 million

    William Blair analyst Jason Ader said it was another beat (the guidance) and raise quarter, with Nutanix “beating consensus across all key metrics, despite ongoing macro pressure.” 

    The future

    Nutanix sees steady demand ahead. The Broadcom-VMware acquisition is still creating uncertainty for VMware customers and where else should a worried VMware customer go but to the safe and steady haven represented by Nutanix? This is a multi-year opportunity for Nutanix, especially with 3-tier (separate server, storage and client GUI access) client-server VMware customers who could embrace HCI for the first time with Nutanix instead of inside the VMware environment. And also, of course, vSphere and VSAN customers who could migrate fairly simply to Nutanix.

    The Cisco alliance, replacing Cisco’s Hyperflex HCI with Nutanix’ offering is a second tailwind, set to become significant revenue-wise in fy2025. Gen AI is a third potential tailwind.

    Nutanix’ presence in the GenAI market is represented by its GPT-in-a-box system on which to run Large Language Models (LLMs). This was announced six months ago. Nutanix still does not support Nvidia’s GPUDirect protocol and thus GPT-in-a-Box looks like a relatively slow deliverer of data to external Nvidia GPU servers.

    Ramaswami said in the earnings call: “Moving on, adopting and benefiting from generating AI, is top of mind for many of our customers. As such, interest remains high in our GPT in-a-Box offering, which enables our customers to accelerate the use of generative AI across their enterprises, while keeping their data secure.” He talked about a federal agency customer for the system, along with other wins, and: “While it’s still early days, and the numbers remain small, I’m excited about the longer term potential for GPT in-a-Box.”

    He added: “We think Nutanix systems could run Gen AI Inferencing on their X86 servers but are not suited to Gen AI training as they have no GPU support. Ramaswami said the GPT-in-a-Box use cases include “document summarization, search, analytics, co-piloting, customer service, [and] enhanced fraud detection.”

    He characterized customer’s AI involvement as being hybrid multi-cloud in style: “It’s starting out with a lot of training happening in the public cloud, but it’s moving towards, okay, I’m going to run my AI, I’m going to fine tune the training, on my own data sets, which are proprietary that I want to keep carefully. And I’m going to have to potentially look at slightly different solutions for inferencing, which are going to be running closer to my edge locations.”

    The proprietary dataset fine tuning could run on-premises if there were GPU servers there. The GPU-in-a-Box supports GPU-enabled server nodes. These are x86 servers with added GPU accelerators, such as HPE’s ProLiant DL320, DL380a and DL395 which come with Intel or AMD CPUs and Nvidia L4/L40/L40S GPUs. It could run in GPU server farms, such as ones operated by CoreWeave and Lambda Labs. The data would need to be sent there and, again, Nutanix does not support GPUDirect

    Comment

    In the Ramaswami era Nutanix is a judicious company. It has not yet seen a need to add GPUDirect support to pump files faster out to GPU servers. This could mean that it is effectively excluded from large-scale Gen AI training and is an on-premises Gen AI inferencing supporter and small-scale, fine tuning trainer instead. 

    BitRipple moves liquid data using RaptorQ

    Profile. Startup BitRipple has a data mover that can deliver large volumes of data in real-time across wireless networks by using RaptorQ erasure coding and a liquid data concept.

    It says its technology is used in a variety of applications, including defense, streaming video, automotive, space satellite communications, and more. BitRipple delivers massive volumes of data at high-speed with robust security, and ultra-low latency.  This positions BitRipple as a high-speed data mover, along with Vcinity and Zettar.

    BitRipple was co-founded  up in April 2020 by CEO Michael Luby,  VP Systems Pooja Aggarwal and VP Engineering Lorenz Minder. Luby spent 5½ years at the International Computer Science Institute in Berkeley, CA, finishing as a Senior Research Scientist in early 2019. Before that he worked at Qualcomm for almost 10 years, becoming VP Technology. Back in 2009 he founded Digital Fountain and was its CTO until early 1998. It developed broadcast and real-time data transport software and was a predecessor of BitRipple.

    RaptorQ is an example of a computer science fountain code or rateless erasure code. Their defining property is that a potentially limitless sequence of encoding symbols can be generated from a given set of source symbols. The original source symbols can ideally be recovered from any subset of the encoding symbols of size equal to or only slightly larger than the number of source symbols, which limits network overhead.

    The abstract of a 1998 ACM document, to which Luby contributed, said: “A digital fountain allows any number of heterogeneous clients to acquire bulk data with optimal efficiency at times of their choosing. Moreover, no feedback channels are needed to ensure reliable delivery, even in the face of high loss rates.We develop a protocol that closely approximates a digital fountain using a new class of erasure codes that for large block sizes are orders of magnitude faster than standard erasure codes.”

    Digital Fountain pioneered RaptorQ use, selling Linux appliances, and was bought by Qualcomm in 2009, which is how Luby ended up in Qualcomm.

    BitRipple describes its technology like this: “The essential idea is that, using RaptorQ, a sender generates and sends liquid packets for each data block to be delivered to a receiver. They are called liquid packets because, like drops of a liquid, they are interchangeable: using RaptorQ, the receiver recovers each block as soon as enough liquid packets arrive at the receiver, independent of which liquid packets arrive.”

    RaptorQ erasure code is specified in IETF RFC 6330. Qualcomm provides a RaptorQ Technical Overview, which says: “Raptor is a forward error correction (FEC) technology implemented in software that provides application-layer protection against network packet loss. … The RaptorQ encoder and decoder software libraries allow streaming and file delivery services to recover data lost in transit and completely reconstruct it, without using a backchannel.”  

    Specifically: “RaptorQ encodes and decodes a block of source data, called a source block, which is partitioned into equal-size pieces of data, called source symbols. The source block size is configured by the application that incorporates the RaptorQ software library based on the application requirements. The RaptorQ encoder generates repair symbols from the source symbols of a source block, where the repair symbols are the same size as the source symbols and the encoded symbols that can be sent consist of the combination of the source symbols and the repair symbols.”

    “Typically, each encoded symbol is sent in an individual packet together with a 32-bit header, called the FEC Payload ID consisting of an 8-bit source block number and a 24-bit encoded symbol identifier (ESI) that allows the receiver to identify the encoded symbol carried in the packet.” 

    YouTube BitRipple video.

    BitRipple says it uses an intelligent delivery protocol which “automatically adjusts on-the-fly the number of liquid packets to send for each block based on continual feedback from the receiver. The number of liquid packets proactively sent for a block is close to the minimal needed to recover the block when packet loss is low, saving on bandwidth. More liquid packets are proactively sent for a block when packet loss is higher, ensuring that enough liquid packets arrive at the receiver to recover the block without re-transmisson delays.”

    A YouTube video shows remote BitRipple-delivered 4K video almost as fast as local playback and much faster than HTTP Live Streaming (HLS). 

    There are three BitRipple products:

    • Mono – data package delivery over a 1-way connection with no back channel from receiver to sender.
    • Tunnel – accelerates data movement between pairs of endpoints over any network.
    • Multipath Tunnel -accelerates data movement between pairs of endpoints utilizing multiple network paths.

    BitRipple says it enable immersive experiences, such as cloud gaming, remote collaboration, augmented and virtual reality, delivered to geographically-distributed endpoints across wireless networks. It is a quiet, relatively low-profile company, making ripples in the data moving space, and not noisy splashes.

    Mechanical Orchard updates mainframe migration with AI

    Two-year-old startup Mechanical Orchard migrates mainframe apps to the public cloud using generative AI technology.

    Mainframe systems are used by 44 out of the top 50 banks, ten out of ten top insurers, and 23 out of 25 top US retailers. Some 71 percent of the Fortune 500 are still reliant on mainframes, and spend in excess of $20 billion a year buying and maintaining them. The mainframe is a restricted and costly environment. Moving apps from it to the public cloud means they can interoperate easily with open source code, get away from historic COBOL programming, become SaaS-style entities, and be more agile and able to develop their previously fossilized code.

    Rob Mee, Mechanical Orchard
    Rob Mee

    Mechanical Orchard was founded in San Francisco in 2022 by three ex-Pivotal Labs guys, CEO Rob Mee, COO Matthew Work, and VP R&D Dan Podsedly. It took $7 million in a seed round that year and has just raised $24 million in an A-round, having demonstrated substantial progress. It has 50 staff and offices in the UK, Ireland, Italy, and Germany. Customers include Omni Logistics.

    ‍It says it uses iterative, AI-enhanced, and reverse engineering approaches to move large enterprise apps off mainframe systems and into the cloud. During a briefing in London, Mee told us that Mechanical Orchard’s method was to look at a mainframe or other legacy system application and migrate it to the public cloud incrementally, module by module. Mee said “there is no silver bullet, and never will be” that can be used to move legacy apps and their components en masse to the public cloud. This migration requires a meticulous, iterative, and incremental approach.

    Mainframes have distinctive hardware and system software architectures that do not translate easily to x86 compute, commodity-based storage, and networking-based infrastructure. There is no straight equivalence between mainframe Direct Access Storage Devices (DASD) and the main SSD and HDD instances available in AWS, Azure, and Google. Mainframe apps could also have code components that are decades old and may not have documented source code.

    A typical customer engagement starts with a pilot and looks like this. It selects a piece or component of the application that has defined inputs and outputs. If it doesn’t have the source code, Mechanical Orchard treats it as a black box with data flows, which it reverse engineers into a cloud-native application. This is loaded into an isolated sandbox and tested with various inputs, amended as needed when the output is different from what is expected, until it parallel runs with the same outputs as the source legacy module and is just as fast if not faster. This pilot can take six to eight weeks. 

    That is then subject to the customer’s judgement as to whether to take it further. If it’s a yes, component by component, the mainframe application is reverse engineered to use cloud-native code based on public cloud compute, networking, and storage instances. It then runs in parallel with the source mainframe component with equivalence monitoring looking for and detecting different outputs that need understanding and correcting, and checking performance.

    Generative AI is being used to help build a visual representation of the mainframe app component and what it does. Mee said AI can synthesize data, identify coding patterns, and speed up code development and documentation. It helps Mechanical Orchard staff be more productive.

    When full equivalence is attained, says the firm, the customer can switch over from the mainframe to the public cloud component. The switchover point varies with the criticality of the software. A financial trading module will be tested far more severely than a retail system for example. AWS, Azure, and Google’s public clouds are supported. In fact, the three partner with Mechanical Orchard because it represents a mainframe app and data on-ramp for them. Mechanical Orchard represents a different approach from using public cloud mainframe emulation, which leaves the source app still frozen.

    This is a service-based operation such as Arthur Andersen or Price Waterhouse might provide. For customers, they see that the Mechanical Orchard process costs money, but when it is complete they have a migrated mainframe app that gives them the ability to reduce mainframe costs and move into a far more scalable, agile, and fast-moving development environment than mainframe land. Mainframe digital transformation moves at glacial speed. Cloud-native development – DevOps with continuous integration and delivery – is much faster and there is a far larger base of skilled coders available. 

    That’s Mechanical Orchard’s would-be USP – dependable mainframe jailbreaking that is carried out proof point by proof point, component by component, until the frozen mainframe app is public cloud resident and ready to be developed as the business needs.

    Bootnote

    Pivotal Labs, an application development, containerization, and Kubernetes business, was founded in 1989. It was acquired by EMC in 2012 and spun-out in 2013. Rob Mee founded Pivotal and became the CEO that year. The company went public in April 2018. VMware completed a $2.7 billion acquisition of Pivotal in December 2019, at which point Mee left.

    Storage news ticker – February 28

    German system house weSystems is working with partner Atempo and its Miria data mover to archive 40 PB of data in the Azure public cloud’s Blob Storage Archive Tier. The archive belongs to a German automotive manufacturer and is growing at 2 PB/month. A September 2023 weSystems case study said: “Until now, our customer, a well-known German automotive supplier, has stored the data volumes accumulating in daily production on an EMC Isilon storage. A single file can reach a size of 500 GB. At regular intervals, the data was transferred to tape systems for backup and long-term archiving.” Now a hybrid backup and archiving system transfers data to central Ceph storage in a secure data centre at high data throughput. Specifically, most of the data is first transferred to weSystems FlexStorage, based on Ceph, in a central German datacenter and stored there for approx. 3 months. Data can be retrieved from it using Miria. In a second step, the data is then transferred to the Azure Archive Tier, as the customer specified, with Miria retrieval again.

     …

    Cloudian has released new open source software that integrates machine learning library PyTorch with local data lakes running on Cloudian HyperStore S3-compatible object storage, eliminating the dependency on a dedicated parallel file system for ML workflows. Data scientists and AI developers can run ML on data resident in local Cloudian object storage, without the need to move and stage the data into another system. The ML tasks can also run on local compute resources such as AWS Outposts and Local Zones. Cloudian is a certified Service Ready partner for AWS Outposts and Local Zones, and is commercially available through the AWS Marketplace.

    Cloudian contributed enhancements to AWS Labs’ open source S3-Connector-for-PyTorch to enable PyTorch ML algorithms to access data in Cloudian’s HyperStore object storage system via the AWS S3 API. The enhanced S3 connector is available from the GitHub repositories of AWS Labs and Cloudian.

    The 2024 AEC Data Insights Report from data collaborator Egnyte highlights the exponential growth in the AEC industry’s transition from local storage to cloud storage and collaboration solutions. Data storage requirements of the companies surveyed grew at an average 50.3 percent CAGR. This rapid growth trend began during the pandemic, but as companies started to see the benefits of the cloud, growth has maintained that pace. Transitioning to the cloud allows companies to leverage technology integrations, artificial intelligence, real-time collaboration, and meet strict regulatory requirements, such as the Cybersecurity Maturity Model Compliance (CMMC) in the United States and the Building Safety Act in the United Kingdom. To download a copy of the report, click here.

    Can Hammerspace’s HyperNAS really add GPUDirect speed to non-GPUDirect-supporting filers? An industry source says GPUDirect does RDMA direct from the storage system into GPU memory. Let’s suppose you add HyperNAS to a Qumulo system, which doesn’t natively support GPUDirect. HyperNAS provides GPUDirect Hammerspace-held metadata client access but doesn’t provide an RDMA path out of the Qumulo system. Its data, the source suggests, wouldn’t flow at RDMA-based GPUDirect speed. A benchmark test would be necessary to verify this claim.

    S3 tiering to tape with Red Hat-acquired NooBaa’s gateway is discussed in a four-part IBM Storage blog by Senior Technical Staff Member Nils Haustein and others. The blog talks about a new and modern S3 object storage services, using open source NooBaa, that can be installed on IBM’s Storage Scale. NooBaa provides S3 endpoints to users and applications, and allows full control over data placement with dynamic policies per bucket or account. The NooBaa-core standalone software package can be deployed on a Storage Scale cluster providing the S3 endpoints while the buckets and objects are stored in Storage Scale file systems. By leveraging the integration of Storage Scale with Storage Archive, S3 buckets and object stored in the file system can be tiered to tape.

    SaaS app backup supplier Keepit has opened a UK headquarters in the City of London and expanded its UK team, led by industry veteran Jerry Mumford. Over the past three years, Keepit has seen 100 percent year-over-year growth in the UK and is on target to repeat this in 2024. Keepit counts The National Gallery and Oxford University Innovation Ltd among its UK customers.

    Tech Radar has written about MaxDigitalData, a GoHardDrive brand, which sells renewed (used and refurbished) and end-of-life disk drives. A 12TB drive with a three-year warranty costs less than $90. A 14 TB SATA, 7,200 rpm drive costs $129.95. Reddit has a thread about the company and its products.

    Nimbus Data CEO Tom Isakovich presented a new FlashRack Turbo all-flash system at an A3 Technology Live event in London. It’s a follow-on from the existing ExaFlash array, and composed of a scaled-out, independent sets of redundant 2RU nodes, which use off-the-shelf NVMe TLC SSDs. There is no high-speed cluster interconnect. A node’s base chassis has two stateless x86 controllers and 24 drives, with two 2RU x 24-drive expansion shelves supported. Max capacity is 4.5 PB raw (to 15 PB usable); raw 1.5 PB/chassis, meaning 62.5TB drives. Asked if these were Solidigm QLC or TLC drives, Isakovich said: ”I can’t talk about that.”

    The software supports block, file, and object access with one tier and logical space, with Ethernet (400g), FC, and InfiniBand connectivity supported. The array does about 50 GBps read and 40 GBps writes. Isakovich called it “the promised dream of the ideal storage platform” and claimed 10 PB of Nimbus costs 80 percent less than 10 PB of VAST. He said Nimbus is 20 years old, privately financed, has 30 staff, 200 customers (95 percent in the US), and 700 deployed systems. Nimbus is working on having its software delivered on an adapter card (DPU-esque) and on an intelligent, Ethernet-connected SSD.

    Onna Technologies enables users to gather, analyze, and act on data to help minimize legal risk. It announced its data management platform has achieved Powered by Oracle Cloud Expertise and is now available in the Oracle Cloud Marketplace. Onna’s scalable platform supports multiple workflows across an organization by offering more than 20 pre-built connectors, multiple export options, and an intuitive interface.

    Disk drive supplier Seagate has launched a new ecommerce website in the US. Customers who register will have first access to special pricing, free shipping offers, and other promotions. At launch, Seagate is offering an exclusive discount on its most popular gaming storage drives including the Xbox Storage Expansion Card for Xbox Series X|S 1TB at $129.99 (regularly $219.99), and Game Drive PS5 NVMe SSD for $89.99 (regularly $124.99). The company plans to expand the website to other regions later this year.

    StorPool Storage has extended the capabilities of its Apache CloudStack integration in light of skyrocketing demand from VMware customers. CloudStack is an open source cloud computing software platform designed to deploy and manage large arrays of virtual machines as a highly available (HA), highly scalable IaaS system. CloudStack’s current KVM-hypervisor host HA feature depends on having NFS primary storage detect which nodes are online, even if other types of primary storage are used, such as faster block-level storage. StorPool extends the HA feature by adding a mechanism that overcomes the NFS primary storage heartbeat dependency to extend KVM host high availability for non-NFS storage in CloudStack environments. Future enhancements will support KVM host HA for hyperconverge infrastructure (HCI) deployments with CloudStack compute and StorPool storage running concurrently on the same servers.

    TerraMaster has launched the F4-212 four-bay private cloud NAS for backup and home multimedia center use. The F4-212 is equipped with an ARM V8.2 Cortex-A55 64-bit 1.7GHz quad-core Realtek 1619B processor and independent NPU, built-in floating point unit (FPU) and NEON SIMD engine, video DSP hardware acceleration. Compared with the previous generation RTD1296, the performance is improved by 40 percent. The F4-212, running the latest TOS 5.1, meets diverse business needs, and enhances speed and security. It backs up data from Windows PC, Mac, mobile, HDD enclosure, and network disk for centralized management.

    Micron touts beefy HBM chips, and hot DRAM, Samsung’s releasing fresh stacked memory tech too

    Micron has launched both a fingernail-sized flash storage device for smartphones and a large high bandwidth memory chip for Nvidia GPUs.

    Flash and DRAM foundry operator Micron builds and sells SSDs. The flash chip is an enhanced version of a UFS 4.0 format device, 9 x 13 mm in size, with up to 1 TB in capacity using Micron’s 232-layer 3D NAND technology. It delivers twice the performance of UFS 3.1 specification devices, pumping out up to 4,300 MBps sequential reads and 4,000 MBps sequential writes. Smartphone apps can be launched in less time and run faster, and the device can be used in automotive applications as well.

    Micron’s Mobile Business Unit GM and corporate VP, Mark Montierth, said in a statement: “Micron’s latest UFS 4.0 solution enables world-class storage performance and reduced power in the world’s smallest UFS package. Supercharged with breakthrough firmware advancements to keep smartphones running like new, Micron UFS 4.0 raises the bar for mobile storage with enhanced performance, flexibility and scalability to accelerate the rollout of generative AI-capable smartphones.” 

    The company claims large language models in generative AI apps can be loaded 40 percent faster than Micron’s prior 176-layer UFS 3.1 product, resulting in a smoother experience when initializing conversations with chatbots. This prior product was physically larger, measuring 11 x 13 mm, and needed more electricity. The new one is said to be 25 percent more power-efficient.

    The chip’s firmware has three new features:

    • High-Performance Mode (HPM): This optimizes performance by prioritizing critical tasks over background ones during intensive smartphone use, with an over 25 percent speed improvement when launching multiple apps over the UFS 3.1 product due to twice-as-fast storage access during heavy usage.
    • One Button Refresh (OBR): Automatically cleaning and optimizing data so smartphones can continue operating in a like-new state. 
    • Zoned UFS (ZUFS): The host can specify different zones where data can be stored, improving the usefulness of the device over time. This ZUFS approach reduces write amplification to extend the flash lifespan, keeping devices feeling like new for longer. 

    Micron is now shipping samples of its upgraded UFS 4.0 drive in capacities of 256 GB, 512 GB, and 1 TB. It announced that both HONOR and Samsung are using its LPDDR5X memory and UFS 4.0 mobile flash storage in their latest AI-driven smartphones, the HONOR Magic6 Pro and the Samsung Galaxy S24 series.

    Nvidia HBM3e

    Micron is building 24 GB HBM3e (High Bandwidth Memory gen 3 extended), with 8 x 3 GB stacked DRAM dies, for use by Nvidia in its H200 Tensor Core GPUs, due to ship next quarter. The H200 is a high-end GPU, a refresh of the existing Hopper architecture H100 intended for AI processing work. It should come with up to 141 GB of HBM3e memory and 4.8 TBps of bandwidth, meaning up to six Micron HBM3e chips per H200.

    We’re told that the pin speed of Micron’s HBM3e device is >9.2 Gbps and its overall speed is >1.2TBps. Micron claims it has around 30 percent lower power consumption than unidentified competitive offerings.

    EVP and Micron Chief Business Officer Sumit Sadana said: “Micron is delivering a trifecta with this HBM3e milestone: time-to-market leadership, best-in-class industry performance, and a differentiated power efficiency profile.

    “AI workloads are heavily reliant on memory bandwidth and capacity, and Micron is very well-positioned to support the significant AI growth ahead through our industry-leading HBM3e and HBM4 roadmap, as well as our full portfolio of DRAM and NAND solutions for AI applications.” Getting Nvidia as a customer is a good win for Micron, and steals a march on rivals Samsung and SK hynix. In its last earnings call, back in December, Micron CEO Sanjay Mehrotra said: “We are on track to begin our HBM3e volume production ramp in early calendar 2024 and to generate several hundred millions of dollars of HBM revenue in fiscal 2024. We expect continued HBM revenue growth in 2025.” The Nvidia deal must have been visible to Mehrotra then.

    Micron will be sampling a larger HBM3e product in March, with a 12-high stack of DRAM chips yielding 36 GB in capacity, and >1.2 TBps performance again.

    But Samsung has announced its own 12-high, 36 GB HBM3e device, with up to 1.28 TBps bandwidth. It builds the 12-layer device using a thermal compression non-conductive film (TC NCF) to squeeze it down to the same height as a 8-layer device. This technique mitigates chip die warping that can come with a thinner die. Sammy says it’s lowered the thickness of its NCF material and achieved the industry’s smallest gap between chips at seven micrometers (µm), and eliminating voids between layers. This results in enhanced vertical density by over 20 percent compared to its HBM3 8-layer product.

    Yongcheol Bae, EVP of Memory Product Planning at Samsung Electronics, said: “This new memory solution forms part of our drive toward developing core technologies for high-stack HBM and providing technological leadership for the high-capacity HBM market in the AI era.”

    Expect more Micron HBM3e details at the Nvidia GTC event on March 18.

    IBM adds real-time malware IO detection to flash drives

    Malware detection
    Malware detection

    IBM has added real-time detection of ransomware and other attacks using machine learning to the firmware of its latest FlashCore Modules (FCMs).

    Update. IBM comments added at end. 4 March 2024.

    These FCMs are proprietary flash drives used in IBM FlashSystem 5000 and Storwize arrays. They come in a U.2 form factor with an NVMe interface. The existing gen 3 FCMs come in 4.8, 9.6, 19.2, and 38.4 TB usable capacity levels, with the 19.2 and 38.4 TB FCMs having a PCIe 4 connection. The other capacities use PCIe gen 3. Onboard compression increases effective capacity by up to 2.3x.

    Sam Werner, IBM Storage Product Management VP, blogs: “The FCM4 technology in new FlashSystem arrays is designed to capture and summarize detailed statistics about every I/O in real time. FlashSystem uses machine learning models to distinguish ransomware and malware from normal behavior, enabling organizations to take immediate action and keep operating in the event of an attack.”

    He says: “Existing IBM FlashSystem products scan all incoming data down to block level granularity without impact to performance as it’s being written, using inline data corruption detection software and cloud-based AI to help identify anomalies that might indicate the start of a cyber-attack, thereby enabling the system to detect, respond, and rapidly recover with immutable copies. The new technology enabled by FCM4 is designed to continuously monitor statistics gathered from every single I/O using machine learning models to detect anomalies like ransomware in less than a minute.”

    The gen 4 FCMs interoperate with IBM Storage Defender software, which incorporates Index Engines’ CyberSense code and Cohesity’s DataProtect offering. Storage Defender uses AI with event monitoring across multiple storage platforms to help detect ransomware, human error, and sabotage.

    Werner says: “FCM works with Storage Defender to provide end-to-end data resilience across primary and secondary workloads with AI-powered sensors that provide earlier notification of cyber threats so enterprises can recover faster.”

    Storage Defender has expanded its threat detection capabilities with the AI-powered FCM hardware “and software sensors that inform an industry leading index of the relative trustworthiness of copies, whether backup or primary snapshots. Additionally, IBM Storage Defender includes sensors developed by IBM Research that are engineered to detect potential threats, such as ransomware, in near real time and raise high fidelity alerts to security tools.”

    IBM has also added “workload and storage inventory management capabilities to IBM Storage Defender to help organizations assess their applications and data so they can properly incorporate all their assets in a business continuity plan to recover a minimum viable company after a cyberattack.”

    In Werner’s view: “Threat actors are now deploying AI-based cyber-attacks, and we must fight fire with fire. Our new FlashCore Module hardware and Storage Defender software both leverage IBM’s AI capabilities to help them better address this challenge.”

    Were standard SSD suppliers such as Micron, Samsung, SK hynix, and Kioxia/Western Digital to add similar AI-scanning of SSD IOs in real time, they would need to send attack alerts upstream to some system management resource to respond to the alerts. IBM’s Storage Defender is that resource for its customers. The commercial-off-the-shelf SSD suppliers don’t have similar functionality unless they partner with an upstream software vendor.

    Bootnote

    Barry Whyte, Principal Storage Technical Specialist and IBM Master Inventor, told us: “The FCM4 are PCIe Gen 4 across all capacities now. They are also now using “Charge-Trap” NAND technology which allows us to use the 176-layer NAND which gives faster programming times, best economics – but of course needs the intelligence in the computation storage that is FCM – no other vendor has anything close!  Also, the FCM are across all NVMe based FlashSystem (we dropped the Storwize name some time ago) so 5200,7300,9500 (the other smaller 5000, 5015, 5045 are still SAS based and can’t use FCMs)”

    Developing AI workloads is complex. Deciding where to run them might be easier

    SPONSORED FEATURE: If artificial intelligence (AI) has been sending shockwaves through the technology world in recent years, the onset of generative AI over the last 18 months has been a veritable earthquake. And for IT leaders looking to harness its potential for their own organisations, the pace of development can feel bewildering.

    Enterprises are racing to leverage their own data to either build their models or repurpose public models already available. But this can pose a significant challenge for the dev and data science teams involved.

    It can also present something of a conundrum for companies that want to keep control of the HPC infrastructure needed to support their AI workloads. AI-enabled applications and services require a far more complex mix of silicon than traditional computing, as well as accompanying storage capacity and connectivity bandwidth to handle the vast amounts of data needed in both the training and inference stages.

    London data centres reflect AI trends

    The potential for enterprise AI innovation and the challenges it presented is reflected by what is happening across colocation giant Digital Realty’s data centre estate in and around London as AI shifts to the top of hosting company’s customer agendas.

    The UK capital and its surrounding areas has a high density of headquarter buildings and R&D offices, not just in financial services, but in other key industry verticals such as pharma, manufacturing, retail, and tech.

    London is attractive because of the UKs political and legal stability, skilled workforce, and advanced tech infrastructure, explains Digital Realty CTO Chris Sharp, making it superb base both for innovation and for deploying AI applications and workloads.

    Many enterprises will be acutely aware of issues around the general importance of data and IP and specific issues around data sovereignty and regulation, he adds.

    “There’s a bit of nuance with training,” Sharp explains. “Nobody knows if it’s going to be able to be done anywhere and then inference has to abide by the [local] compliance [rules].” In addition, there’s an increasing understanding that one model cannot necessarily serve the world: “There’s going to be some regionality, so that will then also dictate the requirement for training facilities.”

    At the same time, these organisations face the same technology challenges as other companies worldwide, particularly when it comes to putting in place and powering the infrastructure needed for AI.

    It’s not enough to simply throw more CPUs at these workloads. One of the challenges with AI and HPC pipelines can be the different types of purpose-built hardware needed to efficiently support the complexity of these applications.

    These range from CPUs to GPUs, even application-specific tensor processing units (TPUs) designed for neural networks, all with subtly different requirements, and all potentially playing a role in a customer’s AI pipeline. “Being able to support the full deployment of that infrastructure is absolutely top of mind,” points out Sharp.

    Moreover, the balance between these platforms is set to change as AI projects move beyond development and into production. “If you take a snapshot, it’s 85 percent training, 15 percent inference today. But over the course of maybe 24 months, it’s 10 times more of a requirement to support inference,” he adds.

    Flexing your AI smarts

    So, the ability to flex and rebalance the underlying architecture as models evolve is paramount.
    There is also the challenge of connecting this vast amount of data and compute together to deliver the AI workload performance levels required. While customers in the UK will have data sovereignty very much in mind, they still need to process workloads internationally when needed. And they may need to tap data oceans around the world. As Sharp says, “How do you connect these things together, because you’re not going to own all the data.”

    But connectivity is not simply an external concern. “Within the four walls of the data centre we’re seeing six times the cable requirements [as] customers are connecting their GPUs, the CPUs, the network nodes. …. so, where we had one cable tray for fibre runs, now we have six times those cable trays, just to enable that.”

    Hanging over all of this are the challenges associated with housing and powering this infrastructure. Just the density of technology required raises floor loading issues, Sharp explains. “The simple weight of these capabilities is massive.” And, as Digital Realty has found working with hyperscale cloud providers, floor loading requirements can increase incredibly quickly as projects scale up and AI technology advances.

    Cooling too is always a challenge in data centres and as far as Sharp is concerned there is no longer a debate as to whether to focus on liquid or air cooling. “You need the ability to support both efficiently.”

    When combined with the sheer density of processing power demanded by AI workloads, this is all having a dramatic effect on power demand across the sector. Estimations published by Schneider Electric last year suggest AI currently accounts for 4.5 GW of demand for data centre power consumption, predicted to increase at a compound annual growth rate (CAGR) of 25-33 percent to reach between 14 GW and 18.7 GW by 2028. That’s two to three times more demand for overall data centre power which is forecast to see a 10 percent CAGR over the same period).

    All of which means that data centre operators must account for “more and more new hardware coming to market, requiring more power density, increasing in square footage required to support these burgeoning deployments.”

    A state of renewal

    That daunting array of challenges has informed the development of Digital Realty’s infrastructure in and around London, and its ongoing retrofitting and optimisation as enterprises scale up their AI operations.

    The company has six highly connected campuses in the greater London area, offering almost a million square feet of colo space. But that doesn’t exist in isolation, with over 320 different cloud and network service providers across the city. “What we’re seeing today is that customers need that full product spectrum to be successful,” Sharp says.

    Liquid cooling is a particular element in its London infrastructure. As liquid is 800 times denser than air, it can have a profound impact on efficiency. Digital Realty’s Cloud House data centre in London draws water from the Millwall dock for cooling, in a system that is up to 20 times more efficient than traditional cooling. Sensors ensure that only the required amount of water is used, and that it is returned to the dock unchanged.

    But this ability to match the demands of corporations in and around London today and for the future also depends on Digital Realty’s broader vision.

    All the power consumed by Digital Realty’s European operation is matched with renewable energy through power purchase agreements and other initiatives, while the company as a whole is contracted for over 1GW of new renewable energy worldwide.

    At a hardware level, it has developed technologies such as its HD Colo product, which supports 70KW per rack, representing three times the requirement of certification for the Nvidia H100 systems which currently underpin cutting edge HPC and AI architectures.

    At a macro level, as Sharp explains, Digital Realty plans its facilities years in advance. This includes “master planning the real estate, doing land banks and doing substations, making sure we pre-planned the power for five to six years.”

    This requires close coordination from the outset with local authorities and utility providers, including investing in substations itself.

    “We work extensively with the utility to make sure that not only the generation is there, but the distribution, and that they fortify the grid accordingly. I think that really allows customers of ours and our up the line suppliers, a lot of time to align to that demand.”

    Cooling, power and infrastructure management complexities

    It might be difficult to decide which is more complex. Developing cooling technologies and power management platforms that keep ahead of rapidly developing AI infrastructure or dealing with utilities and municipalities over a multiyear time horizon.
    But tackling both is crucial as organisations look to stand up and expand their own AI capacity both quickly, and sustainably.

    Sharp cites the example of one European education and research institution that needed to ramp up its own HPC infrastructure to support its AI ambitions, and knew it needed to utilise direct liquid to the chip. It would certainly have had the technical know-how to build out its own infrastructure. But once it began planning the project, it became clear that starting from scratch would have meant a five-to-six-year buildout. And that is an age in the current environment. Moreover, local regulations demanded it reduce their energy footprint by 25 percent over five years.

    But partnering with Digital Realty, Sharp explains, it was able to deploy in one year, and using 100 percent liquid cooling improved its energy efficiency by 30 percent. As Sharp puts it, “It really helped them out rather quickly.”

    Given how quickly the world has changed over the last 18 months, the ability to get an AI project up and running and into production that quickly is much more than a nice to have. For many enterprises, it’s going to be existential.

    “Many AI deployments have failed, because there’s a lot of science and complexity to it,” says Sharp. But he continues, “We spend a lot of time removing complexity.”

    Sponsored by Digital Realty.