Home Blog

Ultimate digital defense: five essentials for your backups

'recovery' key on keyboard
recovery key on keyboard

PARTNER CONTENT: Organizations are taking too many backups. It’s that simple.

While data is undeniably critical, the endless cycle of creating backups of backups to ensure business continuity has spiralled out of control, consuming massive amounts of storage and driving up costs. Backups are necessary, of course, but the proliferation is costly. Yet, with increasing threats like cyberattacks, system failures, and natural disasters, data is constantly at risk. The question remains – how does a backup solution ensure the right balance between storage costs, functionality and the appropriate level of security?

This is the space in which backup targets thrive. Adding a backup target to your backup solution enhances data protection by providing a secure and reliable destination for storing backup data, ensuring quick recovery and safeguarding against data loss or corruption. On top of all of this, it can save you money on storage, too.

This article explores five essential features that your overall backup environment must include to protect your organization’s most valuable asset: its data.

1. Cloud optimization

Not all data is equal, yet often lower value data (e.g. older, less vulnerable) is being stored in expensive cloud tiers. Cloud optimization offers significant benefits by allowing businesses to leverage scalable, low-cost object storage solutions for both structured and unstructured data. When properly implemented, cloud storage provides high scalability, ease of management, and cost-efficiency. By making strategic architectural changes and using technologies which work in synergy with your backup solution, organizations can avoid costly mistakes and optimize data protection. This approach minimizes cloud storage requirements and expenses, ensuring that only essential data is stored in the most cost-effective manner.

2. Security capabilities

Ransomware attacks are on the rise, threatening to cripple businesses by holding critical data hostage. Encryption in flight and at rest, though essential, is no longer enough on its own. Backup data needs increased protection in the form of immutable backup storage. Immutable backups provide robust data protection by ensuring that data cannot be modified, deleted, or encrypted by ransomware, making it invulnerable to cyber-attacks. This unassailable nature of backup data assures IT administrators of data availability for recovery in case of disasters or outages. Consequently, companies face reduced risk of ransomware payments, as cybercriminals have less leverage against organizations with secure and reliable backup systems, prompting them to target less protected entities.

3. Speed of upload and recovery

The speed of backup data uploads matter. Faster speeds mean shorter replication windows, allowing you to perform more frequent backups with minimal disruption. This is especially useful in areas with unreliable WAN connections. Quicker recovery times are essential for meeting Recovery Time Objectives (RTOs), ensuring that operations can resume swiftly after a disruption. This efficiency translates into increased productivity and significantly reduces the risk of data loss, protecting valuable information. In today’s fast-paced digital world, faster backup and recovery speeds give businesses a competitive edge, enhance customer satisfaction, and mitigate financial losses from unexpected downtime.

4. Intelligent protection

Imagine being able to detect unusual deviations in backup data in real time. Often, this is where data is manipulated during an attack. These deviations, while innocuous at first glance, could have a significant impact on backup data and the ability to recover. New intelligent functionalities such as AI-based anomaly detection learn the patterns around backup data flow and flag anomalies. Flagging anomalies at the storage layer offers the ability to get the deepest level of security protection–mandatory in today’s environment.

5. Cost and consumption

Exponential data growth can significantly strain your storage budget and resources. Traditionally, data deduplication has depended on hardware appliances or been restricted to a single vendor’s backup solution. To ensure you’re maximising your budget look for a solution with source-side data deduplication and built-in compression so you can capitalize on storage savings–particularly beneficial for cloud object storage.

In conclusion, an effective backup and recovery storage solution is vital for protecting an organization’s most valuable asset: its data. By incorporating features such as cloud optimization, robust security capabilities, rapid upload and recovery speeds, intelligent protection, and cost-effective consumption strategies, businesses can safeguard their data against threats while optimizing storage costs and maintaining high functionality. These elements ensure that data remains secure, accessible, and efficiently managed, enabling organizations to operate smoothly and resiliently in today’s fast-paced digital environment.

Quest® QoreStor® has the ability to address the key challenges faced by modern organizations, such as yours, in managing backup storage. By optimizing storage, enhancing data protection, and integrating seamlessly with cloud environments, QoreStor provides a comprehensive solution that meets the needs of today’s data-driven businesses. Its advanced security technology, environment versatility, and cost efficiency make it a compelling choice for anyone looking to optimize their storage infrastructure and safeguard their critical data. Learn more by clicking this link.

Contributed by Quest Software.

Technology combinations for complex chip projects

PARTNER CONTENT: Given the size and complexity of modern semiconductor designs, functional verification has become a dominant phase in the development cycle.

Coverage lies at the very heart of this process, providing the best way to assess verification progress and determine where to focus further effort. Code coverage of the register transfer level (RTL) design, functional coverage as specified by the verification team, and coverage derived from assertions are combined to yield a single metric for verification thoroughness.

Coverage goals are usually quite high (95 percent or more) and hard to achieve. Verification engineers spend weeks or months trying to hit unreached coverage targets to ensure that the design is thoroughly exercised and bugs are not missed. Traditionally this has involved a lot of manual effort, consuming valuable human resources and delaying project schedules. Fortunately, in recent years several powerful techniques have been developed to automate the coverage process, achieve faster coverage closure, and end up with higher overall coverage.

A presentation by NVIDIA at the Synopsys Users Group (SNUG) Silicon Valley 2024 event described a project in which the coverage enhancement techniques of test grading, unreachability analysis, and artificial intelligence (AI) were highly successful. The NVIDIA team carefully measured the impact across three generations of related chips, providing an exceptionally quantitative case study. The designs involved were large, with more than 100 million coverage targets. Many blocks were multiply instantiated, with unique tie-offs for each instance.

Project A – the baseline design

On the baseline design, Project A, this design topology made coverage convergence very challenging. The tie-offs left each instance with large unreachable cones of logic whose coverage targets could never be hit by any test. Each instance required its own unique set of coverage exclusions, so each instance had to be signed off for coverage independently. As shown in the following example (Project A) for one set of coverage targets, convergence using a constrained-random testbench was slow and a large manual effort was required to reach coverage signoff.

Click here to view Project A coverage results.

Some important design bugs were not found until late in the project, a cause for concern. The verification engineers wanted to accelerate coverage to find bugs earlier and to reduce the amount of manual effort required. The first technique they tried on the derivative Project B was test grading, available in the Synopsys VCS® simulator. Test grading analyzes the simulation tests and ranks them according to achieved coverage. This enables verification engineers to set up simulation regressions in which the most productive tests run more often, with more seeds, than less productive tests. Coverage converges more efficiently, saving project resources.

Test grading was a good first step, but the team still faced the challenge of the many unreachable coverage targets in the design. They found an effective solution with Synopsys VC Formal and its Formal Coverage Analyzer (FCA) application (app), which determines the unreachable coverage targets in the RTL design. This eliminates the traditional quagmire in which the verification team spends enormous time and resources trying to hit coverage targets that can never be reached.
Formal analysis conclusively determines unreachable coverage targets and removes them from consideration for future simulations. This benefits the overall coverage calculation:

Excluding the unreachable coverage targets boosts total coverage by eliminating apparent coverage holes that are actually unreachable and by reducing the total number of coverage targets to be hit in simulation. This is a completely automated process. The FCA app generates an exclusions file with the specific unreachable coverage points for each unique instance in the design. As shown in graph at the link below, the combination of test grading and unreachability analysis on Project B achieved a major “shift left” in coverage by two key milestones.

Click here to view Project A vs Project B coverage results.

Learnings from Project B

In their SNUG presentation, the NVIDIA engineers reported the following learnings from Project B:

– Focus early on test grading to improve stimulus productivity to hit more coverage
– Focus early on coverage to uncover bugs earlier, which increases design quality and saves integration effort
– Use automatic unreachability exclusion to save manual effort, focus verification efforts on reachable coverage gaps, and find bugs earlier
– Achieve a left shift in coverage and bug finding by applying test grading and unreachability analysis effectively
– Experiment with the tools, learn, and adjust to enhance verification methodologies

After the results of Project B, the verification was eager to try additional techniques to further shift left the verification process. For project C, they experimented with AI-based techniques, starting with the Synopsys VSO.ai Verification Space Optimization solution. It includes a Coverage Inference Engine to help define coverage points based on both simulated stimulus and the RTL design. It also uses connectivity engines and a machine learning (ML) based solver to target hard-to-hit coverage points.

The verification team first tried Synopsys VSO.ai in the late stage of Project C, using a constrained random testbench complaint with the Universal Verification Methodology (UVM). The results over using just test grading and unreachability analysis were impressive: adding VSO.ai achieved 33 percent more functional coverage in the same number of test runs while reducing the size of the regression test suite by 5X. Code coverage and assertion coverage improved by 20 percent in the same number of runs with an impressive 16X regression compression over the baseline.

Using a different set of baseline regression tests, the engineers experimented with the Intelligent Coverage Optimization (ICO) capability in Synopsys VCS. ICO enhances test diversity using reinforcement learning, resulting in faster regression turnaround time (TAT), faster coverage closure, higher achieved coverage, and discovery of more design and testbench bugs. ICO provides testbench visibility and analytics, including stimulus distribution histograms and diversity metrics. It also provides root cause analysis to determine the reasons for low coverage, such as skewed stimulus distribution or over/under constraining.

As shown in the final graph at the link below, applying ICO, VSO.ai, and unreachability analysis achieved 17 percent more coverage in the same number of runs with a 3.5x compression of regression tests compared to the baseline. Four unique bugs were also uncovered.

Click here to view Project C results.

Learnings from Project C

The NVIDIA team reported the following learnings from Project C:

– Better functional, code, and assertion coverage in the same number of runs
– Faster coverage, improved coverage, and better regression compression
– More bugs discovered due to better exercise of the design

The SNUG presentation concluded with a summary of the results from the three chip projects. Unreachability analysis provided the single biggest gain, boosting coverage metrics by 10-20 percent with minimal effort. The combination of technologies resulted in up to 33 percent better functional coverage with 2-7X regression compression on all testbenches. They found that ICO uncovered unique bugs and that VSO.ai could be used across all project milestones.

The recommendation from the NVIDIA verification engineers is that test grading be used from the very beginning of the project to improve stimulus effectiveness. VSO.ai should be used for early milestones, when stimulus is immature, to achieve high regression compression, and continued through late stage milestones for additional compression and for increasing the total coverage.

Finally, ICO and unreachability analysis should be enabled in mid-project to reduce compute resources, left-shift coverage by at least one milestone, and find unique bugs earlier. The combined power of all four technologies will benefit any complex chip project.

Contributed by Synopsys.

Get your data house in order for GenAI

COMMISSIONED: Organizations must consider many things before deploying generative AI (GenAI) services, from choosing models and tech stacks to selecting relevant use cases.

Yet before most organizations begin to tackle these tasks, they must solve perhaps their biggest challenge of all: their data management problem. After all, managing data remains one of the main barriers to creating value from GenAI.

Seventy percent of top-performing organizations said they have experienced difficulties integrating data into AI models, according to recent McKinsey research. These organizations experience issues with data quality, defining processes for data governance and having sufficient training data, McKinsey said. This can increase risks for organizations pursuing GenAI initiatives.

Getting your data house in order is table stakes for fostering AI capabilities while protecting corporate IP. But where do you start? And what data management and governance options are available?

Prioritize data quality and governance

Boosting data quality is a logical starting point. Large organizations are awash in data that could be useful for GenAI models and their resulting applications. However, the quality of data is often too poor to use without some corrections. Data, which is often siloed across different business functions, often includes wrong, outdated or even duplicative data.

This is par for the course in many organizations that have generated enterprise data over the years. However, using such disorganized data can wreak havoc on models, leading to bad outcomes, hallucinations and risk to corporate reputation. Remember, this is your organization’s IP, so you need to protect it.

How you massage your data to get the right outcomes will vary based on your business requirements. However, many organizations opt to collect, clean, preprocess, label and organize their data prior to leveraging it for training models.

Data governance is a critical factor for protecting corporate IP as you build GenAI models and applications. You’ll institute guidelines addressing AI usage within the organization and determine approved AI tools and usage policies.

Key to this is articulating a formal training policy to educate employees on how to use GenAI services ethically and responsibly, as well as the risks associated with inputting sensitive content into restricted gen AI systems.

Ultimately, however, a critical component of a good governance strategy is keeping a human-in-the-loop at all times. After all, isn’t it about time your humans and machines learn to work together

Synthetic data gives you secure options

Cleaning and governing your data will be good enough for many organizations dabbling in GenAI technologies. However, others may need to take a more prescribed approach when it comes to protecting their corporate IP.

For example, some GenAI use cases may be tough to execute as the data can be hard to obtain. And many organizations can’t afford to use their actual data, which may include personally identifiable data. This is particularly true in regulated markets, such as financial services, healthcare and life sciences bound to stringent data protection rules.

As a result, some organizations have turned to GenAI to use synthetic data, which mimics real-world patterns without exposing sensitive personal information. This can help you test data and see potential desirable outcomes.

It isn’t perfect; after all, the data is made up. But it may serve as a reasonable proxy for achieving your outcomes.

The unstructured data challenge

GenAI services produce unstructured data, such as PDFs, audio and video files, complementing the structured data stored in databases. Too many organizations let raw data flow into their lakes without cataloguing and tagging it, which can denigrate data quality.

Organizations typically wrangle the data with disparate tools and approaches, which challenges their ability to scale their initiatives.

To streamline their efforts, more organizations are turning to a data lakehouse, which is designed to work with structured and unstructured data. The data lakehouse abstracts the complexity of managing storage systems and surfaces the right data where, when and how it’s needed.

Dell offers the Dell Data Lakehouse, which affords your engineers self-service access to query their data and achieve outcomes they desire. The solution uses compute, storage and software in a single platform that supports open file and table formats and integrates with the ecosystem of AI and ML tools.

Your data is your differentiator and the Dell Data Lakehouse respects that by baking in governance to help you maintain control of your data and adhere to data sovereignty requirements.

The Dell Data Lakehouse is part of the Dell AI Factory, a fungible approach to running your data on premises and at the edge using AI-enabled infrastructure with support from an open ecosystem of partners. The Dell AI Factory also includes professional services and use cases to help organizations accelerate their AI journeys.

While organizations prefer their GenAI solutions to be plug-and-play, the reality is you’ve got to grab your shovel and come to work ready to dig through your data, prepare it to work with your models and protect it. Is your organization up to the task?

Brought to you by Dell Technologies.

Building blocks of AI: How storage architecture shapes AI success

Storage array
Storage array

COMMISSIONED: On a bustling factory floor, an advanced AI system orchestrates a symphony of robotic arms, each performing its task with precision. This AI-driven automation has revolutionized the manufacturing process, drastically reducing errors and increasing productivity. A core component of this complex AI system is sophisticated storage architecture, and without it none of this powerful automation would be possible, and it ensures the AI system has the data throughput, latency, and scalability to operate seamlessly. Without this robust foundation, the entire operation would grind to a halt. This anecdote underscores the critical role of storage architecture in AI success.

Data throughput and bandwidth: Fueling AI workloads

AI workloads often rely on the ability to process vast amounts of data in real time. High data throughput and bandwidth are essential to ensure that AI algorithms have timely access to the data they need. Modern storage solutions must be capable of supporting these high demands without bottlenecks.

Projections indicate that the global data sphere could grow to 175 zettabytes by 2025, driven largely by AI and machine learning applications. Leveraging technologies such as Non-Volatile Memory Express (NVMe) can be a game-changer to handle such massive volumes of data. NVMe offers superior performance, compared to traditional storage interfaces by reducing latency and increasing data transfer speeds. For AI workloads, this means faster data ingestion, real-time analytics, and more efficient training of machine learning models.

Consider the case of a major healthcare provider that implemented NVMe-based storage to support its AI-driven diagnostic systems. By upgrading their storage infrastructure, they achieved a 40 percent reduction in data processing times, enabling quicker and more accurate patient diagnoses.

Latency: The hidden adversary

While throughput is critical, low latency is equally important in the AI realm. Latency refers to the delay between a request for data and the delivery of that data. High latency can significantly hinder the performance of AI applications, particularly those requiring real-time decision-making, such as autonomous vehicles or financial trading systems.

In the financial sector, for instance, trading firms rely on AI to execute high-frequency trades where milliseconds can mean the difference between profit and loss. Storage solutions optimized for low latency, such as those utilizing NVMe over Fabrics (NVMe-oF), can mitigate these delays. By providing rapid access to data, these solutions enhance the responsiveness and efficiency of AI systems, enabling quicker insights and more agile operations.

A 2023 study by the Enterprise Strategy Group found that companies deploying NVMe-oF experienced up to a 60 percent improvement in application response times, highlighting the significant impact of low-latency storage solutions on AI performance.

Scalability: Growing with AI demands

AI projects often start small but can quickly scale as they demonstrate value. Storage architecture must be able to grow in tandem with these expanding data needs. Scalability involves not just adding more storage capacity but doing so in a way that maintains performance and manageability.

Scale-out storage solutions are particularly effective in this regard. These systems allow for seamless expansion by adding more nodes to the storage cluster, ensuring that performance scales linearly with capacity. This architecture is ideal for AI workloads, which can experience exponential growth in data volumes and processing requirements.

For example, a leading e-commerce platform leveraging AI for personalized recommendations saw its data storage needs double within a year. By adopting a scale-out storage solution, they maintained high performance and user experience, even as their data demands surged. According to Gartner, by 2025, 80 percent of enterprises will have adopted scale-out storage systems to manage their expanding AI workloads.

Data management: Beyond storage

Effective data management goes beyond merely storing data; it involves organizing, protecting, and optimizing data for accessibility and usability. AI applications thrive on high-quality, well-organized data. Hence, storage solutions must incorporate advanced data management features.

Technologies such as automated tiering and intelligent data placement can optimize storage efficiency by ensuring that frequently accessed data resides on high-performance media, while less critical data is stored on more economical tiers. Additionally, robust data protection mechanisms, including snapshots, replication, and encryption, safeguard data integrity and availability.

A prominent case is that of a global logistics company that used automated tiering to improve its AI-driven supply chain management system. By strategically placing frequently accessed data on faster storage tiers, they reduced data retrieval times by 35 percent, resulting in more efficient and reliable operations.

Integration with AI ecosystems

Lastly, the ability to seamlessly integrate with existing AI ecosystems is crucial for the success of storage solutions. AI development frameworks, such as TensorFlow, PyTorch, and Hadoop, have specific requirements and workflows. Storage systems that offer compatibility and optimized performance for these environments can significantly optimize AI operations.

Storage solutions with comprehensive APIs and support for containerized applications that leverage Kubernetes, enable smoother integration and orchestration of AI workloads. This ensures that storage infrastructure does not become a bottleneck but rather a facilitator of AI innovation.

For instance, a tech company implementing Kubernetes for container orchestration found that their storage solution’s integration capabilities reduced deployment times for AI models by 50 percent. This seamless integration allowed data scientists to focus on refining algorithms rather than wrestling with infrastructure issues.

The building blocks of storage architecture – data throughput, latency, scalability, data management, and integration – are fundamental to the success of AI applications. By focusing on these elements, organizations can create a robust and efficient storage infrastructure that unlocks the full potential of their AI initiatives. As AI continues to evolve and permeate various sectors, investing in advanced storage solutions will be essential to staying at the forefront of technological innovation.

The robust storage architecture that underpins the previous AI factory floor automation example can be applied to other AI applications and environments as well. Understanding and optimizing the necessary foundational elements will ensure that your AI projects not only succeed, but thrive in an increasingly data-driven world.

For more information on how Dell PowerScale provides an advantageous foundation for building an integrated and optimized IT infrastructure for AI, please visit us online at www.dell.com/powerscale.

Commissioned by Dell Technologies.

Mastering the AI terrain: Why optimal storage is essential for competitive edge

COMMISSIONED: Enterprises adopting AI to stay competitive must tailor AI models to their needs. This means defining use cases, choosing workflows, investing in the right infrastructure, and partnering for success.

Amidst today’s intense market competition, enterprises seek to leverage AI to gain a strategic advantage. Developing proprietary AI models enables companies to tailor solutions to their unique needs, ensuring optimal performance and differentiation. Starting a project to develop AI models involves navigating a complex landscape of technological challenges and requires careful planning, problem-solving skills, and a strategic approach to AI integration.

In AI development, defining a clear use case is the initial critical step, followed by selecting an AI workflow that ensures efficiency and effectiveness, with tools that are simple, integrated, customizable, scalable, and secure. Performance sizing is key, involving benchmarking and optimizing AI models for speed and accuracy, while balancing other performance metrics. The infrastructure to support AI is extensive, requiring robust data storage, compute resources, data processing, machine learning frameworks, and MLOps platforms. And with investments in AI predicted to reach nearly $200 billion by 2025 according to Goldman Sachs reports, the economic potential is significant and necessitates substantial capital investment. Not to mention, the specialized knowledge required for AI projects often necessitates enlisting external expertise.

Each of these challenges must be carefully considered and addressed to ensure the successful development and deployment of AI models. The following step-by-step approach can help organizations address these challenges.

Step 1: Define your use case

Deploying a Generative AI (GenAI) system successfully involves a series of strategic steps, the first and most crucial being defining a clear use case. This foundational step is about understanding the specific needs and objectives of the business, which will guide the selection of the appropriate GenAI workflow. It’s essential to consider the parts of the organization that will be impacted, identify the end-users, and locate where the necessary data is stored.

Aligning GenAI’s capabilities with business goals, whether it’s generating marketing content, providing digital assistance on a website, creating synthetic data or images, or facilitating natural language code development, helps to ensure that the technology is applied in a way that adds value and drives innovation. The success of GenAI deployment hinges on this alignment, resulting in technology that serves as a powerful tool to enhance business processes, engage customers, and foster growth.

Step 2: Choose your AI workflow

Choosing the right AI workflow is crucial for the success of any AI-driven project. Starting with a clear understanding of the objective and the specific use case will guide selection of the appropriate workflow pattern.

Pre-trained models offer a quick start, as they are ready-made solutions that work out-of-the-box for a variety of tasks. Model augmentation, such as retrieval augmented generation (RAG), involves adding new knowledge to an existing model, allowing it to make informed decisions based on additional data. Fine-tuning is a more in-depth process, where the model’s existing knowledge is refined to improve its performance on specific tasks. Finally, model training from scratch is the most comprehensive approach, involving the creation of a new neural network tailored to the unique requirements of the task at hand. This step-by-step escalation in AI workflow complexity, while requiring additional time and effort to complete, allows for a tailored approach that aligns with the project’s goals and technical needs.

Step 3: Size performance requirements

When planning for AI deployment, sizing performance requirements is critical. The type of model you choose, whether it is a language model like GPT4 or an image-based model like DALL-E and Stable Diffusion, influences your compute and storage needs. Language models, while having a high number of parameters, are more compact, which means they require less storage space but more computational power to process a large number of parameters.

On the other hand, image-based models may have fewer parameters but require more storage due to the larger size of the model itself. This distinction is important because it affects how you architect your system’s infrastructure. For instance, a system designed for language models should prioritize processing power, while one for image-based models should focus on storage capabilities. Compute and storage requirements will vary depending on a model’s architecture and the task it is designed to perform so this needs to be factored into how you architect your entire AI project. Understanding these nuances can lead to more efficient resource allocation and a smoother AI workflow.

Common storage solutions for AI models include many options, each with unique benefits and best use cases. Local file storage is often used for smaller, individual projects due to its simplicity and ease of access. Network-attached storage provides more robust solutions for larger datasets, offering better performance and scalability. Distributed file systems (DFS) are ideal for large datasets that require high availability and fault tolerance, as they distribute the data across multiple machines. Object storage is another choice, especially for cloud-native applications, due to its scalability and performance with substantial amounts of unstructured data. It is important to consider the specific needs of your AI model, such as the size of the model and the number of parameters, to choose the most suitable storage solution.

Step 4: Right size your infrastructure investments

Right-sizing infrastructure investments is a critical step in developing efficient AI systems. It involves selecting the appropriate hardware that aligns with the computational demands of the AI models. For instance, smaller AI models may be able to run on optimized laptops such as Dell Precision workstations, while more complex algorithms require powerful setups, such as those with multiple GPUs like Dell’s XE9640 and XE9680 servers. Dell PowerScale offers a versatile storage solution that caters to various needs, from all-flash arrays designed for high performance to tiered storage that balances cost and scalability.

The main advantages of PowerScale for GenAI applications include its scalability, which allows starting with a small and economical setup that can grow exponentially across different environments. It also offers universal data access which allows data to be ingested, read, and written through multiple protocols. Additionally, PowerScale supports GPUDirect, allowing for high-speed and efficient data access, crucial for intensive tasks like AI training. With high-performance Ethernet and NFS over RDMA, it provides for rapid data collection and preprocessing. Lastly, its multicloud deployment capability is essential for running AI workloads in various settings, whether on-premises, at the edge, or in the cloud, providing flexibility and efficiency in AI infrastructure.

Step 5: Engage Dell resources for help

Engaging Dell resources can significantly streamline the process of integrating advanced technologies into your business operations. With step-by-step guidance, your teams can concentrate on strategic growth and innovation rather than the intricacies of implementation. Dell’s Validated Designs and Reference Architectures provide a solid foundation for building efficient IT solutions and assurance that your infrastructure is optimized for performance and reliability. Additionally, we work with our Dell partners to offer specialized AI Workshops which are designed to bring your team up to speed on the latest in AI developments and applications. For a more tailored approach, Dell Professional Services for GenAI offer expertise in deploying generative AI, helping you to quickly establish a robust AI platform and align high-value use cases to drive tangible business value.

In order to be successful with AI model implementation, you need clear guidance on defining use cases, ensuring that your AI initiatives are aligned with strategic business goals. Our Dell AI solutions are designed for efficiency and effectiveness, featuring tools that are not only simple and integrated but also customizable and scalable to meet the evolving demands of AI projects. Performance sizing resources and best practices available through Dell are streamlined with our advanced benchmarking and optimization capabilities, enhancing the speed and accuracy of AI models. The infrastructure required for AI is robust and extensive, and our solutions encompass high-performance data storage, powerful compute resources, and sophisticated data processing capabilities. Recognizing the need for specialized knowledge, we connect you with industry experts to bridge any gaps in expertise, ensuring that your AI projects are not only successful, but also cutting-edge.

To learn more about how Dell storage can support your AI journey visit us online at www.dell.com/powerscale.

Contributed by Dell Technologies.

GenAI only as good as its data and platforms

COMMISSIONED: Whether you’re using one of the leading large language models (LLM), emerging open-source models or a combination of both, the output of your generative AI service hinges on the data and the foundation that supports it.

The right mix of models combined with the right data, architecture and solutions can provide GenAI services that retrieve critical contextual information at lightning speed while reducing inaccuracies.

Balancing the right mix of technologies and techniques is a delicate dance. Technology implementation challenges, security concerns and questions about organizational readiness rank among the top barriers to success, according to recent Bain & Company research.

What architecture and toolset will help me build and test a digital assistant? How do I safely integrate my enterprise information into models? What platform will provide the best performance? And how do I avoid overprovisioning, a common sin from past deployment efforts?

These are frequently asked questions organizations must consider. There are no cookie-cutter approaches for your organization. Yet early trailblazers have found fortune with certain technologies and approaches.

RAG – fueled by a vector database

From digital assistants to content creation and product design, many corporations choose to tap their own corporate data for GenAI use cases. Among the most popular techniques for this is retrieval-augmented generation (RAG).

An alternative to fine-tuning or pre-training models, RAG finds data and documents relevant to a question or task and provides them as context for the model to provide more accurate responses to prompts.

A crucial ingredient of RAG applications is the vector database, which stores and retrieves relevant, unstructured information efficiently. The vector database contains embeddings, or vector representations, of documents or chunks of text. RAG queries this content to find contextually relevant documents or records based on their similarity to keywords in a query.

The pairing of vector database and RAG is a popular combination. Research from Databricks found that use of vector databases soared 376 percent year over year, suggesting that a growing number of companies are using RAG to incorporate their enterprise data into their models.

Synthetic data protects your IP

You may have reservations about trying techniques and solutions with which you and your team may not be familiar. How can you test these solutions without putting your corporate data at risk?

One option organizations have found success with is synthetic data, or generated data that mimics data from the actual world. Because it isn’t real data it poses no real risk to corporate IP.

Synthetic data has emerged as a popular approach for organizations looking to support the development of autonomous vehicles, digital twins for manufacturing and even regulated industries such as health care, where it helps avoid compromising patient privacy.

Some GenAI models are partially trained on synthetic data. Microsoft’s Phi-3 model was partially trained on synthetic data and open source options are emerging regularly.

Regardless of whether you use synthetic data to test or real corporate IP, content created by GenAI poses risks, which is why organizations must keep a human in the loop to vet model outputs.

To the lakehouse

RAG, vector DBs and synthetic data have emerged as critical ingredients for building GenAI services at a time when 87 percent of IT and data analytics leaders agree that innovation in AI has made data management a top priority, according to Salesforce research.

But where is the best place to draw, prep and manage that data to run through these models and tools?

One choice that is gaining steam is a data lakehouse, which abstracts the complexity of managing storage systems and surfaces the right data where, when and how it’s needed. A data lakehouse forms the data management engine for systems that detect fraud to those that derive insights from distributed data sources.

What does a lakehouse have to do with RAG and vector databases? A data lakehouse may serve as the storage and processing layer for vector databases, while storing documents and data RAG uses to craft responses to prompts.

Many organizations will try many LLMs, SLMs and model classes, each with their trade-offs between cost and performance, and many will attempt different deployment approaches. No one organization has all the right answers – or the wherewithal to pursue them.

That’s where a trusted partner can help.

Dell offers the Dell Data Lakehouse, which affords engineers and data scientists self-service access to query their data and achieve outcomes they desire from a single platform.

Your data is your differentiator and the Dell Data Lakehouse respects that by baking in governance to help you maintain control of your data on-premises while satisfying data sovereignty requirements.

The Dell Data Lakehouse is part of the Dell AI Factory, a modular approach to running your data on premises and at the edge using AI-enabled infrastructure with support from an open ecosystem of partners. The Dell AI Factory also includes professional services and use cases to help organizations accelerate their AI journeys.

Brought to you by Dell Technologies.

Your AI strategy called: It wants you to free the data

Commissioned: The importance of data has never been more salient in this golden age of AI services. Whether you’re running large language models for generative AI systems or predictive modeling simulations for more traditional AI, these systems require access to high-quality data.

Seventy-six percent of organizations are counting on GenAI to prove significant if not transformative for their businesses, according to Dell research.

Organizations teem with sales summaries, marketing materials, human resources files and obscene amounts of operational data, which course through the data center and all the way to the edge of the network.

Yet readily accessing this data to create value is easier said than done. Most organizations lack a coherent data management strategy, storing data in ways that aren’t easy to access, let alone manage. For most businesses, anywhere and everywhere is just where the data ended up.

Think about how many times employees have tried and failed to find files on their PCs. Now multiply that experiences thousands of times daily across an enterprise. Finding information can often feel like looking for a virtual needle in a data haystack.

You probably tried to centralize it and streamline it to feed analytics systems, but without structure or governance, the monster has grown unwieldy. And don’t look now – with the advent of GenAI and other evolving AI applications, your organization craves access to even more data.

Accessible data in the AI age

Maybe you’ve been tasked with activating AI for several business units, with partners in marketing and sales collateral to product development and supply chain operations looking to try out dozens or even hundreds of use cases.

Given the years of data neglect, affording these colleagues access to the freshest data is a great challenge. How do you move forward when these tools require data that must be cleaned, prepped and staged?
As it stands, IT typically spends a lot of time on the heavy lifting the comes with requests for datasets, including managing data pipes, feeds, formats and protocols. The struggle of tackling block, file and other storage types is real.

What IT doesn’t tackle may get left for others to wrangle – the data analysts, engineers and scientists who need high-quality data to plug into AI models. Asking the folks who work with this data to take on even more work threatens to overwhelm and capsize the AI initiatives you may be putting in place.

But what if IT could abstract a lot of that effort, and make the data usable more rapidly to those who need it, whether they’re running LLMs or AI simulations in HPC clusters?

To the lakehouse

Organizations have turned to the usual suspects, including data warehouses and lakes, for this critical task. But with AI technologies consuming and generating a variety of structured and unstructured data sources, such systems may benefit from a different approach: A data lakehouse.

The data lakehouse approach shares some things in common with its data lake predecessor. Both accept diverse – structured and unstructured – data. Both use extract, transform and load (ETL) to ingest data and transform it.

However, too many organizations simply let raw data flow into their lakes without structure, such as cataloguing and tagging, which can lead to data quality issues – the dreaded data swamp.

Conversely, the data lakehouse abstracts the complexity of managing storage systems and surfaces the right data where, when and how it’s needed. As the data lakehouse stores data in an open format and structures it on-the-fly when queried, data engineers and analysts can use SQL queries and tools to derive business insights from structured and unstructured data.

Organizations have unlocked previously siloed data to make personalized recommendations to customers. Others have tapped lakehouses to optimize their supply chains, reducing inventory shortfalls.

Democratizing data insights

While a data lakehouse can help organizations achieve their business outcomes, it shouldn’t be mistaken for a lamp. You can’t plug it in, switch it on and walk away. That’s where a trusted partner comes in.
Dell offers the Dell Data Lakehouse, which affords engineers self-service access to query their data and achieve outcomes they desire. The solution leverages compute, storage and software in a single platform that supports open file and table formats and integrates with the ecosystem of AI and ML tools.

Your data is your differentiator and the Dell Data Lakehouse respects that by baking in governance to help you maintain control of your data and adhere to data sovereignty requirements.

The Dell Data Lakehouse is part of the Dell AI Factory, a modular approach to running your data on premises and at the edge using AI-enabled infrastructure with support from an open ecosystem of partners. The Dell AI Factory also includes professional services and use cases to help organizations accelerate their AI journeys.

How is your organization making finding the needle in the haystack easier?

Brought to you by Dell Technologies.

Pure’s as-a-service portfolio takes STaaS to the max

SPONSORED FEATURE: Pure Storage’s Evergreen subscription services, aimed at making on-premises storage service consumption as much like the public cloud as feasible has been proven across eight hardware generations and over 30,000 controller upgrades for customers.

There is a fair amount of flattering me-too imitation occurring in the storage industry but the Evergreen services portfolio retain their superiority with the latest advances concerning power and rack space customer cost payments.

We can better understand how Pure has led the industry in Storage-as-a-Service (STaaS) terms by taking a look at the origins of the program, rooted in array architectural choices, and then seeing how it has developed.

Image source: Pure Storage

Pure Storage first introduced its Forever Flash architecture and services in 2014 as a response to problems afflicting storage arrays at the time; disruptive hardware and software upgrades, forklift replacements when storage controllers needed changing, and such technology refreshes repeating every three to five  years. The Evergreen concept was to make these problems disappear by designing the hardware and software to support non-disruptive upgrades and then to enable the reliable provision of storage as a service, following the example of the top public cloud providers, with array state monitoring, metering and management.

The hardware innovation was to continue to use a dual-controller architecture but with a significant difference. The traditional active-active design involved both controllers sharing the array’s workload. But they were configured to only operate at 50 percent capacity in terms of CPU performance and memory occupancy. If one of the controllers failed then the other could failover use that spare capacity to take in the entire array’s workload with no disruption of service to the array’s application  users.

However, the disadvantage of this was that when a controller upgrade is needed the entire controller chassis has to be pulled out with the array going offline. Pure’s Prakash Darji, General Manager, Digital Experience, explains: “In the original design of our FlashArray, going back now 14 years,  we didn’t do that. We decided to do an active-passive controller setup, where you’re using 100 percent of one, and 0 percent of the other.” The second controller “is just a hot standby. It’s completely idle and there’s no state in the controller.”

Prakash Darji.

That means that, if the first controller is pulled out for replacement, the second one fails over and there is no disruption of array service whatsoever. When the replacement controller is switched on and ready then the second one is extracted, operations fail over to the new controller, again with no service interruption, and the second one is then replaced. It’s all transparent to the applications and customers using the array.

Darji said: “That design of the hardware and the software of not managing state and an active-passive controller architecture was designed with Evergreen in mind, because evergreen means non-disruptive.  It means in place, everything is in place.”

Extending the usable life of storage arrays

With these hardware and software underpinnings in place the Forever Flash program was launched in 2014 and it extended the usable life of all Pure Storage FlashArrays indefinitely, for both new and existing customers with current maintenance contracts. It was enabled by Pure’s arrays being modular, stateless, and having a software-defined architecture. The program featured flat maintenance pricing, software upgrades, a controller upgrade every three years, both included in the maintenance pricing, and an ongoing warranty for all components, including flash media.

The core principle was that customers could deploy storage once and seamlessly upgrade it in-place for multiple generations, avoiding disruptive “forklift” controller and storage chassis upgrades, and software and data migrations. The roots of the Evergreen program can clearly be seen here and it enabled a subscription model for on-premises storage, similar to the simplicity of cloud services where upgrades happen transparently in the background.

The very first Evergreen service was introduced in 2015, ten years ago, called Evergreen Gold and  built on the existing Forever Flash program. It’s now called Evergreen//Forever.

The Evergreen//One option arrived in 2018 and, although it also related to on-premises storage, the array’s actual ownership was retained by Pure with customers paying – subscribing – for consumption of its services; paying for the storage they use, as with public cloud storage services. This changed the customer’s focus away from specific hardware. They paid for set levels of storage services, not for a specific array model or array configuration.

Customers can scale storage capacity up or down flexibly based on changing needs without over-provisioning. Pure delivers extra capacity to the array as needed if the customer’s workload increases its storage demands.

Satisfying performance and usage SLAs

It also featured non-disruptive upgrades and proactive monitoring, while satisfying performance and usage Service Level Agreements (SLAs). Currently there are ten such SLAs covering guarantee performance, buffer capacity, uptime availability, energy efficiency, cyber recovery, resilience,  site relocation, zero planned downtime,  zero data loss, and no data migration. An SLA could promise 99.5% uptime.

These SLAs are accompanied by Service Level Objectives (SLOs) which are internal goals that Pure needs to reach to meet the SLAs. Eg; 99.7% uptime. The arrays have on-board metering to provide metrics for these, Service Level Indicators (SLIs) with telemetry sending them to Pure so they can be tracked.

Darji explains: “An obligation is ‘we’re going to try’, and agreement is ‘we’re guaranteeing it’ with monetary remediation and no exclusions to the monetary remediation. No fine print.”

Pure customers can see their own SLIs in Pure1, Pure’s cloud-based data management and monitoring platform.

 Darji says: “You actually can’t do Evergreen without Pure1.”

(There is more information about this in a Pure blog.)

Evolving the storage as-a-service proposition

A third Evergreen//Flex option was announced in 2022. It took the as-a-service idea a stage further by having customers own the array but only pay for the capacity they use. It provides a “fleet-level architecture” where performance and unused capacity can be dynamically allocated across the customer’s entire storage fleet as needed by applications and workloads.

In late 2023 Pure announced upgrades to its Evergreen//One and Evergreen//Flex portfolio members paying its customers’ power and rack unit space costs. Pure Storage will make a one-time, upfront payment to cover the power and rack space costs for customers subscribing to Evergreen//One or Evergreen//Flex. The payment can be made directly as cash or via service credits. It is based on fixed rates for kilowatt per hour (kWh) and Rack Unit (RU), proportional to the customer’s geographic location and contract size.

Power and rack space account for roughly 20 percent of the total cost of ownership (TCO) in the storage market.

By covering these costs, Pure eliminates a significant expenditure that still exists with on-premises STaaS deployments. For customers it deals with the problem of managing rising electricity costs and also rack space limitations, and it provides cost savings and aligns with long-term efficiency objectives for customers, helping them optimize  resource and energy efficiency as per their sustainability and net-zero goals.

No other supplier does this. The Evergreen portfolio will continue to develop as Pure views its importance as the same as array hardware and software features. Darji says Pure wants to “stay focused on moving the industry forward to an outcome; that outcome being just predictability and performance and capacity in an ageless way.”

He says: ”If we could eliminate all labour required to run and operate storage that is due north for us. It’s like anything that takes work, minimise it, make it go away and be smart about that.” It’s this strand of thinking that is informing Pure’s Evergreen development roadmap. When we consider that some customers having more than a thousand Pure arrays distributed around their environment the opportunities for fleet management services look attractive.

Sponsored by Pure Storage.

Ransomware thieves beware

SPONSORED FEATURE: You know that a technology problem is serious when the White House holds a summit about it. Ransomware is no longer a simple nerd-borne irritation; it’s an organized criminal scourge. Research from the Enterprise Systems Group (ESG) found 79 percent of companies have experienced ransomware attacks within the last 12 months. Nearly half were getting attacked at least once each month, with many reporting attacks happening on a daily basis.

From the early days of enterprise ransomware, security pros had one common piece of guidance: back up your data. It’s still good advice, even in the era of double-extortion attacks where criminals exfiltrate victims’ information while encrypting it. But there’s a problem: attackers are very aware of your backup systems, and they’re searching for them while also looking for production data to encrypt or exfiltrate.

A typical ransomware attack starts when the attacker gains a foothold, often through phishing emails or exploited/unpatched vulnerabilities. Once inside, attackers aim to locate and encrypt production data to cripple operations.

Increasingly, though, they’re also searching for backup environments and data. If they find them unsecured they’ll encrypt that too, hampering recovery efforts. In fact, some attacks – such as 2021’s REvil attack on Kaseya – target backup systems first to ensure that backups will be useless after the malware scrambles production data.

According to Veeam’s 2023 Ransomware Trends Report, 93 percent of cyber attacks last year targeted backup storage to force ransom payments. Attackers successfully stopped victims’ recovery in three quarters of those cases said the company, which specializes in backup and recovery software and services.

Companies are aware of the problem and are looking for help. The ESG study, which surveyed over 600 organizations, found nearly nine in 10 were concerned that their backups have become ransomware targets.

“Government cybersecurity agencies now tell businesses that they should plan on when, rather than if, they’re breached,” points out Eric Schott, chief product officer at Object First.

Started by Veeam’s founders, Object First is on the front line of the battle to protect backup data with its immutable backup storage appliances. “We understand backups are an early target for recon and subsequent attack,” says Schott.

Object First designed its out-of-the-box immutability (Ootbi) backup storage to integrate with Veeam’s backup software. The immutable storage feature prevents data tampering, even if attackers were to gain access to the object storage buckets or appliance administration.

Zero trust data resilience

Employing immutable storage is part of a strategy that Object First and Veeam developed based on the Zero Trust Maturity Model. This framework, which the U.S. Department of Homeland Security’s Cybersecurity and Infrastructure Security Agency (CISA) introduced in September 2021, follows a gradual 15-year development of zero-trust principles that use the ‘trust no one’ approach to cybersecurity.

Zero Trust focuses on stopping people from compromising systems after they breach initial defenses. At its core is the assumption that you’re already breached (or will be at some point in the future).

“We view system hardening as important, but it is not the same as Zero Trust,” says Schott, explaining why the company chose this approach as a foundational part of its system design.

The Object First and Veeam framework building on that model is Zero Trust Data Resilience (ZTDR). It contains several principles. One is the use of least-privilege access to backup infrastructure, others include end-to-end system visibility and threat intelligence to protect systems from attack, along with the use of automated recovery plans if an attack does occur.

Another important principle is segmentation, which divides the backup infrastructure into distinct, isolated ‘resilience zones’ with its own security controls and policies. This minimizes the attack surface and limits the impact of a single hardware or software compromise.

When applied to backup infrastructure, this multi-layered security approach ensures that a breach in one zone does not compromise the ability to recover the zone, and does not compromise the entire backup infrastructure. For example, primary and secondary backup storage can be placed in separate zones to enhance resilience.

Object First has also used this principle to segment its backup hardware from backup software. This makes it harder for an attacker to move laterally to the backup storage.

“Object First’s appliance is a single-function device, so it is also easier to manage and secure” says Schott. “It makes things simpler for smaller organizations to deploy without security specialists or dedicated IT staff and improves operations in large organizations by reducing administrator overhead.”

Divide and conquer, encrypt and protect

What happens if an attacker does reach Object First’s hardware? This is where Zero trust principles come into play. Object First’s Ootbi (out-of-the-box immutability) appliance is built to ensure that backup data cannot be modified or deleted once it is written. “It’s crucial for protecting data from ransomware attacks and other cyber threats,” Schott adds.

To achieve immutability, Object First based Ootbi on the S3 storage protocol. This includes a feature called Object Lock, which uses a write once, read many (WORM) approach to ensure that written data cannot be modified or deleted after the fact. Users control the time limit on immutability using retention periods in Veeam, and they can apply legal holds to prevent deletion or modification of data until the hold is removed.

Immutability means that even total system compromise won’t enable hackers to delete or scramble your data. “Even if you have full admin credentials and access to every bucket secret, you can’t destroy immutable data,” Schott says.

A hacker with physical access could conceivably take a hammer to the appliance if they want to destroy the data, but that’s where the 3-2-1 backup approach recommended by Veeam and Object First is important. It involves keeping at least three copies of your data, storing them on at least two different types of media, and having one copy stored offsite or in the cloud.

Immutability was a key driver for managed IT service provider Waident Technology Solutions, which tested multiple products before settling on Ootbi to support its customers. This gave the company an on-site primary backup solution that it could combine with off-site backups in the U.S. and Europe.

Scale and grow

Object storage’s architecture provides an optimal platform for backup workflows because it’s not bound by the size limitations associated with file and block storage. It uniquely separates data from metadata, storing each as discrete objects. This architecture allows it to easily scale on demand to accommodate large amounts of data, addressing the needs of modern businesses dealing with swelling data volumes.

Conversely, file and block storage is constrained by hierarchical structures or fixed capacity limits. People wanting to scale block-based storage architectures typically build smaller systems and manage them individually, introducing more management complexity and overhead. Object First joins multiple storage units into a single cluster, allowing scalability and load balancing, without shared storage hardware, or a single distributed database for metadata. This allows an on-premise scaling and performance experience without burdening the administrator to manage separate storage systems.

Object storage is well suited for cloud environments. Its focus on individual data objects supports the distributed, often multi-regional nature of cloud resources in a way that is more difficult for file and block architectures.

One company that relied on this immutability in the cloud was SaaS-based legal practice management company Centerbase. Legal companies are top targets for ransomware because of the sensitive data they hold about their clients.

The company used Ootbi storage for its immutability and Veeam integration. It felt this combination could reduce its Recovery Time Objective (RTO) and Recovery Point Objective (RPO) metrics, helping it to get back up and running more quickly in the event of an attack. After installing Ootbi, it slashed its RPO by 50 percent from eight hours to four, while also improving backup speeds, it reported.

End-to-end encryption excludes exfiltration

Out of the box immutability protects data from malicious encryption or deletion, but that’s not all that ransomware attackers want to do. They increasingly want to steal data, threatening to publish it unless victims pay up. To protect customers from that, Object First relies on another capability in the backup software from Veeam: end-to-end encryption.

Veeam’s end-to-end encryption ensures that all data sent into the backup storage is encrypted, providing an additional layer of protection against data exfiltration. By encrypting data at all locations within the 3-2-1 backup environment, Veeam makes it impossible for attackers to read sensitive data in the highly unlikely event that they’re able to reach it at all.

The Veeam encryption keys can be securely stored within Veeam servers, or within external Key Management Services (KMS) including those stored in the cloud.

Having both on-site and off-site backups with immutable storage and Veeam’s encryption enables busy admins to enforce the same set of operations across both domains for maximum security without complex configuration, Schott explains.

“This level of protection provides a strong deterrent against ransomware attacks, safeguarding businesses and enabling continuity in operations,” he says.

In the face of rising ransomware threats targeting backup data, the combination of Veeam’s end-to-end encryption and Object First’s immutable storage provides an advanced line of defense. To develop an easy approach to zero-trust backup deployment, Object First did a pretty good job of thinking outside the box.

Sponsored by Object First.

Matching the cloud database to real workload needs

SPONSORED FEATURE: It’s easy to see why running enterprise applications in the cloud seems like a no-brainer for chief information officers (CIOs) and other C-suite executives.

Apart from the prospect of unlimited resources and platforms with predictable costs, cloud providers offer managed services for key infrastructure and applications that promise to make the lives of specific tech teams much, much easier.

When it comes to databases, for example, the burden and costs involved in deploying or creating database copies and in performing other day-to-day tasks just melt away. The organization is always working with the most up-to-date version of its chosen applications. Cloud providers may offer a variety of nontraditional database services, such as in-memory, document, or time-series options, which can help the organization squeeze the maximum value out of its data. That’s the headline version anyway.

Database administrators (DBAs) know that the reality of running a complex database infrastructure in the cloud and maintaining and evolving it into the future isn’t so straightforward.

Consider how an enterprise’s database stack will have evolved over many years, even decades, whether on-premises or in the cloud. DBAs will have carefully curated and tailored both the software and the underlying hardware, whether regionally or globally, to deliver the optimal balance of performance, cost, and resilience.

That painstaking approach to hardware doesn’t encompass only compute. It applies just as much to storage and networking infrastructure. International Data Corporation (IDC) research shows that organizations spent $6.4 billion on compute and storage infrastructure to support structured databases and data management workloads in the first half of 2023 alone, constituting the biggest single slice of enterprise information technology (IT) infrastructure spending.

As for the database itself, it’s not just a question of using the latest and greatest version of a given package. DBAs know it’s a question of using the right version for their organization’s needs, in terms of both cost and performance.

That’s why DBAs may worry that a managed service won’t offer this degree of control. It might not support the specific license or version that a company may want to put at the heart of its stack. Moreover, some DBAs might think that the underlying infrastructure powering a cloud provider’s managed database is a black box, with little or no opportunity to tune the configuration for optimal performance with their preferred software.

Many database teams – whether they’re scaling up an existing cloud database project or migrating off-prem – might decide the best option is a self-managed database running on a platform like Amazon Web Services (AWS). This gives them the option to use their preferred database version and their existing licenses. It also means they can replicate their existing infrastructure, up to a point at least, and evolve it at their own pace. All while still benefiting from the breadth of other resources that a provider like AWS can put at their disposal, including storage services like Amazon FSx for NetApp ONTAP.

Costs and control

Though the self-managed approach offers far more flexibility than a managed service, it also presents its own challenges.

As NetApp Product Evangelist, Semion Mazor explains, performance depends on many factors: “It’s storage itself, the compute instances, the networking and bandwidth throughput, or of all that together. So, this is the first challenge.” To ensure low latency from the database to the application, DBAs must consider all these factors.

DBAs still need to consider resilience and security issues in the cloud. Potential threats range from physical failures of individual components or services to an entire Region going down, plus the constant threat of cyberattacks. DBAs always focus on making sure data remains accessible so that the database is running, which means the business’ applications are running.

These challenges are complicated and are multiplied by the sheer number of environments DBAs need to account for. Production is a given, but development teams might also need discrete environments for quality assurance (QA), testing, script production, and more. Creating and maintaining those environments are complex tasks in themselves, and they all need data.

Other considerations include the cost and time investment in replicating a full set of data and the cost of the underlying instances that are powering the workloads, not to mention the cost of the storage itself.

That’s where Amazon FSx for NetApp ONTAP comes in. FSx for ONTAP is an expert storage service for many workloads on AWS. It enables DBAs to meet the performance, resilience, and accessibility needs of business-critical workloads, with a comprehensive and flexible set of storage features.

As Mazor explains, it uses the same underlying technology that has been developed over 30 years, which underpins the on-prem NetApp products and services used in the largest enterprises in the world.

The data reduction capabilities in FSx for ONTAP can deliver storage savings of 65 to 70 percent, according to AWS. Mazor says that even if DBAs can’t quite achieve this scale of compression, scaled up over the vast number of environments many enterprises need to run, the benefit becomes huge.

Another crucial feature is the NetApp ONTAP Snapshot capability, which allows the creation of point-in-time, read-only copies of volumes that consume minimal physical storage. Thin provisioning, meanwhile, allocates storage dynamically, further increasing the efficient use of storage – and potentially reducing the amount of storage that organizations must pay for.

Clones and cores

One of the biggest benefits, says Mazor, comes from the thin cloning capability of NetApp ONTAP technology. Mazor says that it gives DBAs the option to “create new database environments and refresh the database instantly with near zero cost.” These copies are fully writable, Mazor explains, meaning that the clone “acts as the full data, so you can do everything with it.”

This capability saves time in spinning up new development environments, which in turn speeds up deployment and cuts time to market. Mazor notes that, from a data point of view, “Those thin clone environments require almost zero capacity, which means almost no additional costs.”

Moving these technologies to the storage layer, rather than to the database layer, contributes to more efficient disaster recovery and backup capabilities. Amazon FSx for NetApp ONTAP also provides built-in cross-region replication and synchronization, giving DBAs another way to meet strict recovery time objective (RTO) and recovery point objective (RPO) targets.

These data reduction, snapshot creation, thin provisioning, thin cloning, and cross-region replication and synchronization features all help DBAs to decouple storage – and, crucially, storage costs – from the Amazon Elastic Compute Cloud (EC2) instances needed to run database workloads. So, storage can be scaled up or down independently of the central processing unit (CPU) capacity required for a given workload.

With fewer compute cores needed, costs and performance can be managed in a much more fine-grained way. DBAs can reduce the need for additional database licenses by optimizing the number of cores required for database workloads and environments. Of course, going the self-managed route means that enterprises can use their existing licenses.

This comprehensive set of capabilities means DBAs can more accurately match their cloud database infrastructure to real workload needs. It also means the cost of running databases in AWS can be further controlled and optimized. It can all add up to a significant reduction in total storage costs, with some customers already seeing savings of 50 percent, according to AWS.

The decision to operate in the cloud is usually a strategic, C-level one. But DBAs can take command of their off-premises database infrastructure and gain maximum control over its configuration, cost, and performance by using Amazon FSx for NetApp ONTAP.

Sponsored by NetApp.

Redefining multicloud with performance and scalability for AI

SPONSORED FEATURE: There’s a pressing need for efficient new data storage solutions given the growing trend of enterprises now deploying AI-enabled applications.

Where megabyte and terabyte storage loads were once commonplace for mere document and single image-type workloads, petabyte (1K terabytes) and even some exabyte (1K petabytes) jobs are now in production.

Factors that have fueled a boom in AI applications include large language models (LLMs) being used in everything from facial recognition software to recommendation engines on streaming services, all to improve user experiences and business processes. Across industries, there’s a growing need for automation, data analysis and intelligent decision-making. AI can automate repetitive tasks, analyze vast datasets to uncover patterns and make data-driven predictions or recommendations. This translates to potentially increased efficiency, productivity and innovation in various fields.

All of this entails vast amounts of data coming from social networks, GPS transmitters, security cameras, point-of-sale locations, remote weather sites and numerous other sources. This trend demands high-performance storage solutions to handle the large volumes of unstructured data involved in AI training and inferencing which can be spread across both on-premises and cloud environments.

A recent IEEE Spectrum report, “Why AI Needs More Memory Than Ever,” explored the ever-increasing data storage demands of AI systems, particularly focusing on the growing size of LLMs. It suggests that besides the demand for high performance, low power, low cost and high capacity, there is also an increasing demand for more smart management functions in or near memory to minimize data movement. As a result, the trend toward deploying hybrid clouds, where all of this is possible, is getting traction.

Traditionally, AI implementation has been marked by siloed solutions and fragmented infrastructure.

“When your applications and tools are running mostly in the cloud, it’s imperative for users to put their data closer to where these tools and applications run,” says Kshitij Tambe, Principal Product Manager at Dell Technologies. “So now if you have your data sitting on premises, and you are building some of these tools and applications to run in the cloud, then there is a big disparity. If you have one thing running in the cloud and enterprise data in the datacenter, this becomes very problematic. So that’s where the need for these hybrid cloud models will come in.”

Why RAGS add even more data to AI systems

They are powerful and require lots of storage, but the LLMs which provide the foundation of AI applications and workloads can only generate responses based on how they’ve been trained. To address this and ensure access to up-to-date information, some AI systems utilize a process called Retrieval Augmented Generation (RAG). RAG integrates information retrieval with prompts, allowing the LLM to access and leverage external knowledge stores. This approach necessitates storing both the base LLM and the vast amount of data it retrieves for real-time use.

With companies – especially long-established ones – building and using many different types of storage and storage devices over years in datacenter, edge and cloud deployments, it becomes a complex problem to managed data across multiple locations at the same time. What some storage admins wouldn’t give to have a single-screen, real-time look at all a company’s storage workloads – whether in production or not – wherever they are in the world!

That was a pipe dream for the longest time. But perhaps not anymore.

New data management platform and processes have emerged in the last year or so to handle these spread-out, next-generation workloads. One example is Dell APEX File Storage in the Microsoft Azure cloud, a NAS platform built to meet AI capacity, performance and data management requirements spanning multicloud environments which is part of Dell’s AI-Ready Data Platform.

Dell APEX File Storage for Microsoft Azure, which became generally available April 9th, bridges the large gap between cloud storage and AI-driven insights, says Dell. It also allows customers a degree of flexibility in how they pay for the service.

At the heart of Dell APEX File Storage for Azure lies PowerScale OneFS, a high-performance scale-out file storage solution already deployed by more than 16,000 customers worldwide.

By bringing PowerScale OneFS to the Azure cloud, Tambe says: “Dell enables users to consolidate and manage data more effectively, reduce storage costs and enhance data protection and security – all while leveraging native cloud AI tools to arrive at insights faster.”

APEX File Storage for Azure serves as a versatile connector to smooth the transition during cloud transformation and enable secure connections to all storage nodes, no matter what type of storage is utilized. A key bonus: the Microsoft interface and control panels have natural familiarity for IT administrators while the PowerScale OneFS replicates the user experience that storage IT professionals are familiar with on-premises.

The APEX File Storage for Azure solution is based on PowerScale OneFS and validated to work with other Dell solutions such as PowerEdge. APEX configurations and specifications include support for up to 18 nodes and 5.6PiB in a single namespace; no other provider can make this claim, boasts Dell. Thus, Dell APEX File Storage for Microsoft Azure puts its stake in the ground with the assertion that it is the most efficient scale-out NAS solution now in the market.

Analysis conducted by Dell indicates that in comparison to Azure NetApp Files, for example, Dell APEX File Storage for Microsoft Azure enables 6x greater cluster performance, up to 11x larger namespace, up to 23x more snapshots per volume, 2x higher cluster resiliency, and easier and more robust cluster expansion.

“Typically, customers might have three nodes, four or five nodes, but there is flexibility to go all the way up to 18 nodes in a single cluster,” says Tambe. “The new architecture of APEX is such that the larger the cluster size, and the larger your data set, it becomes more and more efficient – efficient in the sense of even by the metric of how much usable space you have in your data set.”

Integration and deployment on Microsoft Azure

As for data management, APEX File Storage for Azure offers a new path with integration of high-performance storage capabilities to deploy on Microsoft’s Azure infrastructure. The idea is to let admins easily move data from on-premises to the cloud using advanced native replication without having to refactor any storage architecture. That can deliver huge time savings which subsequently enable data management capabilities to help organizations design, train and run AI-enabled workloads faster and more efficiently, says Dell.

APEX File Storage for Azure leverages Azure’s cloud infrastructure and functionalities to benefit AI tasks in a few ways. Developing infrastructure for advanced AI models necessitates significant investment, extending beyond powerful compute resources to encompass critical data storage infrastructure. Training datasets can range in size from terabytes to petabytes, and concurrent access via numerous processes. Saving checkpoints which each potentially consist of hundreds of gigabytes is equally vital.

APEX File Storage directly integrates with several of the most common AI tools – including Azure AI Studio, to change the way developers approach generative AI applications and help simplify the journey from concept to production. It’s a developer’s playground for evaluating responses from large language models and orchestrating prompt flows says Dell, ensuring optimal performance and scalability.

And since OneFS supports S3 as an access protocol, getting APEX File storage to work with Azure AI Studio should be easy. Developers can point Azure AI Studio using OneLake Data Gateway directly to a OneFS directory, for example. This allows them to use files on OneFS clusters (AFS or on-prem) without copying the data to Blob Storage thus running fine-tuning of AI models with files remaining in a OneFS filesystem.

For providing scalability, APEX File Storage utilizes Azure’s cloud-native technologies, allowing it to elastically scale storage capacity and performance based on AI workload demands. This helps ensure smooth operation, even when dealing with large datasets used in AI training and processing.

For integration, APEX File Storage integrates directly with the Azure architecture, facilitating data transfer between on-premises and cloud environments. This eliminates the need to redesign storage infrastructure when moving AI workloads to the cloud. This combination creates the foundation for a universal storage layer that simplifies storage management in multicloud environments, says Dell.

For data management and protection, APEX File Storage offers features such as advanced native replication, data deduplication and erasure coding. These functionalities assist with data redundancy, security and efficient storage utilization, which are all crucial aspects of managing large datasets for AI applications.

Dell preceded the Microsoft Azure APEX initiative with an AWS version of the service last year. This stands as an example of Dell’s commitment to offering a wide range of storage and data management options for different cloud platforms to meet customer requirements.

Sponsored by Dell.

Don’t fall for the bring-your-own-AI trap

Commissioned: Generative AI adoption within organizations is probably much higher than many realize when you account for the tools employees are using in secret to boost productivity. Such shadow AI is a growing burden IT departments must shoulder, as employees embrace these digital content creators.

Seventy-eight percent of employees are “bringing their own AI technologies” (BYOAI) to work, according to a joint Microsoft and LinkedIn survey. While the study acknowledges that such BYOAI puts corporate data at risk it downplays the sweeping perils organizations face to their data security.

Whether you call it BYOAI or shadow AI this phenomenon is potentially far worse than the unsanctioned use of cloud and mobile applications that pre-dated it.

As an IT leader, you’ll recall the bring-your-own-device (BYOD) trend that marked the early days of the consumer smartphone 15 years ago.

You may have even watched in horror as employees ditched their beloved corporate Blackberries for iPhones and Android smartphones. The proliferation of unsanctioned applications downloaded from application stores exacerbated the risks.

The reality is that consumers often move faster than organizations. But consumers who insist on using their preferred devices and software ignore integrating with enterprise services and don’t concern themselves with risk or compliance needs.

As risky as shadow IT was, shadow AI has the potential to be far worse – a decentralized Wild West or free-for-all of tool consumption. And while you can hope that employees have the common sense not to drop strategy documents into public GPTs such as OpenAI, even something innocuous like meeting transcriptions can have serious consequences for the business.

Of course, as an IT leader you know you can’t sit on the sidelines while employees prompt any GenAI service they prefer. If ignored, Shadow AI courts potentially catastrophic consequences for organizations from IP leakage to tipping off competitors to critical strategy.

Despite the risks, most organizations aren’t moving fast enough to put guardrails in place that ensure safe use, as 69% companies surveyed by KPMG were in the initial stages of or had not begun evaluating GenAI risks and risk mitigation strategies.

Deploy AI safely and at scale

Fortunately, organizations have at their disposal a playbook to implement AI at scale in a way that helps bolster employees’ skills while respecting the necessary governance and guardrails to protect corporate IP. Here’s what IT leaders should do:

Institute governance policies: Establish guidelines addressing AI usage within the organization. Define what constitutes approved AI systems, vet those applications and clearly communicate the potential consequences of using unapproved AI in a questionable way.

Educate and train: Giving employees approved AI applications that can help them perform their jobs reduces the incentive for employees to use unauthorized tools. You must also educate them on the risks associated with inputting sensitive content, as well as what falls in that category. If you do decide to allow employees to try unauthorized tools, or BYOAI, provide the right guardrails to ensure safe use.

Provide use cases and personas: Education includes offering employees use cases that could help their roles, supported by user “personas” or role-based adoption paths to foster fair use.

Audit and monitor use: Regular audits and compliance monitoring mechanisms, including software that sniffs out anomalous network activity, can help you detect unauthorized AI systems or applications.

Encourage transparency and reporting: Create a culture where employees feel comfortable reporting the use of unauthorized AI tools or systems. This will help facilitate rapid response and remediation to minimize the fallout of use or escalation of incidents.

Communicate constantly: GenAI tools are evolving rapidly so you’ll need to regularly refresh your AI policies and guidelines and communicate changes to employees. The good news? Most employees are receptive to guidance and are eager to do the right thing.

Solutions to help steer you

GenAI models and services are evolving daily, but there are some constants that remain as true as ever.

To deploy AI at scale, you must account for everything from choosing the right infrastructure to picking the right GenAI models for your business to security and governance risks.

Your AI strategy will be pivotal to your business transformation so you should weigh whether to assume control of GenAI deployments or let employees choose their own adventures, knowing the consequences of the latter path.

And if you do allow for latitude with BYOAI, shadow AI or whatever you choose to call it, do you have the safeguards in place to protect the business?

Trusted partners can help steer you through the learning curves. Dell Technologies offers a portfolio of AI-ready solutions and professional services to guide you along every step of your GenAI journey.

Learn more about Dell AI solutions.

Brought to you by Dell Technologies.

Why GenAI has IT in the catbird seat

COMMISSIONED: In baseball, “sitting in the catbird seat” means holding a position of strategic advantage. Like a batter with no strikes and three balls.

The legendary commentator for the Dodgers, Red Barber, borrowed this phrase from the behavior of the gray catbird in the southern United States. Sitting high above the ground, the catbird is ready to capitalize on opportunities. IT administrators are in the same spot, ready to steer their organization to the forefront of Generative AI (GenAI).

Research suggests 82 percent of business leaders believe GenAI will significantly transform their business. Using their organization’s data to create new content, automate processes, and deliver insights with speed and efficiency.

However, the deployment and management of GenAI technologies demands a nuanced understanding of data management, network requirements, and infrastructure scalability. These are domains where Infrastructure Technology teams excel.

IT admins: strategic players in AI initiatives

IT teams are poised to lead the adoption and integration of GenAI technologies across the business for three key reasons:

IT knows how to get to the data

At the heart of GenAI initiatives is the ability to access and manage large volumes of data. Organization specific knowledge holds incredible potential. What begins as “just” data, has the potential to become insights, innovation, and intelligence unique to each business.

Storage admins have long been at home in this complex landscape. They have spent years managing structured and unstructured data. They have learned to excel in a multicloud world, providing access to both on-premises and cloud-based data.

The success of GenAI projects hinges on the efficient retrieval and processing of data. New requirements like real-time access to large volumes of data, and architectures optimized for speed and flexibility will drive the need for new approaches to storage management. Here, storage administrators can spearhead the adoption of innovative tools like the Dell Data Lakehouse to ensure that data in any format is accessible, preprocessed, and primed for effective AI training.

IT knows how to navigate the network

Network administrators have played a crucial role in establishing and maintaining the connectivity framework within organizations since the inception of IT teams. They laid foundational digital highways, enabled seamless communication, data transfer, and access to resources.

If data is the lifeblood of GenAI, then networking is its backbone. GenAI initiatives and the exponential growth of data will drive even more complex and powerful network infrastructure. Software-defined networking, orchestration and advances in congestion control and adaptive routing mechanisms will continue to help fuel this rapid growth. While InfiniBand is most frequently used, Ethernet technologies continue to advance. Analysts forecast that the need for non-proprietary, cost-effective solutions, will drive a 50 percent expansion in the datacenter switch market with 800Gbps making up most AI back-end networks by 2025. Network admins can ensure their organizations are ready for AI infrastructure by strategically learning and deploying the right solution to meet the growing demands of their organization.

IT knows how to scale and virtualize compute

IT teams excel at scaling technology to meet the expanding needs of the business. When management complexity expanded, they drove containerization and virtualization. When their organizations grew, they deployed scalable infrastructure and implemented cloud services.

As GenAI projects find success, the ability to quickly scale is paramount. Transitioning from proof of concept to full-scale production while quickly showing ROI can present a significant challenge. The key for IT teams will be to start with low hanging, high probability use cases. Simultaneously they will need to anticipate growth trajectories and prepare infrastructure to support expansion. Leveraging an agile framework like the Dell AI Factory for NVIDIA provides a highly scalable infrastructure with a flexible microservices architecture.

A call to action

It is time for IT teams to take advantage of their unique position, not as participants, but as leaders. This journey calls for a deepening of skill sets. From mastering data processing for GenAI, to understanding the demands of high-bandwidth GenAI infrastructure, to looking at the datacenter in an entirely new light. The opportunity is ripe for IT professionals to build upon their established expertise, driving not only their careers forward but also positioning their organizations at the forefront of the AI revolution. IT teams do not have to do this on their own. Dell Technologies is also driving these initiatives with education, services, and great solutions. Visit us at Dell.com/AI to learn more.

Brought to you by Dell Technologies.

Modernizing storage for the age of AI

SPONSORED FEATURE: You might have already analyzed AI use cases for artificial intelligence (AI) within your business and identified potential efficiencies, revenue opportunities and more. Now comes the hard part: building an infrastructure that supports your mission. Computing capacity is a crucial part of that portfolio, but companies often overlook another equally important ingredient: storage.

Investing heavily in the latest GPUs or cloud capabilities to give yourself an edge in the training and inference of AI models is important, but it will all be for nought if you can’t feed the beast with the data it needs to deliver results. That’s where scale-out storage technology comes in – to help provide organizations with the answers to the infrastructure questions which this new world of AI is asking.

“Data is a differentiator for companies involved in AI,” says Tom Wilson, a product manager at Dell Technologies focusing heavily on AI workloads, who analogizes data as the fuel, compute as the engine and storage as the fuel tank. “Having a modernized platform that provides the security, storage efficiency, performance, and scale that companies need to use that data in AI workflows is one of our key pillars for PowerScale.”

The benefits of scale-out storage

Wilson is a veteran evangelist for the technology underpinning PowerScale, the Dell file storage solution which has been upgraded to deliver an AI-optimized storage infrastructure with the launch of two new PowerScale F210 and F710 flash arrays. Leveraging the latest PowerEdge hardware and OneFS software, PowerScale is showcased as a key component of an ‘AI-ready data platform’ designed to provide the performance, scale and security needed to help customers build AI-enabled workloads wherever their data lives, on premises, or cloud and cloud-adjacent environments.

Dell was one of the first companies to support the NVIDIA GPUDirect protocol which enables storage systems to rapidly send and receive data without involving the host processor to accelerate AI workloads. Wilson recalls that customers were grappling with the rise of storage volumes thanks to unstructured data even before GPUs and cloud computing had taken AI mainstream, but the surge in demand for AI and generative AI (GenAI) enabled applications has put even more strain on existing storage infrastructure.

“One of the things that we wanted to help solve was how do you manage massive amounts of data predictably, all around the world,” Wilson says. “That’s what led us to create a scale-out file system.”
Traditional scale-up storage can struggle to handle the vast volumes of data needed to feed AI models for a couple of reasons. Firstly, it expands by adding more drives to a single system with its own dedicated head unit. The obvious downside to this approach is limited capacity, as the chassis will eventually run out of space.

The less obvious drawback is limited performance. The head unit, which organizes the storage, will come under an increasing load as the storage volume rises and more disks are added, explains Wilson. The performance that you get with the first few dozen Terabytes in a scale-up system might be great for your needs, but as you add more storage capacity the performance doesn’t increase. At some point says Wilson, the storage workflows might outgrow the throughput that a scale-up system can provide.

Conversely, scale-out storage uses clustered storage nodes, each of which has its own computing and storage capacity. Adding another node to the system boosts the computing capacity of the entire cluster. “So when you add capacity, you aren’t just scaling up by adding drives; you’re adding performance,” he adds.

Inside the PowerScale architecture

PowerScale’s next-generation nodes, the F210 and F710 improve on the previous generation all-flash nodes, leveraging the latest -generation PowerEdge platform to deliver faster computing capabilities in the form of fourth-generation Intel Xeon Sapphire Rapids CPUs. They also feature improved memory speed thanks to the latest DDR5 DRAM options. A faster PCIe Gen 5 bus offers up to quadruple the throughput compared to the PCIe Gen 3 used in previous nodes.

These hardware improvements are especially relevant for AI applications, explains Wilson. For example, the mix of PCIe and SSD interface improvements helps to double the streaming read and write throughput – key performance metrics that affect the phases of the AI pipeline like the model training and checkpointing phases.

The 1U-format systems have also increased their node density by adding the capacity needed to ensure the vast volumes of data that AI requires can be easily accommodated. The F710 features room for 10 drives compared to the F600’s eight, while the F210 doubles capacity with the introduction of the 15Tb drive.

The systems also feature a Smart flow chassis – a piece of IP from Dell’s PowerEdge hardware – that pushes air through the system more efficiently. This helps maintain system reliability while reducing the power used for cooling, explains Wilson – an important consideration in datacenters facing big electricity bills and total cost of ownership challenges in powering the server, storage and network equipment required to get and keep AI workloads running. It contributes to a key efficiency increase figure for the new units – the F710 offers up to 90 percent higher performance per watt compared to the previous generation of the product.

How advanced software complements the hardware

Dell has also updated the PowerScale’s OneFS operating system to take full advantage of the hardware enhancements. Features like thread optimization help to bolster AI performance. Dell reports up to a 2.6 times improvement in throughput in the F710 compared to the F600 when handling the kind of high-concurrency, latency-sensitive workloads that underpin many AI training and inferencing applications, for example.

“The performance improvements of all-flash NVMe drives means that we don’t necessarily need the same level of caching that we used previously,” says Wilson. “OneFS optimizes communications to those NVMe drives, using techniques like read locking. We also write directly from the journal to the drives.”

OneFS 9.6 also added another important capability for AI workloads – the ability to handle AI training and inferencing tasks with hybrid cloud capability. APEX File Storage for AWS was launched with OneFS 9.6, while more recently OneFS 9.8 introduced APEX File Storage for Azure as well – allowing organization even greater flexibility and choice, says Dell. By running OneFS in the cloud, customers can move a subset of the data they need off-premises. They might choose to handle data preparation and cleansing on- premises, for example, and then move the prepped data into the cloud to take advantage of computing capabilities that they don’t have on-site.

The key benefit of running PowerScale in a cloud environment is that customers can take their security model along with them, explains Wilson. They move the data they need using native replication in OneFS, making data available with the same security policies, permissions, and identity management parameters in the cloud as they already have on- premises. They don’t have to refactor their workflows, which means they can quickly move to the next part of the AI pipeline without skipping a beat, while staying compliant with their data privacy and protection policies.

A comprehensive AI infrastructure

PowerScale storage can be optimized for efficiency, performance and cost depending on the specific AI workflow it is destined to support, says Dell (whether model retention, data preparation or large-scale model training or tuning for example). The new units were already producing useful results in field tests with Dell customers by the time they were released for general availability. Alan Davidson, CIO at Broadcom, said that the systems had helped significantly bump up performance in its electronic design automation (EDA) operations.

“Collaborating with Dell means faster innovation for my business. The new Dell PowerScale F710 has exceeded our expectations with more than 25 percent performance improvements in our EDA workloads while delivering improved datacenter sustainability,” he told Dell.

These systems further built out a portfolio that can serve complex AI infrastructure, enhanced by partnerships including that between Dell and NVIDIA. The F710, the first Ethernet-based storage appliance certified by NVIDIA DGX SuperPOD, is a key part of the Dell AI Factory that the company announced with NVIDIA in March. It’s an end-to-end validated combination of Dell infrastructure and NVIDIA GPUs and software that supports the entire generative AI life cycle.

“Nobody is better at building end-to-end systems for the enterprise than Dell,” said NVIDIA CEO Jensen Huang at the company’s GTC 2024 AI developer conference. This combined hardware and software portfolio ties into a range of documentation and architectural guidance from Dell.

“Not only do we have best of breed infrastructure, but we also have the expertise, whether it’s on the services side or in terms of best practices documentation and validated designs and reference architectures,” Wilson says. “We have the complete stack to help customers simplify their AI journeys.”

As they rush to adopt AI, organizations are grappling to manage their infrastructure. Because AI projects are so data intensive, the chances are good that at least part of a company’s AI pipeline will involve on-premises storage. Getting the storage part of the infrastructure portfolio right can eliminate bottlenecks further along in the process as development teams, software engineers, data scientists and others begin to deal with the large volumes and high bandwidth requirements necessary to feed these AI workloads. In this data-laden future, optimized scale-out storage infrastructure increasingly looks like the right approach.

No organization can afford to rest on its laurels when it comes to ensuring the business has the efficient, high-performance infrastructure it needs to build and launch new AI-enabled applications and services. Continuous optimization and upgrades are the norm in IT – and in many cases has been rendered more critical by the recent surge in demand for AI. Dell is expected to keep up its own momentum and announce even more enhancements to its AI-optimized portfolio at Dell Technologies World 2024 to enable customers in this AI era.

Sponsored by Dell.

Demystifying multicloud complexity with a universal storage layer

Three clouds
Three clouds

COMMISSIONED: All organizations today use information technology (IT) to deliver their value. While the intensity and depth of reliance on IT varies across industries and individual entities, software applications serve as the engine of innovation and data serves as the fuel for the engine.

What has changed in the last decade is a dramatic increase in the distribution of locations across which applications and data reside. They used to reside in a climate-controlled datacenter or a dark computer room in a corner, but today IT infrastructure is distributed across private (edge, datacenter, or colocation facilities) and public cloud locations. Organizations need to leverage both, often referred to as multicloud, to enhance customer experiences and optimize the delivery of services. Recent research by TechTarget’s Enterprise Strategy Group (ESG) shows that 88 percent of organizations agree that leveraging multiple public cloud providers delivers strategic benefits; 87 percent stated that their application environment will become distributed across more locations over the next two years. In addition, 86 percent indicated they regularly migrate applications and/or data from on-premises locations to the public cloud.

This increased distribution of IT infrastructure has brought benefits of agility, geographical reach, and democratized access to technology, but it has created challenges. Diverse locations often have fundamentally different configurations and interfaces. This has dramatically increased complexity of operations, fractured information assets into disparate pools of data, and increased the cost of managing and maintaining that infrastructure.

An industry example

Let’s explore the impact of multicloud operations in an example industry. Retail operations are all about selling goods, and the industry has benefited from software-driven automation and intelligence increasingly deployed across distributed locations. Retailers today need to provide an omnichannel experience to meet customers where they are. The IT that supports this has followed. A retailer is likely to use private infrastructure to host its e-commerce platform, CRM systems, and inventory management to remain compliant, and deliver the right performance and reliability for core applications. However mobile application interfaces may be served from a public cloud location. Other common uses for public cloud include access to scalable analytics or GPU as-a-service for AI model training and personalized product recommendations. Bricks-and-mortar stores or kiosks are edge locations that gather and process transactional data.

Most industries have similar requirements that drive the need for multicloud infrastructure, and those requirements are best served by establishing a universal storage layer across locations. But what in the world is a universal storage layer?

To achieve agile, secure, efficient operations across a multicloud data landscape there are two fundamental requirements: common storage services across locations and centralized management. A universal storage layer is created when storage services with the same underlying architecture are deployed across multiple locations. Whether it is block, file or object storage or even protection storage, having a common set of storage services across locations translates to consistency and control. It reduces the need for custom processes to move data between locations, making it easier to take advantage of the inherent benefits of multicloud environments.

Let’s focus on the data storage aspects of the retail scenario. If a retailer is capturing unstructured data such as real time point-of-service streams in a retail edge setting and then gathering it in a central location, they need an enterprise NAS solution in their datacenter with file storage services for the aggregated data. If they want to move that data to another location to analyze it – perhaps leveraging a public cloud service – having the same software-defined file storage in the public cloud creates a universal file storage layer. That universal storage layer eliminates friction, enabling seamless mobility of file data to the right location for storage or processing. Also, common storage services mean better security through consistent controls for things like managing individual storage instances and establishing policies for access control or recovery points.

Centralized storage management drives efficiency by providing a hub for visibility across the multicloud estate and orchestrating movement between locations. The foundation of common storage services creates a universal storage layer that enables consistent, centralized management.

How Dell makes this work

Organizations are embracing multicloud at an unprecedented pace. They want to benefit from geographic flexibility and access to unique services (like data analytics and model training) available in the public cloud. They also continue to depend on the reliability, high-performance, compliance and control of private locations. Dell has the solution to manage the complexities of this diverse multicloud ecosystem.

Dell is recognized as a leader in Gartner’s Magic Quadrants for Primary Storage, File Storage and Data Protection Storage. Our huge and growing customer base in private (on-premises) environments is testimony to decades of leadership in providing storage solutions that are scalable, resilient, secure, and efficient. We’ve now taken these leading software-defined storage solutions for block, file, object, and protection storage and created Dell APEX Storage for Public Cloud. It is a whole family of public-cloud-based storage services designed to work with any workload running in the public cloud providing value-added capabilities not supported by native cloud storage services. This extension of common storage services from on-prem to cloud environments lends customers the ability to use Dell software-defined storage in multiple environments. These common storage services across locations create a universal storage layer, enabling organizations to leverage best-in-class technology wherever their business needs dictate.”

A complementary new offer is Dell APEX Navigator for Multicloud Storage. It is a centralized and streamlined management plane for our storage solutions that provides a simplified experience and reduces manual effort. It accelerates productivity with automated deployment of public cloud storage and integrates with preferred automation tools such as Terraform through standardized APIs. It provides visibility across your multicloud storage estate and is a foundation for secure multicloud operations facilitating Zero-Trust adoption with capabilities like Single Sign On and Federated Identity.

The Enterprise Strategy Group whitepaper analyzes the impact of creating a universal storage layer with Dell on-premises storage and APEX Storage for Public Cloud Storage orchestrated by Dell APEX Navigator for Multicloud Storage. It quantifies the benefits of cost optimization, increased productivity, enterprise-grade resilience, business acceleration and agility. It also lays out a specific example of a three-year savings of 42 percent for multicloud storage infrastructure.

Partners on your multicloud journey

While this blog cites a retail example, distributed applications are the reality across industries. We highlighted the challenges around data storage that arise when an organization starts to enable those applications across private (including datacenter, edge and colocation facilities) and public cloud locations. The solution lies in the implementation of common storage services and centralized management – what we have termed “a universal storage layer” across locations.

Dell’s data-centric multicloud solutions strategy, featuring common storage services that can be implemented across locations to create a universal storage layer, makes it the right partner for your multicloud journey.

To learn more about data mobility and optimizing costs with Dell software-defined storage, APEX Storage for Public Cloud and APEX Navigator, read this ESG report: Measuring the Operational Benefit of Dell APEX Navigator for Multicloud Storage. For more information on APEX Navigator for Multicloud Storage please visit dell.com/navigator. For more information on APEX Block Storage for AWS visit dell.com/apex-block.

Brought to you by Dell Technologies.

Why It’s Time to Rethink Your Cloud Strategy

COMMISSIONED: The decision by the three largest U.S. public cloud providers to waive data transfer fees is a predictable response to the European Data Act’s move to eradicate contractual terms that stifle competition.

A quick recap: Google made waves in January when it cut its data transfer fee, the money customers pay to move data from cloud platforms. Amazon Web Services and Microsoft followed suit in March. The particulars of each move vary, forcing customers to read the fine print closely.

Regardless, the moves offer customers another opportunity to rethink where they’re running application workloads. This phenomenon, which often involves repatriation to on-premises environments, has gained steam in recent years as IT has become more decentralized.

The roll-back may gain more momentum as organizations decide to create new AI workloads, such as generative AI chatbots and other applications, and run them in house or other locations that will enable them to retain control over their data.

To the cloud and back

Just a decade ago, the organizations pondered whether they should migrate workloads to the public cloud. Then the trend became cloud-first, and everywhere else second.

Computing trends have shifted again as organizations seek to optimize workloads.

Some organizations clawed back apps they’d lifted and shifted to the cloud after finding them difficult to run them there. Others found the operational costs too steep or failed to consider performance requirements. Still others stumbled upon security and governance issues that they either hadn’t accounted for or had to reconcile to meet local compliance laws.

“Ultimately, they didn’t consider everything that was included in the cost of maintaining these systems, moving these systems and modernizing these systems in the cloud environment and they balked and hit the reset button,” said David Linthicum, a cloud analyst at SiliconANGLE.

Much ado about egress fees

Adding to organizations’ frustration with cloud software are the vendors’ egress fees. Such fees can range from 5 cents to 9 cents per gigabyte, which can grow to tens of thousands of dollars for organizations working with petabytes. Generally, fees vary based on where data is being transferred to and from, as well as how it is moved.

Regulators dislike fear the switching costs will keep customers locked into the platform hosting their data, thus reducing choice and hindering innovation. Customers dislike these fees and other surcharges as part of a broader strategy to squeeze them for fatter margins.

This takes the form of technically cumbersome and siloed solutions (proprietary and costly to connect to rivals’ services), as well as steep financial discounts that result in the customer purchasing additional software they may or may not need. Never mind that consuming more services – and thus generating even more data – makes it more challenging and costly to move. Data gravity weighs on IT organizations’ decisions to move workloads.

In that vein, the hyperscalers’ preemptive play is designed to get ahead of Europe’s pending regulations, which commence in September 2025. Call it what you want – just don’t call it philanthropy.

The egress fee cancellation adds another consideration for IT leaders mulling a move to the cloud. Emerging technology trends, including a broadening of workload locations, are other factors.

AI and the expanding multicloud ecosystem

While public cloud software remains a $100 billion-plus market, the computing landscape has expanded, as noted earlier.

Evolving employee and customer requirements that accelerated during the pandemic have helped diversify workload allocation. Data requirements have also become more decentralized, as applications are increasingly served by on-premises systems, multiple public clouds, edge networks, colo facilities and other environments.

The proliferation of AI technologies is busting datacenter boundaries, as running data close to compute and storage capabilities often offers the best outcomes. No workload embodies this more than GenAI, whose large language models (LLMs) require large amounts of compute processing.

While it may make sense to run some GenAI workloads in public clouds – particularly for speedy proof-of – concepts, organizations also recognize that their corporate data is one of the key competitive differentiators. As such, organizations using their corporate IP to fuel and augment their models may opt to keep their data in house – or bring their AI to their data – to maintain control.

The on-premises approach may also offer a better hedge against the risks of shadow AI, in which employees’ unintentional gaffes may lead to data leakage that harms their brands’ reputation. Fifty-five percent of organizations feel preventing exposure of sensitive and critical data is a top concern, according to Technalysis Research.

With application workloads becoming more distributed to maximize performance it may make sense build, augment, or train models in house and run the resulting application in multiple locations. This is an acceptable option, assuming the corporate governance and guardrails are respected.

Ultimately, whether organizations choose to run their GenAI workloads on premises or in multiple other locations, they must weigh the options that will afford them the best performance and control.

Companies unsure of where to begin their GenAI journey can count on Dell Technologies, part of a broad partner ecosystem, for help. Dell offers AI-optimized servers, client devices for the modern workplace and professional services to help organizations deploy GenAI in trusted, secure environments.

Learn more about Dell Generative AI Solutions.

Contributed by Dell Technologies.

Generative AI and the wizardry of the wide-open ecosystem

COMMISSIONED: IT leaders face several challenges navigating the growing generative AI ecosystem, but choosing a tech stack that can help bring business use cases to fruition is among the biggest. The number of proprietary and open-source models is growing daily, as are the tools designed to support them.

To understand the challenge, picture IT leaders as wizards sifting through a big library of magical spells (Dumbledore may suffice for many), each representing a different model, tool or technology. Each shelf contains different spell categories, such as text generation, image or video synthesis, among others.

Spell books include different diagrams, incantations and instructions just as GenAI models contain documentation, parameters and operational nuances. GPT-4, Stable Diffusion and Llama 2 rank among the most well-known models, though many more are gaining traction.

Moreover, this “library” is constantly growing, making it harder for IT leaders to keep up with the frenetic pace of conjuring – er – innovation. Kind of like chasing after a moving staircase.

You get the idea. If you’re unsure, you can brush up on Harry Potter books or films. In the meantime, here are three key steps to consider as you begin architecting your AI infrastructure for the future.

Pick models and modular architecture

As IT leaders adopted more public cloud software they realized that a witches’ brew of licensing terms, proprietary wrappers and data gravity made some of their applications tricky and expensive to move. These organizations had effectively become locked into the cloud platforms, whose moats were designed to keep apps inside the castle walls.

If you believe that GenAI is going to be a critical workload for your business – 70 percent of global CEOs told PwC it will change the way their businesses create, deliver and capture value – then you must clear the lock-in hurdle. One way to do this is to pick an open model and supporting stack that affords you flexibility to jump to new products that better serve your business.

Tech analyst Tim Andrews advocates for reframing your mindset from predicting product “winners” to one that allows you to exit as easily as possible. And a modular software architecture in which portions of your systems are isolated can help.

Fortunately, many models will afford you flexibility and freedom. But tread carefully; just as spell books may harbor hidden curses, some models may drain the organization’s resources or introduce biases or hallucinations. Research the models and understand the trade-offs.

Choose infrastructure carefully

GPU powerhouse NVIDIA believes that most large corporations will stand up their own AI factories, essentially datacenters dedicated to running only AI workloads that aim to boost productivity and customer experience. This will be aspirational for all but the companies who have the robust cash flow to build these AI centers.

Public cloud models will help you get up and running quickly, but, if right-sizing your AI model and ensuring data privacy and security are key, an on-premises path may be right for you. Your infrastructure is the magic wand that enables you to run your models. What your wand is made of matters, too.

In the near term, organizations will continue to run their AI workloads in a hybrid or multicloud environment that offers flexibility of choice while allowing IT leaders to pick operating locations based on performance, latency, security and other factors. The future IT architecture is multicloud-by-design, leveraging infrastructure and reference designs delivered as-a-Service. Building for that vision will enable you to run your GenAI workloads in a variety of places.

Know this: With organizations still evaluating or piloting GenAI models, standardization paths have yet to emerge. As you build, you must take care to head off technical debt as much as possible.

Embrace the wide-open ecosystem

While wizards may spend years mastering their spells, IT leaders don’t have that luxury. Eighty-five percent of C-suite executives said they expect to raise their level of AI and GenAI investments in 2024, according to Boston Consulting Group.

It’s incumbent on IT leader to help business stakeholders figure out how to create value from their GenAI deployments even as new models and iterations regularly arrive.

Fortunately, there is an open ecosystem of partners to help mitigate the challenges. Open ecosystems are critical because they help lower the barrier to entry for most mature technology teams.

In an open ecosystem, organizations lacking the technical chops or financial means to build or pay for LLMs can now access out-of-the-box models that don’t require the precious skills to train, tune or augment models. Trusted partners are one of the keys to navigating that ecosystem.

Dell is working with partners such as Meta, Hugging Face and others to help you bring AI to your data with high-performing servers, storage, client devices and professional services you can trust.

Keeping your options open is critical for delivering the business outcomes that will make your GenAI journey magical.

Learn more about Dell Generative AI Solutions.

Brought to you by Dell Technologies.

Developing AI workloads is complex. Deciding where to run them might be easier

SPONSORED FEATURE: If artificial intelligence (AI) has been sending shockwaves through the technology world in recent years, the onset of generative AI over the last 18 months has been a veritable earthquake. And for IT leaders looking to harness its potential for their own organisations, the pace of development can feel bewildering.

Enterprises are racing to leverage their own data to either build their models or repurpose public models already available. But this can pose a significant challenge for the dev and data science teams involved.

It can also present something of a conundrum for companies that want to keep control of the HPC infrastructure needed to support their AI workloads. AI-enabled applications and services require a far more complex mix of silicon than traditional computing, as well as accompanying storage capacity and connectivity bandwidth to handle the vast amounts of data needed in both the training and inference stages.

London data centres reflect AI trends

The potential for enterprise AI innovation and the challenges it presented is reflected by what is happening across colocation giant Digital Realty’s data centre estate in and around London as AI shifts to the top of hosting company’s customer agendas.

The UK capital and its surrounding areas has a high density of headquarter buildings and R&D offices, not just in financial services, but in other key industry verticals such as pharma, manufacturing, retail, and tech.

London is attractive because of the UKs political and legal stability, skilled workforce, and advanced tech infrastructure, explains Digital Realty CTO Chris Sharp, making it superb base both for innovation and for deploying AI applications and workloads.

Many enterprises will be acutely aware of issues around the general importance of data and IP and specific issues around data sovereignty and regulation, he adds.

“There’s a bit of nuance with training,” Sharp explains. “Nobody knows if it’s going to be able to be done anywhere and then inference has to abide by the [local] compliance [rules].” In addition, there’s an increasing understanding that one model cannot necessarily serve the world: “There’s going to be some regionality, so that will then also dictate the requirement for training facilities.”

At the same time, these organisations face the same technology challenges as other companies worldwide, particularly when it comes to putting in place and powering the infrastructure needed for AI.

It’s not enough to simply throw more CPUs at these workloads. One of the challenges with AI and HPC pipelines can be the different types of purpose-built hardware needed to efficiently support the complexity of these applications.

These range from CPUs to GPUs, even application-specific tensor processing units (TPUs) designed for neural networks, all with subtly different requirements, and all potentially playing a role in a customer’s AI pipeline. “Being able to support the full deployment of that infrastructure is absolutely top of mind,” points out Sharp.

Moreover, the balance between these platforms is set to change as AI projects move beyond development and into production. “If you take a snapshot, it’s 85 percent training, 15 percent inference today. But over the course of maybe 24 months, it’s 10 times more of a requirement to support inference,” he adds.

Flexing your AI smarts

So, the ability to flex and rebalance the underlying architecture as models evolve is paramount.
There is also the challenge of connecting this vast amount of data and compute together to deliver the AI workload performance levels required. While customers in the UK will have data sovereignty very much in mind, they still need to process workloads internationally when needed. And they may need to tap data oceans around the world. As Sharp says, “How do you connect these things together, because you’re not going to own all the data.”

But connectivity is not simply an external concern. “Within the four walls of the data centre we’re seeing six times the cable requirements [as] customers are connecting their GPUs, the CPUs, the network nodes. …. so, where we had one cable tray for fibre runs, now we have six times those cable trays, just to enable that.”

Hanging over all of this are the challenges associated with housing and powering this infrastructure. Just the density of technology required raises floor loading issues, Sharp explains. “The simple weight of these capabilities is massive.” And, as Digital Realty has found working with hyperscale cloud providers, floor loading requirements can increase incredibly quickly as projects scale up and AI technology advances.

Cooling too is always a challenge in data centres and as far as Sharp is concerned there is no longer a debate as to whether to focus on liquid or air cooling. “You need the ability to support both efficiently.”

When combined with the sheer density of processing power demanded by AI workloads, this is all having a dramatic effect on power demand across the sector. Estimations published by Schneider Electric last year suggest AI currently accounts for 4.5 GW of demand for data centre power consumption, predicted to increase at a compound annual growth rate (CAGR) of 25-33 percent to reach between 14 GW and 18.7 GW by 2028. That’s two to three times more demand for overall data centre power which is forecast to see a 10 percent CAGR over the same period).

All of which means that data centre operators must account for “more and more new hardware coming to market, requiring more power density, increasing in square footage required to support these burgeoning deployments.”

A state of renewal

That daunting array of challenges has informed the development of Digital Realty’s infrastructure in and around London, and its ongoing retrofitting and optimisation as enterprises scale up their AI operations.

The company has six highly connected campuses in the greater London area, offering almost a million square feet of colo space. But that doesn’t exist in isolation, with over 320 different cloud and network service providers across the city. “What we’re seeing today is that customers need that full product spectrum to be successful,” Sharp says.

Liquid cooling is a particular element in its London infrastructure. As liquid is 800 times denser than air, it can have a profound impact on efficiency. Digital Realty’s Cloud House data centre in London draws water from the Millwall dock for cooling, in a system that is up to 20 times more efficient than traditional cooling. Sensors ensure that only the required amount of water is used, and that it is returned to the dock unchanged.

But this ability to match the demands of corporations in and around London today and for the future also depends on Digital Realty’s broader vision.

All the power consumed by Digital Realty’s European operation is matched with renewable energy through power purchase agreements and other initiatives, while the company as a whole is contracted for over 1GW of new renewable energy worldwide.

At a hardware level, it has developed technologies such as its HD Colo product, which supports 70KW per rack, representing three times the requirement of certification for the Nvidia H100 systems which currently underpin cutting edge HPC and AI architectures.

At a macro level, as Sharp explains, Digital Realty plans its facilities years in advance. This includes “master planning the real estate, doing land banks and doing substations, making sure we pre-planned the power for five to six years.”

This requires close coordination from the outset with local authorities and utility providers, including investing in substations itself.

“We work extensively with the utility to make sure that not only the generation is there, but the distribution, and that they fortify the grid accordingly. I think that really allows customers of ours and our up the line suppliers, a lot of time to align to that demand.”

Cooling, power and infrastructure management complexities

It might be difficult to decide which is more complex. Developing cooling technologies and power management platforms that keep ahead of rapidly developing AI infrastructure or dealing with utilities and municipalities over a multiyear time horizon.
But tackling both is crucial as organisations look to stand up and expand their own AI capacity both quickly, and sustainably.

Sharp cites the example of one European education and research institution that needed to ramp up its own HPC infrastructure to support its AI ambitions, and knew it needed to utilise direct liquid to the chip. It would certainly have had the technical know-how to build out its own infrastructure. But once it began planning the project, it became clear that starting from scratch would have meant a five-to-six-year buildout. And that is an age in the current environment. Moreover, local regulations demanded it reduce their energy footprint by 25 percent over five years.

But partnering with Digital Realty, Sharp explains, it was able to deploy in one year, and using 100 percent liquid cooling improved its energy efficiency by 30 percent. As Sharp puts it, “It really helped them out rather quickly.”

Given how quickly the world has changed over the last 18 months, the ability to get an AI project up and running and into production that quickly is much more than a nice to have. For many enterprises, it’s going to be existential.

“Many AI deployments have failed, because there’s a lot of science and complexity to it,” says Sharp. But he continues, “We spend a lot of time removing complexity.”

Sponsored by Digital Realty.

Generative AI designs: subscribe your way to faster ROI

COMMISSIONED: If the pundit prognostications prove correct, adoption of generative AI in enterprises will mature in 2024. Especially if the money organizations expect to invest in Generative AI (GenAI) pays dividends.

Eighty-five percent of business leaders said they expect to boost their spending on AI and GenAI in 2024, according to Boston Consulting Group.

Of course, spending does not equal results, which will vary per business. As the year progresses, the narrative arcs will become more nuanced.

Yes, some organizations eager to gain a leg up on rivals will race to incorporate or build GenAI products and services. But many others will take a more measured approach.

From uncertainty about where to begin to concerns about the accuracy of large language models, IT leaders are still grappling with the best ways to use GenAI services as well as where to run these data-ravenous workloads.

But grapple they must – their CEOs demand it.

Sixty-eight percent of US CEOs believe AI will significantly change the way their companies create, deliver and capture value in the next three years, according to a PwC poll. “Significantly” in this context is code for digital transformation.

Optimizing workloads in the AI era

For more than a decade the public cloud has been the leading approach for adopting emerging technologies that help facilitate digital transformation. The public cloud helped organizations shrink the time it took to build and launch a prototype.

Most organizations looking to test the GenAI waters face a different calculus today, as the workload location landscape has expanded. The predominantly public cloud and on-premises mixed bag IT leaders lorded over in recent years has ballooned to become multicloud estates.

In these heterogeneous environments, applications run on-premises in datacenters, public and private clouds, colocation facilities and edge devices – call it a multicloud-by-default. Yet at a time when more bosses are seeking operational efficiency and lower IT costs, the happenstance isn’t sustainable.

Organizations today are becoming more prescriptive about their workload location choices, prioritizing performance, latency, security and other variables. Workload placement has become more intentional – a multicloud-by-design model.

And yes, that holds just as true in this new GenAI era. Today 82 percent of organizations prefer to take an on-premises or hybrid approach to building GenAI workloads, according to a Dell survey of IT decision makers. Hybrid meaning a combination of on-premises and public cloud resources.

Seeing the value of bringing AI to their data, some IT leaders are starting their GenAI implementations in-house. Leveraging off the shelf or open source models for security, control and governance over their data and models, IT departments are running pilots in their datacenters before electing to expand them to other locations.

Such leaders value the lower latency associated with running data-intensive models in house, where they can control throughput and resources and watch and refine their model output – while managing costs.

So while the public cloud remains a valuable location for placing GenAI workloads it isn’t the first or only consideration, marking a departure from common IT practices over the past decade. There are more options today than ever and CIOs are taking advantage of them.

The path to GenAI duccess

2023 proved to be a giant learning curve for GenAI, with organizations exploring the art of the possible. GenAI will most certainly mature in 2024; the question is how much.

As it is, most organizations remain unsure of how they can take advantage of GenAI to support their unique business needs, with 66 percent of business leaders ambivalent or dissatisfied with their progress on GenAI and AI initiatives, found BCG.

If you’re an IT leader uncertain how to proceed first identify critical use cases for GenAI, along with technology solutions to operate them. But you needn’t proceed alone.

Trusted partners offering professional services can help you work to bring your GenAI products to fruition. Dell Technologies Validated Designs for GenAI aim to improve your success rate with pre-tested hardware and software solutions.

These DVDs may be purchased as subscriptions via Dell APEX, providing access to agile infrastructure as you test and scale your GenAI projects. This allows you to only pay for what you use, avoiding unpredictable costs while aligning financial and operational needs as your footprint grows.

Moreover, with shadow AI giving rise to security vulnerabilities due to the lack of oversight organizations are seeking greater control over their data. Dell Validated Designs for GenAI delivered via APEX can help you bring AI to your data, while offering you peace of mind as you implement your service.

Whether you are looking to find your path or choose a different route, pick a partner that can help you on your GenAI journey.

Learn more about how Dell APEX for Generative AI can help you bring AI to your data.

Brought to you by Dell Technologies.

Observability is key in the AI era

Commissioned: The adage that you can’t manage what you can’t measure remains polarizing. Some leaders lean on analytics to improve business processes and outcomes. For such folks, aping Moneyball is the preferred path for execution.

Others trust their gut, relying on good old-fashioned human judgement to make decisions.

Regardless of which camp you IT leaders fall into – planting a foot in both is fashionable today – it’s becoming increasingly clear that analyzing the wealth of data generated by your IT estate is critical for maintaining healthy operations.

Analytics in a multicloud world

Analyzing sounds easy enough, given the wealth of tools designed to do just that, except that there is no one tool to measure everything happening in your datacenter. Moreover, more data is increasingly generated outside your datacenter.

The growing sprawl of multicloud environments, with applications running across public and private clouds, on-premises, colocation facilities and edge locations, has complicated efforts to measure system health. Data is generated in multiple clouds and regions and multiple sources, including servers, storage and networking appliances.

Add to that the data created by hundreds or thousands of apps running within corporate datacenters, as well as those of third-party hosts, and you can understand why data volumes are not only soaring but becoming more unwieldy. Data is poised to surpass more than 221,000 exabytes and hit a compound annual growth clip of 21 percent by 2026, according to IDC research.

Mind you those lofty stats were released before generative AI (GenAI) rocked the world last year, with text, video, audio and software code expanding the pools of unstructured data across organizations worldwide. Unstructured data will account for more than 90 percent of the data created each year, the IDC report found. Again – that’s before GenAI became The Thing in 2023.

One approach to measure everything

If only there were an app for measuring the health of such disparate systems and their data. The next best thing? Observability, or a method of inferring internal states of infrastructure and applications based on their outputs. Observability absorbs and extends classic monitoring systems to help IT pinpoint the root cause of issues by pushing intelligence to surface anomalies down to the endpoint.

When it comes to knowing what’s going on in your IT estate, observability is more critical now than ever thanks to the proliferation of AI technologies – GenAI in particular. Like many AI tools, GenAI learns from the data it’s fed, which means trusting that data is crucial.

Companies implementing LLMs or SLMs may use proprietary data to improve the efficacy of their solutions. They also want to prevent data loss and exfiltration. This puts a premium on observability, which provides a sort of God’s Eye View of IT system health.

Observability stacks typically include monitoring tools that continuously track system metrics, logs, traces and events, sniffing out bottlenecks across infrastructure and applications. Data collection tools, such as sensors, software agents and other instruments track telemetry data.

Modern observability stacks leverage AIOps, in which organizations use AI and machine learning techniques to analyze and interpret system data. For example, advanced ML algorithms can detect anomalies and automatically remediate issues or escalate them to human IT staff as necessary.

Moreover, observability and AIOps are taking on growing importance amid the rise of GenAI services, which are black boxes. That is, no one knows what’s happening inside them or how they arrive at their outputs.

Ideally, your AIOps tools will safeguard your GenAI and other AI technologies, offering greater peace of mind.

Observability must be observed

Automation often suggests a certain set it and forget-it approach to IT systems, but you must disabuse yourself of that notion. Even the monitoring must be monitored by humans.
For example, organizations that fail to capture data across all layers of their infrastructure or applications can succumb to blind spots that make them susceptible to system failures and downtime.

Moreover, failing to contextualize or correlate data across different systems can lead to mistaken interpretations and makes it harder for IT staff to pinpoint root causes of incidents, which can impact system reliability.

Finally, it is not unusual for thousands of incidents to be buried in billions of datapoints that an enterprise produces daily. IT departments should prioritize alerts based on their business impact. Even so, without the ability to analyze this data in real-time, generating actionable insights is impossible.

In today’s digital era, organizations can ill afford to suffer downtime. Gauging system health and resolving issues before they can denigrate the performance of IT infrastructure and applications is key.

You can’t watch everything. A partner can.

With so many business initiatives requiring IT resources, it can be hard for IT teams to monitor their own systems. Trusted partners can help you observe IT health.

Dell Technologies CloudIQ portfolio proactively monitors Dell hardware, including server, storage, hyperconverged infrastructure and networking technologies, as well as Dell APEX multicloud systems. Moreover, with its acquisition of Moogsoft Dell is rounding out its AIOps capabilities, supporting a multicloud-by-design strategy for running systems on premises, across clouds and on edge devices.

Sure, going with your gut works for certain things, but would you want to take that tack when it comes to your IT systems? For such a critical undertaking, holistic measurement is the key to managing your IT systems.

Learn more about Dell’s Cloud IQ portfolio by clicking this link.

Brought to you by Dell Technologies.

Safeguarding against the global ransomware threat

SPONSORED FEATURE: Ransomware is used by cybercriminals to steal and encrypt critical business data before demanding payment for its restoration. It represents one of, if not the most, serious cybersecurity threat currently facing governments, public/private sector organizations and enterprises around the world.

Infosec experts warn that data loss from attacks is often irreversible, even when the ransom is paid. Sterling Wilson, Data Resilience Strategist at Object First — the provider of Ootbi (Out-of-the-Box-Immutability), the ransomware-proof backup storage appliance purpose-built for Veeam — believes that data is one of the most precious assets available, and as such, must be tightly protected.

“Ransomware means the bad guys can sit by and get tens of millions of dollars from organizations that are literally dead in the water without their data,” he explains. “In all verticals, for all companies, data is their lifeblood.”

The potentially devastating consequences of ransomware attacks were highlighted by the UK House of Commons/House of Lords Joint Committee on the National Security Strategy in a December 2023 report, A hostage to fortune: ransomware and UK national security.

“Due to its potential ability to bring the UK to a standstill, ransomware has been identified by UK authorities as the number one cyber threat to the nation,” the report read.

“A mature and complex ecosystem has evolved, involving an increasingly sophisticated threat actor; ransomware is also now marketed as a service, which can be purchased by the uninvolved e.g., criminal gangs, making it more widely available to those who wish to inflict harm for profit. Past attacks have shown that ransomware can cause severe disruption to the delivery of core government services, including healthcare and child protection as well as ongoing economic losses.”

In a November 2022 report, the Royal United Services Institute (RUSI) warned that the impact of a ransomware attack expanded beyond the financial terms of the ransom payment to encompass other potentially serious factors, including business interruption and privacy liability costs, as well as the expense of hiring incident response firms, negotiators and crisis managers.

Lindy Cameron, CEO of the UK National Cyber Security Centre, which is part of GCHQ, adds that ransomware can be “truly devastating” for victims: “Attacks can affect every aspect of an organization’s operation, hitting finances, compromising customer data, disrupting operational delivery, eroding trust and damaging reputations. The impact will be felt in the short and long term, particularly when organizations are unprepared. Recovery is often lengthy and costly.”

Assume the worst and plan ahead

Object First’s Wilson warns that it’s really “not a matter of if, it’s a matter of when” a cyberattack attack will come. He explains that company cybersecurity teams need to assume that they will be targeted and develop a plan to secure their backup data before they get attacked.

“It used to sound so alarmist when we said this 18 to 24 months ago, but now we’re seeing the sense of alacrity with which all of these invasions are happening,” Wilson advises. “You know what’s coming. The first thing is to have a plan. Treat your data, your backups, and your infrastructure like it already has an attacker inside, and treat your most valuable possessions as such.”

Wilson describes the different types of backups that have historically been commonly deployed. The first of these is direct attached storage (DAS), which involves nothing more than a bunch of disks attached directly to the main servers. While this is the easiest option, it provides absolutely zero data protection in the event of a ransomware attack, as the bad guys will immediately own the backups.

The second option is to backup to separated physical media, such as tape and deduplication devices. While more secure, this approach takes inordinate amounts of time to invoke data recovery. The third option described by Wilson is to backup to an on-premises Hardened Linux Repository, which is a relatively more secure option as it uses some facets of object immutability.

However, these technically complex solutions are not completely immutable and require considerable expertise to set up and maintain. Offsite cloud backup is another option, but Wilson points out that recovering your whole company’s data from the cloud will usually be too slow for practical purposes and is subject to expensive egress fees.

Zero Trust principles apply

To address the shortcomings of these approaches, Wilson stresses the importance of following zero trust principles: “We’ve gotten convoluted messages on the ideas of zero trust, but the facets of zero trust are very succinct. You assume a breach, and you make sure you check and authenticate at every step of the way. You cannot trust the user who has authenticated at the edge of your network to move laterally through your network. You must make sure that every single place, every application, every point of data ingress and egress is secure.”

These zero trust security principles were fundamental to Object First when it engineered its secure backup storage appliance dubbed Ootbi (which stands for Out-of-the-Box Immutability). This ransomware-proof and immutable out-of-the-box solution delivers secure on-premises backups that are immutable by default and protected from cybercriminals.

Object First has a long-established and close relationship with Veeam, a provider of software that delivers secure backup and fast, reliable recovery solutions. Consequently, the Ootbi solution has been designed to synchronize perfectly with Veeam V12 direct-to-object storage configuration powered by the Smart Object Storage (SOS) API. The Ootbi locked-down Linux-based storage appliances, which incorporate a hardened operating system with no root or backend access by design, can scale linearly, supporting backup speeds up to 4.0 Gigabytes per second with up to half a petabyte of storage space.

“Ootbi provides the security that we need today,” Wilson says. “It is secure and simple to operate and deploy: from taking the appliance out of the box to running the first backup takes less than 15 minutes. It’s powerful. We grab the data in the best way using Veeam SOS storage API and place it into a unit that is completely secure using S3 object storage and all the facets of object immutability.”

According to Veeam’s 2023 Data Protection Trends Report, 85 percent of global organizations have suffered at least one cyberattack in the preceding twelve months; an increase from 76 percent experienced in the prior year. Conducted by an independent research firm, this survey polled 1,200 IT leaders whose companies suffered at least one ransomware attack in 2022 — including 350 in Europe. The research offers the following recommendations to maximize protection against ransomware:

– Immutable storage within disks, clouds and air-gapped media to ensure survivability.
– Hybrid IT architectures for recovering to alternative platforms like any other backup/disaster recovery strategy.
– Staged restorations to prevent re-infection during recovery.

Giving customers the data protection they need

A real-world example of the power of Object First’s Ootbi platform can be seen in its deployment by Prodatix, a US Veeam certified engineering company specializing in data backup, recovery and ransomware protection. The company provides a range of services including data management, Veeam consulting, Veeam licensing and backup appliances optimized for Veeam. Recognizing the critical importance of deploying a robust on-premises and immutable storage solution to ensure secure backup storage and resilience against ransomware attacks, Prodatix teamed up with Object First in 2022 as a beta partner.

Prodatix noted that Ootbi is purpose-built for Veeam and designed to create a seamless customer and partner experience. “Ootbi by Object First made sense as we know the market needs immutable on-premises storage, but we did not want to get into that business of building complex, large capacity immutable appliances,” said Prodatix’s vice president of technical sales, Matt Bullock. “Ootbi makes immutable storage simple, and we were excited by the 100 percent Veeam focus as we are a 100 percent Veeam shop.”

Having a demo by the Ootbi engineers really helped Prodatix to see the power of the appliance. And the company really appreciated the modified Linux platform that Object First developed to handle some of the immutable storage realities in Veeam’s VBR 12.
Bullock explained that Ootbi allows Prodatix to offer a robust immutable storage solution for Veeam and give customers the data protection they need.

“We present Ootbi as a necessary part of a data protection strategy and recovery plan,” Bullock said. “We also explain that the cloud is great, but you have to be able to restore from on-premises during and after a cyberattack to ensure you are reducing downtime.”

Ease-of-use and deployment are key advantages for Object First’s Oobti platform: “Object First’s Veeam-specific, easy to use, and stable immutable storage appliance ‘easy-button’ is the most impressive for the customers. Busy IT teams need a turn-key immutable appliance that works the first time, and every time, without a lot of time and interaction needed with the appliance.”

Looking to the future, Wilson cautions that infosec pros should brace themselves for a new wave of AI-generated attacks. “Security now touches everything,” he warns. “In the very same way that we leverage AI for some really cool things, it’s also being leveraged by the bad guys.”

In the future, we’re likely to see more AI and more impersonators pretending to be legitimate actors. “If someone has access to something, someone can impersonate them and get to that data,” points out Wilson. “But at Object First, we don’t offer that root access.”
And so that’s why it never has been more important to make sure that you’re keeping your backup data on-premises on storage appliances that are completely immutable out of the box.

Sponsored by Object First.

Effective storage strategies for a new year

WEBINAR: We all want flexible, cheap storage in the face of fast-growing data volumes. Blink and there’s more. A hundred thousand times more every second according to some estimates.

So making sure that data storage works efficiently for every company is central to any enterprise IT strategy worth its salt. But to do that requires a comprehensive understanding of what’s required coupled with the right storage solution to meet a specific business need.

StorPool says it has a fresh, ground-breaking approach to the conundrum that scopes the newest developments in block, file and object storage. The company has just published its ‘2024 Block Data Storage Buyer’s Guide’, a free reference resource designed to provide clear, practical, and proven guidance for IT practitioners and business leaders with the responsibility for deploying all types of enterprise-grade data storage solutions.

Find out more by joining our webinar on 25th January 4pm GMT/11am ET/8am PT. You’ll hear StorPool experts Alex Ivanov (Product Lead) and Marc Staimer (President, Dragon Slayer Consulting) explore the key challenges for organizations looking to upgrade their block data storage solutions in 2024, and share their insight into how to tackle real-world storage problems.

Alex and Marc will advise on how to avoid common pitfalls when devising an effective storage strategy and empower data storage buyers to compare and contrast different vendor block data storage systems while building a suitable infrastructure able to handle any amount of data growth on the horizon.

Deluged by data in 2024? This is one you really should take a look at. Sign up to watch the webinar here and we’ll send you a reminder when it’s time to log on.

Sponsored by StorPool.

Discover the ‘2024 Block Data Storage Buyer’s Guide’

Webinar: We consume data, and we create it – exponentially. The expectation is that we will be using 180 zettabytes of data globally by 2025.

This means that what we do with data, how we store and access it, is a vitally important part of enterprise IT strategy. A comprehensive understanding of the optimal approach to data storage for any business or organization is essential, and so too is ensuring the right storage solution is in place to meet specific business needs.

Join our webinar on 25th January at 4pm GMT/11am ET/8am PT to hear how StorPool can offer an up-to-date approach based on the latest block storage, file storage and object data storage technologies. The company has just published its ‘2024 Block Data Storage Buyer’s Guide’ a free reference resource that provides clear, practical, and proven guidance for IT practitioners and business leaders with the responsibility for deploying all types of enterprise-grade data storage solutions.

StorPool experts Alex Ivanov (Product Lead) and Marc Staimer (President, Dragon Slayer Consulting) will use the webinar to explore the key challenges for organizations looking to upgrade their block data storage solutions in 2024 and share insight on how to tackle real-world storage problems. They’ll also offer tips on how to avoid common pitfalls when devising an effective storage strategy and comparing and contrasting different vendor block data storage systems.

Sign up to watch our ‘Radical Practical Approach to Buying Block Data Storage’ webinar on 25th January here and we’ll send you a reminder when it’s time to log in.

Sponsored by StorPool.

2024 IT management trends: It’s generative AI and everything else

Commissioned: As the curtain falls on 2023, IT organizations are looking toward the new year with a mix of renewed enthusiasm and cautious optimism. The enthusiasm stems from the arrival of generative AI services a year ago.

Generative AI (GenAI) has emerged as perhaps the biggest productivity booster for knowledge work since the proliferation of word processing and spreadsheet software in the 1990s. In elevating customer operations, sales and marketing and software engineering, GenAI could add up to $4.4 trillion annually in productivity value to the global economy, according to McKinsey Global Institute.

The cautious optimism comes from IT leaders’ opportunity for modernizing corporate datacenters to accommodate GenAI and other data-hungry workloads. It’s complex work but executed well it can help organizations improve application performance and drive operational efficiency while curbing costs.

Divining proclamations from a proverbial crystal ball is an imprecise exercise. Few people expected GenAI would democratize AI for employees. But it did – and I’m here for it.

Without further ado, check out these IT management trends for 2024.

GenAI drives workload placement decisions

GenAI can shave hours off tasks that knowledge workers complete each day to do their jobs, potentially transforming industries. This is a big reason 52 percent of IT organizations are already building or deploying GenAI solutions, according to a Generative AI Pulse Survey conducted by Dell Technologies earlier this year.

In 2024, GenAI will accelerate workload placement trends, with organizations reckoning with how and where to run large language models (LLMs) that fuel their GenAI applications. Some IT decision makers will choose public services.

Others will run open source LLMs on premises, which will afford them the control as well as the ability to right-size workloads using domain-specific implementations and server clusters and other reliable infrastructure provided by partners.

Bringing AI to their data thusly will help organizations dictate security policies and access, create guardrails that reduce reputational risk and enjoy cost efficiencies.

Multicloud management becomes more seamless

Craving flexibility, more organizations will further abstract software functions from operating environments to run workloads in locations of their choosing.

This entails moving some storage options they traditionally run on premises into public clouds. This provides IT staff with more data mobility and consistency in managing their environments.

Others will make it possible for developers to access their preferred cloud services on premises. For instance, organizations might build custom GenAI chatbots to surface corporate information.

These approaches are part of a broader trend of trying to manage multicloud environments as seamless systems. This means treating the entire infrastructure estate as one entity to deliver greater operational efficiency and business value.

The edge of operation consolidation

Infrastructure that supports edge environments has historically been highly fragmented, with organizations stitching together solutions they hope will keep applications running at near real-time. Reducing latency has also been a big bugbear.

In 2024, you’ll see more organizations embrace edge operations approaches that simplify, optimize and secure deployment across complex multicloud estates, ensuring better uptime and service.

This will accelerate innovation across retail (smart shelves, anyone?), healthcare, automotive, agriculture, energy and several other sectors. Moreover, more organizations will explore how to extend GenAI applications to the edge to cultivate business value.

As-a-Service, at your service

Most IT staff are as comfortable with public cloud experiences as they are one-day shipping. Both offer agility and rapid service. Yet it’s also true that most organizations are weary of wrangling cloud services from different vendors, as well as the unpredictable costs associated with consuming said services.

In 2024, more organizations will seek to enjoy the same pay-as-you-go subscription models for infrastructure services but delivered on premises to their datacenters or colos of their choosing.

Such as-a-Service solutions balance flexibility with control, helping IT leaders pay only for what they require to run their business. This will help curb rising costs associated with resource-intensive workloads – such as GenAI and HPC apps – while affording IT more control over how it consumes compute and storage.

Multicloud-by-Design will evolve

Over the years, organizations have watched their applications sprawl across a number of operating locations, based on requirements for performance, latency, security, data portability or even whims.

As such, most IT organizations run apps on premises, in public and private clouds, in colos and at the edge – a kind of de facto multicloud estate. The location variance will grow significantly, with 87 percent of organizations expecting their application environment to become even more distributed over the next two years, according to a report – Unlocking the Power of Multicloud with Workload Optimization – published by the Enterprise Strategy Group in May 2023.

In 2024, more IT leaders will build multicloud-by-design estates, or intentionally constructed architectures intended to improve application performance and operation efficiency. This will also help meet regulatory requirements, control and secure assets and optimize costs.

Also: Given the large volumes of data they create GenAI apps will have an outsize influence over how IT leaders design their infrastructure, including shepherding staff as they build and train LLMs.

The key takeaway

You may have noted that GenAI is the thread woven throughout these trends. In fact, the most disruptive force in technology in 2023 will also remain the hottest workload in 2024.

IT leaders will have some critical decisions to make about what GenAI applications they run, as well as whether to operate them internally, externally or across multiple locations.

This will require careful consideration of the compute and storage, as well as the architecture that will situate and run them. A multicloud-by-design approach to IT infrastructure provides a smart, responsible path. And a trusted partner can light the way along their journey.

You don’t need a crystal ball to see that.

Brought to you by Dell Technologies.

Generative AI: It’s not just for the big guys

COMMISSIONED: Being stuck in the middle is no fun.

Just ask the titular character Malcom, of the TV series “Malcolm in the Middle,” (2000-2006) who struggles to stand out among his four brothers. In the earlier sitcom, “The Brady Bunch,” (1969-1974) Jan Brady feels overshadowed by big sister Marcia and younger sis Cindy.

Polite (or impolite) society has a name for this phenomenon, which describes the challenges children sandwiched between younger and elder siblings feel within their families: Middle child syndrome.

Reasonable minds differ on the legitimacy of middle child syndrome. Is it real or perceived and does it matter? Even so, it can be hard to compete with siblings – especially brothers or sisters who enjoy the lion’s share of success.

The middle children of the global economy

As it happens, the global economy has its own middle children in the form of small- to medium-sized businesses, which find themselves competing with larger enterprises for talent, capital and other vital resources.

Yet like their larger siblings, SMBs must innovate while fending off hungry rivals. This dynamic can prove particularly challenging as SMBs look to new technologies such as generative AI, which can be resource intensive and expensive to operate.

No organization can afford to overlook the potential value of GenAI for their businesses. Seventy-six percent of IT leaders said GenAI will be significant to transformative for their organizations and most expect meaningful results from it for within the next 12 months, according to recent Dell research.

Fortunately, SMBs wishing to capitalize on the natural language processing prowess of GenAI can do so – with the right approach: Using a small language model (SLM) and a technique called retrieval augmented generation (RAG) to refine results.
You may have noticed I called out an SLM rather than a large language model (LLM), which you are probably more accustomed to reading about. As the qualifiers imply, the difference between the model types is scale.

LLMs predict the next word in a sequence based on the words that have come before it to generate human-like text. Popular LLMs that power GenAI services such as Google Bard and ChatGPT feature hundreds of billions to trillions of parameters. The cost and computational resources to train these models is significant, likely putting building bespoke LLMs out of reach for SMBs.

SMBs have another option in building small language models (SLMs), which may range from a hundred million to tens of billions parameters and cost less to train and operate than their larger siblings.

SLMs can also be more easily customized and tailored for certain business use cases than LLMs. Whereas LLMs produces long form content, including whole software scripts and even books, SLMs can be used to build applications such as chatbots for customer service, personalized marketing content such as email newsletters and social media posts and lead generation and sales scripts.

Even so, whether you’re using an LLM or an SLM, GenAI models require enough computational resources to process the data, as well as data scientists to work with the data, both of which may be hard for SMBs to afford. And sure, organizations may use a pre-trained model but it will be limited by the information it knows, which means its accuracy and applicability will suffer.

RAG fine-tunes models with domain knowledge

Enter RAG, which can add helpful context without having to make big investments, thus democratizing access for SMBs. RAG retrieves relevant information from a knowledge repository, such as a database or documents in real time, augments the user prompt with this content and feeds the prompt into the model to generate new content. This helps the model generate more accurate and relevant responses for the information you wish your model to specialize in.

For example, at Dell we show organizations how to deploy RAG and Meta’s LLama2 LLM to retrieve domain-specific content from custom PDF datasets. The output was used to show how an organization might theoretically use RAG and an LLM to train a help-desk chatbot.

SMBs can use an SLM with RAG to build a more targeted and less resource-intensive approach to GenAI. Effectively, the combination affords them a very accurate tool that delivers more personalized information on their company’s data – without spending the time and money building and fine-tuning a custom model.

Getting started with RAG may seem daunting to SMBs but organizations can repurpose a server, a workstation or even a laptop and get started. They can pick an open-source LLM (such as LLama2) to begin the process. Dell calls this the GenAI easy button.

That way organizations can bring the AI to their data, keeping control of their sensitive corporate IP while freeing up IT resources as they innovate.

SMBs play an important role in the economy by contributing to innovation. Yet too often they’re relegated to Malcom or Jan status – the oft-underestimated and neglected middle children of the global economy.

By combining the right approach and technology tools, SMBs can leverage GenAI to accelerate innovation, enabling them to better compete and woo new customers – rather than feeling lost in the middle of the corporate pack.

To learn more, visit dell.com/ai.

Brought to you by Dell Technologies.

What is the best choice of storage in the container era?

ADVERTORIAL: Containerized deployment of applications has become increasingly popular. In fact, the evolution of container technologies has witnessed many ups and downs over the past 20 years.

Containers isolate application runtime environments based on process management. Although server virtualization emerges as another technical roadmap and gives birth to cloud service models such as infrastructure as a service (IaaS) and platform as a service (PaaS), container technologies have always maintained a competitive edge. The launch of Docker in 2013 is a major milestone. Docker allows applications to be packaged once and deployed multiple times without repeated coding. This capability delivers the high-level of application portability needed for agile development and elastic deployment in the Internet era.

The Kubernetes project released in 2015 provided users with a universal container management tool that eliminated the weakest link in container commercialization.

However, commercial container deployment still faces numerous challenges.

Challenges in container storage

According to the user survey report released by the Cloud Native Computing Foundation (CNCF) in 2020, 78 percent of users have deployed or are about to deploy stateful applications in containers in application scenarios like databases, message queues, logs, and data analysis. The report also showed that storage has become a major factor that affects the commercial deployment of containers.

Data source: CNCF SURVEY 2020

Prolonged data replication: A cluster is typically deployed across at least three nodes, with each node retaining at least one data copy. In the event of a failure, containerized applications can be switched over or a secondary node can be added. This causes time-consuming continuous data synchronization between nodes.

Lack of service continuity solutions: Currently, most mission-critical enterprise services are stateful applications that require high levels of disaster recovery (DR). However, the container storage interface (CSI) specifications lack comprehensive DR standards and strong support for DR, and therefore are unable to ensure enterprise service continuity.

Poor adaptation to persistent storage: The native Kubernetes does not support stateful storage. As a result, the use of persistent storage has always been a major challenge in the container field.

Storage selection for container environments

Facing a variety of storage devices, enterprises need to choose the right one for their container environments. Professional storage is recommended for enterprise-level containerized applications, regardless of whether they are stateful or not. This is because professional storage not only facilitates data sharing, cluster expansion, and fault tolerance but also ensures service availability, allowing enterprises to fully harness the potential of container technologies.

Among numerous choices of container storage, NAS stands out in this area as the optimal data foundation for containers.

Enterprise-level NAS storage delivers excellent performance and supports data sharing and access between multiple containers, making it ideal for a wide range of application scenarios, including SQL and NoSQL databases, data analysis, continuous integration and continuous delivery (CI/CD), high-performance computing, content management, data processing, and machine learning.

Huawei’s solution for container storage

As a founding member of the Open Container Initiative (OCI) as well as a founding member and platinum member of the CNCF, Huawei ranks No. 1 in Asia and No. 4 in the world in terms of contribution to the Kubernetes community. Huawei is also one of the world’s first Kubernetes Certified Service Providers and has always been committed to promoting container technologies and applications.

To help enterprises better address container storage challenges, Huawei has launched an industry-leading container storage solution that fully adapts to industry CSI standards while also incorporating additional enterprise-level enhancements.

Huawei’s container storage solution

Huawei CSI: Enterprise-Level Enhancement of CSI Capabilities: In addition to supporting standard CSI functions such as topology, static volume, dynamic volume, ephemeral volume, volume snapshot, and volume clone, Huawei CSI has expanded QoS, multi-tenant, quota, and active-active storage capabilities. Moreover, Huawei CSI has further improved disk scanning and concurrent processing of RESTful commands to achieve efficient provisioning and scheduling of storage resources, delivering 30 percent higher performance of batch container deployment than the competition. Huawei CSI is also compatible with multiple mainstream ecosystems.

Huawei Container Disaster Recovery (CDR): All-in-One Container Application Backup and DR: As previously discussed, general CSI standards fail to meet enterprise-level DR requirements. So, Huawei CDR has risen to the occasion. This CDR component innovatively uses resource groups to automatically associate container-related resources for an application, achieving application-level data DR and backup. It also works with BCManager, Huawei’s management software for business continuity, to make DR and backup easier and more intelligent.

Huawei Container Storage Monitor (CSM): Bidirectional Visualization Between Container Applications and Storage Resources: In traditional scenarios, containers and storage resources cannot be effectively associated for efficient management, greatly hindering resource planning and routine O&M. By integrating Prometheus and Grafana, two mainstream tools in the industry, Huawei CSM implements bidirectional visualization between managed objects and mappings and provides a unified resource view to help users simplify O&M.

OceanStor Dorado: Industry-Leading All-Flash NAS Storage: Finally, Huawei OceanStor Dorado NAS all-flash storage outperforms its rivals by 30 percent in container application scenarios thanks to its cutting-edge features such as the unique full-mesh controller architecture, distributed RAID 2.0+, and smart disk enclosures shared between controller enclosures. Furthermore, it delivers device-, data-, and application-level security and reliability thanks to its end-to-end three-level security and reliability design, one-of-a-kind NAS active-active cluster architecture, and proprietary four-layer ransomware protection solution powered by machine learning and intelligent algorithms. These advantages combine to maintain enterprise service continuity without any data loss, facilitating digital transformation across a wide range of industries.

The best choice for container storage

Huawei OceanStor Dorado has been named a Leader in Gartner’s Magic Quadrant for Primary Storage for 8 consecutive years thanks to its exceptional product capabilities and the overwhelming praise it has received from customers and partners.

Powered by OceanStor Dorado, Huawei’s container storage solution further improves data replication performance, service reliability, and application deployment efficiency in container environments, helping enterprise customers smoothly implement intelligent digital transformation.

This makes Huawei storage the best choice for any enterprise looking at container solutions.

Click here for more information about Huawei’s container storage solution.

Contributed by Huawei.

Building cyber resilience with data vaults

SPONSORED: In August 2023, Danish hosting subsidiaries CloudNordic and AzeroCloud were on the receiving end of one of the most serious ransomware attacks ever made public by a cloud services company. During the incident, CloudNordic suffered a complete encryption wipe-out that took with it applications, email services, websites, and databases, and even backup and replication servers. In a memorably frank admission, the company said that all customer data had been lost and would not be recoverable.

To the hundreds of companies Danish media reported as having lost data in the incident, this must have sounded incredible. Surely service providers are supposed to offer protection, not even greater vulnerability? Things were so bad, CloudNordic even offered customers last resort instructions on recovering lost website content through the Wayback Machine digital archive. The company reportedly refused to pay a ransom demanded by the attackers but even if it had paid there is no guarantee it would have made any difference.

Ransomware attacks are a dime a dozen these days and the root causes are various. But the assumption every customer makes is that behind a service provider’s virtual machine (VM) infrastructure is a comprehensive data protection and disaster recovery (DR) plan. Despite the common knowledge that ransomware targets backup and recovery systems, there is still a widespread belief that the same protections will always ride to the rescue and avoid catastrophic data loss. The CloudNordic attack is a warning that this isn’t always the case. Doubtless both companies had backup and data protection in place, but it hadn’t been enough.

“The attack and its outcome is not that extraordinary,” argues Kevin Cole, global director for technical product marketing at Zerto, a Hewlett Packard Enterprise company. “This probably happens more than we know. What’s unusual about this incident is simply that the service provider was open about the fact their backups had been attacked and deleted.”

This is what ransomware has done to organizations across the land. Events once seen as extreme and unusual have become commonplace. Nothing feels safe. Traditional assumptions about backup and data resilience are taking a battering. The answer should be more rapid detection and response, but what does this mean in practice?

The backup illusion

When responding to a ransomware attack, time is of the essence. First, the scale and nature of the incursion must be assessed as rapidly as possible while locating its source to avoid reinfection. Once this is within reach, the priority in time-sensitive industries is to bring multiple VM systems back online as soon as possible. Too often, organizations lack the tools to manage these processes at scale or are using tools that were never designed to cope with such an extreme scenario.

What they then fall back on is a mishmash of technologies, the most important of which is backup. The holes in this approach are well documented. Relying on backup assumes attackers haven’t deactivated backup routines, which in many real-world incidents they manage to do quite easily. That leaves offline and immutable backup, but these files are often old, which means that more recent data is lost. Even getting that far takes possibly days or weeks of time and effort.

Unable to contemplate a long delay, some businesses feel they have no option but to risk paying the ransom in the hope of rescuing their systems and data within a reasonable timescale. Cole draws a distinction between organizations that pay ransoms for strategic or life and death reasons – for example healthcare – and those who pay because they lack a well-defined strategy for what happens in the aftermath of a serious attack.

“Organizations thought they could recover quickly only to discover that they are not able to recover within the expected time window,” he explains. “They pay because they think it’s going to ease the pain of a longer shutdown.”

But even this approach is still a gamble that the attackers will hand back more data than will be recovered using in-house backup and recovery systems, Cole points out. In many cases, backup routines were set up but not properly stress tested. Under real-world conditions, poorly designed backup will usually fall short as evidenced by the number of victims that end up paying.

“Backup was designed for a different use case and it’s not really ideal for protecting against ransomware,” he says. “What organizations should invest in is proper cyber recovery and disaster recovery.”

In the end, backup falls short because even when it works as advertised the timescale can be hugely disruptive.

Achieving ransomware resilience

It was feedback from customers using the Zerto solution to recover from ransomware that encouraged the company to add new features tailored to this use case. The foundation for the Zerto solution is its continuous data protection (CDP) technology, with its replication and unique journaling technology, which reached version 10 earlier this year. Ransomware resilience is an increasingly important part of this suite, as evidenced by version 10’s addition of a real-time anomaly system that can detect that data is being maliciously encrypted.

Intercepting ransomware encryption early not only limits its spread but makes it possible to work out which volumes or VMs are bad and when they were infected, so that they can be quickly rolled back to any one of thousands of clean restore points.

“It’s anomaly and pattern analysis. We analyze server I/O on a per-volume basis to get an idea of what the baseline is at the level of virtual machines, applications and data,” explains Cole. “Two algorithms are used to assess whether something unusual is going on that deviates from this normal state.”

An important element of this is that Zerto is agentless which means there is no software process for attackers to disable in order to stop backup and replication from happening behind the victim’s back.

“It sounds small but it’s a really big advantage,” says Cole. “Many ransomware variants scan for a list of backup and security agents, disabling any they find protecting a VM. That’s why relying on a backup agent represents a potential weakness.”

A second advanced feature is the Zerto Cyber Resilience Vault, a fully isolated and air-gapped solution designed to cope with the most serious attacks where ransomware has infected the main production and backup infrastructure. Zerto stresses that this offers no point of compromise to attackers – replication from production via a ‘landing zone’ CDP mirror happens periodically via an FIPS-validated encrypted replication port rather than a management interface which might expose the Vault to compromise.

The possibility of a total compromise sounds extreme, but Cole points out that the use of this architecture is being mandated for financial services by the SEC in the U.S., and elsewhere by a growing number of cyber-insurance policies. The idea informing regulators is that organizations should avoid the possibility of a single point of failure.

“If everything blows up, do you have copies that are untouchable by threat actors and which they don’t even know exist?,” posits Cole. “In the case of the Cyber Resilience Vault, it’s not even part of the network. In addition, the Vault also keeps the Zerto solution itself protected – data protection for the data protection system.”

Ransomware rewind

The perils of using backup as a shield against ransomware disruption are underscored by the experience of TenCate Protective Fabrics. In 2013 this specialist manufacturer of textiles had multiple servers at one of its manufacturing plants encrypted by the infamous CryptoLocker ransomware. This being the early days of industrial ransomware, the crippling power of mass encryption would have been a shock. Tencate had backups in place but lost 12 hours of data and was forced to ship much of its salvageable data to a third party for slow reconstruction. In the end, it took a fortnight to get back up and running.

In 2020, a different version of CryptoLocker returned for a second bite at the company, this time with very different results. By now, Tencate was using Zerto. After realizing that one of its VMs had been infected, the security team simply reverted this to a restore checkpoint prior to the infection. Thanks to Zerto’s CDP, the total data loss was a mere ten seconds and the VM was brought back up within minutes.

According to Cole, TenCate’s experience shows how important it is to invest in a CDP that can offer a large number of recovery points across thousands of VMs with support for multi-cloud.

“Combined with encryption detection, this means you can quickly roll back, iterating through recovery points that might be only seconds apart until you find one that’s not compromised.”

While loss of service is not the only woe ransomware causes its victims, the inability to run applications and process data is where the immediate economic damage always begins. For the longest time, the only remedy was to keep the attackers out. But when those defenses fail as they surely will one day, it is better to fail in style, says Cole.

“The choice is not just backup as people have come to know it,” he advises. “Continuous data protection and isolated cyber recovery vaults are the future anyone can have now.”

Sponsored by Zerto.

How to stop ransomware thieves WORMing their way into your data

SPONSORED FEATURE: Most of us dislike cyber criminals, but not many of us dislike them quite as much as Anthony Cusimano. The director of technical marketing at storage company Object First was on the sharp end of an identity theft attack after his details were leaked in the massive 2017 Equifax breach. Thieves armed with these details SIM-jacked his phone, used it to authenticate into his PayPal account, then stole money from Cusimano and his family.

“I became passionate about security for both individuals and businesses,” he says.

The attack inspired Cusimano to join the battle against cyber crime and move increasingly into more cybersecurity-focused roles. Today, he spends his working day at Object First helping customers understand the importance of protecting their data from a range of attacks.

Object First specialises in protecting data from encryption by ransomware crooks. Its solution, Ootbi, is designed specifically to work with Veeam backup solutions, providing extra protection in the form of data immutability.

The company was founded by Ratmir Timashev and Andrei Baronov, who started Veeam as a backup company for VMware virtual machines in 2006 and then expanded quickly, building it first into a multi-faced backup solution and then into a data management empire.

However, one thing that the two didn’t have was a purpose-built object storage system for Veeam. They wanted a hardware appliance that would work seamlessly with their backup software, providing customers with a way to easily store backup data on their own premises, fed directly from Veeam’s system. They had specific requirements in mind, the most important of which was to make that backup data tamper-proof.

Timashev and Baronov understood the security risks facing stored data and backups. They had made great progress getting companies to back up their data properly in the first place by creating automated solutions that made it more convenient.

Nice bit of backup data you have there

Then, along came the spectre of ransomware. Beginning as badly-coded malware released ad hoc by individuals or small groups, it exploded into a sophisticated business model with professionally written code.

As more victims hit the headlines, the spread of ransomware hammered home the need to back up your data.

Then, the crooks started coming for the backups.

Data backups were a form of business risk to these new, grown-up ransomware gangs. Like any business, they sought to eliminate the risk. They did it by seeking out backup servers and encrypting or deleting those, too, leaving victims more inclined to pay them.

One answer to this is write once read many (WORM) disks, or storage taken offline. WORM disks can’t be overwritten, but they are expensive and difficult to manage. Offline hard drives or tape must be connected to the system and then disconnected when the backups are complete, all in the hope that ransomware doesn’t target them while they’re online.

In search of indelible data

Instead, Object First wanted a system that combined the advantages of both; the immutability of a WORM disk with the convenience of online backup storage that could stay permanently connected to the network. And, naturally, they wanted a solution built specifically for Veeam.

This is what prompted them to begin creating Ootbi (it stands for ‘out of the box immutability’) three years ago, which eventually led to Object First.

“Ootbi is based on the idea of resiliency domains”, explains Cusimano. “You treat every single software stack you have as an individual resiliency domain. If one gets compromised, you still have the others to lean on and recover from.”

One component of this is the 3-2-1-1-0 rule: this means, storing three copies of your data, in addition to the original, across two media types, one of which must be off-site. Ootbi satisfies both of these by storing one in the cloud and the other on the customer’s premises on its own appliance’s NVME flash storage.

That leaves another one and a zero. The zero refers to zero errors, meaning that the storage solution must check that the data is clean going in so that you’re not restoring garbage later. The one means that one of the copies must be kept offline, or air-gapped, so that no one can tamper with it.

Ootbi didn’t air-gap this data by taking it physically offline. It wanted to handle the offline storage within its own network-connected appliance for maximum efficiency and user convenience.

“How do we make something where the backup lands on a box and there is no digital way that data can be removed from the box once it gets there?” says Cusimano. “That’s what we built.”

The inner workings of immutability

To build an immutable but connected backup appliance, Object First began by locking down the box as much as possible. Any attacker hoping for privilege escalation on the Linux-based product has a surprise in store: there’s no basic or root account that is accessible to users on its hardened version of their customized Linux OS.

Unsurprisingly given its name, Object First also opted for native object storage out of the box with its appliance. Whereas file and block-based storage models tend to store data in hierarchical structures, object storage stores data as uniquely-identifiable units with their own metadata in a single bucket.

Object storage has its historical drawbacks, the main one being its slower speed relative to file and block approaches. However, this is a backup appliance rather than a transactional one, and in any case it uses extremely fast NVME flash for write caching.

Because it’s built exclusively for Veeam, the technology also takes advantage of some proprietary work that Veeam did in building its data communications on the Amazon S3 API andVeeam’s SOS (Smart Object Storage) API. That enables the backup appliance to eke more performance out of Amazon’s cloud-hosted Simple Storage Service than other solutions can, Cusimano says. Ootbi also avoids any compression or de-duplication overhead because Veeam already takes care of those tasks.

Tight integration gives Ootbi support for all Veeam functionality, including simple backup, restore, disaster recovery, Instant Recovery, SureBackup, and hybrid scenarios. The appliance can run failed Instant Recovery workloads directly from backup within minutes, according to Object First.

Object storage also scales quickly and simply thanks to the GUID object labelling. This makes it good at scaling to handle large amounts of static, unstructured data.

“Because the concept was created in the last 20 years, it doesn’t have the kind of baggage that that file or block carries,” he adds.

The company not only configured its own hardened Linux distribution but also its own customized file system that communicates using the S3 API, which while developed by Amazon is now available as an open protocol.

“We’ve modified our own file system and we’ve created our own object storage code base,” Cusimano says. “That’s proprietary, so we’re running our own special sauce on this very normal box.”

The S3 API enabled Object First to take advantage of object lock. This introduces write-once-ready-many (WORM) immutability to stop an attacker doing anything even if they did somehow compromise the box. Explicitly built for object storage, it has two modes: governance, and compliance.

Governance mode prevents people overwriting, deleting, or altering the lock settings of a stored object unless they have special permissions. Compliance mode, which is the only mode used in Ootbi’s immutable storage, prevents any protected object from being altered or deleted by anyone for the designated retention period (set by the user in Veeam Backup and Recovery).

Software is key

The hardware is effectively a JBOD appliance, with up to ten 16Tb hard drives, another hot spare drive, and a 1.6Tb NVME that acts as a data cache. The hard drives form a RAID 6 array, storing data parity information twice, so that data is recoverable even if two disks fail. This gives customers up to 128Tb of available backup capacity, along with fast data reading thanks to multi-disk striping.

Data arrives from Veeam through two 10Gbit/sec NICs and lands on the NVME cache, which provides a 1Gb per second write speed per node.

The system is designed with expandability in mind. Customers can build a cluster of up to four Ootbi appliances, adding nodes when necessary. This not only increases capacity, but also speed, as each appliance’s built-in NIC provides another 1Gb/sec of write speed. It only supports a maximum four-node implementation today, but that’s because the company is a small startup focusing on its first sales. The design of its software architecture will allow it to increase that threshold as demand comes in from customers, Cusimano says.

Object First also tailored the system for usability, with an interface that relatively non-technical people can use.

“There’s no operating system updates. There’s nothing they have to do to make this thing work. You plug it in, you rack and stack the box, you hook it up to your network. You go through two different NIC configurations inside of a text user interface, give it a username and password, and you’re configured,” Cusimano says. The system automatically optimises its storage, minimising the amount of on-site storage expertise that customers need.

Data backups alone aren’t a gold-plated protection against more modern ransomware business models. Double-extortion ransomware gangs will steal your data even if they can’t encrypt it, meaning that restoring scrambled files will only solve half of your problems.

With that said, backup protection forms a critical part of a multi-layered defence-in-depth solution that should include employee awareness, anti-phishing scans and malware protection. It will enable you to continue operating after a ransomware attack, making that data immutability worth every penny of your investment.

Sponsored by Object First.

The generative AI easy button: How to run a POC in your datacenter

Commissioned: Generative AI runs on data, and many organizations have found GenAI is most valuable when they combine it with their unique and proprietary data. But therein lies a conundrum. How can an organization tap into their data treasure trove without putting their business at undue risk? Many organizations have addressed these concerns with specific guidance on when and how to use generative AI with their own proprietary data. Other organizations have outright banned its use over concerns of IP leakage or exposing sensitive data.

But what if I told you there was an easy way forward already sitting behind your firewall either in your datacenter or on a workstation? And the great news is it doesn’t require months-long procurement cycles or a substantial deployment for a minimum viable product. Not convinced? Let me show you how.

Step 1: Repurpose existing hardware for trial
Depending on what you’re doing with generative AI, workloads can be run on all manner of hardware in a pilot phase. How? There are effectively four stages of data science with these models. The first and second, inferencing and Retrieval-Augmented-Generation (RAG), can be done on relatively modest hardware configurations, while the last two, fine-tuning/retraining and new model creation, require extensive infrastructure to see results. Furthermore, models can be of various sizes and not everything has to be a “large language model”. Consequently, we’re seeing a lot of organizations finding success with domain-specific and enterprise-specific “small language models” that are targeted at very narrow use cases. This means you can go repurpose a server, find a workstation a model can be deployed on, or if you’re very adventurous, you could even download LLaMA 2 onto your laptop and play around with it. It’s really not that difficult to support this level of experimentation.

Step 2: Hit open source
Perhaps nowhere is the open-source community more at the bleeding edge of what is possible than in GenAI. We’re seeing relatively small models rivaling some of the biggest commercial deployments on earth in their aptitude and applicability. The only thing stopping you from getting started is the download speed. There are a whole host of open-source projects at your disposal, so pick a distro and get going. Once downloaded and installed, you’ve effectively activated the first phase of GenAI: inferencing. Theoretically your experimentation could stop here, but what if with just a little more work you could unlock some real magic?

Step 3: Identify your use cases
You might be tempted to skip this step, but I don’t recommend it. Identify a pocket of use cases you want to solve for. The next step is data collection and you need to ensure you’re grabbing the right data to deliver the right results via the open source pre-trained LLM you’re augmenting with your data. Figure out who your pilot users will be and ask them what’s important to them – for example, a current project they would like assistance with and what existing data they have that would be helpful to pilot with.

Step 4: Activate Retrieval-Augmented-Generation (RAG)
You might think adding data to a model sounds extremely hard – it’s the sort of thing we usually think requires data scientists. But guess what: any organization with a developer can activate retrieval-augmented generation (RAG). In fact, for many use cases this may be all you will ever need to do to add data to a generative AI model. How does it work? Effectively RAG takes unstructured data like your documents, images, and videos and helps encode them and index them for use. We piloted this ourselves using open-source technologies like LangChain to create vector databases which enable the GenAI model to analyze data in less than an hour. The result was a fully functioning chatbot, which proved out this concept in record time.



Source: Dell Technologies

In Closing

The unique needs and capabilities of GenAI make for a unique PoC experience, and one that can be rapidly piloted to deliver immediate value and prove its worth to the organization. Piloting this in your own environment offers many advantages in terms of security and cost efficiencies you cannot replicate in the public cloud.

Public cloud is great for many things, but you’re going to pay by the drip for a PoC, it’s very easy to burn through a budget with users who are inexperienced at prompt engineering. Public cloud also doesn’t offer the same safeguards for sensitive and proprietary data. This can actually result in internal users moving slower as they think through every time they use a generative AI tool whether the data they’re inputting is “safe” data that can be used with that particular system.

Counterintuitively, this is one of the few times the datacenter offers unusually high agility and a lower up front cost than its public cloud counterpart.

So go forth, take an afternoon and get your own PoC under way, and once you’re ready for the next phase we’re more than happy to help.

Here’s where you can learn more about Dell Generative AI Solutions.

Brought to you by Dell Technologies.

Allocating AI and other pieces of your workload placement puzzle

COMMISSIONED: Allocating application workloads to locations that deliver the best performance with the highest efficiency is a daunting task. Enterprise IT leaders know this all too well.

As applications become more distributed across multiple clouds and on premises systems, they generate more data, which makes them both more costly to operate and harder to move as data gravity grows.

Accordingly, applications that fuel enterprise systems must be closer to the data, which means organizations must move compute capabilities closer to where that data is generated. This helps applications such as AI, which are fueled by large quantities of data.

To make this happen, organizations are building out infrastructure that supports data needs both within and outside the organization – from datacenters and colos to public clouds and the edge. Competent IT departments cultivate such multicloud estates to run hundreds or even thousands of applications.

You know what else numbers in the hundreds to thousands of components? Jigsaw puzzles.

Workloads Placement and… Jigsaw Puzzles?

Exactly how is placing workloads akin to putting together a jigsaw puzzle? So glad you asked. Both require careful planning and execution. With a jigsaw puzzle – say, one of those 1,000-plus piece beasts – it helps to first figure out how the pieces fit together, then assemble them in the right order.

The same is true for placing application workloads in a multicloud environment. You need to carefully plan which applications will go where – internally, externally, or both – based on performance, scalability, latency, security, costs and other factors.

Putting the wrong application in the wrong place could have major performance and financial ramifications. Here are 4 workload types and considerations for locating each, according to findings from IDC research sponsored by Dell Technologies.

AI – The placement of AI workloads is one of the hottest topics du jour, given the rapid rise of generative AI technologies. AI workloads comprise two main components – inferencing and training. IT departments can run AI algorithm development and training, which are performance intensive, on premises, IDC says. And the data is trending that way, as 55 percent of IT decision makers Dell surveyed cited performance as the main reason for running GenAI workloads on premises. Conversely, less intensive inferencing tasks can be run in a distributed fashion at edge locations, in public cloud environments or on premises.

HPC – high-performance computing (HPC) applications ALSO comprise two major components – modeling and simulation. And like AI workloads, HPC model development can be performance intensive, so it may make sense to run such workloads on premises where there is lower risk of latency. Less intensive simulation can run reliably across public clouds, on premises and edge locations.

One caveat for performance-heavy workloads that IT leaders should consider: Specialized hardware such as GPUs and other accelerators is expensive. As a result, many organizations may elect to run AI and HPC workloads in resource-rich public clouds. However, running such workloads in production can cause costs to soar, especially as the data grows and the attending gravity increases. Moreover, repatriating an AI or HPC workload whose data grew 100x while running in a public cloud is harsh on your IT budget. Data egress fees may make this prohibitive.

Cyber Recovery – Organizations today prioritize data protection and recovery, thanks to threats from malicious actors and natural disasters alike. Keeping a valid copy of data outside of production systems enables organizations to recover lost or corrupted due to an adverse event. Public cloud services generally satisfy organizations’ data protection needs, but transferring data out becomes costly thanks to high data egress fees, IDC says. One option includes hosting the recovery environment adjacent to the cloud service – for example, in a colocation facility that has a dedicated private network to the public cloud service. This eliminates egress costs while ensuring speedy recovery.

Application Development – IT leaders know the public cloud has proven well suited for application development and testing, as it lends itself to the developer ethos of rapidly building and refining apps that accommodate the business. However, private clouds may prove a better option for organizations building software intended to deliver a competitive advantage, IDC argues. This affords developers greater control over their corporate intellectual property, but with the agility of a public cloud.

The Bottom Line

As an IT leader, you must assess the best place for an application based on several factors. App requirements will vary, so analyze the total expected ROI of your workloads placements before you place them.

Also consider: Workload placement is not a one-and-done activity. Repatriating workloads from various clouds or other environments to better meet the business needs is always an option.

Our Dell Technologies APEX portfolio of solutions accounts for the various workload placement requirements and challenges your organization may encounter as you build out your multicloud estate. Dell APEX’ subscription consumption model helps you procure more computing and storage as needed – so you can reduce your capital outlay.

It’s true: The stakes for assembling a jigsaw puzzle aren’t the same as allocating workloads in a complex IT environment. Yet completing both can provide a strong feeling of accomplishment. How will you build your multicloud estate?

Learn more about how Dell APEX can help you allocate workloads across your multicloud estate.

Brought to you by Dell Technologies.

Five steps to mastering multicloud management

Three clouds
Three clouds

COMMISSIONED: As an IT leader, you have the daunting task of managing multicloud environments in which applications run on many public clouds and on-premises environments – and even the edge.

Operating in this cloud-native space is like conducting a symphony orchestra but with a major catch: You’re managing multiple orchestras (cloud providers and on-premises systems), rather than one. Each orchestra features its own musicians (resources), unique instruments (services) and scores (configurations). With so many moving parts, maintaining harmony is no trivial pursuit.

Take for example the relationship between different parts of a DevOps team. While developers build and test containerized applications in multiple public clouds, operators are playing a different tune, focused on scaling on-premises systems. These teams accumulate different management interfaces, technologies and tools over time. The result? Layers upon layers of complexity.

This complexity is then magnified when we consider data management challenges. Shuttling data between public clouds and on-premises systems becomes harder with the increase of data gravity, and even container orchestration tools, such as Kubernetes, require deep technical skills when it comes to managing persistent storage. And the data reflects this struggle. According to a recent VMware State of Kubernetes research report published in May, 57 percent of those surveyed cited inadequate internal experience as a main challenge to managing Kubernetes.

Achieving consistent performance, security and visibility across the entire symphony – your IT infrastructure – remains a hurdle when you account for the different instruments, music sheets and playing styles.

A playbook for managing multicloud and Kubernetes storage

There is no silver bullet for managing storage and Kubernetes in multicloud environments. But here are some steps you can take for running modern, cloud-native apps across public clouds and on-premises systems alike.

– Educate. Your IT operations and DevOps teams must learn about the various container and storage services available across public cloud and on-premises systems, as well as how to manage the technologies and tools that make them hum.

– Standardize. You’re not going to have a single tool to handle all of your needs, but you can simplify and streamline. For instance, standardizing on the same storage can help you reduce complexity and improve efficiency across multiple clouds and on-premises environments, making it easier to respond nimbly to shifts in your multicloud strategy. Pro tip: Examine where applications and data are living between on-premises systems and public clouds and whether those workload placements align with your goals.

– Automate. One of IT’s greatest magic tricks is automating manual tasks. Why manage containers and storage piecemeal? Tools exist to help IT operations staff automate chores such as provisioning storage, deploying containers and monitoring performance.

– Test. Testing and deploying your applications frequently is crucial in multicloud environments with numerous interdependencies. Not only will this help you detect and fix problems, but it will also ensure that your applications are compatible with various cloud and on-premises systems.

– Manage. Pick a multicloud management platform that empowers your team to consolidate activities from discovery and deployment to monitoring and day-to-day management from a single experience spanning multiple public clouds. Streamlining processes will make IT agile when responding to business needs.

Hype versus reality

The last step is critical. Yes, the multicloud management platform provides one unified experience for managing multiple resource types. But the onus is on you to select the right partners who can deliver on their promises and make life easier for your staff, rather than moving your processes to a system that doesn’t meet your IT service requirements.

And while there is talk of a supercloud, or one platform architecture to rule all clouds, solutions are emerging to help you manage your multicloud estate today.

Dell APEX Navigator for Multicloud Storage, available later this year, is a SaaS tool designed to help IT and DevOps teams provide deployment, management, monitoring and data mobility for APEX block and file storage across multiple public clouds and move data between on-premises systems and public clouds.

Dell APEX Navigator for Kubernetes, available next year, is designed to simplify management of Kubernetes persistent storage, enabling storage administrators and DevOps teams to deploy and manage Dell’s Kubernetes data storage software at scale across both on-premises and cloud systems.

Ultimately, these tools, which are integrated into the Dell APEX Console, help customers reduce time switching between different management experiences – so they can focus more on innovation or higher-level business tasks.

As an IT leader, the conductor of your own multicloud symphony, making sure all the orchestras in your IT infrastructure play in tune and at the right tempo is paramount – with consistent governance monitoring and management.
Just as a skilled conductor must master how to blend each orchestra’s unique sounds to create a unified masterpiece, managing a multicloud environment requires you to master each cloud provider’s offerings and each on-premises systems operations.

To orchestrate them effectively, you need a strategy that helps you maintain coherence and consistency as well as reliability and performance.

What steps will you take to master your multicloud symphony orchestra?

Click here to learn more about Dell APEX Console.

Brought to you by Dell Technologies.

Unlocking the potential of multicloud

SPONSORED POST: Imagine a stable, unified infrastructure platform and future proofed storage capacity that shrinks the configuration overhead, minimizes downtime and allows the IT team more time to anticipate and proactively address potential problems before they occur without being inconvenienced by having to address out of hours support issues?

Sound too good to be true? Well it’s not, according to this Modern Multicloud webinar from Dell – Achieve Greater Cost Control and IT Agility by Leveraging Dell APEX. In it you’ll hear Chris Wenzel, senior Presales manager at Dell Technologies, talk about how organizations across the world are finding ways to address their rising IT costs using unified infrastructure solutions.

Proving that point are Neil Florence, Head of Systems and Infrastructure for UK construction and engineering company NG Bailey and Neil’s colleague Stephen Firth, NG Bailey’s Infrastructure Manager. They outline how the company has worked hard to transform its own infrastructure and get its business onto a really stable platform over the last few years, having updated all of its core processes to help simplify its operations.

NG Bailey was accelerating the rate of SaaS adoption but worried that its traditional three tier approach to datacenter infrastructure wouldn’t fit the new model or help the business grow. To that end, and having taken the decision not to move everything into the public cloud due to associated cost risks, a natural evolution towards a multicloud environment was already underway.

At that point NG Bailey sought help and advice from Dell Technologies and Constor Solutions, a Dell Technologies IT solutions provider, including a strategic session at its Innovation Centre in Limerick which helped it understand how different Dell solutions could help the organization.

You can hear Neil and Stephen explain where Dell APEX Flex on Demand fits in helping to bring cost effective unified storage and data protection controls across multiple clouds, whether Microsoft Azure, AWS, Google Cloud and others – and whatever the applications and data they happen to host.

Watch the video here.

Learn more about the business value of Dell APEX here.

Sponsored by Dell.

Optionality: The key to navigating a multicloud world

Commissioned: Cloud software lies at the heart of modern computing revolution – just not in the way you might think.

When most people think of cloud computing, they think of the public cloud, which is fair play. But if you’re like most IT leaders, your infrastructure operations are far more diverse than they were 10 or even five years ago.

Sure, you run a lot of business apps in public cloud services but you also host software workloads in several other locations. Over time – and by happenstance – you’re running apps on premises, in private clouds and colos and even at the edge of your network, in satellite offices or other remote locations.

Your organization isn’t unique in this regard. Eighty-seven percent of 350 IT decision makers believe that their application environment will become further distributed across additional locations over the next two years, according to an Enterprise Strategy Group poll commissioned by Dell.

It’s a multicloud world; you’re just operating in it. But you need options to help make it work for you. Is that the public cloud, or somewhere else? Yes.

The many benefits of the public cloud

Public cloud services offer plenty of options. You know this better than most people because your IT teams have tapped into the abundant and scalable services the public cloud vendors offer.

Need to test a new mobile app? Spin up some virtual machines and storage, learn what you need to do to improve the app and refine it (test and learn).

What about that bespoke analytics tool your business stakeholders have been wanting to try? Assign it some assets and watch the magic happen. Click some buttons to add more resources as needed.
Such efficient development, fueled by composable microservices and containers that comprise cloud-native development, is a big reason why most IT leaders have taken a “cloud-first” approach to deploying applications. It’s not by accident that worldwide public cloud sales topped $545.8 billion in 2022, a 23 percent increase over 2021, according to IDC.

The public cloud’s low barrier to entry, ease-of-procurement and scalability are among the chief reasons why organizations pursuing digital transformations have re-platformed their IT operating models on such services.

The public cloud’s data taxes

You had to know a but is coming. And you aren’t wrong. Yes, the public cloud provides flexibility and agility as you innovate. And yes, the public cloud provides a lot of options vis-a-vis data, analytics, IoT and AI services.

But the public cloud isn’t always the best option for your business. Like anything else, it’s got its share of drawbacks, namely around portability. As many IT organizations have learned, getting data out of a public cloud can be challenging and costly.

In fact, many IT leaders have come to learn that operating apps in a public cloud comes with what amounts to data taxes. For one, public cloud providers use proprietary data formats, making it difficult to export data your store there to another cloud provider, let alone use it for on-premises apps.

Then there are the data egress fees, or the cost to remove data from a cloud platform, which can be exorbitant. A typical rate is $0.09 per gigabyte but the more data you want to move, the greater the financial penalty you’ll incur.

Finally, have you tried to remove large datasets from a public cloud? Okay, then you know how hard and risky it is – especially datasets stored in several locations. Transferring large datasets courts network latency that impinges application performance. Moreover, because your apps depend on your datasets the more you offload to a public cloud platform the greater the gravity of that data and thus the harder it is to move.

The sheer weight of data gravity is a major reason why so many IT leaders continue to run their software in public clouds, regardless of other available options. After a time, IT leaders feel locked-in to a particular cloud platform(s).

Rebalancing, or optimizing for a cloud experience

Such trappings are among the reasons many organizations are rethinking the public “cloud-first” approach and taking a broader view of optionality.

Many IT departments are assessing the best place to run workloads based on performance, latency, cost and data locality requirements.

In this cloud optimization or rebalancing, IT organizations are deploying apps intentionally in private clouds, traditional on-premises infrastructure and colocation facilities. In some cases, they are repatriating workloads – moving them from one environment to another.

This multicloud-by-design approach is critical for organizations seeking the optionality to move workloads where they make the most sense without sacrificing the cloud experience they’ve come to enjoy.

The case for optionality

This is one of the reasons Dell designed a ground-to-cloud strategy, which brings our storage software, including block, file and object storage to Amazon Web Services and Microsoft Azure public clouds.

Dell further enables you to manage multicloud storage and Kubernetes container deployments through a single console – critical at a time when many organizations seek application portability and control as they pursue cloud-native development.

Meanwhile, Dell’s cloud-to-ground strategy enables your organization to bring the experience of cloud platforms to datacenter, colo and edge environments while enjoying the security, performance and control of an on-premises solution. Dell APEX Cloud Platforms provide full-stack automation for cloud and Kubernetes orchestration stacks, including Microsoft Azure, Red Hat OpenShift and VMware.

These approaches enable you to deliver a consistent cloud experience while bringing management consistency and experience data mobility across your various IT environments.

Here’s where you can learn more about Dell APEX.

Brought to you by Dell Technologies.

Social media’s reaction to generative AI proves just how valuable data is

Mosaic AI
Mosaic AI

Commissioned: It’s not a revolutionary statement to say data has value – how many times have we heard it’s the new oil? But what might not be readily apparent is how much. Up until this point, the vastness of data – and lack of tools to efficiently parse it – has made it almost impossible to analyze at scale. But that’s all changing thanks to generative AI (GenAI), as well as broader AI breakthroughs on the horizon. The bigger question is, how do we measure it? How do we begin to understand it? What examples can we point to that show exactly how massive the value of data has become?

The good news is the internet, possibly the largest source of data value on earth, provides a great case study in data, its usage, and its exponentially growing value. Let’s take a look at why the internet might be the most visible and timely battleground to analyze and learn just how critical data will be in this new AI era.

The ebb and flow of the internet

We started with links and search engines. Early portals gave way to Google search dominance – a page with a single search bar that took you where you wanted to go. This soon gave way to social media platforms that brought your attention back to centralized hubs where you could scroll through curated lists of content that were brought to your feed. In an effort to create freedom from platforms and their algorithms and democratize access and ownership of digital content, we saw the rise of Web3 and the promise of the internet moving back towards decentralization. However, the Web3 movement relied on advanced technical skills and a trust in a broad community that had some high-profile challenges in recent years.

Could GenAI be the next iteration of the Web? A Web4?

GenAI has burst onto the scene and has become a sensation that every organization is now looking at. It’s changing how developers and ops folks are working. Every organization is rushing to explore how it can impact their business, touted in many areas. Let’s unpack how it changes the game for the internet.

GenAI offers ease of use. It’s why it’s captured the attention of users and organizations everywhere, and why it offers the potential to reshape the internet. Because GenAI suggests a return to a centralized web consumption experience, it’s not surprising that search engines are jumping in on this trend and trying to bolster their positions. Simply put, it’s doing something Web3 never did: making it much easier for users to get to what they want.

Here’s how it works: instead of going to a recommended destination where they must identify the answer themself, the user simply asks the GenAI application a question and the answers are brought to them in a conversational way. This is a potentially earth-shattering level of change to the internet – for the user to be able to interrogate the data and land on exactly what they were looking for is very elusive. This suggests the new Web4 era will be centered around personalization and the ability to interact with data similar to how we interact as humans. It is also creating the battle lines for what the future of the internet could look like, by bypassing the existing players and their platforms and surfacing up results in conversational responses.

GenAI training and internet scraping

When building and training a foundation model for GenAI, data quantity and quality is critical to achieving the best outcomes. As such, the internet was one of the earliest places that many of these AI models looked to. Where else could you get so much data from so many active participants? Plus, it was data that was freely available and, in many cases, the users of these platforms were actively giving their data away. But it gets murky because data ownership and whether there were intellectual property protections in play still is not readily understood nor legally tested.

Add to that many of the early players in this space were startups with a “move fast and break stuff” mentality, and AI and the internet very much is the wild west. We’ve seen early battlegrounds drawn by companies like Getty Images, who see GenAI images that pull from their archives as derivative works and are arguing these tools are more like Napster than something novel and new. But perhaps an area that is sparking the biggest seismic shifts is in social media, where typically, the platforms are granted some level of access and assigned rights via terms of service and offer a free platform as a result. The challenge here is that many had built extensions to their platforms using APIs or allowed unfettered access to users, and now they see GenAI as a massive threat to their efforts and valuation.

Social media strikes back

Recently, we’ve seen many social media companies go on the offensive, restricting access to GenAIs and leaving their users in the crossfire. It’s not hard to see why this is occurring; these AIs can collect and consume a sea of content and provide a contextual search of that content that is completely personalized.

What a value proposition for an audience. With things like GitHub Copilot, a simple GenAI prompt can surface documentation or code snippets in seconds. Stack Overflow, Reddit and Twitter have all moved to start charging for API access to their site’s content. It makes sense; if a perceived competitor is gathering and using all your data to gain a competitive advantage, why wouldn’t you seek compensation or limit access? Reddit has also recently moved to restrict third-party apps, and interestingly, that has sometimes put them at odds with their community moderators.

Then there’s Twitter. Over the 4th of July weekend, the social media platform began to temporarily limit user access to its content. We’re also beginning to see some of this battle spill over into other web properties, such as news outlets and GenAI powered search defeating paywalls. In a world where content and data become products themselves, we will continue to see this tug of war. Decisions are being made that are shaking the internet to its very foundation, and the sole reason for any of this is just how precious this data is.

What the internet can tell you about your own data

When thinking about your enterprise or business data versus what might commonly be found on the internet, consider this: your data’s probably way more valuable. We’re seeing tech behemoths fight tooth and nail to protect their data and IP, even when it’s user-generated and potentially available in many other platforms and forms, or full of low-quality data like spam and bot networks. With GenAI, the proportion of data that holds value has increased exponentially. This means that organizations may have to re-evaluate their existing notions on data because GenAI has changed the equation.

If social media companies work this hard to limit access and bolster their competitive position with data, leaders embracing AI must take similar steps. Lean into the world of ambiguity – the use cases may not yet be obvious – but the answer will lie in the data and retention policies must change as a result. Consider where your data sits and how best to get AI to it, because data gravity will still play a role. Understand how it’s used. Data and Intellectual Property leakage must be avoided; the data itself and the potential training it offers a foundation model could potentially reduce your differentiation.

Ultimately, what we are watching now in social media will play out repeatedly in other spaces; data will be the great differentiator. In this space, that means taking a hard look at your AI solutions and ensuring you are limiting areas of exposure. We’re rapidly approaching a world where every organization is data-driven and using AI. This means it will be more important than ever to protect your most valuable asset – your data – and never outsource your core competencies.

Learn how GenAI and automation lower the bar for data center management in our latest podcast, The Great Equalizer: GenAI and AI Transforming the Data Center and learn how to bring generative AI to your organization.

Brought to you by Dell Technologies.

Unsure about your generative AI plans? Start with your data

Commissioned: Call it an overnight success years in the making. Generative AI (GenAI) has exploded on the scene as one of the most critical areas organizations need to invest in. Decades of AI innovation paired with a natural language interface created the magic that has captured our collective attention. This is a wave for the ages. In oceanography there’s a term for these rare and unusually large phenomena: “rogue wave,” an outlier where the height is greater than twice the significant wave height. It’s an apt term for what we’re seeing with GenAI, and the challenge with these waves is they often materialize out of nowhere. Who had GenAI as the most important investment opportunity for CIOs in 2023 on their bingo card?

Catching the wave versus being caught under it

A boat can float on the water; it can also sink in it. Technology leaders everywhere, as captains of their respective boats, are rushing to stay ahead of this potentially crushing wave. The challenge is where to start? A decade ago, IT leaders worried about shadow IT projects springing up outside of governance; today they have a new concern: shadow AI.

At the same time, they’re also seeing opportunities. Forward looking and agile organizations are eagerly piloting GenAI solutions, often in the software development space where things like GitHub Copilot have really captured the mindshare of developers. What’s interesting is that while 70 percent of developers surveyed by Stack Overflow say they are already using or plan to use GenAI in coding, only 42 percent trust the accuracy of the output in their tools. This tells us we’re still relatively early in the GenAI cycle, but the rate of acceleration might keep some leaders up at night.

While there’s a broader conversation to be had on developing a winning GenAI strategy and who should be leading the charge, today we’ll focus on the role data will play and why it’s critical to your AI strategy. The good news? You can take steps to address it now.

Data’s value is only increasing with GenAI

As organizations become increasingly digital, every activity in their business is effectively exhausting data. Now not all data is equal; some data is inherently more valuable. For example, data that’s proprietary or hard to acquire, private or confidential data and, of course, data that’s useful to the business. This is where it gets tricky, though, because today very little of this data is meaningfully used. For instance, only 26 percent of ITDMs say all innovation efforts are based on data insights. This is because historically the volume of data has greatly surpassed our ability to analyze it.

Enter GenAI. Now imagine for a second that same corpus of data can be parsed with AI in minutes, if not seconds (depending on how much power you’re putting behind it). Suddenly, the long tail of data which was previously incomprehensible, has now been unlocked. This puts us in a quandary because existing logic on data retention and the understanding of how data can be used, or even its useful longevity, has now completely changed.

Getting your data house in order

As you start your GenAI journey, you must start first by getting your data house in order. Because you might not know what questions to ask your data today, but at some point you will – and if you haven’t retained that data, those answers will always elude you.  You can begin building the strategy and experimenting, but getting a handle on your data now is key. With that in mind, here are four areas to consider:

• Collection: digital end points or data creation areas are currently generating data? Are they connected? Is the data just living on an edge device somewhere never being utilized?

• Curation: Do you have a way to tag and classify data so its value and usage restrictions are known? Do you know who would benefit from having access to this data, and have you also appropriately applied labeling to understand privacy, governance, and intellectual property protections?

• Storage: Is there a process for centralizing this data and ensuring access? Have you looked into connecting different IT environments, potentially clouds? Have you established an effective data tiering strategy to align with the data’s value and lifecycle?

• Protection: Can you protect against a data loss or ransomware event? Can you ensure you’re meeting your governance and data sovereignty requirements? Do you understand the risks with various data and have you built an approach to mitigate them and secure access?

Walk before you run

I know it’s an incredibly exciting time and the way GenAI has captured our mindset has been very consumer oriented. There’s a lot of work for IT to do to catch up. In many ways it’s like in the “bring your own device” wave of innovation; it was very easy for consumers to move quickly. There’s no existing legacy infrastructure or data to consider; simply buy a smartphone and you’re up and running. But as we saw in that time period, the backend requirements to enable work from anywhere in a secure manner took years to catch up. Avoid the temptation to barrel right into something without first setting the stage for your success. While you build the strategy for the next decade, be sure to triage your existing environment as well.

Learn more about our Dell APEX portfolio of as-a-service offerings and how Project Helix helps organizations leverage generative AI to drive emerging use cases and deliver value.

Brought to you by Dell Technologies.

Generative AI has democratized AI – what does this mean for COEs?

Commissioned: To centralize or decentralize? That was once the salient question for many enterprises formulating a strategy for deploying artificial intelligence. Whether it was nobler to support a singular AI department or suffer the slings and arrows that accompany distributed AI projects.

Tectonic shifts in technology can render such debates moot. Is that happening now, as generative AI catalyzes creativity in businesses, enabling employees to create texts, images and even software code on the fly?

Some of these experiments are useful; others not so much. There will be more failures along the path to innovation, which is littered with the bones of fallen tech projects.

What is crystal clear: Generative AI has democratized AI consumption for enterprises in ways that previous AI applications could not. The genie, in its many forms and functions, has shot out of the bottle.

The way it was

To streamline and curate our AI competency or allow projects to roam unchecked and hope for the best? It’s a fair question, with mixed approaches.

Over the years, some organizations consolidated AI capabilities in one department, often established as an AI center of excellence (COE). The COE was often composed of database engineers, data scientists and other specialists trained in querying machine learning (ML) models.

The inverse of COEs was highly decentralized. In classic, do-it-yourself fashion, business leaders experimented with some tools in the market on AI projects that might eventually foster innovation. Naturally, these projects tended to be more rudimentary than those created by COE members.

Both approaches had their pros and cons.

Centralizing AI functions afforded organizations the ability to dictate strategy and policy and control costs, thereby reducing risks. But COEs’ dedication to rigorous processes had its drawbacks. Typically, the COE received specifications and built a deliverable over several months. Over a long enough timeline, the goal posts moved. As data grew stale, the output rarely resembled the desired outcome.

Conversely, distributed AI functions granted business experts the freedom to quickly experiment and explore so that data remains fresh and current. Projects may have lead to some insights that were harder to cultivate in an AI COE, which lacked the domain expertise of a business line.

However, ad-hoc efforts often resulted in projects with no demonstrable ROI for the business. And lacking the kind of guardrails present in a COE, these efforts were often risky to the business.

How organizations approached AI varied from business to business, based on leaderships’ philosophies and appetite for risk, which are informed by internal capabilities and competencies.

Generative AI changed the paradigm

The arrival of generative AI clarifies the question of whether to centralize or distribute AI functions.

Today, average Joes and Janes interface directly with AI technologies using natural human language rather than special tools that query AI models.

Knowledge workers create cogent texts using Google Bard and ChatGPT. Graphic designers craft new image content with DALL·E and Midjourney. Software developers write basic apps with Copilot or Codeium.

Increasingly, employees layer these capabilities, creating mashups of text, graphic and code creation technologies to generate marketing content, analytics reports, or other dashboards – without the help of data experts who might spend months putting something more sophisticated together.

To be clear, generative AI cannot replace the expertise offered by AI COE specialists. It can’t teach somebody the intricacies of TensorFlow, the magic of Kafka, or other sophisticated tools used to query AI and ML models – yet.

Generative AI has democratized content creation as much as smartphones have facilitated access to information to anyone on the go – anywhere in the world.

Thinking through the implications

IT departments often hold the keys to many technologies, but generative AI is a different animal, requiring IT leaders to consider the impact of its use within the department and across the broader business.

As with technologies that are new to your business, you’ll huddle with C-suite peers on rules and guardrails to make sure the business and its employees are covered from a compliance, risk and security standpoint. And you’ll guard against potential lawsuits alleging content created by generative AI tools infringes on intellectual property rights and protections.

Yet this may be easier said than done for many organizations.

Fewer than half of U.S. executives surveyed by KPMG said they have the right technology, talent and governance to implement generative AI. Moreover, executives plan to spend the next 6 to 12 months increasing their understanding of how generative AI works and investing in tools. This is critical for the C-suite and board of directors, according to Atif Zaim, National Managing Principal, Advisory, KPMG.

“They have a responsibility to understand how generative AI and other emerging technologies will change their business and their workforce and to ensure they have sustainable and responsible innovation strategies that will provide a competitive advantage and maintain trust in their organization,”

To be sure, the democratization of generative AI means your rivals have ready access to these tools, too. Take care not to lose the name of action.

How will your organization use these emerging technologies to future proof your business and gain competitive advantages?

Learn more about our Dell APEX portfolio of as-a-service offerings and how Project Helix accelerates business transformation and unleashes productivity with trusted AI.

Brought to you by Dell Technologies.

A last line of defense against ransomware

'recovery' key on keyboard
recovery key on keyboard

Sponsored Feature: The impact of cyberattacks around the world continues to escalate at an alarming rate, even after reaching “an all-time high” last year, new research warns. The latest IBM Cost of a Data Breach Report 2022 estimates that the average cost of a data breach hit $4.35 million in 2022, with some 83 percent of organizations found to have had more than one breach during the year. Business downtime can have devastating impacts on reputation and financial performance, and sixty percent of the surveyed organizations stated that they had been forced to hike the price of their services or products because of a data breach.

Ransomware cyber-crime remains “by far” the most common tactic employed by cyber criminals, with this type of attack launched approximately every 11 seconds according to calculations from Cybersecurity Ventures. These attacks can involve encrypting an organization’s backup data then demanding a fee for its decryption. Payment of these ransoms, David Bennett, Object First’s CEO, told El Reg, has been widespread among those desperate to get compromised businesses back up and running when they are unable, or don’t have sufficient time, to hunt down and restore previous versions of their data held on premises or in the cloud.

And the impact of the global ransomware problem is huge. IBM’s study estimates that, for critical infrastructure organizations (which the report defines as financial services, industrial, technology, energy, transportation, communication, healthcare, education and public sector industries), the cost of an attack is an eye-watering $4.82 million, considerably above the average. Some 28 percent of these were found to have been targeted with a destructive or ransomware attack, while 17 percent experienced a breach because of a business partner being compromised.

Bennett doesn’t mince his words when describing the potentially devastating impact of such ransomware attacks: “What can go wrong for your company if you are hit by a ransomware attack? The answer is everything. It doesn’t matter whether you’re a small company or a big company.”

“You need to consider all aspects of your business operations,” he continues. “How do you collect money from your customers, for example? How do you pay your employees? What happens if your ability to pay your employees or pay your suppliers is impacted? If you can’t collect cash, or you can’t pay cash, your business is out of business. And of course, there are all the other essential functions outside of just the financial operating ability of the company to consider.”

Ransomware scaling up and out

The scale of the danger is evidenced by the wide variety of global companies and organizations that have fallen victim to ransomware attacks over recent years. The WannaCry strain is reported to have affected telco Telefónica and other large Spanish companies for example, as well as the British National Health Service (NHS), FedEx, Deutsche Bahn, Honda and Renault, as well as the Russian Interior Ministry and the Russian telecom company MegaFon.

However, Object First’s Bennett notes that reported attacks represent just the tip of the ransomware iceberg given that so many organizations are typically very reluctant to admit being hit: “Unless they have a legal requirement, organizations are concerned about disclosing their vulnerability to cyber attacks,” he says. “The risk of reputational damage that can come from a data loss incident is enormous.”

Even the most robust traditional IT security measures – such as intrusion prevention, network protection, VPNs, DNS, and endpoint protection – cannot guarantee that mission crucial systems will remain beyond the reach of determined cyber criminals.

“It doesn’t matter what security systems you have put in place,” explains Bennett. “In the end, cyber attacks like ransomware are inevitable. The only way to deal with the situation is to be fully and properly prepared and know how to recover. The absolute minimum that any organization needs to have is an effective disaster recovery strategy.”

Three to one rule aids protection

Adhering to best practice may also mean abiding by the “three-to-one rule” which stipulates there are always three copies of data: one in production, one stored on different types of media, and one type which is immutable. This is exactly where, according to Bennett, Object First’s solutions can help by protecting object storage assets and making it impossible for ransomware-touting cybercriminals to encrypt an organization’s backup data.

Object First’s Ootbi platform has an unusual name, but is an abbreviation of “Out of the Box Immutability”, Bennett explains. It’s designed to be a ransomware-proof and immutable out-of-the-box solution which can deliver secure, simple and powerful backup storage for mid-enterprise organizations.

“We actually provide an immutable copy of our customers’ backups,” says Bennett. “This is important because historically, immutable copies were media like tape, or optical devices. And a lot of people offloaded their data to public cloud. You can have an immutable copy in the public cloud, be that AWS or Azure. But what happens if you must recover data that supports business critical systems from the cloud? It doesn’t happen very fast, and can you afford all that down time? If you have to rebuild all your systems it can take weeks or even months, and you must pay hefty egress fees.”

What Ootbi enables is a local immutable copy on-prem with a data separation from the storage control layer, which is the backup and recovery software, on Object First hardware. By separating the two and hosting them on-prem, customers can quickly recover, without high cloud egress fees.

Avoid paying the ransom

Bennett points out that, at a corporate level, Object First enjoys a long and close relationship with Veeam, a leading provider of modern data resiliency software and systems designed to deliver secure backup and fast, reliable recovery solutions. Ootbi is specifically designed for Veeam and Veeam only, which creates the experience of a near integrated storage appliance that combines all the necessary software, hardware and data management in one package. So, if a company already knows Veeam, they can implement an immutable storage solution without any previous experience.

Ootbi’s three-year subscription model with 24/7 support included means no surprises with fees in the long term. By delivering cheaper-than-cloud costs in a far more secure on-prem package, Object First argues that its customers can avoid paying the ransom and save money while doing it. Capacity and performance of the Ootbi locked-down Linux-based appliances scale linearly, supporting backup speeds up to 4.0 Gigabytes per second with up to half a petabyte of storage space.

This range of features made Object First’s Ootbi the perfect solution for Mirazon, a North America-based IT consulting specialist that helps customers with everything from basic helpdesk to complete infrastructure redesigns through managed services, consulting, and product sales.

“Mirazon required a solution that would not allow backup copies to be deleted or encrypted should they, or their customers, fall victim to a ransomware attack,” stated Brent Earls, chief technology officer, Mirazon.

The company also recognized the need to be able to properly secure not only its primary data, but also its backup data. As a result, it sought an effective ransomware-proof solution that would be simple to deploy and manage. Earls explains that cloud-based backups, while flexible, can be limited by bandwidth constraints: “Scale-out backup repositories are the only way of getting back-ups into the cloud, storage has either had to be re-architected to follow for smaller repositories to sync only critical data to the cloud, or the entire backup repository had to go all at once, which again, causes bandwidth issues,” he said.

“Implementing an on-premises device would solve the limited bandwidth issue while also eliminating the unpredictable and variable costs of the cloud.”

Choosing a solution that was both incredibly robust, but also simple to use and deploy was very important for Mirazon: “There was no existing product on the market that would solve the immutable issue that was also simple to deploy and operate which also fell within budget — until now. Object First has given us confidence that the solution will deliver everything it said it would.”

This focus on ease of deployment and use was also an overarching priority for Object First’s solutions, according to Bennett: “I‘ve been in the data storage industry for 20 years and the industry hasn’t really moved forward with the times. Historically, products are hard to use, hard to manage, hard to set up and you will need a storage management team to deal with managing backups. How many companies have a separate data storage team these days?”

“From opening the box to racking and stacking and getting it set up takes only 15 minutes. But in practice, most users have been able to do it in well under 15 minutes,” Bennett explains. “The hardest part is lifting the unit out of the box because it’s 180 pounds. But actually, implementing and getting up and running is super simple.”

Sponsored by Object First.

Generative AI and automation: accelerating the datacenter of the future

Commissioned: In the age of automation and generative AI (GenAI), it’s time to re-think what “datacenter” really means. For those who have become heavily invested in public cloud, the datacenter might not be the first place you think of when it comes to automation and GenAI, but these technologies are rapidly changing what is possible in all environments.

Ten or fifteen years ago, when businesses started bypassing IT by swiping credit cards and setting developers loose on cloud resources, the public cloud was absolutely the right move. In most large organizations, internal customers were often ignored, or their needs were not being fully met. They wanted flexibility, they craved scalability and they needed a low up-front cost to allow incubation projects to flourish.

If time stood still, perhaps the dire prognosticators of the datacenter’s end would have been right. I myself was quite the cloud evangelist before learning more about the other side of the fence. So why hasn’t this extinction-level event come to pass? Because the datacenter has adapted. Sure, there are “aaS” and subscription models now available on-premises; but the real stabilizing force has been automation.

Which brings us to the story of the day: GenAI and how it can augment automation in the datacenter to be an experience nearly on par with the public cloud. Before we get there though we need to look at the role automation and scripting have played in the datacenter. We’ll start by explaining some essentials, then we’ll unpack why automation and GenAI have changed what is possible on-premises.

Cloud operating model and infrastructure as code

Let’s start with the basics: the foundation of cloud was infrastructure as code and the idea of consuming IT as a Service. Your developers never had to talk to a storage admin, IT ops person, or the networking team to rapidly spin up an environment and get to work. This should be table stakes in 2023, and the good news it’s entirely possible to build it for yourself. Adopting this operational model means IT is leveraging policies and processes alongside automation to remove friction from the environment.

Visual representation of the end experience when you’ve automated a cloud operating model

Automation toolsets and telemetry data

Today there are many automation, management and telemetry/AIOps products available that provide unparalleled control and insights into datacenters. Data is the foundation AI and of managing a datacenter effectively. The control and visibility now in datacenters is often a superset of what can be achieved in the public cloud – although the hyperscalers have done a great job in that department as well. Given the cloud’s multitenant nature, cloud providers must obscure some of the operational knowledge to keep every customer secure. This results in architectural decisions that limit how some monitoring systems can be deployed and what data can be collected. One important are of focus is ensuring that you’re heavily integrating these solutions, embracing automation and infrastructure as code, measuring/monitoring everything and using a cohesive workflow for all your roles.

Visual representation of a common automation/management stack

The next wave of IT automation with GenAI

This brings us to the next evolution of the datacenter incorporating GenAI. Let me share a fun story about a past role where the client made the marketing consultant build an HCI deployment hands-on lab for physical and virtual infrastructure, and then didn’t provide any subject matter experts to help. If it’s not clear, that marketing consultant was me, and it was probably one of the most challenging projects I’ve ever worked on. I used code snippets and YouTube tutorials to get to the foundation of how to do such a task. I spent weeks assembling the puzzle, figuring out how each puzzle piece fit together. By some miracle I actually managed to get it right, even though I didn’t know much about coding. Anyway, here’s wonderwall… I mean here’s GenAI doing that.

GenAI is the Search Engine and code assembly machine we were looking for

Now mind you in my hands-on-lab, I was doing a lot more than just installing Windows Server, but there is no doubt in my mind if I asked it to provide the rest of that process, it could. What’s so important is that with the infrastructure-as-code mentality, and in new environments where developers may not be familiar with these types of calls or runbooks, GenAI is a new ally that can really help. Many people don’t realize access to common infrastructure scripts is prevalent – and oftentimes it’s written by the tech companies themselves. Both hardware and software vendors have large runbook repositories, sometimes it’s just a matter of finding them: enter GenAI. Another important consideration is that the infrastructure itself is intelligent and secure. These commands can be pushed out to thousands of servers for remote management purposes. This greatly lowers the bar on managing your environment.

GenAI and process building

One of my favorite customer engagement stories might sound a little long in the tooth – somewhat like those stories of being lost or unable to reach someone that are unfathomable to those who grew up with smartphones. We hear a ton of talk about containers, but when I broached this topic with one customer, he said, “I can’t even keep my VMware admins 18 months, what makes you think I could ever do containers?” This is something I’ve thought a lot about and it’s probably the biggest challenge with technology: if I don’t have the skillset, how could I possibly onboard it? Enter GenAI’s next incredible friction reducer: writing or finding documentation.

In just two prompts we have a routine and highly valuable process documented and ready to use

We’ve long had access to an incredible amount of information, however previously there’s been no ability to parse it all. This all changes with GenAI. Now, instead of navigating search and sifting through code repositories, a simple natural language query or prompt yields exactly the documentation needed. Instead of hours of looking for answers, extensive documentation is at your fingertips in minutes. This completely destroys any barriers to embracing technology. Imposter syndrome, skill gaps, and switching costs: you’re on notice.

Thousands of possibilities but AI Ops is next

I want to acknowledge the wealth of ways this technology can help us run a datacenter. Probably the next one to add significant value is AI Ops. That rich telemetry data can tell us a lot but also tends to have a signal-to-noise ratio problem. We’re simply generating too much data for human beings to analyze and comprehend it all. By pushing this data into GenAI and using natural language as an interface, we will extend insights to a broader audience and make it possible to ask questions we may never have thought of when looking at charts and raw data. The mean time to resolution will plummet when we use this kind of data. But there is one massive drawback, which brings us to our final point.

GenAI and automation change what’s possible, but we must use it carefully

Two of the major challenges with GenAI must be addressed. They are: Intellectual Property (IP) leakage and its ability to “hallucinate” or make things up. Let’s unpack each and determine how to embrace the technology without stumbling during implementation.

First, let’s discuss IP leakage. In any scenario where data is being sent to GenAI models that are delivered as a service, we risk leaking IP. Much like the early days of public cloud and open S3 buckets, early experimenters in their misuse or misunderstanding, created risk for their companies. The best way to counter this is to have a centralized IT strategy, insert them into your common workflows or development pipeline, and lastly prioritize building your own GenAI on-premises for highly sensitive data that cannot go to a AIaaS which is constantly learning off your data.

The other benefit of bringing a large language model (LLM) in house is you can also make it more precise and put guardrails on it. This makes the responses it generates more precise and in context of your own business. The guardrails can also stop some of the “hallucinating” i.e. when the GenAI is compelled to answer but provides inaccurate and/or made-up information to comply with the request. This is a common problem with GenAI. The reality is these tools are all still in their infancy. Just as most would work testing into their release pipeline, this too is an area where more rigor should be placed prior to pushing to production. I’m a big proponent of human in the loop, or human assisted machine learning, as a way to reduce mistakes with AI.

The future is automated

The datacenter is here to stay, but it can be radically transformed with GenAI and automation. These tools can augment our workflows and help IT Ops and developers achieve superhuman capabilities, but they are not a direct replacement for people. As you roll out your AI and automation strategies it’s important to think about what you’re trying to accomplish and what level of automation your organization is comfortable with. The future is bright and the ability to innovate anywhere is now a reality.

Learn how our Dell APEX portfolio helps organizations embrace a consistent cloud experience everywhere so they can adopt technologies like AI and accelerate innovation.

Brought to you by Dell Technologies.

Lenovo retools storage portfolio for AI

Lenovo has released 21 storage products, including a liquid-cooled HCI system, in a major storage portfolio refresh aimed at customers building AI-focused infrastructure.

Scott Tease

Scott Tease, Lenovo’s Infrastructure Solutions Product Group VP and GM, talked of helping “businesses harness AI’s transformative power,” stating: “The new [systems] help customers achieve faster time to value no matter where they are on their IT modernization journey with turnkey AI solutions that mitigate risk and simplify deployment.” 

Marco Pozzoni, Director, EMEA Storage Sales at Lenovo’s Infrastructure Solutions Group, stated: “Our largest-ever data storage portfolio refresh delivers the performance, efficiency, and data resiliency required by Data Modernization and Enterprise AI workloads.”

Lenovo says  “businesses are moving to a disaggregated infrastructure model, turning compute, storage, and networking into on-demand, shared resources that can adapt to changing needs to boost scalability, efficiency, and flexibility.” New AI Starter Kits for Lenovo Hybrid AI Platform are pre-configured and validated infrastructure systems comprising compute, storage, GPUs and networking. The storage elements are new ThinkSystem Series Storage Arrays, OEMed from NetApp, delivering unified file, block, and object storage using “high-performance SSD flash technology.”

The ThinkSystem DG and DM arrays feature include new AI-powered autonomous ransomware protection – ONTAP ARP/AI – plus encryption for greater data protection and synchronous replication with transparent failover.

Marco Pozzoni.

There are five areas of Lenovo storage product portfolio change:

  • ThinkAgile SDI V4 Series provides full-stack, turnkey solutions that simplify IT infrastructure and accelerate computing for data-driven reasoning.
  • New ThinkSystem Storage Arrays deliver up to 3x faster performance while reducing power consumption – comparing the DE4800F to the DE4000 – and providing up to 97 percent energy savings and 99 percent density improvement for the DG7200 over a Lenovo device featuring 10K HDDs for smaller data center footprints and lower TCO and when upgrading legacy infrastructure.
  • New converged ThinkAgile and ThinkSystem hybrid cloud and virtualization systems deliver flexible and efficient independent scaling of compute and storage capabilities, reducing software licensing costs for additional storage, but not core-based licenses, up to 40 percent. A ThinkAgile Converged Solution for VMware brings together the enterprise-class features of ThinkAgile VX Series – lifecycle management and operational simplicity – , with the ThinkSystem DG Series storage arrays data-management to provide a unified [private] cloud platform supporting hybrid storage workloads.  
  • New ThinkAgile HX Series GPT-in-a-Box solutions featuring Lenovo Neptune Liquid Cooling leverage the industry’s first liquid cooled HCI appliance to deliver turnkey AI inferencing, from edge to cloud, yielding up to 25 percent energy savings for the HX V4 over the HX V3 generation.
  • New AI Starter Kits for Lenovo Hybrid AI Platform deliver a validated, flexible, and easy on-ramp for enterprise inferencing and retrieval-augmented (RAG) workflows.

Lenovo partnered Nutanix in 2023 in order to sell GPT-in-a-Box-based ThinkAgile HX Series systems. There are no details available about the liquid-cooled system, with Lenovo saying it is afull-stack generative AI solution [and]  jumpstarts AI integration using a repeatable solution that lowers energy costs for faster data-powered AI reasoning.” It has “up to 2.6x higher transaction rate and up to 1.4x faster input/output operations per second (IOPs) compared to traditional cooling methodS,” specifically comparing ThinkAgile SDI V4 Series to 3 racks of Dell VxRail Cascade Lake models.  It also has a more compact form factor with increases of up to 1.4x in Container/VM density over its ThinkAgile SDI V3 Series predecessor

As the disk drive manufacturers stopped selling 10,000rpm disk drives in the 2018-2019 era and capacities were limited at up to 1.8TB or so, the comparison with the modern DG7200 using high-capacity 3.5-inch drives is not that impressive

Lenovo says that its “ThinkAgile SDI V4 Series and new ThinkSystem Storage Arrays deliver full-stack, turnkey AI-ready infrastructure for enterprises beginning AI implementations while delivering faster inferencing performance for LLM workloads.” The primary AI use cases include AI inferencing, RAG, and fine-tuning AI models, while “the offerings are purpose-built for a broad range of enterprise AI and data modernization workloads, from databases, ERP and CRM, to virtualization, hybrid cloud, backup and archive.” AI model training is not a primary AI use case.

Lenovo is in the process of buying Infinidat and that will give it an entry into the high-performance enterprise block storage array market. It will also provide it with its own storage IP and a different, enterprise-focussed, go-to-market sales channel.

You can dig around in Lenovo’s storage web pages here to find out more about today’s storage news.

WEKA reshuffles C-suite amid AI push

WEKA has appointed new chief revenue, product, and strategy officers.

The company supplies high-performance, parallel access file system software for enterprise and HPC, with AI and agentic AI workloads now a major focus. It lost senior execs in December 2024, when president Jonathan Martin, CFO Intekhab Nazeer, and CRO Jeffrey Gianetti left. Hoang Vuong became the new CFO that same month. WEKA then laid off some employees – sources estimated around 50 – as it said it was restructuring for the AI era, having 75 open positions at the time.

Now WEKA has hired Brian Froehling as chief revenue officer, Ajay Singh as chief product officer, and reassigned Nilesh Patel, prior CPO, to chief strategy officer and GM of Alliances and Corporate Development.

Brian Froehling, WEKA
Brian Froehling

CEO and co-founder Liran Zvibel stated: “WEKA is operating at the forefront of the AI revolution … Ajay, Brian, and Nilesh each bring deep domain expertise and seasoned leadership experience that will be instrumental in accelerating our global sales execution, scaling go-to-market strategies with our strategic partners, and advancing product development to provide the foundational infrastructure needed to power the agentic AI era into WEKA’s next growth phase – and beyond.” 

Froehling was most recently CRO at cloud-native logging and security analytics company Devo Technology, responsible for sales, customer success, training, and operations. Before that, he served as EVP of global sales at video cloud services supplier Brightcove, driving $220 million in revenue. He also held senior sales leadership roles at CA Technologies and Pivotal Software, where he helped scale go-to-market teams and played a role in Pivotal’s IPO.

Froehling said: “This is a once-in-a-generation opportunity to empower AI-driven organizations with market-leading solutions.”

Ajay Singh, WEKA
Ajay Singh

Singh comes from being Snorkel AI’s CPO. Snorkel’s product tech focuses on accelerating the development of AI and ML apps by automating and streamlining data labeling and curation to build AI training datasets faster. He was also CEO and co-founder of Zebrium, which developed software using ML and generative AI to analyze software logs for the root cause of outages. Zebrium was acquired by ScienceLogic in April 2024. Before that, he was Head of Product at Nimble Storage, leading from concept through its IPO and eventual acquisition by HPE, and a product management director at NetApp.

Singh said: “WEKA has been trailblazing in AI data infrastructure innovation with the ultra-powerful WEKA Data Platform, which dramatically speeds model training and inference by orders of magnitude.”

Nilesh Patel, WEKA
Nilesh Patel

Patel’s lateral move indicates that alliances and partnerships are going to play an important role at WEKA. He said: “We’re at a pivotal moment in the evolution of AI, and WEKA is powering many of the world’s largest AI deployments to supercharge innovation. This is an incredible opportunity to shape WEKA’s path forward and work closely with key partners that will help us scale AI solutions and fuel sustained hypergrowth.”

Possible partnership and alliance opportunities exist in the AI large language model, retrieval-augmented generation, vector database, and AI data pipeline areas.

Several other storage suppliers are intensely focused on storing and supplying data for AI training and inference workloads, including DDN, Pure Storage, VAST Data, and NetApp, which is developing an AI-focused ONTAP variant. Dell, which has a strong AI focus for its servers, has said it is parallelizing PowerScale. WEKA has to show that its customers can develop and run AI applications faster and better by using WEKA’s software over alternatives.

Another recent WEKA exec-level change was Sean Hiss, VP of GTM Operations and chief of staff to the CEO, who resigned in February, having joined WEKA in October 2021 from Hitachi Vantara.

Scality and Veeam bundle backup and object storage in single-server appliance

Scality has combined its ARTESCA object storage with Veeam’s Backup & Replication in a single software appliance running on commodity x86 servers.

Having a single server with co-located, integrated Scality and Veeam software eliminates the need for separate (physical or virtual) infrastructure for Veeam, reducing deployment complexity, time, and cost by up to 30 percent, the vendor claims. The server hardware can come from suppliers such as Dell, HPE, Lenovo, and Supermicro.

Andreas Neufert, Veeam
Andreas Neufert

Veeam’s Andreas Neufert, VP of Product Management, Alliances, stated: “This innovative solution simplifies the deployment of our industry-leading data resilience software alongside Scality’s robust object storage, making it easier for organizations to enhance their cyber resilience. Scality has integrated Veeam into this new solution, combining our strengths to empower our joint customers to create secure defenses against cyber threats while optimizing their backup operations.”

The competition to provide object storage-based Veeam backup targets is intensifying. Cloudian, Scality, and MinIO object storage can be targets for Veeam backups, as can the Ootbi hardware/software appliance from Object First.

Scality reckons its single box Veeam-ARTESCA software appliance offers better cyber resilience than an Ootbi-type alternative because ARTESCA + Veeam bring backup and storage together in a single, secure, unified environment that reduces malware exposure.

The ARTESCA OS is a dedicated, security-optimized Linux distribution developed and maintained by Scality with only the minimum required packages, and no root or superuser privileges. Veeam runs inside this hardened environment, reducing the total attack surface compared to traditional Veeam-on-ESXi deployments.

Artesca plus Veeam

There is internal-only communication between Veeam and the ARTESCA storage components, ensuring there are no exposed S3 endpoints that can be accessed externally, no need for external DNS resolution, and no IAM access keys or secret keys shared outside the appliance. Another point is that running backup and storage on a single server reduces hardware requirements and integration effort, while providing unified, single-pane-of-glass monitoring and management, minimizing the need to switch between separate interfaces.

Erwan Girard, Scality
Erwan Girard

Erwan Girard, Scality chief product officer, said: “The unified software appliance marks a major milestone in our partnership with Veeam. By combining ARTESCA’s security and simplicity with Veeam’s industry-leading data resilience solutions we’re enabling organizations to build unbreakable defenses against cyber threats while optimizing backup operations without compromising performance.”

The Scality ARTESCA + Veeam unified software appliance will initially be available as a single node, configurable from 20 VMs/TBs to hundreds of VMs/TBs. It will need to go multi-node to provide more backup storage capability.

Note that object storage does not natively deduplicate stored data, with backup data deduplication being carried out by the backup software. Non-object-based backup target systems from Dell (PowerProtect), ExaGrid, HPE (StoreEver), and Quantum (DXi) provide their own deduplication capabilities.

Customers will be able to purchase the combined Scality-Veeam offering from their channel partners. Scality will provide the partners with documentation and tooling to install the software appliance on one of a number of pre-validated hardware configurations.

If you’re interested in accessing the new ARTESCA + Veeam unified software appliance, you can submit a request here. The appliance can be viewed at the VeeamON 2025 conference in San Diego, which runs today and tomorrow.

Are AFAs right for VDI?

Generic shot inside a datacenter of storage flash arrays

Partner Content In any VDI deployment, storage is crucial for success and user experience. It ensures performance and responsiveness while supporting scalability and reliability. Storage bottlenecks can undermine the best VDI strategies, especially during peak times like boot storms and login surges. Consequently, storage choices are central to VDI design discussions, with organizations typically selecting between a traditional virtual SAN (vSAN) or an external all-flash array (AFA).

While a vSAN may appear to be the more cost-effective option — especially in VDI environments where per-desktop economics are critical — many organizations are forced to abandon it. Traditional vSAN implementations often fall short in real-world VDI scenarios, struggling with high I/O demands during boot storms, and they lack the flexibility required to deliver their promised price advantage. vSAN’s layered architecture introduces latency, reduces responsiveness, and lacks the efficiency for consistent user experiences. As a result, IT teams often resort to dedicated all-flash arrays to overcome these limitations despite the added cost and infrastructure complexity. Though this approach addresses the performance gap, it creates new challenges in management, scalability, and long-term cost control.

The cost problem with external all-flash arrays

Traditional external all-flash arrays, though powerful, are typically costly and often become a significant portion of the overall VDI budget. Beyond the initial purchase, this expense also encompasses ongoing support contracts, expansion shelves, and frequent hardware upgrades to maintain performance.

For VDI deployments, where storage performance can make or break the user experience, IT often finds itself buying expensive storage equipment designed to handle infrequent peak loads. Consequently, organizations overspend on storage resources.

Complexity and separate infrastructure management

External all-flash arrays demand their own management expertise, separate from the rest of the VDI environment. Storage admins must manage the arrays, monitor performance, troubleshoot bottlenecks, and tune storage configurations independently from the virtualization and application layers. This creates operational silos and complicates VDI management.

Every change to the external array—whether an upgrade, firmware update, or configuration adjustment—requires coordination across storage, virtualization, and networking teams. The result is longer downtime, an increased risk of misconfigurations, and more frequent user disruptions.

Lack of VM and application awareness

A critical shortcoming of external arrays in VDI environments is their lack of direct integration and visibility into the virtual machines they support. Without VM-level awareness, these storage systems cannot automatically optimize storage resources based on real-time application and desktop needs. Instead, IT administrators must manually tune performance, which is complex, time-consuming, and reactive rather than proactive.

For example, external arrays do not automatically understand how to distribute resources best to prevent bottlenecks during boot storms or login events. Administrators must continuously monitor and adjust storage policies manually, resulting in inefficiencies and a higher potential for performance degradation.

Vendor lock-in risks

Investing heavily in external all-flash storage arrays often leads to vendor lock-in. Each vendor uses proprietary technologies and unique management interfaces. Consequently, IT teams become heavily dependent on a single vendor, limiting flexibility and reducing negotiating power for future expansions or upgrades.

As storage requirements grow or evolve, organizations face substantial hardware refresh costs or forced upgrades to maintain compatibility and performance. This situation limits an organization’s ability to adapt quickly to new requirements or take advantage of more cost-effective storage technologies as they become available.

Next-generation vSAN: A better alternative

A next-generation vSAN eliminates many drawbacks of external all-flash arrays. A modern vSAN architecture:


• Delivers High Performance: A next-generation vSAN integrates storage services directly into the hypervisor, eliminating unnecessary software layers and reducing I/O latency. This streamlined architecture increases efficiency, eliminates boot storm concerns, and delivers performance that rivals—and sometimes exceeds—external all-flash arrays.

• Reduces Costs: By using internal storage and supporting a wide range of industry-standard flash drives across multiple vendors, a next-generation vSAN enables organizations to avoid expensive, proprietary hardware. This flexibility allows IT teams to deploy high-performance storage at a fraction of the cost, often reducing capital and operational expenses by up to 10x.

• Simplifies Management: Integrated storage eliminates the need for separate infrastructure and specialized storage skills. Administrators manage the system from within the virtualization console, streamlining operations and allowing them to seamlessly manage VDI storage alongside their other virtualization tasks.

• Provides VM-level Awareness: Because a next-generation vSAN operates at the hypervisor layer, it inherently understands VM workloads. This awareness enables automatic resource allocation, proactive performance optimization, and rapid response to changing demands, improving the user experience during boot storms or peak usage periods.

• Avoids Vendor Lock-in: Using commodity hardware and integrated storage software, organizations gain flexibility in their hardware purchasing decisions. They can select best-of-breed components based on cost, performance, and support without being locked into a single vendor's proprietary ecosystem.

• Enhances Scalability: Expanding storage is as simple as adding additional standard servers with internal drives, allowing organizations to scale incrementally and cost-effectively without the significant capital investments required by external arrays.

Conclusion

While external all-flash arrays can deliver performance for VDI, they come with high costs, increased complexity, vendor lock-in risks, and a lack of VM-level optimization. A next-generation vSAN solution integrated into the virtualization layer addresses these issues head-on, providing a simplified, cost-effective, scalable, and VM-aware storage solution ideal for modern VDI deployments.

VergeOS is one example of a next-generation vSAN approach. It integrates storage directly into the hypervisor, delivers VM-level performance insights, and scales using standard hardware without needing separate storage infrastructure. Organizations looking to maximize their VDI investment should strongly consider a next-generation vSAN strategy like VergeOS to reduce cost and complexity while improving performance and resiliency.

If you are looking for an alternative to VMware or Citrix, watch VergeIO’s on-demand webinar: Exploring VDI Alternatives.

Partner content provided by VergeIO.

Hitachi Vantara adds AI ransomware detection from Index Engines

Hitachi Vantara is joining Dell, IBM, and Infinidat in using Index Engines’ AI-powered ransomware detection software.

The CyberSense product from Index Engines uses AI and machine learning analysis to compare unstructured data content as it changes over time to detect suspicious behavior and cyber corruption. Index Engines claims a 99.99 percentccuracy rating and flags anomalies and indicators of corruption. The full-content analytics approach is said to be better able to detect malware-caused corruption than just checking metadata. Hitachi Vantara is integrating CyberSense with its Virtual Storage Platform One (VSP One) products.

Octavian Tanese, Hitachi Vantara
Octavian Tanase

Hitachi Vantara chief product officer Octavian Tanase stated: “IT complexity, cyber threats, and sustainability challenges continue to put enterprises under extreme pressure. With VSP One’s latest enhancements, we are eliminating those roadblocks by delivering a unified, automation-friendly platform with guaranteed performance, resilience, and efficiency built in. This is more than just data storage – it’s a smarter, more sustainable way to manage enterprise data at scale.”

The company said it is adding three guarantees:

  • Performance Guarantee: Applications run at predictable, high-performance levels with minimal intervention by meeting workload demands with confidence through guaranteed minimum performance levels across all VSP One Block platforms, backed by EverFlex from Hitachi. Service credits apply if performance targets aren’t met.
  • Cyber Resilience Guarantee: Organizations can mitigate downtime and data loss after cyberattacks with protections enabled by immutable snapshots and AI-driven Ransomware Detection powered by CyberSense. If data can’t be restored, Hitachi Vantara provides expert incident response and up to 100 percent credit of the impacted storage volume.
  • Sustainability Guarantee: Helps businesses track and optimize energy consumption and contributes to a lower CO₂ footprint by up to 40 percent with VSP One’s energy-efficient architecture and reporting. A power efficiency service level agreement (SLA) ensures improved cost efficiency and environmental responsibility.

These join three existing Hitachi Vantara guarantees. A 100 percent data availability guarantee provides uninterrupted access to critical business data. The Hitachi Vantara Effective Capacity Guarantee provides 4:1 data reduction and maximizes storage efficiency. The Modern Storage Assurance Guarantee delivers continuous innovation with perpetual, non-disruptive upgrades to new controllers, ensuring a path to the future of storage without needing to repurchase capacity.

Index Engines says Hitachi VSP Ransomware Detection, powered by CyberSense, detects corruption early, isolates threats, and guides IT to the last known clean backup so recovery is fast, precise, and trusted.

Index Engines says CyberSense supports NIST CSF (Cyber-Security Framework) functions by embedding data integrity analytics and recovery assurance into the storage product layer:

  • Identify: CyberSense indexes and analyzes metadata and content at scale, helping organizations understand what data they have and identify potential risks related to data corruption or tampering.
  • Protect: While Hitachi provides immutable storage and encryption, CyberSense adds another layer by continuously validating the integrity of backup data, protecting against undetected ransomware dwell time.
  • Respond: With audit trails and forensic-level reporting, CyberSense enables quick investigation and targeted recovery, reducing the time to respond to a data breach.
  • Recover: CyberSense identifies the last known clean copy of data and enables precise, surgical recovery, supporting fast, verified restoration in line with NIST’s Recover function.
Jim McGann, Index Engines
Jim McGann

Jim McGann, VP of strategic partnerships at Index Engines, said: “The addition of VSP One’s Cyber Resilience Guarantee, including Ransomware Detection powered by CyberSense, equips organizations with the intelligence and automation needed to strengthen their cyber resilience. By integrating advanced tools like VSP One and CyberSense, IT teams can streamline recovery workflows, minimize downtime and validate the integrity of critical data with greater confidence to minimize the impact of an attack.”

Hitachi Vantara gets up to the minute ransomware detection technology while Index Engines notches up a fourth OEM for its software. McGann will now be turning his attention to uncommitted potential OEMs, such as DDN and Pure Storage, but not NetApp, which has its own ransomware detection tech, or HPE, which offers Zerto ransomware detection.

For more information on VSP One and the expanded SLA guarantees, visit the Hitachi Vantara website.

Chinese researchers claim fastest memory in graphene-based paper

Chinese researchers have developed a new type of storage-class memory based on graphene, claiming it outperforms both SRAM and DRAM in speed.

A team of eight researchers are said to have developed the tech at Shanghai-based Fudan University, which made the strong claim: “It became the fastest semiconductor charge storage device currently known to humanity.”

The basic data transfer rate of this storage-class memory, named PoX, is reported to be 400 picoseconds (a picosecond is a trillionth of a second), with the device reportedly able to switch state 2.5 billion times a second. This would be quicker than the fastest SRAM, which can operate at less than 1,000 picoseconds but is volatile, and DDR5-6000 DRAM, which can have a 12,000-picosecond write time. The PoX device is fabricated using 2D graphene and is described in an open access Nature paper, “Subnanosecond flash memory enabled by 2D-enhanced hot-carrier injection.”

The paper claims: “We report a two-dimensional Dirac graphene-channel flash memory based on a two-dimensional enhanced hot-carrier-injection mechanism, supporting both electron and hole injection.” 

The two-dimensional Dirac graphene-channel refers to a single layer of graphene with the carbon atoms bonded in a hexagonal lattice. The electrons in this lattice are called charge carriers. They behave in such a way as to provide high electrical conductivity and mobility. 

A hot-carrier-injection mechanism involves charge carriers, such as these electrons or holes (electron vacancies), being accelerated enough by a relatively strong electrical field to be injected into an adjacent structure such as an insulator or floating gate.

NAND flash memory uses Fowler-Nordheim (FN) tunneling and hot-carrier-injection mechanisms to write and erase data by moving charge carriers (electrons or holes) across an insulating barrier into a floating gate or charge-trapping layer. However, the speed is restricted to a range of 10-100 μs (10 million to 100 million picoseconds). The restriction is caused by the height of the energy barrier between the silicon substrate (e.g. floating gate) and the tunnel silicon dioxide. 

By exploiting the atomically thin properties of 2D materials, the researchers managed to find a combination with a much lower barrier height and consequently faster charge carrier injection process. They call this 2D-enhanced hot-carrier injection (2D-HCI). ”Utilizing the 2D-HCI mechanism, we developed sub-1-ns flash memory,” they write.

They “fabricated graphene flash memory based on a hexagonal boron nitride (hBN), hafnium dioxide (HfO₂), and aluminum oxide (Al₂O₃) memory stack,” with the HfO₂ being a trapping layer and the carriers injected through the hBN layer. Here’s an image of their fabricated structure:

The researchers said they tried this out with a two-dimensional tungsten diselenide semiconductor material as well as the Dirac graphene, but it exhibited scattering, causing it to be less efficient. Diagrams in the paper show this:

A fourth diagram shows the setup:

The researchers say their device “still has a large memory window after ten years at room temperature” using statistical extrapolation, and “the device can repeatedly switch between two states and work well within 5.5 x 10⁶ cycles.”

They conclude: “Our findings provide a mechanism to achieve sub-1-ns program speed in flash memory, providing a path to achieve high-speed non-volatile memory technology.”

Liu Chunsen, a researcher on the team, said: “The two-dimensional super-injection mechanism pushed the non-volatile memory speed to its theoretical limit, redefining the boundaries of existing storage technologies.”

Zhou Peng, another research team member, added: “Our technology breakthrough is expected to not only reshape the global storage technology landscape, drive industrial upgrades, and foster new application scenarios, but also provide robust support for China to lead in relevant fields.”

That is, if the device can be cost-effectively fabricated, and remain stable, which would be a crucially important point. As far as we can tell, there has yet to be a method that produces graphene on a larger scale cost effectively.

The Nature paper was submitted in September last year, accepted in February this year, and published this month. It looks solid, having passed through Nature’s review process.

It’s clearly still at the research phase, but if this storage-class memory tech stands up, and can be cost-effectively fabricated, it could potentially reshape the memory and SSD landscape. Let’s see how the DRAM, SRAM, and NAND fab suppliers react.

Bootnote

The PoX moniker comes from the Chinese term for the technology; 破晓 or Pòxiǎo which translates to breaking dawn.

How Pure Storage developed the FlashBlade//EXA system

Interview. What led up to Pure Storage developing a new disaggregated system architecture for its FlashBlade//EXA?

Pure Storage was founded as an all-flash storage array (AFA) vendor, surviving and prospering as a wave of similar startups rose, shone briefly, and then were acquired or failed as more established competitors adopted the technology and closed off the market to most new entrants.

In the 2010-2015 era and beyond, the mainstream enterprise incumbents including Dell, Hitachi Vantara, HPE, IBM, NetApp, and Pure ruled the AFA roost with what was essentially a dual-controller and drives architecture with scale-up features. There was some limited scale-out, with examples such as Isilon, but for true scale-out you had to look to the HPC space with parallel file system-based suppliers such as DDN.

Then, in 2016, a disaggregated canary was born in the AFA coal mine: VAST Data emerged with separately scale-out metadata controllers and scale-out all-flash storage nodes talking to each other across an NVMe RDMA fabric. There was a lot more clever software in VAST’s system yet the base difference was its disaggregated design, aided and abetted by its single storage tier of QLC flash drives. 

Moving on, the AI boom started and gave an enormous fillip to VAST Data and parallel file system AFA vendors such as DDN. Only their systems could keep GPU server clusters running GenAI training workloads without data I/O waits. The incumbent enterprise AFA vendors added Nvidia GPUDirect support to speed their data delivery to GPU servers but could not match the data capacities enabled by the disaggregated scale-out and parallel file system vendors. Consequently, VAST Data and DDN prospered in the Nvidia-dominated GPU Server mass data training space.

This caused a revaluation among other suppliers. HPE adopted a VAST-like architecture with its Alletra storage. Dell assembled Project Lightning to add parallelism to its PowerScale/Isilon storage, while NetApp created the ONTAP for AI project. And Pure? It announced its FlashBlade//EXA a couple of weeks ago.

This has a clever twist in that the existing dual-controller FlashBlade system is used as the metadata controller layer system with separately scaled-out QLC flash storage nodes accessed across an RDMA fabric. We spoke to Pure’s VP for Technology, Chadd Kenney, to find out more.

Blocks & Files: FlashBlade//EXA seems like a radical revision of the FlashBlade architecture, giving it a parallel processing style of operation. Would it be possible to do the same sort of thing with FlashArray, and is that a silly question?

Chadd Kenney: No, it’s an interesting one. So let me give you some background on kind of where EXA came about, which may help you understand the logic behind why we decided to build it. It was a very fun project. I think we spent a lot of time thinking about what architectures existed in the market today and which ones did we want to take additional values from, and then what core competences did we have that we could potentially apply to a new solution. 

And so this was an interesting one because we had a couple different options to play with. The first is we obviously could make bigger nodes and we could break away from the chassis kind of configuration and the bigger nodes would give us more compute. We’d have ability to scale them in a different manner. And that gave us a pretty sizable performance increase, but it wasn’t big enough.

Customers were starting to ask us for, it’s funny, we would get in conversations with customers, they would ask us for 10 terabytes or 20 terabytes per second of performance. So we were kind of scratching our head saying, wow, that’s a lot of performance to deliver there. And so we started to think then later about how decoupling the infrastructure could be an interesting possibility, but how would we do it? 

Then we went back and forth on what the core competence that we were trying to solve for customers was. And the one thing we continued to hear was metadata performance was incredibly problematic and in many cases it was very rigid in the way that it actually scaled. So as an example, there are alternate PNFS or Lustre-like solutions that have tried to solve this before where they disaggregated metadata out. The downside was that it wasn’t scalable somewhat independently in a very granular fashion so that I could say I want to have a massive amount of metadata and a tiny bit of capacity, or I want to have massive capacity or a little tiny bit of metadata.

And so that intrigued us a little bit to understand, OK, so how could we actually deliver this in a different mechanism? 

The second thing we started to think about was, if you think about what FlashArray started with, it’s a key-value store that was highly optimized to flash in order to access data. We then took that same exact key-value store and FlashBlade and somewhat scaled it out across these nodes and we realized, OK, there’s a couple of core competencies to this. 

One is incredibly fast with multi-client connections. And then the second part was that we had an object count that was in the 40 quadrillion level. I mean it was ridiculous level for the amount of objects. So as we started to think about that, we said, well, what if we use the metadata engine that was core to FlashBlade and just kept that intact and then with the amount of objects that we could scale, we could infinitely scale data nodes.

We actually don’t even know what the quantity of data nodes we could get to was. So then we had to test to say, well, could these data nodes scale in a linear fashion? And so when we started adding data nodes in this decoupled way we decided to go with PNFS because that was what customers were asking for initially from us. 

Although I think ObjectStore with S3 over RDMA will probably be the longer term approach, PNFS was kind of like what every customer is asking us about. And so when we started building this, we started to realize, oh my God, as we add nodes, we are seeing exactly a linear performance rise. It was between 85 to a hundred gigabytes per second. And we got really excited about the fact that we could actually build a system that could take very different client profiles of access patterns and linearly scale the overall bandwidth.

And so then we all of a sudden got really excited about trying to build a disaggregated product. And the first logical thing we thought about was, well, we’ll build our own data nodes. And we’ve talked about doing those for quite some time and we were back and forth on it. The hyperscale win kind of gave us some ideas around this as well. And so we then looked at talking to customers – how would you consume this? And what was interesting is, most said, I already have a massive data node infrastructure that I’ve invested in. Could I just use those instead? 

And we said even competitive nodes, comically enough, we could potentially use, we just have to meet the minimum requirements, which are pretty low. And so we decided let’s go to market first with actually off-the-shelf whatever nodes that meet the requirements and Linux distribution and we’ll lay a small little set of packages to optimize the workflows, things like our rapid file toolkit and a bunch of other stuff that were on top of it. 

It got really exciting for us that all of a sudden we had this thing we could bring to market that would effectively leapfrog the competition in performance and be able to achieve some of these larger GPU cloud performance requirements that were almost unattainable by anybody without specific configurations out there.

Blocks & Files: I think I’m asking a dumb question by asking if FlashArray could use this same approach because that’s basically like asking could the compute engine in Flash array use the same approach.

Chadd Kenney: It’s an interesting concept because I think we’re always open to new concepts and ideas and we try to prove these out, especially in the CTO office, we spent a lot of time conceptualising could we change the way that we construct things. I mean the one thing that changed with FlashBlade that was different was we went to this open ecosystem of hardware. And so there is a different mode of operation than what we typically build. We typically build these gorgeous, elegant systems that are in highly simple that anyone can get up and running in near minutes. EXA is obviously a little bit different, but what we realized is the customers wanted that build-it-themselves thing. They kind of liked that, whereas in the enterprise model, they’re not that into that. They really want just the more black box, plug it in and it works.

Blocks & Files: An appliance?

Chadd Kenney:  Yes. I think with FlashArray, the one thing that’s tough about it is it’s just so good at what it does in its own ecosystem and we’ve kind of built now every possible tier. In fact, you’ll see some announcements come out from us around Accelerate timeframe … But we’re going to take somewhat of an EXA-level performance in FlashArray too. And so we really like the way it’s built, so we haven’t really found a mechanism of disaggregating it in any way that made sense for us. And I think you’ll see there are maybe new media types we’ll play with that we will talk about later. But QLC is still very much our motion for this way of doing multi-tiers. So I think FlashArray is somewhat going to stay where it is and we just love it the way it is to be honest. There hasn’t really been demands to shift it.

Blocks & Files: Do you see the ability to have existing FlashArrays occupy the same namespace as EXA so the data on them can be included in EXA’s namespace?

Chadd Kenney: Yeah, really good question. I think what you’re going to start seeing from us, and this will be part of the Accelerate announcement as well, is Fusion actually is going to be taking the APIs across products now. 

So for Object as an example, you may or may not see Object show up on the alternate product in the future here. And so the APIs will actually be going through Fusion for Object and then you can declare where you want it to actually land based upon that. So you’ll start to see more abstraction that sits on top of the systems that will now dictate things like policy-driven provisioning and management where consumers very likely won’t even know if it lands on FlashArray or FlashBlade or not. For certain protocols. Obviously for Block, it’s likely going to land on a FlashArray, but they may not know which FlashArray it lands on. 

So some of the cool stuff I’m working on the platform side of the house is trying to take the telemetry data that we have, build intelligence for recommendations and then allow users to interact with those recommendations both through policy-driven automation but also through natural language processing with some of the Copilot capabilities we’re building.

And then inside the platform we’re starting to build these workflow automations and then recipes that sit outside of that to connect to the broader ecosystem. So we will see our platform thing come to life around Accelerate. I won’t steal too much thunder there, but that’s where we’re really trying to go. It’s like instead of actually trying to bridge namespaces together at the array side, we’re trying to build this all into Fusion to actually have Fusion dictate what actually lands on which system.

Blocks & Files: Let’s imagine it’s six months down the road and I’m a Hammerspace salesperson. I knock on the door of one of your customers who’s aware of all this stuff going on and say, I’ve got this brilliant data orchestration stuff, let me in and I’ll tell you about it. Would your customer then be able to say to the Hammerspace guy, sorry, I don’t need you. I’ve already got that.

Chadd Kenney: Yes. So one thing that we embedded into EXA that we didn’t really talk a lot about is we use FlexFiles as well. So data nodes can reside anywhere and those all can be part of the same namespace. And so we didn’t go too big on this initially because, if you think about the core use case, honestly most people are putting it all in the same datacenter right next to the GPUs. 

So I think over the longer term you’ll start to see more geo-distribution. With FlashBlade as a whole, that is a plan. And as we build pNFS into the core FlashBlade systems, you’ll start to see FlexFiles and multi-chassis be kind of stitched together as one common namespace. So I think we’ll have similar use cases. Of course we’re all-flash and they can use other media types out there. So they may get into use cases maybe that we’re not so tied to; ultra archive and those types of things. But I think as we get into 300 terabytes and then 600 terabytes and the incremental growth of capacity, we’ll start to break into those similar use cases as them.

Scale Computing picks Veeam as its bestie backup supplier

Edge, ROBO and mid-market hyperconverged infrastructure biz Scale Computing, is launching support for its Scale Computing Platform (SC//Platform) within the Veeam Data Platform.

Veeam now describes its backup and recovery software as a ‘platform’, saying it goes beyond backup and recovery by offering data management and cyber-resilience. The resilience relies on things like malware detection, backup immutability and incident response.

Jeff Ready.

Scale’s co-founder and CEO Jeff Ready claimed in a statement “The partnership between Scale Computing and Veeam delivers the best of both worlds: streamlined, autonomous IT infrastructure from Scale Computing and the industry’s most trusted data resilience platform from Veeam.”

Customers “have long asked for deeper integration with Veeam,” he added.

Scale’s HyperCore hypervisor scheme treats its software, servers and storage as an integrated system with self-healing features, and has its own snapshot capability with changed block tracking. Scale’s SW pools storage across tiers with its SCRIBE (Scale Computing Reliable Independent Block Engine). Snapshots use SCRIBE to create near-instant, point-in-time, consistent backups of VMs, capturing virtual disks, VM configuration, and, with Guest Agent Tools installed, the running OS and applications.

We envisage SCRIBE-based snapshots being used by Veeam as its basic unit of Scale backup, the entry point for Scale’s customer data into its data protection system.

Customers can, we’re told, choose from a range of Veeam-compatible backup targets, including object storage, tape, and cloud, and use full VM and granular file recovery from SC//Platform to any supported environment. They can also migrate and restore workloads between SC//Platform, VMware vSphere, Microsoft Hyper-V, and major public cloud environments.

Shiva Pillay.

Shiva Pillay, SVP and GM Americas at Veeam, said in a statement: “This collaboration with Scale Computing further strengthens Veeam’s mission to empower organizations to protect and ensure the availability of their data at all times and from anywhere, delivering cyber recovery and data portability across a purpose-built platform tailored for the unique needs of edge IT.”

Support for SC//Platform in the Veeam Data Platform is expected to be generally available in Q4 2025.

Live demonstrations of the Scale Computing solution will be featured at VeeamON 2025, taking place April 21-23, 2025 at booth #G5 in San Diego, CA. An online brochure about the two supplier’s integration can be accessed here.

Seagate claims spinning disks beat SSDs on carbon footprint

A Seagate Decarbonizing Data report says energy usage is a top concern for more than half of business leaders and better use of disks is a wise datacenter choice.

Jason Feist, Seagate
Jason Feist

The report cites a forecast by Goldman Sachs that global power demand from datacenters will increase by as much as 165 percent by 2030, compared with 2023. With this in mind, it says that rising data volumes, slowing power efficiency gains, and increasing AI adoption are putting pressure on organizations to manage carbon emissions, infrastructure expansion, and total cost of ownership (TCO) at the same time.

Jason Feist, SVP of cloud marketing at Seagate, says in the report: “Datacenters are under intense scrutiny – not only because they support modern AI workloads, but because they are becoming one of the most energy-intensive sectors of the digital economy. This calls for a fundamental shift in how we think about data infrastructure – not as a trade-off between cost and sustainability, but as an opportunity to optimize for both.” 

What does Seagate think a fundamental shift consists of? Getting rid of spinning disks? Not at all. The report includes a table showing the embodied carbon of the three main types of storage media – disk, SSD, and tape:

Seagate figures

It concludes:

  • SSDs have the highest embodied carbon, both in total and per TB, making it the most carbon-intensive option among the three storage media.
  • Hard drives exhibit the least carbon footprint, both in total and on a per-TB basis, offering the most carbon-efficient sustainable storage solution.
  • LTO tape shows moderate embodied carbon, but its annual impact is higher than that of hard drives.

The report suggests three strategic pillars for building a more sustainable data future:

  • Technological Innovation: Advances in computational power, storage areal density, and energy-efficient technologies like liquid/immersion cooling and HVAC systems can significantly lower energy consumption and carbon emissions, effectively managing the growing demand profile. 
  • Commitment to life cycle extension and circularity: Refurbishing, reusing, and maintaining storage equipment extends lifespan and reduces waste. Real-time environmental monitoring and transparent reporting can foster accountability across the datacenter environment. 
  • Share accountability across the ecosystem: Achieving meaningful emissions reduction – across Scopes 1, 2, and 3 as outlined in the report – requires collaboration across the entire value chain, including vendors, suppliers, and cloud service providers.
Seagate graphic
Decarbonizing Datareport graphic

An example of tech innovation, Seagate says, is its HAMR-based Mozaic 3+ disk technology, now in volume production. This enables up to three times more capacity in the same footprint than a 10 TB drive, and reduces a drive’s embodied carbon by more than 70 percent per terabyte. It also lowers cost per terabyte by 25 percent, according to IDC.

Feist says: “Sustainability cannot be solved in isolation. A holistic approach spanning infrastructure, life cycle management, and industry-wide accountability could ensure that the growth of AI and datacenter operations does not come at the expense of the environment.”

Download the Decarbonizing Data report here.

Bootnote

Seagate produces and sells its own Nytro SSD products for datacenter use. Flash drive-based vendor Pure Storage has a different point of view, as you might expect. It takes a system-level view rather than per-drive. Pure suggests that, for a 1 exabyte deployment over ten years with a five-year life cycle for the HDDs and a ten-year life cycle for DirectFlash Modules, it finds that the HDD system emits 107,984 metric tons of carbon whereas a Pure-based system emits 14,779 tons. You can find more details here.

Hammerspace secures $100M to chase AI-driven data growth

Data orchestrator Hammerspace has ingested $100 million in funding to accelerate its global expansion.

Up until two years ago, Hammerspace, founded in 2018, was financed by founder and CEO David Flynn along with cash from a group of high net worth individuals, and long-term investing sources. It raised $56.7 million in an A-round in 2023, saying it planned to expand its sales and marketing business infrastructure. Now it has raised another $100 million as it races to capitalize on anticipated demand for AI, with customers demanding widespread, rapid access to data – a challenge Hammerspace says it can solve through orchestration.

David Flynn, Hammerspace
David Flynn

Flynn stated: ‘AI isn’t waiting. The race isn’t just about raw throughput – it’s about how fast you can deploy, move data, and put your infrastructure to work. Every delay is unrealized potential and wasted investment. We built Hammerspace to eliminate friction, compress time-to-results, and significantly increase GPU utilization. That’s how our customers win.”

In June last year, Flynn was looking forward to Hammerspace becoming cash flow-positive and thinking about a possible IPO in the next 18 to 24 months. He said: “We are building a very financially disciplined company, but we are reaching a point where we have to grow a lot faster to get in front of the opportunities being created by AI.” But he cautioned: “It’s a tough market for an IPO right now, with budgets going down in some areas, and Nvidia sucking the air from everybody else at the moment. It was a tough time to envisage an IPO as Nvidia was so dominant in AI.”

Ten months later, Nvidia’s dominance has only grown, the AI surge continues, and unpredictable tariff shifts have made short-term business planning increasingly difficult. Hammerspace has decided to go for AI-fueled growth and needs fresh funding to achieve it.

This B-round was led by Altimeter Capital with participation from Cathie Wood’s ARK Invest, and a combination of new and existing investors. ARK is an existing investor, having participated in the 2023 A-round.

These investors have bought into Flynn’s view that the AI wave needs a unified data infrastructure that allows AI agents to query and process as much of an organization’s information assets as possible. That means the organization must have complete visibility and control over its data estate – regardless of protocol or location – all accessible through a single pane of glass. This is Hammerspace’s Global Data Environment.

Flynn said: “We didn’t build orchestration for the sake of it. We orchestrate data to the GPU faster regardless of where it is physically stored. We instantly assimilate data from third-party storage so it’s ready to process faster. We deploy and scale easily and quickly so our customers can achieve their business outcomes faster. 

We expect storage vendors like Arcitecta, Dell, DDN, NetApp, Pure Storage, VAST Data, and WEKA to vigorously challenge Hammerspace and fight hard for the enterprise AI data infrastructure business.

Pure debuts RC20 FlashArray//C with reused controllers for smaller deployments

Pure Storage has introduced a more affordable, low-capacity FlashArray//C model, the RC20, built with reconditioned controllers and aimed at edge deployments and smaller workloads.

The FlashArray//C series uses Pure’s proprietary DirectFlash Module (DFM) drives built from QLC flash chips. They are currently at the R4 (release 4) level with the new //RC20 added to the existing C50, C70, and C90 models, as the table indicates:

Pure’s reconditioned controller CPUs and PCIe bus plus the reduced drive drive bays mark out the RC20.

Shawn Hansen, Pure VP and GM of the FlashArray business unit, blogged about the new smaller FlashArray box, saying: “This capacity-optimized all-flash system delivers enterprise-class performance, reliability, and agility at capacities and prices that are more accessible for edge deployments and smaller workloads.”

It has a 148 TB entry-level capacity compared to the prior entry-level //C50’s 187 TB. This could consist of 2 x 75 TB DFMs, whereas the C50 uses 2 x 75 TB and 1 x 36 TB DFMs but Pure can mix and match drive capacities as it sees best.

Hansen says: “We heard loud and clear that SMBs and smaller deployments or scenarios with remote office/branch office (ROBO) were missing out. They were looking for the full capabilities of the Pure Storage platform but with smaller capacities and at a competitive price point.”

The “R” in the RC20 name denotes “the use of factory-renewed controllers, which in conjunction with other new parts, such as new chassis, delivers a product that provides many of the same benefits as other members of the FlashArray//C family, all while reducing e-waste and aligning with our commitment on delivering a sustainable data platform.”

The RC20 can be non-disruptively upgraded to larger products in the FlashArray//C family and/or to next-generation controllers in the future. A Pure chart depicts this, though it lacks numerical values on its axes:

Hansen is keen to show the low electricity needs of the RC20, as expressed in a table:

No pricing details were provided. Check out a FlashArray//RC 20 datasheet here.

Lightbits: record Q1 with rise in sales and deal sizes

Privately-owned Lightbits says it has broken growth records in the first 2025 quarter with a business surge – though we have no way of verifying this claim.

It supplies a disaggregated virtual SAN, block storage accessed by NVMe/TCP that runs either on-prem or in the Azure, AWS, or Oracle clouds, using ephemeral instances to deliver ultra-high performance.

There was, we’re told, a 4.8x increase in software sales, a 2.9x increase in average deal size, a jump in new customers, and a 2x year-over-year license increase. We have nothing to compare this to in terms of actual baseline numbers but it sounds impressive. New customers came from financial services, service providers, and e-commerce organizations with analytics, transactional, and mixed workload environments that need massive scale, high-performance, and low-latency access storage.

Eran Kirzner, Lightbits
Eran Kirzner

CEO and co-founder Eran Kirzner stated: “The quarter close marked significant progress financially and strategically … We now service Fortune 500 financial institutions, as well as some of the world’s largest e-Commerce platforms and AI cloud companies.” 

Lightbits cited AI cloud service providers Crusoe and Elastx and cloud company Nebul as recent customer wins. It says its storage software scales to hundreds of petabytes and delivers performance of up to 75 million IOPS and consistent sub-millisecond tail latency under a heavy load.

CRO Rex Manseau said: “We’re seeing a consistent pattern of engagement with customers finding that other software-defined storage can only accommodate low and middle-tier workloads. They adopt Lightbits for tier 1 workloads, and then we move downstream to their utility tier, as well. And customers seeking VMware alternatives like Lightbits for its seamless integrations with OpenShift and Kubernetes to enable their infrastructure modernization.” 

The intention from Lightbits management is “to expand its global install base, prioritizing key markets across the Americas and Europe, and other high-growth regions.” It has agreements coming this year as well as new products to enable a broader workload coverage area.

Lightbits has raised a total of $105.3 million in funding since being founded in 2016. The most recent C-round in 2022 brought in $42 million.