Home Blog Page 373

Compute Express Link triumphs in the post-PCIe bus war

The Compute Express Link (CXL) bus has won the post-PCIe war and will enable disaggregated systems technology.

Alex McDonald, EMEA Chair of the SNIA, told a press briefing in London last week that AMD, ARM, and IBM have joined Intel aboard the CXL bus technology bandwagon. That’s all four main CPU vendors.

Only interconnect

We looked CXL a year ago and saw it then as a fourth future high-speed interconnect technology alongside the Gen-Z Consortium, OpenCAPI and CCIX initiatives.

CXL has the full backing of Intel and we suggested “that the sooner CCIX, Gen-Z and OpenCAPI combine the better for them. An Intel steamroller is coming their way and a single body will be harder to squash”.

They haven’t combined and in our view Gen Z, OpenCAPI and CCIX are dead in the water. They have not attracted the same degree of CPU manufacturer support – which means CXL has won the bus war.

Any post-PCIe bus technology for interconnecting processors to DRAM, FPGAs and other dedicated processors must have the support of the server CPU suppliers. It’s a sine qua non.

CXL bus features

Composable system vendors such as Liqid that use the PCIe bus are nicely placed to move into CXL bus technologies and increase the granularity of their composed systems. Blocks & Files thinks all composable systems vendors will have to support the CXL bus because of this.

In his presentation, McDonald discussed CXL bus features which make it attractive to CPU vendors and how the CXL bus could help shatter or disaggregate the integrated server (compute + DRAM + storage + networking) model which is already under attack from GPUs as well as assorted ASICs and FPGAs.

Disaggregated systems, featuring computational storage, persistent memory and server composability, all need a high-speed bus to work. This requires better-than-PCIe technology because today’s PCIe bus is too slow and does not support distributed memory.

McDonald said current servers have memory directly connected to the CPU with storage connected over buses such as SAS and SATA and, lately, the PCIe bus using NVMe. That’s the left side of the diagram below.

The right side of the diagram shows the CXL bus being used to link processors to memory, including persistent memory and NVMe, hooking up storage with a storage controller / CPU present. This storage controller can be a dedicated storage processor carrying out some form of computational storage, offloading the host server CPU.

The CXL bus can be used in systems with distributed memories and its cache coherency ensures that item updates in one part of the distributed memory system are rippled through to the copies, or caches. The current PCIe bus is not cache coherent. (Cache coherency applies to two or more memory resources, such as main memory and a cache, holding copies of data that are identical or coherent. With multi-core CPUs each core has its own cache. A cache coherent system causes cached copies of data in main memory to be invalidated when the main memory item is updated, forcing the caches to be reloaded from main memory.)

When PCIe Gen 5 with its 32Gbit/s bandwidth arrives the CXL bus protocol can run across it. The PCIe 5.0 spec was published in May last year. PCIe 4.0’s spec was finalised in 2017 and systems featuring the technology are just about to hit the market in 2020. The earliest PCIe Gen 5 systems, and hence CXL, will arrive is 2021. The general idea is that PCIe 5.0 + CXL will be used for higher performance data centre servers, while PCIe 4.0 is relegated to lower-performance servers and desktop/notebook/workstation systems.

CXL protocol elements

There are three sub-protocols in the CXL scheme and they can be used at the same time on a CXL wire. They are CXL.io, CXL.cache and CXL.memory. A diagram shows how data and processing logic blocks are connected in the CXL world;

CXL.io is basically PCIe Gen 5 and all PCIE services will work. CXL.memory enables a host CPU to access persistent memory while CXL.cache connects a host CPU to cached memory in external processing devices such as accelerators like smart NICs, GPUs, FPGAs, ASICs,  dedicated storage processors – think Pensando and Pliops. It can also be used to link computational storage devices to a host server – think Eideticom, NETINT Technologies, NGD Systems and ScaleFlux.

Lastly, CXL.memory can interconnect memory buffers when memory is expanded, providing direct memory access (DMA).


Infinidat exhibits six exabytes – good that!

Infinidat, the high-end disk-based storage array maker, today announced it has deployed more than six exabytes of storage capacity, up from 5EB six months ago.

The company anticipates continued demand for petabyte scale systems to accelerate growth in customer numbers and total storage deployment. It notes that shipping “over 1EB in less than six months is 50 per cent faster growth than the nine months prior”. (Incidentally, 1EB can represent 83,333 x 12TB disk drives; or 100 arrays at 10PB/array average; or 50 arrays at 20PB/array.)

The company argues that all-flash arrays (AFA) at scale are prohibitively expensive and in its release today pointed out that its 6EB of storage capacity is more than the top eight all-flash array (AFA) vendors shipped in 2019.

Infinidat’s marketers appear to be comparing total capacity shipped for their company with only one year’s capacity shipped by its AFA competitors. No need for that, when it is demonstrably growing so fast.

Capacity for growth

We have charted Infinidat’s deployed capacity using its own statements (e.g. 3.7 EB in September 2018); 

A fairly linear capacity deployed growth line since October 2017, but with a recent acceleration

In November 2018 Infinidat CTO Brian Carmody told Blocks & Files: “Our average customer has 7.3PB of InfiniBox, and our largest have over 100 PB.” A 10PB/array average looks to be not unreasonable 14 months later.

InfiniBox upgrade

Infinidat today updated InfiniBox Storage Software to Release 5.0.11. The upgrade adds replication with Active-Active Consistency Groups, support for the FIPS-140-2 federal data protection standard and makes changes for the next generation of its DRAM/NAND Neural Cache engine. This speeds read and write data accesses.

Scality preps scale-out file system on Azure Blob Storage

Scality is developing a scale-out file system (SOFS) for Microsoft Azure Blob storage.

The company is building a cloud-native scale-out file system in the Azure Cloud using Microsoft Azure Cosmos DB to store file system metadata and file data payloads stored as objects in the Azure Blob service.

Paul Speciale, Scality chief product officer, told a press briefing in London last week that “Microsoft does not have a big fat file system in Azure” for unstructured data.

He said Azure Blob is designed for large amounts of unstructured object data, not files. Scality’s idea is to offer file services to hold large volumes of unstructured file data on Azure Blob storage as a lower-cost, high-capacity file storage tier.

According to Scality such a file system must support:

  • Thousands of clients in a scale-out fashion,
  • Deliver a minimum of 350MB/sec on single file transfers,
  • Support SMB and be Blob API-compatible,
  • Run in a customer’s cloud subscription,
  • Object storage economics.

Scality RING has scale-out file system capability and Scality is already familiar with Azure Blob Storage, Speciale said.

For example, RING has a certified Azure Blob API and can provide local storage for an Azure Stack on-premises system. Scality’s Connect for Azure Blob Storage enables any application that works with Amazon S3 to support Azure Blob Storage. It uses Azure Blob Storage API calls to map S3 Buckets and Objects to Azure Containers and Blobs.

Scality RING is already used outside Azure to provide scale-out file system resources with 10PB and 20PB SMB shares and up to 1P per day ingest rates. Customers include more than 30 hospitals, several national libraries and a US Department of National Energy Lab.

SOFS in Azure

Scality SOFS supports SMB with scale-out functionality and will host within the customer’s Azure account. It is namespace compatible with Azure Blob and S3 APIs, and there is policy-based tiering between Azure Block Blob tiers (Premium, Hot, Cool, Archive).

Initial performance tests of the Scality Blob SOFS, using the Azure Blob Premium service ($0.20/GB), show 400MB/sec single stream read and write to a single client over SMB. Cache reads reach 800MB/sec.

Blocks & Files expects Scality SOFS for Azure Blob storage to be announced later this year. Scality is also developing a single management facility for multiple clouds and multiple RINGS.

Scality is working with Microsoft to identify early customers in oil and gas seismic exploration, biopharma (genomic research simulations) and financial services (fraud detection, tradings).

Plans are also afoot for an Azure Marketplace presence alongside Scality’s Zenko Azure Connect, which provides S3 to Blob translation.

Azure and files

There are three other file offerings for Azure, none of which satisfy a requirement for storing massive amounts of unstructured data.

Microsoft Azure Files offers fully-managed cloud file shares accessible via the SMB protocol. There are premium shares for performance-sensitive, IO intensive workloads and standard shares for general purpose file storage.

Microsoft-owned Avere Systems supports a file system in Azure. According to Speciale, Microsoft positions Avere as a cloud edge service, focused on getting data into the Azure cloud.

Azure NetApp Files provides NetApp’s ONTAP file services on the Azure cloud. This is designed for running mission-critical apps in Azure, not for storing large amounts of unstructured data.

Scality SOFS would complement these file services.

Rubrik claims $600m revenue run-rate

Data management startup Rubrik this week said it achieved a $600m annualised gross bookings run rate for the fiscal 2020 year ended January 31.

The company’s fy17 total bookings run rate approached $100m and the fiscal ’18 amount was around $300m. Rubrik did not provide a run rate for its fiscal ’19 year and declined to do so when we asked. The fy18 number though had more than doubled by the end of fy 2020.

About half of the fiscal ’20 bookings came from customers who spent $1m or more on Rubrik products and services to date. Subscriptions accounted for nearly half of bookings by the end of the year.

Rubrik said it finished fiscal ’20 with more than 2,500 customers. It has previously reported relative customer number growth, saying it was 4x in fiscal ’18 and 7x in fiscal ’17.

Rival Cohesity achieved a $200m run rate in its fiscal 2018. The company trumpeted another good year for FY 2019 ended July 31, 2019. However, Cohesity did not reveal a revenue run-rate for the year, presumably because it was navigating the migration to a software-only business and therefore no longer has hardware revenues to tot up.

Despite the apparent disparity in size between the two companies, Rubrik fared less well than Cohesity in the GigaOm Radar for Hybrid Cloud Protection, published this week. By our reading of this market landscape, report author Enrico Signoretti considers Rubrik as less mature than Cohesity and both are behind Commvault. But Rubrik is making the transition from  feature-led supplier to more of a platform play, according to Signoretti.

This image has an empty alt attribute; its file name is Radar-screen-gigaom.jpg

Size does not matter for the GigaOm analyst, who “does not take into account the market share of each vendor, in order to “give a clearer view of the potential of each vendor. It is a forward-looking market landscape representation that leaves more room for innovation and differentiation, instead of weighting established positions”.

Rubrik buys Opas AI

The company last week announced the acquisition of Opas AI, a company specialising in artificial intelligence for root cause analysis. Terms were undisclosed.

Vinod Marur, senior vice president of engineering at Rubrik, said in a blog post announcing the news: “With Opas AI’s technology, we see a near-term opportunity to enhance our product offerings to provide our customers more proactive alerting and shorter problem resolution cycles when faults occur. This will in turn translate into increased availability of data and applications across any cloud or infrastructure.”

Toshiba selects Showa Denko platters for MAMR HDDs

Showa Denko (SDK), the Japanese chemicals and electronics company, among other things, makes platters for disk drive makers such as Toshiba. A 2019 annual results report reveals it is making MAMR (Microwave-assisted Magnetic Recording) platters for Toshiba.

MAMR drives have read-write heads beaming microwaves at bit value locations on the platter to make the recording medium more receptive to write impulses from the head. They are an alternative to HAMR (Heat-assisted Magnetic Recording) drives which use laser-generated heat to accomplish the same purpose.

Toshiba 16TB MG08 disk drive.

SDK has also developed HAMR platter technology. Seagate is relying on in-house manufactured HAMR technology to take its disk drive capacities past 16 to 18TB and onwards to 40TB and beyond. Current PMR (Perpendicular Magnetic Recording) technology cannot get past a 16TB/disk barrier.

Western Digital is starting to deploy in-house developed elements of MAMR technology in its 18TB and 20TB drives.

Toshiba has previously indicated it is leaning towards using MAMR technology. This SDK report shows that it is serious about a MAMR push. The SDK MAMR platters hold 2TB each. The SDK report say they have been adopted by Toshiba Electronic Devices & Storage Corporation for use in MAMR-technology-based 18 terabyte near-line HDD.”

SDK began shipping the 2TB MAMR platters to Toshiba last year. This implies that Toshiba could announce an 18TB nearline disk drive later his year. Its current largest capacity disk is the 16TB MG08, which was announced in September 2019.

Shingled magnetic recording diagram

Blocks & Files expects a Toshiba MG09 18TB drive announcement later this year, possibly with an enhanced capacity shingled media drive with 20TB. These use the same platters but drive software causes wider write tracks to be partially overlapped so that more, and narrower, read tracks can be added to the platters.

Kioxia NVMe SSDs use PCie 4.0 for big speed increase

Kioxia has launched two much faster than average NVMe SSDs – the CD6 and CM6, – which use the PCIe 4.0 bus to reach 1 million IOPS and beyond.

The CD4 and CD6 are enterprise class flash drives, designed for data centre use. The CD4 is single-ported and the CM6 is dual-ported, and therefore faster. They succeed Kioxia’s CD5 and CM5 drives.

Kioxia (formerly called Toshiba Memory) said they are the first data centre PCIe v4.0 drives to ship and their performance numbers (see below) suggest that servers fitted with PCIe v4.0 will get data in and out of memory much faster than current PCIe v3 servers.

The PCIe 4.0 bus delivers up to 64GB/sec, double PCIe gen 3’s 32GB/sec maximum. Now we wait for PCIe 4.0-compliant servers to ship with these drives.

The CD6 and CM6 come in mixed-use and read-intensive versions and are built from Kioxia’s BiCS 4 NAND – 96-layer 3D NAND formatted as TLC (3bits/cell). The read-intensive versions have higher capacities and lower endurance ratings as the mixed-use models have capacity set aside for over-provisioning (replacing worn-out cells). 

The CM6 has a maximum 30.72TB capacity, which is a third more than Western Digital’s DC HC650 3.5-inch 20TB disk drive.

Kioxia CM6 and CM6 capacities and “up to” performance numbers. TBD means “to be defined.”

The headline numbers are up to 6.9GB/sec sequential read bandwidth and 1.4 million random read IOPS.

This is faster than Gigabyte’s Aorus PCIe 4.0 SSD with its 5GB/sec sequential read speed.

Compared to Kioxia’s CM5 drive (PCIe v3.0, 64-layer TLC NAND) the CM6 has up to 106 per cent more bandwidth than the CM5, 134 per cent more random read IOPS, 386 per cent more random write IOPS and 67 per cent lower latency.

Both drives come in a U.3 format, meaning they will fit in tri-mode 2.5-inch drive bays. These support SATA, SAS and NVMe connectivity. The drives are self-encrypting to a FIPS standard.

Western Digital is a partner of Kioxia in building NAND chips, so we can anticipate the company to announce PCIe 4.0 SSDs soon.

Commvault looms large on hybrid cloud data protection radar screen

A GigaOm analyst has positioned hybrid cloud data protection vendors in the market, with a radar screen diagram to indicate strengths and focus.

Fourteen vendors are covered in the GigaOm Radar for Hybrid Cloud Data Protection. They are Actifio, Acronis, Clumio, Cohesity, Commvault, Delphix, Druva, HYCU, IBM, Rubrik, StorageCraft, Veeam, Veritas and Zerto.

By our reckoning the GigaOm Radar shows Commvault is the strongest supplier. Also Rubrik is considered less mature than Cohesity but is transitioning from feature-lead supplier to more of a platform play. Acronis and Veeam are given slightly higher ratings than Rubrik as challengers moving to leaders’ status.

According to report author Enrico Signoretti, the “general market direction is trending towards additional services built on top of data protection. Data protection is becoming instrumental in collecting and consolidating data across the entire organisation (no matter where it is created or stored) with the goal of reusing it for other purposes.”

GigaOm Radar details

GigaOm’s Radar is a four-circle, four-axis, four-quadrant diagram. The circles form concentric rings and a supplier’s status – new entrant, challenger, or leader – is indicated by placement in a ring.

There is a progression, with new entrants growing to become challengers and then, if all goes well, leaders.

 The inner white area is for mature and consolidated markets, with very few vendors remaining and offerings that are mature, comparable, and without much space for further innovation. This is not the case with data protection.

Signoretti does not take into account the market share of each vendor, in order to “give a clearer view of the potential of each vendor. It is a forward-looking market landscape representation that leaves more room for innovation and differentiation, instead of weighting established positions.”

There are four orthogonal axes, each with a progression from low (outer ring) to high (innermost ring);

  • Maturity – the maturity and solidity of the product, user acceptance of the solution, and overall ecosystem sustainability,
  • Horizontal platform play – products that can face a broader range of challenges, with a comprehensive feature set and an extensive ecosystem,
  • Innovation – differentiation of the solution, innovative technical aspects, and overall vendor approach to the market,
  • Feature play – focus of the solution in terms of single differentiating features and technical aspects of the product. It usually includes new vendors that are defining new product categories, niche players, and point solutions.

Slow, fast and out-performer arrows are placed on the diagram by each vendor, indicating their current position, direction and speed of progress. And, yes, out-performers are faster moving than fast movers.

Signoretti aims to provide readers a better understanding of a technology, to help them evaluate it, and explore the market to find the best products for their organisation. The Radar should be considered a companion piece to two other GigaOm reports: ‘Key Criteria to Evaluate Hybrid Cloud Data Protection’ and ‘Vendor Profiles for Key Criteria’.

We understand that Dell EMC’s Data Protection Suite may be added to a forthcoming edition of the Radar for Hybrid Cloud Data Protection report.

Comment

GigaOm’s radar screen is an alternative representation to Gartner’s classic four-box Magic Quadrant, Forrester’s Wave, IDC’s Marketscape, and Coldago’s Map, which are shown below, clockwise from the top left;

The radar screen seems to combine graphic elements of Forrester’s Wave and Gartner’s MQ. By having four axes it has more classification criteria by which to position vendors. These four axes form two pairs; innovation vs maturity and horizontal platform vs feature play.

Vendors have to be placed with reference to these, which is how they end up as points in the quadrants. You cannot have a vendor simultaneously rated high on maturity and innovation or on a horizontal play and a feature play in the GigaOm universe.

HPE Primera gains Veeam V10 support

HPE has announced Veeam support for its Primera high-end storage array and for disaster recovery data copies made by its StoreOnce software.

Veeam already supports HPE’s 3PAR and Nimble storage arrays and HPE’s hyperconverged SimpliVity line. With this week’s announcement of Veeam Availability Suite V10, the two companies have added Primera storage arrays to the mix.

With the VAS v10 support Primera users can recover data at granular levels and in minutes from storage snapshots made in virtualized and physical environments. StoreOnce is HPE”s deduping backup storage target appliance.

Veeam’s StoreOnce support provides a single interface for data protection and includes Catalyst Copy support. This stores, replicates and archives additional copies of data for disaster recovery. According to HPE the process does not affect backup service level agreements or constrain server CPU resources.

The HPE Veeam announcement does not mention S3 Object Lock support for immutable archive copies of data. We’re checking with HPE to clarify this.

Veeam v10 will be available for HPE Primera and StoreOnce customers this month.

SK hynix sends 800 workers home to self-quarantine against coronavirus

Coronavirus

SK hynix, the Korean flash and DRAM maker, has sent home 800 workers from its Incheon campus in South Korea as a self-quarantine precaution against the COVID-19 coronavirus.

The company said a new recruit has showed symptoms of pneumonia and was transported to a nearby hospital for tests. The recruit had been in close contact with a coronavirus patient in Daegu, 300 kilometres southeast of Seoul, where South Korea’s health authorities  have reported more than 10 COVID-19 cases.

SK hynix has shut down its in-house clinic, visited by the new recruit, and also closed its education centre, sending 280 new recruits home.

The company’s Incheon campus is 70 kilometres south of Seoul and employs some 15,000 workers. SK hynix said the self-quarantine measure has not affected operations of the chip fabs on the site.

Cohesity branches out data management software to ROBO and the edge

Cohesity has upgraded its data management software to cover remote office – branch office (ROBO) and edge IT sites, in a single environment along with central data centres.

Cohesity’s hyperconverged secondary storage platform offers backup, golden master file copies for test and dev, compliance and other data users, archiving, tiering to the cloud and general file and object storage.

Its market sweet spot to date has been hybrid, covering business data centres and the public cloud. Now it is extending from the data centre base to branch offices and data-generating IoT edge locations. The software runs as a virtual appliance or on certified HPE and Cisco servers.

Vineet Abraham, Cohesity SVP for engineering and product management, said in a statement: “By offering the enterprise-class features of Cohesity software in a cost-effective, plug-and-play solution, we are empowering organisations to bring their data centre, cloud, and edge together on a single platform.”

Backup to the Edge

Cohesity ROBO and edge users get instant mass restores to recover an entire branch when needed. Cohesity’s dedupe and compression reduces bandwidth utilisation when sending/receiving data to/from central or cloud sites. Cohesity also suggests its file and object services can replace local Windows file servers.

For Cohesity the edge includes branches of national banks, retailers, chain restaurants, rental car agencies, warehousing and distribution, pharmaceutical, and global IT services. Data from these sites typically need to be managed, protected, and secured locally, without dedicated IT staff to handle on-site administration.

Cohesity cites IDC research that shows more than half of enterprise data will be created and processed outside the data centre or cloud by 2022. According to the company, this underscores the need for a unified approach covering ROBO, the edge, enterprise data centres and public cloud.

Cohesity’s ROBO offering will be generally available to customers from Cisco and HPE by Spring 2020.

Wasabi intros price-fixing – in a good way – for cost-conscious cloud storage customers

Wasabi, a cloud storage startup, has devised a pricing model whereby customers buy capacity up-front in return for lower charges.

Wasabi’s Reserved Capacity Storage (RCS) provides enterprises with price predictability and one-time billing. The deal sees storage costs cut by up to 27 per cent compared with pay-as-you-go when customers commit to fixed price terms of up to five years. Customers pay only for storage that is reserved; there are no fees for data egress, API requests or data retrieval.

David Friend, CEO of Wasabi, gave out a quote: “With most of the world’s data still stored in on-premises hardware, we want to make it easy for people to migrate their data to the cloud and start saving enormous amounts of money in the process.”

Cloud storage customers like predictability

Wasabi said a 100TB/five year RCS plan costs under $40,000 and includes technical support. By contrast, a five-year, on-premises 100TB hardware appliance deal from an unnamed vendor lists at $130,000. This rises to more than $240,000 with maintenance over five years and does not include data centre or people costs.

Wasabi said its research shows many customers prefer the predictability of a fixed price purchase order. The company thinks RCS will be useful for backups, second copies of data, and archives. Also there is a make-margin bonus for channel partners, who can resell the service as a ‘boxed cloud’ SKU.

Wasabi claims its service is faster and 80 per cent cheaper than Amazon S3. The company was founded in 2017 and it said it has 10,000 customers worldwide and has deployed more than 100 petabytes to date.


Cloudian and Veeam build ransomware data recovery service atop Amazon S3 Object Lock

Cloudian and Veeam have introduced Amazon’s S3 cloud Object Lock on-premises to combat ransomware attacks.

In the event of an attack, ransomware-locked files can be replaced by Veeam backup data stored on the Cloudian HyperStore appliance, which sits in the data centre behind the firewall. These will be free of the ransomware infection, according to Cloudian, which said its storage has US government-certified data immutability.

Danny Allan, Veeam CTO, said in a prepared quote: “Preventing a ransomware attack has become increasingly difficult, so we’re focused on ensuring users can quickly and easily recover from such an attack – that’s the peace of mind that we’re delivering with on-premises S3 Object Lock in the new Veeam Availability Suite V10.”

S3 Object Lock

Launched yesterday, Veeam Availability Suite (VAS) V10 introduced S3 Object Lock for backup data, with backups to an S3 target made immutable through a software setting. Cloudian’s HyperStore is on-premises object storage which supports the Amazon S3 APIs. Cloudian supports Object Lock functionality and is now a VAS V10 Object Lock target device.

S3 Object Lock provides a virtual air-gap that logically disconnects stored objects from connected commuter systems. In Object Lock’s Compliance Mode, stored data can’t be overwritten or deleted by any user including the root account in AWS within the retention period. There is no way to change the retention mode or to shorten the retention period.

According to Cloudian, restores from HyperStore are faster than via the public cloud because an on-premises object store generally has a faster network link.

The company said the Veeam partnership means a HyperStore Object Lock system can be a platform for ransomware protection-as-a-service. Also in the case of VMware Cloud Providers, it can deliver single-point management via VMware vCloud Director.

Ransomware attacks are increasing

Cloudian quotes an US inter-agency report that reveals 4,000 ransomware attacks “have occurred daily since January 1, 2016. This is a 300 per cent increase over the approximately 1,000 attacks per day seen in 2015.”

Preventing users from disclosing account passwords to phishing emails and other preventative measures are useful hygiene but are not 100 per cent effective. According to the US federal agencies, ransomware attacks can be expected to surmount password and other barriers.

Organisations should prepare accordingly and take the following preventative measures:

  • Back up data regularly and verify the integrity of those backups and test the restoration process to ensure it is working.
  • Conduct an annual penetration test and vulnerability assessment.
  • Secure their backups – ensure backups are not connected permanently to the computers and networks they are backing up. Examples are securing backups in the cloud or physically storing backups offline. Some instances of ransomware have the capability to lock cloud-based backups when systems continuously back up in real time, also known as persistent synchronization. Backups are critical in ransomware recovery and response; if you are infected, a backup may be the best way to recover your critical data.