Home Blog Page 253

Infinidat strengthens InfiniGuard with automated network fencing and more throughput

Infinidat has added automated fenced-in networks to its InfiniGuard data protection array to safeguard forensic analysis of malware attacks, and doubled throughput with new B4300 hardware. The resulting software is branded InfiniSafe, and is claimed to enable a new level of enterprise cyber-resiliency.

The InfiniGuard system is a mid-range F4200 or F4300 InfiniBox array front-ended by three stateless deduplication engines (DDEs) and running InfiniGuard software providing virtual tape library functionality (disk-to-disk backup) as a purpose-built backup appliance (PBBA). It supports Commvault, IBM Spectrum Protect, Networker, Oracle, Veeam, Veritas and other software products backing up heterogeneous storage vendor’s kit.

Stan Wysocki, president at Infinidat channel partner Mark III Systems, provided an announcement quote: “What I’m most excited about is providing our customers a comprehensive storage-based ransomware solution that combines air-gapped immutable snapshots with an automated fenced-in network to determine safe recovery points, and then delivering near-instantaneous recovery.”

InfiniGuard arrays are preinstalled in a 42U rack with four drive enclosures preconfigured. There are InfiniGuard B4212N and B4260N systems, shipping from April 2021, based on the Infinidat mid-range F4200 array. The B4300s are based on the F4300 InfiniBox – a go-faster F4200 upgrade. Some differences between two now discontinued B4200 products and two B4300 products are listed in the table below:

The InfiniGuard software currently offers immutable snapshots, a virtual air-gap and virtually instantaneous recovery, based on all disk drive spindles being used in the restore operation. It can also replicate an InfiniGuard system’s contents to a remote InfiniGuard system.

Performance boost

The new B4300 hardware increases the DDE CPU power by moving from six-core to 20-core CPUs and doubling DDE memory. The network link between the DDEs and the backend storage is 16Gbit/s FC, and a new 25GbitE Ethernet frontend option has been added.

InfiniGuard software has been updated to use all the cores and the result is a throughput jump from 74TB/hour to 180TB/hour. This can reduce backup windows by 50 per cent or so, as well as shortening remote replication time. This means you get to a fault-tolerant (replication complete) state faster.

Recovery time is also reduced. Infinidat can now demonstrate the recovery of a 1.2PB backup data set in 13 minutes and 30 seconds, at which point it’s available to the backup software to do a restore. Eric Herzog, Infinidat’s CMO, tells us “No one else in the PBBA space can do a full recovery in less than 30 minutes.”

Competitor HPE’s fastest StoreOnce deduping PBBA, the 5660, has a 104TB/hour throughput rating. Dell EMC’s high-end PowerProtect DP8900 runs at up to 94TB/hour. InfiniGuard B4300 is almost double that speed. However, In January 2019, ExaGrid CEO Bill Andrews claimed his firm’s EX6300E PBBA scaled out two a 400TB/hour ingest rate. Its EX84 currently scales out to 32 systems and a 488TB/hour ingest rate in the EX2688G configuration. That beats InfiniGuard’s B4300s in sheer speed but is not comparing apples to oranges in terms of system configurations. InfiniGuard looks to be one of the fastest single node, disk-based, PBBAs out there, if not the fastest.

Pure’s FlashBlade offers up to a claimed 270TB/hour recovery performance from its all-flash datastore. A single node all-flash VAST Data/Commvault system offers up to 144TB/hour restore speed. Four nodes offer up to 576TB/hour and eight nodes 1.15PB/hour; this is what scale-out design is supposed to do.

Network fencing

The latest InfiniGuard release adds support for fenced and isolated networks. Network fencing isolates a device, such an InfiniGuard array from its connecting network. That means backups can be test restored inside a closed and isolated environment to find the latest known good backup and not affect primary systems at all. This means InfiniGuard customers can disregard any compromised backups and only recover good data.

Herzog suggests that several competing PBBA vendors have cyber-resilient systems roughly similar to InfiniGuard but they can require two separate systems. With its network fencing InfiniGuard can provide the cyber resilience in a single system.

Storage news ticker – February 8

W. Thomas Stanley.

Cloud file data manager (sync and sharer) Nasuni has a new board member: W. Thomas Stanley, the president and CRO for Chainalysis, CRO at Tanium before that, and a 13.5-year NetApp veteran, leaving as SVP and GM for the Americas in September 2019. He joined Nasuni’s board, taking the second independent seat, in November last year. Paul Flanagan, Nasuni CEO said “He will play a key role in helping guide Nasuni in the next phase of our growth as we continue to disrupt the file storage and data management market with a cloud approach.”

The other independent board member is Joel Reich who was chief product officer at NetApp for 16 years. 

Qumulo says it had a good 2021, achieving its best-ever quarterly results in overall billings, software subscription billings, cloud billings, and customer adoption. It increased its billings by 75 per cent from Q3 to Q4 alone. Qumulo added more than 200 customers in the year and its customers have stored over 2 exabytes (EB) of data, more than doubling last year’s record. Its latest Qumulo Core v5.0 release features Qumulo on Azure as a Service, Qumulo Shift for Amazon S3 to export/import data to/from AWS S3 while leaving data in its native object format, Qumulo on HPE and Supermicro all-NVMe, NFSv4.1 Support and NVMe-Hybrid to enable the performance of NVMe at the cost of dense disk.

Ghazal Asif.

Data protector/manager Rubrik, appointed Ghazal Asif as VP of global partners and alliances to drive continued growth and engagement with current and future partners. Asif joins Rubrik from Google where she was the head of channel partners in EMEA for Google Customer Solutions, leading channel go-to-market strategy and execution for the region. 

Kaseya’s Unitrends business launched ION and ION+ data protection appliances. They are smaller than rackmount products and the ION is a silent all-NVMe SSD device featuring automated disaster recovery (DR) testing, immutable storage, and AI-based ransomware detection. The ION+ has more powerful CPUs and additional memory. It comes in a small tower form factor with the ION’s software plus full-service backup and recovery. Both the ION and ION+ can be used with Unitrends Forever Cloud for offsite immutable storage and Unitrends Disaster Recovery as a Service (DRaaS) for hands-free failover to eliminate downtime.

IBM constructs CyberVault around new FlashSystem family

IBM is fortifying its new PCIe 4-enhanced FlashSystem family with CyberVault features designed to detect, prevent and recover from malware attacks.

Update; gen 3 FlashCore Module information added. More performance data added. 9 February 2022.

FlashSystem data is actively monitored in real-time and ransomware recovery accelerated by using validated restore points. Two new mid-range and high-end FlashSystems get PCIe 4 connectivity along with new controller CPUs, and have both capacity and speed increases.

David Chancellor, director enterprise systems for IBM biz partner Gulf Business Machines, said “Cyber resilience is clearly a top priority for our customers. [IBM Cyber Vault’s] ability to help dramatically reduce recovery times is exactly what cyber resilience teams need to keep the business running.”

IBM mainframe CyberVault graphic.

IBM did not say how CyberVault’s active monitoring and restore point validation work. Its mainframes have a Z CyberVault feature with active monitoring and safeguarded copies and FlashSystem CyberVault is based on this scheme and uses regularly taken safeguarded copy point-in-time snapshots. CyberVault automatically scans the copies created regularly by Safeguarded Copy looking for signs of data corruption introduced by malware or ransomware. It uses standard database tools and automation software.

FlashSystem rack and (inset bottom to top) 5200, 7300 and 9500 models.

New hardware

There are three members in the FlashSystem family, all running the same Spectrum Virtualize code: 

  • Existing 5200 – 1U – 12x NVMe drives in base controller chassis
  • New 7300 – 2U – 24x NVMe drives and replacing the 7200
  • New 9500 – 4U – 48x NVMe drives and replacing the 9200

The 7300 supports dual-port and hot-swappable 4.8, 9.6, 19.2 and 38.4TB gen-3 FlashCore Modules (FCMs – drives). The 19.2 and 38.4B FCMs have a PCIe 4 connection. Commercially available NVMe SSDs with 1.92, 3.84, 7.68, 15.36 and 30.72TB capacities are supported in dual-port, hot-swap form. It also supports SCM drives. The expansion chassis support the use of 2.5-inch and 3.5-inch disk drives as well, caused by convergence of the FlashSystem and Storwize product ranges.

Italicised columns are older systems included for reference.

The gen 3 FlashCore Module has an onboard processor (computational storage) to handle background storage tasks and uses industry standard QLC (4 bits/cell) and faster SLC (1 bit/cell) NAND technology to address performance and capacity requirements. At 38.4TB raw capacity (or 116TB effective), this makes it possible for IBM to deliver 1.1PB effective capacity per single rack unit in the FlashSystem 5200, 7300 and 9500 products.

IBM Master Inventor Barry Whyte said the gen 3 FCM’s “internal computational storage capabilities have been enhanced with a new “hinting” interface that allows the FCMs and the Spectrum Virtualize software running inside all FlashSystem products to pass useful information regarding the use, heat, and purpose of the I/O packets. This allows the FCM to decide how to handle said I/O, for example, if its hot, place it in the SLC read cache on the device. If its a RAID parity scrub operation, de-prioritise ahead of other user I/O. Single FCMs are now capable of more IOPs than an entire storage array from a few years ago!”

Tom Richards, systems and storage practice lead at IT consultancy Northdoor, told us the gen-3 FCMs have improved performance and double the Drive Write Per Day (DWPD) rating over standard SSDs. The in-line compression hardware increases the effective capacity of the larger modules to 3:1. FCM above the entry 4.8TB modules in the previous generation were limited to just over 2:1 compression. The new FCM 4.8, 9.6, 18.9 and 38.4TB modules can achieve hardware-accelerated compression of up to 22, 29, 58 and 116TB effective capacities respectively.

Richards told us the new FCMs will be supported in both of the newly announced arrays, along with the existing 5200 arrays (although mixing of old and new FCM types will not be supported within the same array).

The 9500 supports the same FlashCore Module and commercial SSD capacities as the 7300, plus Storage Class Memory with support for 1.6TB drives specified. The 9500 also supports SAS SSDs but not disk drives.

Performance

The FlashSystem 9500 offers twice the performance, twice the number of drives, and twice the connectivity of the 9200. IBM claims it provides performance gains of 50 per cent, 2x increase in connection options, and a 4x increase in workload support, but neglects to say what product is being used in the comparison. A customer said that the Cyber Vault FlashSystem reduced overall cyber attack recovery time from days to hours with a “comparable DS8000 function.”

An IBM spokesperson said the 7300 delivers 45GB/sec bandwidth, 3.5 million cache hit IOPS, and 580,000 IOPS with a 70 per cent read, 30 per cent write, 50 per cent cache workload. The faster 9500 outputs 100GB/sec bandwidth, 8 million cache hit IOPS, and 1.6 million 70/30/50 IOPS.

Whyte said: “I’ve been calling this box “the beast” and for any ‘The Chase’ fan’s I’m not referring to Mark Labbett, but the new top of the line FlashSystem. With 48 x NVMe slots, 12 x Gen4 PCIe Host Interface slots (48 x 32Gbit/s FC, 12 x1 00GbitE, 24 x 25GbitE), 4 x PSU, 4 x Battery, 4 x boot devices – this is the largest and most powerful FlashSystem we’ve produced, and, based on the performance measurements, it has to be one of, if not the fastest, highest bandwidth single enclosure systems in the world!”

Connectivity

FlashSystem 7300 Model 924 systems include eight 10Gb Ethernet ports as standard for iSCSI connectivity and two 1Gb Ethernet ports for service technician use. These models can be configured with up to three I/O adapter features to provide up to 24x 32Gb FC ports with SCSI and FC-NVMe support, or up to 12x 10/25Gb or 100Gb Ethernet ports with iSCSI and NVMe RDMA support.

FlashSystem 9500 systems can be configured with twelve I/O adapter features to provide up to 48x 32Gb FC ports, up to 20x 10/25Gb Ethernet (iSCSI, NVMe RDMA capable) ports, and up to 12x 100Gb Ethernet (iSCSI, NVMe RDMA capable). Two 1Gb Ethernet ports are provided for management and service actions.

According to Richards, “Both of the new FS7300 and FS9500 include the option for 100GbitE iSCSI host connectivity and, crucially, the ability to drive these host connections at full speed thanks to the new packaging at PCIe 4,” (in the 9500).

The two systems can both have end-to-end NVMe connectivity. Both come with utility model options which delivers a variable capacity system, where billing is based on actually provisioned space above the base. The base subscription is covered by a three-year lease that entitles you to utilise the base capacity at no additional cost. If storage needs increase beyond the base capacity, usage is billed based on the average daily provisioned capacity per terabyte, per month, on a quarterly basis.

IBM says that its customers can be helped to achieve sustainability goals because of the new FlashSystem’s greater performance and capacity in a smaller and more energy efficient footprint than prior models.

The FlashSystem 7300 requires IBM Spectrum Virtualize licensed machine code level 8.5 or later, for operation and the 9500 requires equivalent Spectrum Virtualise software. Get a FlashSystem 7300 datasheet here and a 9500 datasheet here.

Bootnote: IBM has refreshed its SAN Volume Controller (SVC) product, giving it a CPU upgrade with 2 x 24-core Intel Ice Lake processors, making it more powerful. The SV3 node hardware is essentially based on the same node canister used in the 9500.

Cool it: Optane needs a DRAM fix, but legal restrictions may have hindered partner progress

A claim that Optane cells can’t be immediately read because they need to cool after being written has been rebuffed by Intel, but it declined to comment on suggestions contractual restrictions hindered Optane’s adoption by third parties.

The suggested delayed read time cell cooling issue emerged after talking to sources who were testing early 3D XPoint technology, and who had in turn spoken with both Micron and Intel Optane product managers and engineers.

Optane is Intel’s 3D XPoint technology based on phase-change memory cells. The chips were manufactured by Micron in its Lehi fab, until Micron pulled the plug and walked away from 3D XPoint sales, marketing and manufacturing.

Intel graphic.

The Intel technology pitch around Optane was that it is cheaper than DRAM but not much slower, while being costlier than NAND but much faster. However, it has struggled to find its niche in the market and a new reason has been identified/suggested for that: the phase change memory cell inside it needs to be heated during the write process and has to cool down afterwards. That means content, once written, can’t be immediately read.

Therefore, Intel had to add DRAM cache to an Optane DIMM/SSD to fix the issue, raising its cost and complexity and making its price/performance ratio increasingly problematic. 

Intel, via a spokesperson, said “The assertion that Optane 3D XPoint memory needs to cool down after content has been written is incorrect. Optane memory can be read immediately after it has been written, and this does not drive any special DRAM caching circuitry.”

The reason for the DRAM caching was due to the DRAM-Optane speed difference, according to Intel. “The access latency of Optane 3D XPoint memory is inherently longer than DRAM and to mitigate this impact, Optane is utilised in ‘Memory Mode’. In this mode, a smaller set of DRAM DIMMs are configured as a cache for the larger Optane memory space. The net performance, at a solution level, is 90–100 per cent of the performance of DRAM-only implementation.”

Legal restrictions

We also heard that lawyers protecting Intel and Micron commercial interests put obstacles in the way of third parties needing to talk to their engineers and product people when wanting to use 3D XPoint technology in their own systems. One person told us his people could talk to Micron engineers about 3D XPoint but couldn’t tell Intel they were talking to Micron. They could also talk to Intel XPoint engineers but couldn’t tell Micron about that.

That meant that (a) they couldn’t talk to the Intel and Micron engineers in one room at the same time, and (b) they received mixed messages. For example, Intel product managers said XPoint development was progressing well while Micron engineers said it was meeting lots of problems. Our source wondered if Intel’s XPoint commercial management people knew of this, and whether messages were going upwards as they should.

This made them seriously doubt XPoint’s prospects. If this description of the situation is accurate it represents, in B&F‘s view, product development lunacy. The engineers/product managers were in a product development race but had to try to run while using crutches.

The industry consultant told us “Micron never liked 3DXP. It doesn’t solve any Micron issues, it is very complex to bring a new memory to market and Intel, Micron, Numonyx had been working on it for 20 years.

“Intel liked it for system reasons. Intel needs to get large amounts of memory, preferably non-volatile, to systems. This helps keep the storage component from slowing down compute. Two-level memory has been a plan from Intel for 10+ years. Intel can solve a lot of system problems with this.

“As a result of the above, Intel pushed Micron into doing 3DXP and come up with a Micron can’t lose finance model. Basically, Intel would pay for cost plus large overhead for bits shipped.”

“Micron wanted to convert the factory to DRAM/3D NAND … Intel said ‘No’, wanted to keep it for possible ramp up to support needs. Intel did a weird loan to IMFT to pay for upgrades and capital.

“Micron annexed the fab as allowed by agreement with the idea they could sell bits to Intel at a high price and still be able to sell SSDs. Intel’s ramp plans failed and they didn’t need the bits so they cancelled orders. Micron found the fast SSD business is not big  enough. At this point, the two companies’ relationship was falling apart and they each blamed the other for financial issues.

“Intel had data embargoes, refused to let companies have samples, made them work at Intel sites, etc.”

He said “At one point, there were no Micron engineers working on 3DXP at all in development. Intel was supplying all the work at the Micron plant.”

The Intel spokesperson told us: “We aren’t going to comment on rumour or speculation.”

MIcron’s XPoint SSD-only restriction

Intel and Micron had, we’re told, a contractual agreement that Intel would sell both Optane SSDs and DIMMs but Micron could only sell SSDs. As we know, with hindsight, Micron never took advantage of this to properly push its own QuantX brand XPoint SSDs into the market.

The industry consultant said “Intel only strategically wanted DIMMs. The market for SSD was and is limited. DIMMs potentially could be a multi-billion dollar business. They had some agreements on where to focus and to allow Intel to drive the DIMM market. This limited Micron’s ability to see any large market. Again, the fast SSD market is very small and takes a lot of work. That’s why Micron gave up.

“The only ‘big win’ for Micron would be when the DIMM market took off, and Micron became second source and no one else could source it. Intels plans were that 50 per cent of the server boards had 3D XPoint. This never got above two per cent … and Micron couldn’t sell any SSDs, and Intel didn’t want any bits from Micron, so they took a huge underload charge on their part of the fab. At that point Micron realised there was no success ever for the market or fab.”

The spokesperson said: “We are not providing additional details about specific business agreements.”

Optane and 3D XPoint’s market success

This background conflict/stand-off between Intel and Micron, together with the need to add DRAM to Optane drives (possibly to overcome a write cooling problem), slowed Optane’s market penetration to the point where, today, its niche in the memory-storage hierarchy has shrunk to a very small size. Within that niche it is still important, but the question is if the niche is big enough to sustain a commercial Optane manufacturing and development operation.

We’d love to know if Intel offered the Optane business to SK hynix, along with its SSD operation, and if SK hynix declined. 

An industry consultant we contacted, who wishes to remain anonymous, said “I have heard rumours that Intel offered it to SK hynix. The issue is that Intel wants to be all-controlling, so they said ‘you take it, we have all control, we have [right of first refusal] on all bits, you can’t license it… etc.’ 

“Hynix refused. Classic Intel: they think they have invented the greatest thing ever and try to control everyone. The end result is a lack of partnership, which kills them.”

In his view, “I have said the best option by far is for Intel to sell Optane to a Chinese company. Tsinghua, or another one who needs memory – let them throw capital at it. You can structure it so that the US government has no say in it. The issue is that Intel can’t control them and it will kill sales since China has political issues now.

“The end result is that the technology is a niche (big niche). Much slower than DRAM. much more expensive than NAND. It is not a game changer. Intel says it’s like when SSDs came out but the fact is that SSDs didn’t become ubiquitous until Samsung, Toshiba, etc sold them to everyone.”

Once more the Intel spokesperson said: “We aren’t going to comment on rumour or speculation.”

In the future we’ll no doubt get to read “spill the beans” stories about XPoint and Optane development. I for one can’t wait.

Ondat backs Trousseau secrets manager for Kubernetes as open source project goes live

Kubernetes stateful app platform supplier Ondat is helping to protect sensitive data in containerised environments with the open source Trousseau product safeguarding the keys needed to access the data.

There is no standard way in the Kubernetes world to protect access to sensitive (secret) data, with the result that many have composed their own ways. With enterprises using Kubernetes to run more and more stateful applications, safeguarding sensitive data is becoming more important.

The project lead for Trousseau is Ondat principal cloud architect Romuald Vandepoel, who said in a statement: “There have been previous projects that attempted to solve this problem, but they required adding lots of components. Naturally, security teams didn’t like that approach because it introduced additional complexity making security more difficult. Secrets management has always been one of the most difficult issues in Kubernetes and Trousseau Vault integration provides the long-sought answer to that problem.”

Ondat diagram

Trousseau uses Kubernetes etcd to store API object definitions and states. The Kubernetes secrets are shipped into the etcd key-value store database using an in-flight envelope encryption scheme with a remote transit key saved in a key management system (KMS).

Secrets protected and encrypted with Trousseau and its native Kubernetes integration can connect with a KMS to secure database credentials, a configuration file or Transport Layer Security (TLS) certificate that contains critical information and is accessible by an application using the standard Kubernetes API primitives.

Any user/workload can leverage the native Kubernetes way to store and access secrets in a safe way by plugging into any KMS provider, like Hashicorp Vault (Community and Enterprise editions), using the Kubernetes KMS provider framework. Users can transition among Kubernetes platforms using the consistent Kubernetes API.

Trousseau is currently being rolled out in a production customer implementation on Suse Rancher Kubernetes Engine 2, leveraging Ondat as the data management platform, along with Hashicorp Vault. 

For more information, read How to keep a secret secret within Kubernetes, and join the Data on Kubernetes Meetup Unravel the Key to Kubernetes Secrets workshop on February 16.

The project is maintained by Trousseau-io.

NetApp beefs up its Astra Control plane for Kubernetes apps

NetApp has added a host of features to Astra Control – its control plane for managing K8S apps supporting more distributions, cloud block stores, adding Operator support and better data protection.

Astra Control is an app-aware control plane that protects, recovers, and moves data-rich Kubernetes workloads in both public clouds and on-premises. It provides Kubernetes workload data protection, disaster recovery and migration using NetApp ONTAP array snapshot, backup, replication, and cloning technology.

There are two versions:

  • Astra Control Service – a fully managed service operated by NetApp;
  • Astra Control Center – a customer-managed software suite with the same rich functionality.  

Update details

Extended cloud storage – Astra Control now supports Azure Disk and Google Persistent Disk cloud block stores in addition to Azure NetApp Files and Cloud Volumes Service in Azure and Google Cloud. You can protect, recover, and move existing and new applications backed by Azure Disks and Google Persistent Disks, as long as they are accessed using the respective CSI drivers with support for snapshots and cloning functionality.

Support for these two block storage providers also enables complete application-data management for your applications that use a combination of file and block storage, like Azure NetApp Files/Cloud Volumes Service and Azure Disks/Google Persistent Disks.

Data management Operators – Users can protect and move apps that are deployed and managed using Operators, in addition to to Helm and Labels. Astra Control automatically discovers Operator-deployed applications, with their custom resources and the associated controllers. You can then do application-aware snapshots, backups, restores, and clones, as you can with any other application. 

Extended K8S support – The Rancher Kubernetes Engine (RKE) and community Kubernetes platforms are supported, in addition to OpenShift Container Platform (OCP). You can protect, recover, and move your K8s apps running on RKE and community Kubernetes with NetApp ONTAP as the storage provider.  

OpenShift – The Astra Control Center Operator is now certified with Red Hat’s OpenShift. It is jointly supported by Red Hat and NetApp, and is monitored and updated automatically to reduce interoperability failure and security risks.

Execution hooks feature – Astra Control provides automatic freeze/thaw logic for popular apps like PostgreSQL, MariaDB, MySQL, and Jenkins, which it discovers and protects automatically. It supports crash-consistent snapshots for all other applications. Execution hooks enable Astra Control to provide custom freeze/thaw logic for in-house developed apps before and after taking snapshots to ensure application consistency. Astra Control can take app-consistent snapshots and backups of any application by using this execution hooks feature.

Restore in place – Astra Control can restore K8s apps in place from an existing snapshot or a backup within the original namespace within the same cluster where the application resides. This means you can recover from service disruption scenarios like accidental or malicious data corruption or deletion, a failed application upgrade, and other similar issues. It allows you to replace your existing app and associated data with a previous instance of the same app and its associated data that you can select from one or more previously recorded application-aware snapshots or backups. 

NetApp says that a preview release of its Astra Data Store (ADS) – a Kubernetes-native, shared file, unified data store for containers and VMs – is now available. ADS was announced in October last year.

This video walks you through the end-to-end procedure to deploy Astra Data Store preview in your Kubernetes cluster.

Free trials are available for both Astra Control Service and Astra Control Center.

Azure offers free inward migration with Data Dynamics and Komprise

Microsoft’s Azure public cloud is providing free inward data migration courtesy of deals with Data Dynamics and Komprise.

An Azure Storage Blog by Karl Rautenstrauch, Microsoft principal program manager for Storage Partners, says these deals with “unstructured data management partners … help you migrate your file data to Azure Storage at no cost!” It’s recommended for use by customers with 50TB or more of data to migrate. Users with less data can use tools such as AzCopy, rsync, or Azure Storage Explorer.

He adds: “We intend this new program to help our customers and partners migrate from on-premises and non-Azure deployments of Windows or Linux File Servers, Network Attached Storage, and S3-compliant object stores to Azure Blob Storage, Azure Files, or Azure NetApp Files.”

This program “is a complement to the Azure Migrate portfolio which many Azure customers have used to automate and orchestrate the migration of servers, desktops, databases, web applications, and more to Azure.”

Customers who take up the program “will be given an onboarding session to learn how to use the software and will receive access to the support knowledgebase and email support for the chosen ISV and up to two support phone calls. We have also co-authored ‘Getting Started Guides’ and our ISVs have created ‘How-To’ videos to help you quickly begin your migration,” writes Rautenstrauch. The program “does not include professional services to help you configure the software beyond the onboarding session, Getting Started Guides and How-To videos.”

Data Dynamics uses its StorageX product while Komprise supplies its Elastic Data Migration (EDM) offering. EDM was launched in March 2020 and takes NFS/SMB/CIFS file data and moves it across a network to a target NAS system, or via S3 to object storage systems or the public cloud.

B&F diagram.

Komprise says EDM eliminates cost and complexity in managing file data by providing analytics-driven data migration to Azure:

  • It provides analytics across existing NAS (e.g. NetApp, Dell, Windows) to identify which data sets to migrate and to which tier of Azure;
  • Systematically migrates files and scales elastically according to the distribution of the shares, directories and files;
  • Ensures data integrity by migrating all file attributes and permissions with full MD5 checksums on every file.

Customers can upgrade to the full product, Komprise Intelligent Data Management, which means they can transparently tier across Azure Storage platforms, cutting up to 70 per cent of cloud costs.

Find out more about the Azure and Komprise migration deal here. Let’s see if Amazon Web Services and the Google Cloud Platform follow in Azure’s footsteps.

Storage news ticker – February 7

Storage
Ticker tape women in Waldorf Astoria

Computational storage startup Eideticom has been assigned two patents: 10,996,892 for “Apparatus and Method for Controlling Data Acceleration” and 11,231,868 for “System and Method for Performing Computational Storage Utilizing a Hardware Accelerator.”

Lee Caswell.

Lee Caswell (Mr. HCI at VMware) has joined Nutanix as SVP for Product Marketing. He resigned from VMware as its VP for Marketing in December 2021 to begin a “new adventure.” He reports to ex-VMware guy Rajiv Ramaswami, Nutanix’ CEO, who left VMware in December 2020.

The latest Toshiba re-organisation plan envisages a two-way split. Japan Semiconductor and Toshiba Electronic Devices and Storage would be spun-off into one entity. A second Toshiba entity would own Toshiba’s 40.6 per cent stake in NAND flash and SSD maker Kioxia, which is destined for an IPO. If the plan gains board approval it will go to a June 23 AGM for shareholder approval.

UK-based replicator WANdisco has adopted a four-day working week with the weekend now consisting of Friday, Saturday and Sunday. According to the Financial Times story, CEO and co-founder David Richards said that staff productivity had become higher during the COVID-19 pandemic, as a result of them working from home. Staff can choose an alternative weekday off and their salaries will not be affected by the move.

Tier-2 public cloud object storage provider Wasabi announced a partnership with Vultr, a cloud infrastructure vendor, to deliver an infrastructure-as-a-service (SaaS) offering claimed to be easier to use and cheaper than hyperscale cloud providers’ services. Customers can use Vultr compute with Wasabi storage to run web apps, big data analytics, media processing and other data-intensive workloads with highly predictable and simplified pricing, meaning no hidden fees. Customers can transfer data at no cost between their Wasabi storage and Vultr virtual machines and bare metal servers.

Should Western Digital acquire Kioxia? Objective Analysis consultant Jim Handy sees downsides in his latest blog. WD, if it acquired Kioxia, would then have to bear the cost of excess NAND chip production from the Kioxia-WD joint venture-owned foundries, and this affects its profitability.

100 per cent of VAST Data‘s customers have a positive perception of VAST. This is shown by a Gartner Peer Insights report and is said to be the first time this has happened to a file and object storage supplier in such a report.

WEKA and Cisco funding contribution: Now you see it, now you don’t

Cheshire Cat

On January 4, WEKA announced Cisco was among contributors to its $73 million funding round. Now Cisco’s name is off the list – for the latest round, at least.

The original release read: “WEKA, the data platform for AI, today announced that Hitachi Ventures led its recent round raising $73 million in funding, which brings the total amount raised to $140 million. Other investors participating in this round were strategic investors, including Hewlett Packard Enterprise, Nvidia, Micron, and Cisco, and financial investors including MoreTech Ventures, Ibex Investors, and Key 1 Capital.”

Spot the difference with a corrected version issued on February 5: “WEKA, the data platform for AI, today announced that Hitachi Ventures led its recent round raising $73 million in funding, which brings the total amount raised to $140 million. Other investors participating in this round were strategic investors, including Hewlett Packard Enterprise, Nvidia, Micron, and Digital Alpha, and financial investors including MoreTech Ventures, Ibex Investors, and Key 1 Capital.”

Enter Digital Alpha as a strategic investor, which now contributes a quote: “As Digital Infrastructure investors, we see Enterprise AI as a highly attractive segment,” says Rick Shrotri, managing partner at Digital Alpha, “and we are delighted to be investors in WEKA’s market-leading AI data platform.” 

What is Digital Alpha?

It describes itself on its website as a premier alternative asset manager focused on digital infrastructure and lists Cisco as a key partner.

Digital Alpha website strategy graphic.

Rick Shrotri is the founder and managing partner at Digital Alpha, and says on LinkedIn that he is the “leader of a private investment fund targeting digital infrastructure worldwide, with proprietary access to deal flow and a knowledge base backed by Cisco Systems, Inc. and other corporate partners.”

He founded Digital Alpha in Feb 2017. Before that he spent almost ten years at Cisco, finishing up as managing director and global head of business acceleration. In that role he was said to have “created and led a global initiative to bring third party equity capital to invest in attractive opportunities from Cisco’s most coveted customers.”

In a WeLink press release Digital Alpha Advisors describes itself as an investment firm focused on digital infrastructure and services required by the digital economy, with a strategic collaboration agreement with Cisco Systems. As part of this agreement, Digital Alpha has preferred access to Cisco’s “pipeline of commercial opportunities requiring equity financing.” This is not mentioned on the revised WEKA release.

In April last year, Digital Alpha announced the closing of Digital Alpha Fund II, over its initial hard cap with over $1 billion in commitments. The fund was oversubscribed.

Digital Alpha is, effectively, a Cisco investment business. That’s why it is classed as a strategic investor, not just as another contributing VC.

Fractal Index Tree

Fractal Index Tree – a data structure used for indexing in databases, which combines elements of traditional tree-based indexing methods like B-trees with concepts from fractal geometry to enhance performance, especially in large-scale, multidimensional databases. Here’s how it generally works:

Key Characteristics:

  1. Fractal Nature:
    • The term “fractal” here refers to the self-similar structure of the tree. Each node in the tree can be seen as a smaller version of the entire tree, allowing for local decisions that mimic global behaviors. This self-similarity helps in scaling the structure efficiently across different levels of data granularity.
  2. Adaptive Indexing:
    • Unlike traditional static trees where the structure is fixed once data is inserted, fractal index trees adapt dynamically. They can reorganize themselves based on data distribution, query patterns, and updates, leading to better performance over time.
  3. Search Efficiency:
    • The tree structure is designed to reduce the average path length to find data, which is particularly useful in handling large datasets or high-dimensional data where traditional methods might perform poorly. The structure might involve:
      • Variable node sizes: Nodes can have different capacities based on the data they hold or the level in the tree they are at.
      • Multiple paths: Unlike a strict B-tree where each node has one parent, fractal trees might allow for multiple paths or connections between levels for quicker access.
  4. Cache Efficiency:
    • Fractal trees are designed with modern hardware in mind, particularly cache hierarchies. They are structured to minimize cache misses by keeping frequently accessed data closer to the root or in more cache-friendly patterns.
  5. Concurrency and Transactions:
    • These trees often include mechanisms to handle concurrent accesses better than traditional trees. They might use techniques like copy-on-write for updates, which allows for more efficient locking strategies or even lock-free operations in some scenarios.

Practical Applications:

  • Databases: Particularly in scenarios where data is frequently updated, or where queries span multiple dimensions or very large datasets. Systems like TokuDB (now part of Percona Server for MySQL) have implemented fractal tree indexing.
  • Big Data: For managing indexes in big data platforms where scalability, speed, and the ability to handle vast amounts of data are crucial.
  • Geospatial Indexing: Where the spatial nature of data can be better managed through fractal-like structures, enhancing query performance across spatial dimensions.

Limitations:

  • Complexity: The adaptive nature and complex structure mean that implementing and maintaining fractal index trees can be more challenging than simpler tree structures.
  • Overhead: There can be additional overhead in terms of memory usage and processing for managing the tree’s dynamic nature.

In essence, a fractal index tree employs principles of fractals to create a more adaptive, efficient, and scalable indexing mechanism for database queries, particularly in environments where data size, dimensionality, and update frequency challenge traditional indexing approaches.

vLLM

vLLM –  virtual Large Language Model (LLM). The vLLM technology was developed at UC Berkeley as “an open source library for fast LLM inference and serving” and is now an open source project. According to Red Hat, it “is an inference server that speeds up the output of generative AI applications by making better use of the GPU memory.”

Red Hat says: “Essentially, vLLM works as a set of instructions that encourage the KV (KeyValue) cache to create shortcuts by continuously ‘batching’ user responses.” The KV cache is a “short-term memory of an LLM [which] shrinks and grows during throughput.”

REST

REST – An appliacyion programming interface (API) the REST API (Representational State Transfer Application Programming Interface) is a web service that follows REST principles to enable communication between clients and servers over the internet. REST APIs allow applications to request and exchange data using standard HTTP methods.

REST API features

  1. Stateless – Each request from a client to the server must contain all necessary information, and the server does not store client state.
  2. Resource-Based – Resources (e.g., users, products, orders) are identified by URLs (Uniform Resource Locators).
  3. Standard HTTP Methods:
  • GET – Retrieve data from the server.
  • POST – Create new resources.
  • PUT – Update existing resources.
  • DELETE – Remove resources.

REST APIs typically exchange data in JSON (JavaScript Object Notation) or XML (Extensible Markup Language), with JSON being the most common. They also have a uniform interface and follow consistent patterns in API design, making it easy for developers to use.