iXsystems is taking simplified service delivery to a new level by enabling a curated set of third-party services to run directly on its TrueNAS arrays. TrueNAS already provided multi-protocol unified storage to include file, block and S3-compatible object storage. Now pre-configured plugins converge additional services onto TrueNAS for simple hybrid cloud enablement.
Persistent Memory is bringing a revolution in performance, cost and capacity that will change server, storage system, data center and software design over the next decade. This article describes some ways storage vendors are integrating persistent memory into enterprise storage systems in 2019.
The SNIA Persistent Memory Summit held in late January 2019 provided a good view into the current state of industry. Some key technologies and standards related to persistent memory are moving forward more slowly than expected. Others are finally transitioning from promise to products. This article summarizes a few key takeaways from the event as they relate to enterprise storage systems.
Companies of all sizes pay more attention to their backup and recovery infrastructure than perhaps ever before. While they still rightfully prioritize their production infrastructure over their backup one, companies seem to recognize and understand that can use backups as more than just insurance policies to recover their production data. This is resulting in cutting edge innovations such as analytics, microservices, and scalable storage finding their way into backup solutions in general and backup appliances specifically.
Dell EMC announced that it will soon add Optane-based storage to its PowerMAX arrays, and that PowerMAX will use Optane as a storage tier, not “just” cache. This statement implies using Optane as a storage tier is superior to using it as a cache. But is it?
The first movie I remember seeing in a theater was 2001: A Space Odyssey. If you saw it, I am guessing that you remember it, too. At the core of the story is HAL, a sophisticated computer that controls everything on a space ship en route to Jupiter. The movie is ultimately a story of artificial intelligence gone awry.
Virtualization largely shaped the enterprise data center landscape for the past ten years. Hyper-converged infrastructure (HCI) is beginning to have the same type of impact, re-shaping the enterprise data center to fully capitalize on the benefits that virtualizing the infrastructure affords them. Enterprises considering HCI as a replacement for existing core data center infrastructure should give special attention to how the solution implements quality of service technology.
The ratification in November 2018 of the NVMe/TCP standard officially opened the doors for NVMe/TCP to begin to find its way into corporate IT environments. Earlier this week I had the opportunity to listen in on a webinar that SNIA hosted which provided an update on NVMe/TCP’s latest developments and its implications for enterprise IT. Here are four key takeaways from that presentation and how these changes will impact corporate data center Ethernet network designs.
In 2019 the level of interest that companies expressed in using artificial Intelligence (AI) and machine learning (ML) exploded. Their interest is justifiable. These technologies gather the almost endless streams of data coming out of the scads of devices that companies deploy everywhere, analyze it, and then turn it into useful information. But time is the secret ingredient that companies must look for as they look to select an effective AI or ML product.
Across more than twenty years as an IT Director, I had many sales people incorrectly tell me that their product was the only one that offered a particular benefit. Did their false claims harm their credibility? Absolutely. Were they trying to deceive me? Possibly. But it is far more likely they lacked accurate and up-to-date information about the current capabilities of competing products in the marketplace. Their competitive intelligence system had failed them.