SYS-CON MEDIA Authors: Pat Romanski, Gary Arora, Zakia Bouachraoui, Yeshim Deniz, Liz McMillan

Blog Feed Post

VMware vVOLS – More Than Just Individual LUNs?

During the VMware keynote session today there was a minor discussion on the upcoming concept of VMware vVOLS.  Today, a virtual machine sits on a VMFS created on a storage LUN or on an NFS share.  An individual virtual machine consists of many files and in the case of VMFS-based VMs, is sitting on a piece of storage that is potentially shared with other virtual machines.  For a couple of reasons that’s not a great thing; firstly if the array is used to replicate the VMFS (either locally or remotely) then all the VMs within that VMFS get replicated.  That can be wasteful and overly complex to manage.  Second,  from a storage array perspective, the LUN is the lowest level of granularity in terms of performance and QOS and as the storage array has no way to determine the individual contents of the LUN, it can’t prioritise workload by VM.

vVOLS are the answer to overcoming the shared VMFS issue.  The vVOL (a bit like a Hyper-V VHD) becomes the single container for storing the entire contents of a VM, including all metadata associated with it.  Having a lower level of granularity means that a storage array that is vVOL-aware can replicate just that virtual machine and can give it a specific level of performance.

I have no insight into how VMware and the storage vendors intend to implement vVOLS, but I can see two options.

NFS – On NAS shares, a vVOL could simply be a single file with metadata to identify it as a vVOL.  The storage system simply manages this file (being aware it is a VM), providing all the features of prioritised access, replication and so on.  The internal format of the file would determine the VM contents, presumably with some header contents to store metadata and the remainder consisting of pages of data representing FBA blocks of the logical disk, much as a VHD works today. As the VM grows, the file grows.

Block – On block storage arrays, a vVOL can simply be a LUN.  Today, LUNs can be created thin provisioned on most storage arrays, so a vVOL can be created as a thin provisioned LUN at the maximum size permitted by the underlying storage, sitting within a thin pool.  This allows the vVOL to grow as necessary.  QOS can be applied easily to an individual LUN.  However block-based storage has more issues.  Firstly, there is usually a limit to the number of LUNs that may be created on an array and this could be a limiting factor.  Second, LUNs presented over both iSCSI and Fibre Channel use the SCSI protocol referencing a target and a device (LUN), with a limit on the number of devices on each target. Although vSphere 5 allows 256 targets per HBA there is a limit of 256 LUNs per host, far too low to be practical in terms of using a single LUN for each vVOL.  This restriction, and the inherent problems in doing a discover of 1000′s of LUNs using the SCSI protocol, means that as currently defined, one vVOL per LUN won’t work.  This has to be the main area on which the storage vendors are focusing, namely how to overcome the issues of SCSI, which is embedded in iSCSI, FCoE and Fibre Channel.

Options

NFS seems like a simple option to implement.  Perhaps we’ll see that as a first step.  However, remembering that EMC owns VMware, then block is bound to be treated with equal priority.  To make vVOLs work, the storage vendors will have to either fix the SCSI issue with clever discovery and mapping techniques, or come up with a totally new way of interfacing with objects on the array.  One suggestion was to use object-based storage.  Today those platforms use REST protocols over HTTP, which is both unreliable for high-volume I/O and doesn’t easily allow for sub-object updating.  In any case, this would mean throwing out all of the existing IP and investment in current technology, which is not going to happen.

The Architect’s View

vVOLs make complete sense in order to scale virtual machine growth.  However today’s storage protocols cause significant issues in achieving vVOL granularity.  Storage vendors won’t throw out their existing architecture, but will most likely modify their hardware implementations in some way.  Yet again, NFS could serendipitously overtake block as the preferred vVOL platform.

Comments are always welcome; please indicate if you work for a vendor as it’s only fair.  If you have any related links of interest, please feel free to add them as a comment for consideration.

Read the original blog entry...

Latest Stories
While a hybrid cloud can ease that transition, designing and deploy that hybrid cloud still offers challenges for organizations concerned about lack of available cloud skillsets within their organization. Managed service providers offer a unique opportunity to fill those gaps and get organizations of all sizes on a hybrid cloud that meets their comfort level, while delivering enhanced benefits for cost, efficiency, agility, mobility, and elasticity.
Isomorphic Software is the global leader in high-end, web-based business applications. We develop, market, and support the SmartClient & Smart GWT HTML5/Ajax platform, combining the productivity and performance of traditional desktop software with the simplicity and reach of the open web. With staff in 10 timezones, Isomorphic provides a global network of services related to our technology, with offerings ranging from turnkey application development to SLA-backed enterprise support. Leadin...
DevOps has long focused on reinventing the SDLC (e.g. with CI/CD, ARA, pipeline automation etc.), while reinvention of IT Ops has lagged. However, new approaches like Site Reliability Engineering, Observability, Containerization, Operations Analytics, and ML/AI are driving a resurgence of IT Ops. In this session our expert panel will focus on how these new ideas are [putting the Ops back in DevOps orbringing modern IT Ops to DevOps].
Darktrace is the world's leading AI company for cyber security. Created by mathematicians from the University of Cambridge, Darktrace's Enterprise Immune System is the first non-consumer application of machine learning to work at scale, across all network types, from physical, virtualized, and cloud, through to IoT and industrial control systems. Installed as a self-configuring cyber defense platform, Darktrace continuously learns what is ‘normal' for all devices and users, updating its understa...
Enterprises are striving to become digital businesses for differentiated innovation and customer-centricity. Traditionally, they focused on digitizing processes and paper workflow. To be a disruptor and compete against new players, they need to gain insight into business data and innovate at scale. Cloud and cognitive technologies can help them leverage hidden data in SAP/ERP systems to fuel their businesses to accelerate digital transformation success.
Concerns about security, downtime and latency, budgets, and general unfamiliarity with cloud technologies continue to create hesitation for many organizations that truly need to be developing a cloud strategy. Hybrid cloud solutions are helping to elevate those concerns by enabling the combination or orchestration of two or more platforms, including on-premise infrastructure, private clouds and/or third-party, public cloud services. This gives organizations more comfort to begin their digital tr...
Most organizations are awash today in data and IT systems, yet they're still struggling mightily to use these invaluable assets to meet the rising demand for new digital solutions and customer experiences that drive innovation and growth. What's lacking are potent and effective ways to rapidly combine together on-premises IT and the numerous commercial clouds that the average organization has in place today into effective new business solutions.
Keeping an application running at scale can be a daunting task. When do you need to add more capacity? Larger databases? Additional servers? These questions get harder as the complexity of your application grows. Microservice based architectures and cloud-based dynamic infrastructures are technologies that help you keep your application running with high availability, even during times of extreme scaling. But real cloud success, at scale, requires much more than a basic lift-and-shift migrati...
David Friend is the co-founder and CEO of Wasabi, the hot cloud storage company that delivers fast, low-cost, and reliable cloud storage. Prior to Wasabi, David co-founded Carbonite, one of the world's leading cloud backup companies. A successful tech entrepreneur for more than 30 years, David got his start at ARP Instruments, a manufacturer of synthesizers for rock bands, where he worked with leading musicians of the day like Stevie Wonder, Pete Townsend of The Who, and Led Zeppelin. David has ...
Darktrace is the world's leading AI company for cyber security. Created by mathematicians from the University of Cambridge, Darktrace's Enterprise Immune System is the first non-consumer application of machine learning to work at scale, across all network types, from physical, virtualized, and cloud, through to IoT and industrial control systems. Installed as a self-configuring cyber defense platform, Darktrace continuously learns what is ‘normal' for all devices and users, updating its understa...
Dion Hinchcliffe is an internationally recognized digital expert, bestselling book author, frequent keynote speaker, analyst, futurist, and transformation expert based in Washington, DC. He is currently Chief Strategy Officer at the industry-leading digital strategy and online community solutions firm, 7Summits.
Addteq is a leader in providing business solutions to Enterprise clients. Addteq has been in the business for more than 10 years. Through the use of DevOps automation, Addteq strives on creating innovative solutions to solve business processes. Clients depend on Addteq to modernize the software delivery process by providing Atlassian solutions, create custom add-ons, conduct training, offer hosting, perform DevOps services, and provide overall support services.
Contino is a global technical consultancy that helps highly-regulated enterprises transform faster, modernizing their way of working through DevOps and cloud computing. They focus on building capability and assisting our clients to in-source strategic technology capability so they get to market quickly and build their own innovation engine.
When applications are hosted on servers, they produce immense quantities of logging data. Quality engineers should verify that apps are producing log data that is existent, correct, consumable, and complete. Otherwise, apps in production are not easily monitored, have issues that are difficult to detect, and cannot be corrected quickly. Tom Chavez presents the four steps that quality engineers should include in every test plan for apps that produce log output or other machine data. Learn the ste...
Digital Transformation is much more than a buzzword. The radical shift to digital mechanisms for almost every process is evident across all industries and verticals. This is often especially true in financial services, where the legacy environment is many times unable to keep up with the rapidly shifting demands of the consumer. The constant pressure to provide complete, omnichannel delivery of customer-facing solutions to meet both regulatory and customer demands is putting enormous pressure on...