We are excited to announce that on September 12th, 2014, we will be migrating the CTO blog to a new version of the blogging platform.
@richardmcdougll

Richard McDougall

vSphere Storage, Big Data

Biography

Richard McDougall is the CTO for Storage and Availability at VMware. He is responsible for the technical strategy for core vSphere storage and application storage services, including Big Data, Hadoop.

Prior, as the Chief Performance architect Richard drove the performance strategy and initiatives to enable virtualization of high-end mission critical applications on VMware products.

Prior to joining VMware, Richard was a Distinguished Engineer at Sun Microsystems. During his 14 years at Sun, he was responsible for driving high performance and scalability initiatives for Solaris and key applications on the Sun platform. He served on the central software platform architecture review committee, and also drove the early resource management initiatives for Solaris. Recognized as an operating system and performance expert, he developed several technologies for the Solaris operating system and co-authored several books—including “Solaris Resource Management”, “Solaris Internals” and “Solaris Performance and Tools”.

Richard holds several patents in the area of performance instrumentation, algorithms and distributed file system technologies.

Posts by Richard McDougall

@richardmcdougll

The Dawn of Virtual SAN

March 12, 2014
By
The Dawn of Virtual SAN

Today I’m proud to announce the availability of VMware Virtual SAN 5.5. This milestone represents a disruptive event for the storage industry and a major achievement for VMware’s strategy. We are bringing together the key third block in the trio of virtualized compute, network and storage. With the fusion of these three resources upon industry standard server components, we can now fully realize a true software-defined data center. This fusion represents a significant milestone in simplification. To date, we have thought of these resources, including storage, as a manifestation of the physical resources that we configure and expose to the virtualization platform. Think of this as bottom-up management of these resources. With Virtual SAN, we invert this model. Just as we did for CPU,...

Read more

@richardmcdougll

Storage and Big Data in 2014

December 20, 2013
By
Storage and Big Data in 2014

Last year, I made several predictions about storage and big data – including one about the mad dash for software-defined storage (SDS). Well that’s certainly turned out to be true, with literally dozens of new products hitting the market this year with various attributes of software-defined storage! VMware also made some big news in this space with the launch of VSAN Beta at VMworld. The tricky part is identifying what qualifies for legitimate software-defined storage. We’re seeing both hardware and software-only products being positioned as software-defined. The best way to think about this space is that hardware-based storage can be software-defined if it has a fully software-controllable API that allows simplification of provisioning and management. The other, arguably more important part of software-defined storage...

Read more

@richardmcdougll

We need your input: Big Data Survey

December 9, 2013
By
We need your input: Big Data Survey

  Our Big Data team is running a survey to gather information on the key facets of your Big Data and Hadoop environments. Please grab some time to provide us your inputs. More details are available here …    

Read more

@richardmcdougll

Is Your Cloud Ready for Big Data?

October 29, 2013
By
Is Your Cloud Ready for Big Data?

This week is Big Data Week in NY, if your into big data it’s one of the coolest places to be right now. At the center of stage is O’Reilly Strata Conf and Hadoop World, with an array of other data events happening around the venue — a data sensing lab , a NY big data meet-up, and many others listed here . I’m happy to see that we’re seeing continued increase in momentum of using virtualization as a platform for Hadoop and big data. At VMworld 2013, I shared the stage with Fedex who described that they are seeing significant benefits from their virtual Hadoop platform — they cited simplified systems management, the ability to rapidly deploy new Hadoop...

Read more

@richardmcdougll

VMware’s Strategy for Software-Defined Storage

August 26, 2013
By
VMware’s Strategy for Software-Defined Storage

What if the concepts Abstract, Pool, and Automate are applied to Storage? The foundation of the modern data center is pooled, automated resources that are sufficiently abstracted from physical hardware leading to significant gains in operational efficiency and resource utilization efficiency. This is true for almost every virtualization and cloud platform today. The advantages come from many dimensions, most significantly the ability to radically simplify the development and operational management of applications in the datacenter. By pooling resources, we no-longer need to deal with individual servers or specific hardware configurations. Rather we can provision by policy onto a pooled set of resources, and then eliminate much of the ongoing management through automation. Overloaded servers and hotspots are automatically rebalanced, and resources are re-allocated where...

Read more

@richardmcdougll

Big Data @ VMworld

August 26, 2013
By

I certainly hope you’re as amped about VMworld as I am.  VMware’s signature event kicks off at the Moscone Center in San Francisco and the team here has been working hard to bring some new products your way.  One of those new products is vSphere Big Data Extensions.  For those of you curious about what Big Data Extensions is capable of, and how virtualizing Hadoop will help your organization, be sure to check out all the Big Data sessions this week. On Monday, August 26, come check out my talk with Chris Greer from FedEx on virtualizing Big Data and Cloud Scale Apps. The session begins at 12.30pm. You can also find Jayanth Gummaraju from VMware and Sasha Kipervarg from Identifed give their talk...

Read more

@richardmcdougll

Big Data Extensions for vSphere

June 26, 2013
By
Big Data Extensions for vSphere

This week marks the one year birthday for Project Serengeti, first released at the Hadoop Summit in 2012. Today we’re announcing vSphere Big Data Extensions Beta. I’d also like to take this opportunity to update out thoughts about the big data program and highlights for the support of Big Data in a virtual environment. Our overall mission is to maximize the combination of virtualization and Hadoop as the premier platform for Big Data. Through virtualization, we allow deploying of a variety of big data workloads on a common infrastructure, enabling rapid provisioning and cost savings through shared hardware clusters. We realize these capabilities by optimizing and enabling the vSphere platform for big data, by building a layer of extensions between Hadoop and virtualization and...

Read more

@richardmcdougll

Pivotal, Big Data and VMware

April 24, 2013
By
Pivotal, Big Data and VMware

  It’s great to see the public launch of Pivotal today. The mission — to build a new platform for a new era — is bold but appropriately targeted at some of the biggest fundamental changes in application technologies. Pivotal is now a separate entity, bringing several teams and technologies from both VMware and EMC — including Greenplum’s Hadoop (now Pivotal HD), Greenplum Database (fused with Hadoop as a new database known as HAWQ), CETAS, Pivotal Labs, Gemfire in-memory database, the Spring Application Framework and the Cloud Foundry PaaS platform. The goal of the platform is to enable the new wave of predictive big data applications — those which pull in vast quantities and sources of data — including high rate real...

Read more

@richardmcdougll

Expanding the Virtual Big Data Platform

April 2, 2013
By
Expanding the Virtual Big Data Platform

Today we are releasing a new set of capabilities in  Serengeti  0.8.0, which extends the reach of partner supported Hadoop versions and capabilities. In addition, we are broadening the reach of Serengeti into mixed workload configurations, enabling provisioning of an HBase cluster in this release. As I’ve discussed in previous posts , most big-data environments consist of a mix of workloads. Serengeti’s mission is to enable as many of the big-data family of workloads into the same theme park, all running on a common shared platform. Supporting mixed workloads is a key capability for big-data. In my customer discussions I see a mix of Map-Reduce, HBase, Solr, numerical analysis (R and SAS), and increasingly more of the Big SQL engines such...

Read more

@richardmcdougll

2013 Predictions for Big Data

December 18, 2012
By
2013 Predictions for Big Data

Over the last few years we’ve seen a frenzy of interest and buzz around the area of Big Data. Beyond the hype, there is a solid base of growing use cases, which are becoming center stage to most businesses. 2011 was the year of awareness. There was a great amount of sharing from the early core developers of the analytic platforms – showing the rest of the world the capabilities of the tools and platforms that had been developed for special purpose high scale analytics. The big names at the core of open source analytics development include Facebook, eBay, Linkedin, Twitter – all blazing the trail with new approaches. These companies brought along with them a new and expanding interest in leveraging the same...

Read more