The EMC CTO Ambassadors program was launched in April of this year and I introduced the program here. The main goal of the program is to create a global volunteer team to message EMC strategy to employees, customers, partners, and the industry. The program has been widely successful. Today we have over 30 CTO Ambassadors operating around the globe. We will have delivered over 100 customer meetings by the end of this year. After six months of operation I have learned of the need and power that a formal message feedback loop provides.
One of the unique components of our evangelist program is we capture and analyze feedback captured during each of our presentation engagements. Each CTO Ambassador presenter is accompanioned by a scribe that captures the audience comments, feedback, and action items. This feedback is entered into our database. On a quarterly basis this information is analyzed at the global level and by local geography. We look for the trends. This is used for input to tune our messages, and identify new message opportunities. In addition, this feedback is review by EMC senior management and product managers.
EMC employee's can review the previous quarters analysis on our team collaboration site here. The CTO Ambassador is supported by our Office of the CTO VisionX team. I recently had the opportunity to interview EMC's VisionX leader, Steve Todd.
Launching the CTO ambassador program has been a great learning experience that is benefiting EMC. Ultimately the industry benefits when EMC is communicating our strategy. Whether you agree with the strategy or not the dialogue ultimately benefits the industry. We are going to continue to grow the program. I am already partnering with John DeWolfe to support the EMC Field Briefing program and greater engagement through social media. If you are an EMC employee, and are interested in joining the EMC CTO Ambassador program, more information available on our group collaboration site here.
The VMware announcement of their converged infrastructure offering EVO(lution) at VMWorld2014 generated a lot of discussion. Converged Infrastructure solutions have been available for several years. EMC has two successful converged infrastructure solution offerings:
Many customers I work with are deploying or are planning to deploy converged infrastructure solutions. Many analysts predict this market will grow at greater than 50%. Gartner recently released their first Converged Infrastructure Magic Quadrant report.
I talk with many EMC customers and partners. There are several common sentiments that I have noticed. The first is there are four primary use cases customers are targeting for Converged Infrastructure solutions:
Most customers believe industry communized hardware is sufficient, except for general-purpose workloads supporting applications with high availability requirements. For high availability client-server applications customers continue to demand best of breed hardware components. These customers are supporting classic client-server applications where infrastructure outages = application outage = resume generating event. EMC has been very successful with our VCE, and VSPEX Converged Infrastructure offerings for this use case. For applications like virtual desktop, and next generation mobile, and big data analytics applications that can easily and quickly restart workloads the Common Modular Building Block and Application Specific Building Block CI architectures are preferred. Clearly there is no single Converged Infrastructure architecture to meet all application needs.
The second requirement most customers have is Converged Infrastructure has to simple to deploy, with short time to value. Converged Infrastructure must be delivered ready for production. At EMCWorld this year, we deployed our Enterprise Hybrid Cloud solution on VCE Vblock in 24 hours. VMware EVO: Rail is promising it will be able to serve virtual machines 15 minutes after power on. Automated workflows simplifying hardware discovery, configuration, and integration into existing customer networks is expected.
The third requirement is built in operations automation that is extensible via simple API. This automation is really at three layers:
Hardware automation is needed to accelerate deployment and mask variability. For example, as vendors change CPU, and disk drives models the installation, provisioning, and monitoring process should not change. There are a couple different ways CI vendors are approaching this requirement. VCE provides the Vision software which provides an API enabled abstraction layer to the underlying Vblock hardware. It is a passive solution that allows your preferred workflow engine to integrate to drive installation, provisioning, and monitoring. Changes to hardware components and hardware software versions are abstracted from your favorite workflow software by VCE Vision. The other approach is to tightly control the underlying hardware components and build a custom installer for the specific hardware. This is an area where I think CI vendors will be able to differentiate with some offering great hardware component flexibility.
Infrastructure provisioning is most often provided by existing commercial products like VMware vRealize Suite or Microsoft Systems Center. Not all customers need the full functionality of these tools so many CI vendors are providing simplified workflow solutions. VMware EVO: Rail will come with the simplified workflow system for provisioning, and monitoring VM's. This workflow tool updates vCenter via its API so there is one system of record and you can still use the full vCenter functionality for more complex tasks. A preview of this solution is available here.
Application provisioning is typically handled by existing commercially available for tools. Application-Specific CI solutions will provide a tool optimized for the application. As customers start to leverage CI solutions to support Platform as a Service (PAAS), solutions like Cloud Foundry application provisioning and updates will be optimized by the PAAS solution. Integration between PAAS solutions and Infrastructure provisioning systems is beginning as evidenced by the Kubernetes, Diego, and Fargo projects. This will make it easier for applications to be deployed on a wide variety of CI, or hybrid cloud systems.
It is clear that the time for Converged Infrastructure systems have arrived especially for the general purpose, remote office, and standard application services. Converged Infrastructure provides a simpler and faster opportunity to provide IT infrastructure. I believe the biggest evolution in this space will be improvements to the automation with new software. Today there are four main Converged Infrastructure architectures and customers need to understand the differences and carefully plan their architecture that will likely leverage a combination of these systems. Everyone I speak with is trying to simplify and minimize the time to value with their IT infrastructure investments. The days of spending six months or longer standing up new IT infrastructure is over.
We hosted our second annual CTO Roundtable at VMworld2014. The EMC Office of the CTO hosts these type of events to interact directly with our customers, and partners. Meeting with customers and partners is required for our team.
Our CTO Roundtable @VMWorld2014 event was our largest ever. Fourteen customers and partners from the financial, healthcare, manufacturing, and IT service provider industries met with EMC Global CTO, John Roese during lunch on the third day of VMworld2014. We wanted to meet after most of the announcements planned for the show were public. We kicked off the event with remarks by John describing his two years at EMC. He reviewed the three businesses: Pivotal, VMware, and EMCII that constitute EMC today. Over the past two years we have re-organized EMC into a federation of three companies with new leadership for each. John believes customers want choice in their IT stack; one size does not fit all. With the federation structure each of the EMC's businesses is motivated to create the best products for their target market while also collaborating with the other EMC product teams. John pointed to the growth of the Cloud Foundry Foundation that governs the direction of the industry leading Platform as a Service solution. If Cloud Foundry's governance continued exclusive to Pivotal it would have never achieved the broad industry support some of which are VMware, and EMCII competitors. In order for Pivotal to grow and our customers to benefit from wide industry support the Cloud Foundry Foundation needed to be created and leverage support from competitors to other parts of the EMC federation companies. John did acknowledge that EMC needed to do a much better job communicating the federation strategy to the industry. EMC will be much more pro-active in engaging the industry about our federation strategy. Shortly after VMworld2014, EMC CEO, Joe Tucci gave an extensive interview to ComputerWorld describing the federation and EMC strategy here. John told our audience they should expect us to release a series of product bundles that leverage products from across the EMC federation companies as turnkey solutions. Many of these solutions were featured prominently at the VMworld Solution Exchange as EMC Federation solutions.
During the roundtable discussion our customers and partners discussed the challenges next generation Mobile and Big Data applications were having on their current IT infrastructure. In the past, well run IT organization implemented Application Portfolio Management often with governance models that drove application consolidation to reduce costs. Today many IT organizations are seeing application portfolio growing rapidly. If IT tells the business they can't deploy a new app, the business will turn to an external service provider and deploy it rapidly for little initial investment. This is resulting in IT having to maintain more applications and with more diverse IT infrastructure optimized for different application types. This is driving a greater focus by IT to introduce operations automation and abstraction technologies like Docker, and Cloud Foundry. These abstraction technologies allow developers to build applications independent of the underlying physical infrastructure. This is makes applications and data more portable while allowing IT to automate deployment, management, and monitoring leveraging API's and built in automation functions of these tools. John cautioned the audience to be diligent inspecting these technologies to keep proprietary hooks from being added. Long term this will hurt the industry and customers.
The group was eager to discuss the potential impact to infrastructure availability with the increased use of commodity hardware. EMC has provided industry leading quality and service over the years enabling IT to deliver ever increasing availability to their customers. The audience asked John how EMC will enable availability and service as they transition to purchasing software from EMC that will be run on commodity hardware. Many analogies were made to the transition the industry experienced with virtualizing compute using VMware. Better protection and availability features are added to the software like VMware High Availability, Fault Tolerance, vMotion, and Distributed Resource Scheduling to support larger fault domains and less reliable hardware. In the long term the IT infrastructure became more reliable.
The last topic the group discussed was the struggles many are starting to experience with traditional data backup paradigms. The volume of digital data they are managing is exploding. Many were starting to experiment with snapshots and remote replica's in place of traditional backups. John discussed need for consolidation of IT meta data tracking all copies of primary data. Several months ago EMC consolidated all its data protection products including backup tools (Networker, Avamar, and Data Domain) with our replication tools (RecoverPoint, and VPLEX) into a single business unit. We need to consolidate our data protection metadata to provide a single view of all copies of data. EMC is starting to make substantial progress with our Data Protection Advisor which presents a consolidated view of your backup and replica that can be used for restore. Also EMC ProtectPoint is an important new technology. We added intelligence to our VMAX to automatically copy primary data to protection storage without moving the data through the compute and application layer. This architecture greatly improves backup performance, and provides a consolidated meta data repository of all copies of your primary data.
The feedback from the participants was outstanding. Customers appreciated the opportunity to interact with EMC Global CTO, John Roese in a small setting as well their industry peers. This is also a great opportunity for EMC to learn and validate our direction with our customers and partners. We run several of these events regularly throughout the year. Feedback from these events is captured and analyzed. The topics our customers wanted to discuss including EMC Federation Strategy, and impacts of next generation applications is consistent with the trends we've seen throughout this year. The need for new data protection paradigms has been a consistent theme most recently. The impact to IT availability with the reliance on commodity hardware was new and likely created with the focus on the software defined data center theme at VMworld2014. The feedback from these events is one of the ways that EMC validates our strategy, and product plans. These types of events make us more effective and a better partner to our customers.
VMWorld2014 just wrapped up. As in the past it will be remembered as one of the premier IT conference of the year. This year there were over 22,000 attendees and every IT vendor you can think of was represented in the Solutions Exchange. There were plenty of opportunities to catch-up and expand your professional network and get hands on experience with labs.
Once again VMWorld started with the 5K fun run and the biggest vOdgeball tournament to date. Nothing is better than catching up with industry colleagues and then getting to hit them with a ball. Most importantly we raise $20K for the Wounded Warrior Project and remembered our friend Jim Ruddy who passed away recently in a tragic car accident.
This is the first VMWorld that didn’t include a major release of the vSphere hypervisor. I think this is a great example of VMware’s maturity and the industry transition from virtualized data center to software defined data center architectures. This year the focus was on three topics: Software Defined Data Center, Hybrid Cloud, and End User computing.
On the first day, VMware CEO Pat Gelsinger challenged the industry to aggressively transitioned from silo’d traditional infrastructure to automated, flexible cloud architectures. He asserted that the technology was ready and VMware has your back, providing the products, and solutions you need. Today your IT infrastructure must be able to support the AND:
* client-server AND Mobile, Big Data, Web applications
* IT secured with operational control AND developer need for On-demand, elastics IT capacity
* On premise AND Off premise IT
VMware made several new products, and updates to most other existing product portfolio. In this post I will summarize the announcements and in future posts will dive into the details. The EVO hyper-converged product announcement generated the most attention on the first day. Two products were introduced, EVO: Rail and EVO: Rack. The EVO products will provide customers with simple, and fast cloud deployment solutions. The EVO will be sold and supported by VMware hardware partners using many of the familiar VMware virtualization, management, and orchestration software with a new EVO Manager workflow automation solution that will simplify initial setup and VM deployment. More information on VMware EVO: Rail program can be found here.
In keeping with enabling IT to support the AND, VMware announced greater support for OpenStack. VMware announced VMware Integrated OpenStack (VIO). VMware claims VIO will simplify the deployment and operation of Apache compliant OpenStack while giving developers the open, API to control their VMware infrastructure. I thought this was an interesting way for VMware to change the conversation from OpenStack vs. VMware to embracing the OpenStack community while validating their membership in the Openstack Foundatation.
The third major announcement was a joint strategy with Docker to optimize container support on VMware. The major benefit for customers is developers can build their applications in Docker containers and IT can deploy them in their existing virtual infrastructure leveraging existing management, compliance, and security processes. VMware is adding enhancements to vCloud Suite to make it at least as easy to deploy containers in a virtual server infrastructure as bare metal with the promise of a more secure, and available infrastructure. VMware’s Project Fargo significantly reduces child VMs startup time and reduces VM storage and memory footprint. This is one of the more promising announcements especially for customers developing next generation mobile, big data, and web applications.
VMware renamed the year old public cloud service, vCloud Hybrid Service to vCloud Air. During the first day keynote Bill Fathers updated everyone on the impressive build out of vCloud Air data center locations. In addition to the impressive growth, I think there were two particularly interesting new offerings announced:
The On Demand Service is available in beta and will allow customers to use a credit card to pay based on per minute metering. Previously the vCloud service required a PO and a minimum service commitment. This will be attractive to individual developers. VMware announced the availability of the service in beta here.
VMware is entering the storage as a service market. The vCloud Air Object Storage will leverage EMC ViPR and include lifecycle management, versioning and large object features that will simplify and reduce management overhead and for highly available hybrid application deployments.
On the second day VMware reviewed their busy year of End User Computing growth. They highlighted the new Horizon Workspace Suite release, along with the successful integration of the AirWatch, and Desktone acquisitions. VMware also demonstrated how they plan to leverage the latest EUC acquisition CloudVolumes to provide real time delivery of applications, user settings, and data to any virtual desktop. Most interesting to me was the announcement of a collaborative effort with NVIDIA and Google to deliver high performance virtual desktops with support for graphics intense applications. This solution will leverage VMware Blast Performance and NVIDIA Grid vGPU technology. This could be especially interesting for the healthcare industry with their need for high performance graphics support for diagnostic services.
VMWorld2014 was particularly impressive because there was no major vSphere release to anchor the show. Instead VMware showed its growing maturity as a mission critical enterprise software company that not only provides the industry’s best hypervisor but is now providing the cloud automation, management software to enable their customer to deploy software defined data center architectures. VMware showed it intends to provide critical services to next generation infrastructure and application developers with the VMware Integrated OpenStack offering and collaboration with Docker to make SDDC infrastructures an optimal platform for containers. In addition to optimizing their SDDC functionality, VMware has continued to expand the number of their public cloud locations while adding the On Demand service offering. VMware combined all this new capability with the integrated and improvements of the Horizon Workspace Suite there may not have been time to to launch a new vSphere release.
All the official EMC activities for VMworld are summarized on the Everything VMware at EMC community site. This year EMC will be hosting our hands on labs in our show floor booth. A complete list of the vLabs is available at: https://community.emc.com/events/2827
We are again hosting the CTO Roundtable session. We have a dozen thought leading CTO’s meeting for lunch to discuss topics the industry is struggling to address. Again this year EMC Global CTO, John Roese will be hosting the event and me and EMC VP, of Innovation, Steve Todd will be facilitating the event. This year we are pleased to have VMware CTO, Ben Fathi will be participating in the session as well. This was a great event last year and based on the confirmed attendee’s it should be even better.
This year in addition to great show content I am looking forward to the #vBrownbag talks scheduled for Sunday. This event has been scheduled as an un-conference with short presentations from many of the VMware community experts. I am looking forward to these sessions in particular:
The complete speaker schedule is available here. It is nice that I can attend these sessions before the years #v0dgeball tournament. This year I am part of Chad Sakac’s EMC team. This is always a great opportunity to catchup with VMware community friends while throwing dodge balls at them! We are again raising money for the Wounded Warriors Project. This year we will also remember our good friend Jim Ruddy. This year the vSpecialist team has chosen to honor Jim by naming their team “The RuddyDucks” and created these cool shirts that are available for purchase at https://www.booster.com/jamesruddy. All the proceeds will be donated to Jim’s family.
This year there seems to be a general renewed focus on content at VMworld. I’ve noticed that there are a number of new content centered events like the #vBrownbag talks and product briefing sessions in addition to all of the great conference sessions. I plan on posting a number times during the show next week. I look forward to seeing everyone at the @vodgeball tournament and the vUnderground party this year.
In 2010 when I joined the EMC vSpecialist team I was paired with Jim Ruddy. Jim was based in Connecticut and I was in upstate NY. Jim and I came into our roles with different professional experiences. Jim had a ton of implementation experience and I had been responsible for running data centers and large IT infrastructure teams. I had already been at EMC nine years. We complimented each other well and our manager, Bernie Baker mentored us to work together and make the role what was needed for our territory. Jim and I spoke or worked side-by-side everyday for four years working with customers, partner, and our EMC sales teams to drive the adaption of server virtualization and cloud.
2013 NE vSpecialist Team
Early in 2013, I was asked to build an engineering team for our new Open Innovation Lab service working for Bala Ganeshan. Jim was the first person I recruited. Like Bernie, Bala challenged us to make this service what was needed. We were back together and working on EMC’s first Hadoop solutions.
2013 NYSE - Launching EMC Hadoop Starter Kit 2.0
Jim always had a smile, and never had a bad word about anyone. He loved to tell you about his latest skiing trip to the hill near his house with his sons, Forest and Dylan. Most recently he would update everyone on his baby Liam and new wife Stefanie. Jim had just bought a new house and we were talking about the type of lawn mower he needed. I still have the picture he sent me of the new riding mower as it was delivered. The deciding factor on the mower was the 360-degree turning capability. He was as interested in the mower as any new technology. That was Jim, always curious, always interested in learning about something new. He loved to share his knowledge directly or through his blog.
Late last Thursday evening I received a phone call no one ever wants to get. There was a terrible car accident and Jim didn’t survive. I can still remember literally dropping the phone.
The next two days we will be saying goodbye to Jim. It has been a tough week. I miss the daily chats about a new technology, updates on his family, or a new feature he discovered with his mower.
While we will be saying goodbye this week to Jim we will not forget him. After the accident a memorial fund was setup and the response has been tremendous. Reading the comments from people that had worked with Jim, competed with Jim for business, or didn’t know him but appreciated his blog has been touching. I’d recommend reading it here for how anyone would like to be remembered. It is good to know so many others enjoyed Jim as I had.
VMworld 2013 - Launching EMC Hadoop Starter Kit
Jim was not perfect and he would be the first to tell you. I will choose to remember him by trying to smile more, appreciate my family more, and to become even more committed to learning new things. This week his family and many, many friends will say goodbye, but we will not forget Jim Ruddy. #RememberRuddy
In my previous post I described the storage requirements for your enterprise content repository or Data Lake:
In my experience data maintained in your Data Lake will have these six storage criteria:
The data criteria will likely change based on the analysis being performed at any given time and not necessarily the age of the data so your storage architecture will most likely consist of a federation of arrays, storage media, and access protocol requirements.
Today I commonly see four common storage components used to support most Data Lake’s solutions.
These components can be federated to leverage your existing storage investments, and meet your data criteria.
Many customers start with deploying the Hadoop File System software that comes with their Hadoop distribution on direct attach server storage hardware or enterprise storage arrays. This has many advantages including:
I have seen many customers hit the wall with this solution at about 30TB of data. The HDFS standard requires three full copies of the data for redundancy so the raw infrastructure capacity is around 100TB. The operational overhead of replacing failed hardware, backup, and the environmental requirements for this much infrastructure is often not sustainable. Most customers begin to look at enterprise infrastructure solutions as the next logical step.
Many customers have large amounts of existing file data they want to be able to analyze. If this data is stored on an enterprise storage array like EMC Isilon the customer can enable HDFS access to this data for analytics. More information on this capability can be found here.
This option has many advantages including:
In addition, enterprise storage arrays provide all the data efficiency, and data security, and operational benefits for Big Data they have been providing for other applications for years including:
As I discussed previously we have build a deployment guide to enable customers to reliably, and rapidly deploy this solution. The guides are available here.
While utilizing HDFS natively on a storage array has many benefits, many customers want the flexibility to use any enterprise storage array. EMC ViPR is one example of a new storage virtualization technology that can be used to provide HDFS data services support for almost any major enterprise storage array or direct attach server storage. This has many advantages for Big Data analytics users:
In addition, a storage virtualization software layer still allows the enterprise storage array to provide all it’s traditional benefits including:
I believe storage virtualization is the most interesting Data Lake storage opportunity. It combines the benefits of enterprise storage with the flexibility of hardware choice. As I discussed previously we have build a deployment guide to enable customers to reliably, and rapidly deploy this solution. The guides are available here.
Appliances solution can be limiting for some customers. Appliance solutions by definition are highly predictable, fully integrated stacks, which tend to lag in the introduction of new features and functionality. The Big Data analytics space is evolving quickly with new HDFS features and functions being introduced every few months. Appliance upgrades can delay access to these new functions so you will want to consider how quickly you want access to the latest and greatest HDFS features.
In conclusion, each of these storage options has strengths and weaknesses. Most customers use a combination of these components in addition to an in memory data grid service to build a fully functional Data Lake solution. Each of these storage options can be federated together from the beginning or as needed to meet your needs.
As next generation analytics is driving new storage service requirements. The main data enablers of next generation analytics discussed in my previous post (here):
These requirements required new storage architectures that create enterprise content repositories, or data lakes. An enterprise data lake has three critical storage requirements. The first requirement is data lake storage architecture’s must scale easily and without service interruption. I see many customers’ content repositories starting in the hundreds of gigabytes and growing to multiple petabytes quickly. The ability to add capacity in modular chuncks quickly and transparent to applications is critical.
The second requirement is your dat lake storage must be optimized for low cost per giga byte. Several data efficiency services such as data de-duplication, and compression are critical. In addition, leveraging large capacity storage media will provide a low cost per gigabyte. Data Lake storage costs are typically targeted at about $.01 per gigabyte for primary storage.
The third critical storage requirement is your data needs to accessible
via multiple storage protocols. For example, if you have applications generating and storing data via NFS, that same data will need to be accessible by analytics applications using next generation analytics tools like Hadoop. Your data lake should be able to accept and serve data to your entire application portfolio via their preferred storage protocol: SQL, JSON, file, block, and object, with the ability to integrate whatever is next.
These requirements provide support for the data volumes, and variety that next generation analytics require. In order to be able to support analytics for velocity data your data lake architecture will need to include an in memory data grid (IMDG) service. In memory data grid technologies pool the NVRAM of a server cluster to provide a high speed storage tier that is able to handle rapid data ingestion and near real time analysis of data. Data generated from high transaction financial systems or sensors will quickly saturate traditional storage architecture. IMDG will provide a high IO caching layer and will copy data from IMDG on periodic basis to and from the lower cost storage areas of your data lake for persistence and deep analytics. Your data lake storage architecture will look like this:
Data Lake storage architectures most often require a combination of storage platforms with file, and database system software like Hadoop File System (HDFS), and in memory data grids (IMDG) to meet the diverse requirements of next generation analytics application. As you design your data lake you will need to use a modular approach to allow you to easily adjust the mix of high speed, expensive capacity (IMDG) , and low cost, highly scalable capacity based on workload demand. Application access to data in different areas of your data lake should be transparent to all your applications using this type of architecture.
In my next post I will discuss the storage technologies that should be used to support an enterprise content repository or Data Lake storage architecture.
A big part of my role at EMC these days is engaging with our customers and partners to discuss the industry trends and EMC strategy. Over the past couple months just about every customer wants to talk about their Big Data project. In addition to my anecdotal market analysis, IDC predicts the IT market for Big Data Solutions will grow from $10.2 billion in 2013 to over $53 billion by 2017. That is a 5x growth. No question that Big Data is one of the hottest IT opportunities today.
Before thinking about solutions, I think it is important to understand the enablers of next generation analytics or Big Data to use the popular term. Analytics has been an important IT capability for many years. Who doesn’t have one or more data warehouses, and data marts to support manufacturing, finance, sales, and marketing today? So what is new and enabling the next generation Big Data analytics opportunity?
I believe there are four enablers that have combined to reach maturity to create the perfect storm. The first enabler is the return on investment for storing and maintaining complete data sets for analytics is positive. The continued reduction of storage hardware costs, and improved data efficiency capabilities like de-duplication, and compression have become pervasive. In the most recent EMC Digital Universe research, IDC predicts the amount of digital data generated in each of the next two years will double, and will continue to double every two years for the rest of the decade.
The second enabler is the rapid growth of data from network-connected devices such as sensors. The EMC Digital Universe study predicts that 10% of all digital data will be generated by network connected devices by the end of the decade. Today, network-connected devices generate about 2% of all digital data. Smart companies like GE, are leveraging this data to provide differentiated services. A GE Wind Turbine contains about 20,000 sensors that generate 400 data points per second. The sensor data is analyzed in near real time to maximize the efficiency. This data is also stored and used for deeper analytics to improve maintenance and parts replacement. A GE Wind Turbine is more efficient and has higher availability enabled by next generation analytics.
The third enabler is the increase in unstructured data types such as images, video, and audio. The EMC Digital Universe study predicts 30% of all digital data by the end of the decade will be security related with the majority being images, video, and audio. Traditional analytics have not leverage these data types.
The forth enabler is a new set of analytics tools designed specifically to analyze large amounts of data, both structured (i.e. log data) and unstructured (images). Tools such as Hadoop, and Splunk were designed to analyze large data sets. These new analytics tools have been created through the Open Source community and have a low initial cost to get at least get started. ext generation analytics has become affordable for big companies but smaller companies are also using these tools to find new revenue, and provide differenaited services.
These are not the only four enablers of next generation analytics but I think they are the most impactful. I only see next generation analytics gaining more momentum in the coming months. I will share my thoughts on the IT requirements for Big Data services over the next few days.
EMC World 2014 (5/5 - 5/8) is quickly approaching. Once again I will be attending. The agenda is familiar to those that have attended previously. There will be a number of keynote presentations from EMC company leaders including Joe Tucci, David Gulden, Pat Gelsinger, and Paul Maritz. This year I am especially looking forward to the Isilon business unit keynote presentation on Wednesday (5/10). I hear the theme is “Grown Up Big Data” with a couple special partner guests.
The EMC World schedules and on demand video’s of all the keynote presentations will be available on the EMC Community Network here.
In addition to the keynotes there will be awesome hands on vLabs again and this year they have added guided hands on vLabs. The complete schedule is available here.
This year I have an active roll this year presenting EMC's HDFS Strategy breakout session with industry Hadoop thought leader, Jim Ruddy. We will be joined by our Hadoop mentor, Pivotal’s Dan Baskette. We have a great presentation planned packed with examples and a demo of the EMC-VMware-Pivotal Big Data stack in action from a recent customer pilot. The session will be presented on Monday (5/8) morning 8:30 in Delfino 4001A. We are the opening act for first EMC World keynote by Joe Tucci. The second presentation will be on Wednesday (5/10) at 4pm in Marcello 4401A. That’s right, we are the opening act for EMC’s Customer Appreciation event with Imagine Dragons. That's my story and I'm sticking to it.
In addition, to the breakout sessions I am working with Mona Patel, and Jim Ruddy at the EMC Big Data kiosk in the solutions arcade. When I’m not at the Big Data Kiosk or attending sessions I will be in the Social Village. There are two meet ups I am looking forward to the most:
I am already excited to leave and get the event started. For me I am particularly excited to see the amount of Big Data content this year. Big Data has arrived. EMC World is always a great opportunity to mix learning about our new solutions, and visit with friends. I hope to see you there.