|By Ignacio M. Llorente||
|March 6, 2014 03:21 PM EST||
In our last post "OpenNebula vs. OpenStack: User Needs vs. Vendor Driven" we stated that"OpenStack penetration in the market is relatively small compared with the investment made by vendors and VCs". We have received several emails from people asking for the numbers that support this statement. This conclusion arises from the comparison between OpenNebula and OpenStack user base, a well as between the resources invested in development and marketing by each of them.
OpenStack is experiencing explosive growth in the number of developers, with more than 200 companies contributing code, 15,000 people and 850 companies involved according to its web site, and almost 1,000 developers involved in its latest release. However, the number of users and the size of the deployments are not that impressive, at least compared with this software development force.
Let us compare the user base of OpenNebula and OpenStack by using their latest surveys:
- According to the most recent OpenStack user survey (November 2013), they received 827 responses, and 387 were deployments. In the 80% of these deployments the number of nodes was below 100, and only 11 deployments with more than 1,000 nodes (hypervisors).
- On the other hand, in the latest OpenNebula survey (November 2012), OpenNebula received 2,500 responses, 820 of these were deployments. In the 70% of these deployments the number of nodes was below 100 nodes, and 99 deployments have more than 500 nodes (hypervisors).
We avoid giving references to featured users, both projects could put on the table good references of large-scale cloud deployments. The surveys show that OpenNebula and OpenStack are achieving a similar level of deployment. However, OpenStack presents a ratio 1/40 between deployments in the survey and number of people involved, a ratio 1/3 between deployments and developers, and a ratio 1/2 between deployments and companies involved. Not every company contributed to the survey?.
We could also use the volume of web searches according to Google Trends to compare the impact of both projects. The ratio in the number of searchers between OpenNebula and OpenStack during the last 12 years is 1/20. This mainly reflects the successful marketing of OpenStack. OpenNebula mainly invests its resources in developing technology and serving its users, being really vendor agnostic and free of marketing.
There is also a quarterly comparative analysis of the community activity (mailing lists traffic mostly) of the four main open-source cloud management platforms: OpenStack, OpenNebula, Eucalyptus and CloudStack. The number of threads and participants in OpenStack is one order of magnitude higher than in OpenNebula. This mostly reflects a higher number of developers. Moreover, it is also worth noting that development coordination in OpenNebula is done through a redmine portal and not through a mailing list.
We conservatively estimate the investment in OpenStack in approximately 300 million per year:
- OpenStack Havana involved 950 developers almost completely hired by vendors. This is approximately $150 Million per year
- OpenStack Foundation budget is approximately $10 Million per year
- Just seven of the many start-ups involved in OpenStack have raised $120 million from VC. Assuming this is for 3 years. This is approximately $40 million per year
- There are other direct costs from other many companies, there are almost 1,000 companies involved, that are also allocating resources to development, training, documentation..., a big overhead in indirect costs, and of course opportunity costs.
So $300 million per year is a good conservative estimate. We have seen other estimations above $0.5 billion per year, some reaching to $1 billion per year. In any case, over a few years, it's billions. Are these companies getting this money back?. I see VC's starting to ask "Where's our future money?". Summarizing, a relatively small user base, and so penetration in the market, compared with the investment made by vendors and VCs. OpenNebula, with a budget at least two orders of magnitude lower, is achieving a similar user base. You can draw your own conclusions.
SYS-CON Events announced today that CodeFutures, a leading supplier of database performance tools, has been named a “Sponsor” of SYS-CON's 16th International Cloud Expo®, which will take place on June 9–11, 2015, at the Javits Center in New York, NY. CodeFutures is an independent software vendor focused on providing tools that deliver database performance tools that increase productivity during database development and increase database performance and scalability during production.
Feb. 28, 2015 11:00 AM EST Reads: 2,384
The IoT market is projected to be $1.9 trillion tidal wave that’s bigger than the combined market for smartphones, tablets and PCs. While IoT is widely discussed, what not being talked about are the monetization opportunities that are created from ubiquitous connectivity and the ensuing avalanche of data. While we cannot foresee every service that the IoT will enable, we should future-proof operations by preparing to monetize them with extremely agile systems.
Feb. 28, 2015 11:00 AM EST Reads: 1,113
There’s Big Data, then there’s really Big Data from the Internet of Things. IoT is evolving to include many data possibilities like new types of event, log and network data. The volumes are enormous, generating tens of billions of logs per day, which raise data challenges. Early IoT deployments are relying heavily on both the cloud and managed service providers to navigate these challenges. Learn about IoT, Big Data and deployments processing massive data volumes from wearables, utilities and ot...
Feb. 28, 2015 11:00 AM EST Reads: 970
Even as cloud and managed services grow increasingly central to business strategy and performance, challenges remain. The biggest sticking point for companies seeking to capitalize on the cloud is data security. Keeping data safe is an issue in any computing environment, and it has been a focus since the earliest days of the cloud revolution. Understandably so: a lot can go wrong when you allow valuable information to live outside the firewall. Recent revelations about government snooping, along...
Feb. 28, 2015 11:00 AM EST Reads: 6,866
In his session at DevOps Summit, Tapabrata Pal, Director of Enterprise Architecture at Capital One, will tell a story about how Capital One has embraced Agile and DevOps Security practices across the Enterprise – driven by Enterprise Architecture; bringing in Development, Operations and Information Security organizations together. Capital Ones DevOpsSec practice is based upon three "pillars" – Shift-Left, Automate Everything, Dashboard Everything. Within about three years, from 100% waterfall, C...
Feb. 28, 2015 11:00 AM EST Reads: 2,595
The explosion of connected devices / sensors is creating an ever-expanding set of new and valuable data. In parallel the emerging capability of Big Data technologies to store, access, analyze, and react to this data is producing changes in business models under the umbrella of the Internet of Things (IoT). In particular within the Insurance industry, IoT appears positioned to enable deep changes by altering relationships between insurers, distributors, and the insured. In his session at @Things...
Feb. 28, 2015 10:30 AM EST Reads: 2,360
There has been a lot of discussion recently in the DevOps space over whether there is a unique form of DevOps for large enterprises or is it just vendors looking to sell services and tools. In his session at DevOps Summit, Chris Riley, a technologist, discussed whether Enterprise DevOps is a unique species or not. What makes DevOps adoption in the enterprise unique or what doesn’t? Unique or not, what does this mean for adopting DevOps in enterprise size organizations? He also explored differe...
Feb. 28, 2015 10:00 AM EST Reads: 2,287
“In the past year we've seen a lot of stabilization of WebRTC. You can now use it in production with a far greater degree of certainty. A lot of the real developments in the past year have been in things like the data channel, which will enable a whole new type of application," explained Peter Dunkley, Technical Director at Acision, in this SYS-CON.tv interview at @ThingsExpo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Feb. 28, 2015 10:00 AM EST Reads: 3,189
SYS-CON Events announced today that Intelligent Systems Services will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. Established in 1994, Intelligent Systems Services Inc. is located near Washington, DC, with representatives and partners nationwide. ISS’s well-established track record is based on the continuous pursuit of excellence in designing, implementing and supporting nationwide clients’ mission-cri...
Feb. 28, 2015 10:00 AM EST Reads: 1,137
PubNub on Monday has announced that it is partnering with IBM to bring its sophisticated real-time data streaming and messaging capabilities to Bluemix, IBM’s cloud development platform. “Today’s app and connected devices require an always-on connection, but building a secure, scalable solution from the ground up is time consuming, resource intensive, and error-prone,” said Todd Greene, CEO of PubNub. “PubNub enables web, mobile and IoT developers building apps on IBM Bluemix to quickly add sc...
Feb. 28, 2015 10:00 AM EST Reads: 4,589
The major cloud platforms defy a simple, side-by-side analysis. Each of the major IaaS public-cloud platforms offers their own unique strengths and functionality. Options for on-site private cloud are diverse as well, and must be designed and deployed while taking existing legacy architecture and infrastructure into account. Then the reality is that most enterprises are embarking on a hybrid cloud strategy and programs. In this Power Panel at 15th Cloud Expo (http://www.CloudComputingExpo.com...
Feb. 28, 2015 10:00 AM EST Reads: 3,942
Data-intensive companies that strive to gain insights from data using Big Data analytics tools can gain tremendous competitive advantage by deploying data-centric storage. Organizations generate large volumes of data, the vast majority of which is unstructured. As the volume and velocity of this unstructured data increases, the costs, risks and usability challenges associated with managing the unstructured data (regardless of file type, size or device) increases simultaneously, including end-to-...
Feb. 28, 2015 09:45 AM EST Reads: 1,955
Sensor-enabled things are becoming more commonplace, precursors to a larger and more complex framework that most consider the ultimate promise of the IoT: things connecting, interacting, sharing, storing, and over time perhaps learning and predicting based on habits, behaviors, location, preferences, purchases and more. In his session at @ThingsExpo, Tom Wesselman, Director of Communications Ecosystem Architecture at Plantronics, will examine the still nascent IoT as it is coalescing, includin...
Feb. 28, 2015 09:45 AM EST Reads: 684
The cloud has transformed how we think about software quality. Instead of preventing failures, we must focus on automatic recovery from failure. In other words, resilience trumps traditional quality measures. Continuous delivery models further squeeze traditional notions of quality. Remember the venerable project management Iron Triangle? Among time, scope, and cost, you can only fix two or quality will suffer. Only in today's DevOps world, continuous testing, integration, and deployment upend...
Feb. 28, 2015 09:45 AM EST Reads: 2,192
The excitement around the possibilities enabled by Big Data is being tempered by the daunting task of feeding the analytics engines with high quality data on a continuous basis. As the once distinct fields of data integration and data management increasingly converge, cloud-based data solutions providers have emerged that can buffer your organization from the complexities of this continuous data cleansing and management so that you’re free to focus on the end goal: actionable insight.
Feb. 28, 2015 09:30 AM EST Reads: 1,593