|By Michael Bushong||
|July 30, 2014 06:00 AM EDT||
The primary indicator of success is success. That is to say that the number one thing people look to as a predictor of future performance is past performance. In the product space, this means that things like adoption are important as much for what it signals to other people as they are for bottom line revenues. And this is true even in the open source world.
As SDN speeds its way towards mainstream adoption, this means that projects like OpenDaylight will need to establish early on that they are not only deployable but also deployed.
Open source adoption
People frequently point to Linux as an example of an open source project that has seen wide adoption. But even Linux adoption did not happen overnight. It took more than a decade to see growth. And if you look at RedHat as an indicator of when that growth spawned commercial success, you have to extend all the way out to 2012 before the first $1B fiscal year.
The point here is not that Linux was not successful but rather that it took time to become successful. And the more success there was, the more success there tends to be. The rise of RedHat has enabled an acceleration of Linux deployments, in part because of an improved support model but in part because it represents a visible measure of commercial viability.
OpenDaylight and adoption timelines
Now consider that Linux was largely unknown and had virtually no expectations around it when it was created. There was no market that was waiting for it to hit. There were not industry players banking on its commercial success. There was not an entire technology movement underway dependent in part on the rise of a vendor-neutral platform.
The conditions under which OpenDaylight has been incubated are markedly different. And that means the expectations are different. Imagine how OpenDaylight would be evaluated if it took more than a decade to reach any kind of adoption. The pundits would not be kind, the customers would not be happy, and the companies expecting OpenDaylight to contribute to their commercial success would not be satisfied. OpenDaylight simply has to accelerate adoption.
What the bulls would say
Those bullish on OpenDaylight will tell you that conditions are certainly different. Open source is a better understood beast than it was in the early 90s. The lessons learned by those that championed Linux should result in faster adoption for projects that follow, and having the very group responsible for Linux (the Linux Foundation) at the helm only makes those lessons easier to put to use. There is an entire consortium of industry giants and would-be giant slayers who are building products and an ecosystem around OpenDaylight. Marketing efforts are helping drive awareness in both the vendor and user communities.
There are absolutely reasons to believe that adoption will happen faster than it did the first go-around.
What the bears would say
But there is a case to be made for the bears as well. SDN is more than a new technology; it’s a new architecture. Migration between architectures is far more disruptive, and thus more risky. The only way to mitigate risk is to move even slower, waiting for others to pave the way. And with a much lower volume of customers to pull from, this means that there will be fewer success stories early on and less overall experience to rely on. On top of that, while the consortium of companies is building products, they continue to sling their legacy portfolios that compete with the very thing they are collaborating on. Can they possibly be expected to push forward aggressively?
The missing ingredient
So what is missing for OpenDaylight to be successful?
In a word, deployments. But how do solutions get deployed? In the networking world at least, the answer is that they are pushed by the people building and ultimately selling them. Whether that’s the vendor itself or the resellers working on its behalf, there is someone on the end of the sales cycle who is explaining to the customer why and how they should deploy the solution.
Who is going to do that for OpenDaylight?
Right now, the answer is unclear. The most obvious answer is that OpenDaylight needs a RedHat to help speed deployments. In the OpenDaylight case, RedHat seems like the likely company to be RedHat. They are already OpenDaylight contributors, and they understand the business model well enough that they should be able to take a page from their own playbook.
But RedHat doesn’t own the networking channel or have the networking street cred. It’s not that they cannot be successful, but it will take more than RedHat to sell OpenDaylight.
Of course, the individual vendors all have a stake in OpenDaylight as well. Maybe they will make up the salesforce? Perhaps. But there is a challenge here. OpenDaylight is not really a revenue generator (at least not right away and not directly). Individual salespeople are compensated on the revenue they bring in. They don’t have a personal incentive to promote an open source project. More tactically, even if they wanted to, they aren’t fully trained on how it works and how they ought to be selling it. And even then, whatever they do know will be specific to the context in which the rest of their product catalog functions. A huge part of the value of OpenDaylight is that it works in heterogeneous environments and has technologies contributed from a bunch of different players. No salesperson is ever going to promote those aspects as aggressively as their own products.
What is needed?
If the problem is similar to a sales problem, then the solution will resemble a sales solution. Adoption will hinge on marketing to drive awareness and field enablement to drive sales capability. The first one is already being done with great effect, but the second one is missing. It’s hard to enable a salesforce that doesn’t really exist.
My suspicion is that the very thing that makes OpenDaylight powerful from a development perspective will swoop in to help out here: namely, the open source community. If community members who are leading adoption become active ambassadors for OpenDaylight, they can take the role of a Systems Engineer (SE) and help speed along deployments.
Engagements will be a little bit tough. Pairing ambassadors with active opportunities is non-trivial because it requires the customer to seek counsel from an ambassador they do not know while being presumably in a sales cycle that is led by vendors who are pushing alternative solutions. Fortunately, the biggest thing OpenDaylight can help do here is right up its alley: provide transparency. If customers are active in registering OpenDaylight opportunities, the Linux Foundation can pair ambassadors with those seeking guidance.
The bottom line
The industry needs a neutral point of control, and having every company reinvent and maintain a common platform is silly. Open source is a great way to advance the industry while limiting overlapping investment on the vendor side. But for adoption to take place, adoption has to happen. OpenDaylight can do something to solve this chicken-and-egg problem. Ultimately, if OpenDaylight is successful at providing opportunity transparency to its community, everyone benefits.
[Today’s fun fact: A group of geese on the ground is a gaggle, but a group of geese in the air is a skein. This is to flocking hard to remember.]
Today’s connected world is moving from devices towards things, what this means is that by using increasingly low cost sensors embedded in devices we can create many new use cases. These span across use cases in cities, vehicles, home, offices, factories, retail environments, worksites, health, logistics, and health. These use cases rely on ubiquitous connectivity and generate massive amounts of data at scale. These technologies enable new business opportunities, ways to optimize and automate, along with new ways to engage with users.
Oct. 8, 2015 03:45 PM EDT
The buzz continues for cloud, data analytics and the Internet of Things (IoT) and their collective impact across all industries. But a new conversation is emerging - how do companies use industry disruption and technology enablers to lead in markets undergoing change, uncertainty and ambiguity? Organizations of all sizes need to evolve and transform, often under massive pressure, as industry lines blur and merge and traditional business models are assaulted and turned upside down. In this new data-driven world, marketplaces reign supreme while interoperability, APIs and applications deliver un...
Oct. 8, 2015 03:30 PM EDT Reads: 211
The Internet of Things (IoT) is growing rapidly by extending current technologies, products and networks. By 2020, Cisco estimates there will be 50 billion connected devices. Gartner has forecast revenues of over $300 billion, just to IoT suppliers. Now is the time to figure out how you’ll make money – not just create innovative products. With hundreds of new products and companies jumping into the IoT fray every month, there’s no shortage of innovation. Despite this, McKinsey/VisionMobile data shows "less than 10 percent of IoT developers are making enough to support a reasonably sized team....
Oct. 8, 2015 03:30 PM EDT Reads: 129
The IoT market is on track to hit $7.1 trillion in 2020. The reality is that only a handful of companies are ready for this massive demand. There are a lot of barriers, paint points, traps, and hidden roadblocks. How can we deal with these issues and challenges? The paradigm has changed. Old-style ad-hoc trial-and-error ways will certainly lead you to the dead end. What is mandatory is an overarching and adaptive approach to effectively handle the rapid changes and exponential growth.
Oct. 8, 2015 03:30 PM EDT
You have your devices and your data, but what about the rest of your Internet of Things story? Two popular classes of technologies that nicely handle the Big Data analytics for Internet of Things are Apache Hadoop and NoSQL. Hadoop is designed for parallelizing analytical work across many servers and is ideal for the massive data volumes you create with IoT devices. NoSQL databases such as Apache HBase are ideal for storing and retrieving IoT data as “time series data.”
Oct. 8, 2015 02:45 PM EDT Reads: 492
Clearly the way forward is to move to cloud be it bare metal, VMs or containers. One aspect of the current public clouds that is slowing this cloud migration is cloud lock-in. Every cloud vendor is trying to make it very difficult to move out once a customer has chosen their cloud. In his session at 17th Cloud Expo, Naveen Nimmu, CEO of Clouber, Inc., will advocate that making the inter-cloud migration as simple as changing airlines would help the entire industry to quickly adopt the cloud without worrying about any lock-in fears. In fact by having standard APIs for IaaS would help PaaS expl...
Oct. 8, 2015 02:30 PM EDT Reads: 644
There are so many tools and techniques for data analytics that even for a data scientist the choices, possible systems, and even the types of data can be daunting. In his session at @ThingsExpo, Chris Harrold, Global CTO for Big Data Solutions for EMC Corporation, will show how to perform a simple, but meaningful analysis of social sentiment data using freely available tools that take only minutes to download and install. Participants will get the download information, scripts, and complete end-to-end walkthrough of the analysis from start to finish. Participants will also be given the pract...
Oct. 8, 2015 02:15 PM EDT Reads: 222
SYS-CON Events announced today that ProfitBricks, the provider of painless cloud infrastructure, will exhibit at SYS-CON's 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. ProfitBricks is the IaaS provider that offers a painless cloud experience for all IT users, with no learning curve. ProfitBricks boasts flexible cloud servers and networking, an integrated Data Center Designer tool for visual control over the cloud and the best price/performance value available. ProfitBricks was named one of the coolest Clo...
Oct. 8, 2015 01:00 PM EDT Reads: 759
Organizations already struggle with the simple collection of data resulting from the proliferation of IoT, lacking the right infrastructure to manage it. They can't only rely on the cloud to collect and utilize this data because many applications still require dedicated infrastructure for security, redundancy, performance, etc. In his session at 17th Cloud Expo, Emil Sayegh, CEO of Codero Hosting, will discuss how in order to resolve the inherent issues, companies need to combine dedicated and cloud solutions through hybrid hosting – a sustainable solution for the data required to manage I...
Oct. 8, 2015 01:00 PM EDT Reads: 472
NHK, Japan Broadcasting, will feature the upcoming @ThingsExpo Silicon Valley in a special 'Internet of Things' and smart technology documentary that will be filmed on the expo floor between November 3 to 5, 2015, in Santa Clara. NHK is the sole public TV network in Japan equivalent to the BBC in the UK and the largest in Asia with many award-winning science and technology programs. Japanese TV is producing a documentary about IoT and Smart technology and will be covering @ThingsExpo Silicon Valley. The program, to be aired during the peak viewership season of the year, will have a major impac...
Oct. 8, 2015 01:00 PM EDT Reads: 254
Apps and devices shouldn't stop working when there's limited or no network connectivity. Learn how to bring data stored in a cloud database to the edge of the network (and back again) whenever an Internet connection is available. In his session at 17th Cloud Expo, Bradley Holt, Developer Advocate at IBM Cloud Data Services, will demonstrate techniques for replicating cloud databases with devices in order to build offline-first mobile or Internet of Things (IoT) apps that can provide a better, faster user experience, both offline and online. The focus of this talk will be on IBM Cloudant, Apa...
Oct. 8, 2015 12:45 PM EDT Reads: 507
WebRTC is about the data channel as much as about video and audio conferencing. However, basically all commercial WebRTC applications have been built with a focus on audio and video. The handling of “data” has been limited to text chat and file download – all other data sharing seems to end with screensharing. What is holding back a more intensive use of peer-to-peer data? In her session at @ThingsExpo, Dr Silvia Pfeiffer, WebRTC Applications Team Lead at National ICT Australia, will look at different existing uses of peer-to-peer data sharing and how it can become useful in a live session to...
Oct. 8, 2015 12:00 PM EDT Reads: 603
As a company adopts a DevOps approach to software development, what are key things that both the Dev and Ops side of the business must keep in mind to ensure effective continuous delivery? In his session at DevOps Summit, Mark Hydar, Head of DevOps, Ericsson TV Platforms, will share best practices and provide helpful tips for Ops teams to adopt an open line of communication with the development side of the house to ensure success between the two sides.
Oct. 8, 2015 12:00 PM EDT Reads: 574
SYS-CON Events announced today that IBM Cloud Data Services has been named “Bronze Sponsor” of SYS-CON's 17th Cloud Expo, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. IBM Cloud Data Services offers a portfolio of integrated, best-of-breed cloud data services for developers focused on mobile computing and analytics use cases.
Oct. 8, 2015 11:00 AM EDT Reads: 727
"Matrix is an ambitious open standard and implementation that's set up to break down the fragmentation problems that exist in IP messaging and VoIP communication," explained John Woolf, Technical Evangelist at Matrix, in this SYS-CON.tv interview at @ThingsExpo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Oct. 8, 2015 07:00 AM EDT Reads: 5,865
WebRTC has had a real tough three or four years, and so have those working with it. Only a few short years ago, the development world were excited about WebRTC and proclaiming how awesome it was. You might have played with the technology a couple of years ago, only to find the extra infrastructure requirements were painful to implement and poorly documented. This probably left a bitter taste in your mouth, especially when things went wrong.
Oct. 8, 2015 06:00 AM EDT Reads: 759
The broad selection of hardware, the rapid evolution of operating systems and the time-to-market for mobile apps has been so rapid that new challenges for developers and engineers arise every day. Security, testing, hosting, and other metrics have to be considered through the process. In his session at Big Data Expo, Walter Maguire, Chief Field Technologist, HP Big Data Group, at Hewlett-Packard, will discuss the challenges faced by developers and a composite Big Data applications builder, focusing on how to help solve the problems that developers are continuously battling.
Oct. 8, 2015 04:00 AM EDT Reads: 488
Nowadays, a large number of sensors and devices are connected to the network. Leading-edge IoT technologies integrate various types of sensor data to create a new value for several business decision scenarios. The transparent cloud is a model of a new IoT emergence service platform. Many service providers store and access various types of sensor data in order to create and find out new business values by integrating such data.
Oct. 8, 2015 04:00 AM EDT Reads: 543
WebRTC converts the entire network into a ubiquitous communications cloud thereby connecting anytime, anywhere through any point. In his session at WebRTC Summit,, Mark Castleman, EIR at Bell Labs and Head of Future X Labs, will discuss how the transformational nature of communications is achieved through the democratizing force of WebRTC. WebRTC is doing for voice what HTML did for web content.
Oct. 8, 2015 03:00 AM EDT Reads: 1,376
Developing software for the Internet of Things (IoT) comes with its own set of challenges. Security, privacy, and unified standards are a few key issues. In addition, each IoT product is comprised of at least three separate application components: the software embedded in the device, the backend big-data service, and the mobile application for the end user's controls. Each component is developed by a different team, using different technologies and practices, and deployed to a different stack/target - this makes the integration of these separate pipelines and the coordination of software upd...
Oct. 8, 2015 03:00 AM EDT Reads: 282