|By Stephen Walli||
|December 12, 2012 10:15 AM EST||
Software is surprisingly dynamic. All software evolves. Bugs are found and fixed. Enhancements added. New requirements are discovered in using the software. New uses are found for it and it is shaped to those new uses. Software solutions that are useful and used must by their very existence evolve. Well organized open source software communities create the right conditions to make this dynamism successful.
The world continues to embrace and adopt free and open source licensed software across the board. Vendors and OEMs, their IT customers, governments and academics are all using, buying and making open source software, and often all three at once.
Using and buying liberally licensed open source software, i.e., consuming such software, are relatively straight forward affairs. You buy a product based on open source licensed software pretty much the way you buy other software, evaluating the company producing the products and services against your own IT requirements and managed procurement risk profiles. You don't procure Red Hat Linux server software differently than you historically bought Solaris or might buy Microsoft Windows Server systems.
Using open source software (as opposed to buying a product) adds additional considerations based on evaluating the strength of the community around the open source project and the costs of supporting that choice either through the development of in-house expertise (likely supported by joining the project's community) or the hiring of external expertise. You look at a project's how-to documentation and tutorials, forum and email list activity, and IRC channels. You consider the availability of contracting support from other knowledgeable sources around the community. These considerations really don't change whether the open source software to be used is tools and infrastructure systems or developer libraries and frameworks. These considerations scale with use from individuals and the amount of time they have to spend solving their problem all the way up through company IT departments wanting to use open source licensed software and the time and money trade-offs they're willing to make.
Once one starts to make open source software, i.e. producing it, a different set of considerations arise. There are really two scenarios for producing open source:
- One can contribute to an existing project, adding value through bug fixes and new functionality (and possibly non-software contributions like documentation and translations).
- One can start a new open source project, which means organizing the infrastructure, developing the initial software, and providing for the early community.
The motivation in the first case of contributing to an existing open source project is simple. People generally start using open source software before they become contributors. People use software because it solves a problem they have. Once they use the software for a while, they will generally encounter a bug, find a change they want to make, or possibly document a new use case. If the user is comfortable with making software changes and the project community has done a good job of making it easy to contribute, then contributions can happen.
While it would be easy to simply make the necessary change and ignore the contribution, living on a personal forked copy of the software comes at a cost. Others' enhancements and bug fixes aren't seen and shared by installing newer versions of the software, and one needs to re-patch the software with one's own changes and fixes if one does try to move to a newer version. It is far better to contribute one's changes back to the project community if feasible, working with the committers to ensure its contributed correctly and patched into the main development tree. The onus is on the community to make it easy to contribute, but it's on the contributor to contribute correctly. The cost of living on a fork gets worse over time as the forked branch drifts further away from the mainline development of the project. It is well worth the investment to contribute.
This brings us to the "making" open source software case of starting one's own project.
First, it all starts with software. You must consider the software itself around which a project and its community is to be built. The software must "do" something useful from the beginning. Open source software developer communities are predominantly a discussion that starts with code, and without the code their is no discussion. Even when a fledgling community comes together to discuss a problem first with an eye to building the solution together, sooner or later someone needs to commit to writing the first workable building software that will act as a centre of gravity for all other conversations.
If an existing body of software is to be published into an open source community then there needs to be certain considerations with respect to the ownership and licensing. Software is covered by digital copyright and someone owns that copyright. To publish existing software requires the owners to agree to the publication and licensing as open source. The weight of existing code and its cultural history need to be considered, and may effect the early project community.
The crucial question becomes "why" open source? What motivates the publication of software under an open source license? Why share the software? Why choose NOT to commercialize it. (There are a number of important reasons not to commercialize or keep the software proprietary.)
The economics of collaboratively developing software is compelling. Writing good software is hard work. Managing the evolution of software over time is equally hard work. Sharing good software and collaboratively developing and maintaining it distributes the costs across a group. Publishing the software as open source, and building a development community (however small) is motivated by a desire to evolve the software and share the value and to be open to the idea that others in the community will join in sharing their domain expertise, learning the software's structure, and sharing the costs of evolution.
The economics are also asymmetric. For a contributor the contribution may represent a small bit of expertise from the contributor (e.g. a single bug fix or particular application of an algorithm that they personally understood), but the contributor is rewarded with the community investment of the entire package of software at relatively small personal cost. Likewise, the contribution is valuable to the software's developer (and user) community at large without necessarily carrying the costs of the contributor as a full time member in the developer community. (Indeed a single contribution may have been the only value the contributor had to give in this instance.)
Motivation to develop an open source community to evolve the software is an essential factor, but so too are knowledge of the problem domain, and the internal knowledge of the software needed to anchor the community. The essential motivation to share the software as open source supports the commitment and investment to maintain enough domain expertise and software knowledge to keep the community going and growing. Without all three factors it is difficult for the community to evolve the software and thrive.
One of the first structural considerations needs to be which open source software license to attach to the project. There are an array of licenses that have been approved by the Open Source Initiative as supporting the Open Source Definition, but there are really just a few that typically need consideration, and we'll discuss those at length in another post. The important thing to realize when choosing a license is that it doesn't just outline the legal responsibilities for how the software is shared, but it also outlines the social contract for how the community will share.
The next structural consideration for a community is to choose a tool platform to support collaborative development. This is the hub for activity for managing source code versions, distributing built software, handling the lines of communications, and logging issues and bugs. There are a number of free forge sites (e.g., Codeplex, Google Code, GitHub, SourceForge), and the tools all exist as open source themselves if a project wanted to develop and manage its own site.
The last structural consideration involves deciding what sort of community one wants to develop. What sort of governance will be required and when will certain things need to be instituted. There are two very good books available in this space:
Contribution is the life blood of an open source software community. It leads to new developers joining the project and learning enough to becoming committers with the responsibility for the code base and its builds. Its what makes the shared economic cost work for all. But as already stated, contributors generally start as users of the software. This means that a project community hoping to attract contributors first needs to attract users. The project's initial participants need to build a solid onramp for users that can then become contributors by making the software easy to "use", ensuring it's discoverable, downloadable, easily installable, and quickly configurable.
Not all users will contribute. Some may never push the software enough to need to make a change. It simply solves the problems they need to solve. Of those that contribute, some will contribute in very simple ways, reporting bugs for particular use cases. Others may contribute more, and this is where the second onramp needs to be developed by the community. Contributors need to know what sorts of contributions are encouraged, how to contribute, and where to contribute. If code contributions are to be encouraged, having scripts and notes on building the software and testing the baseline build make it easy for potential contributing developers to get involved.
So building an open source software project follows a pattern:
- There needs to be useful software, at least a seed around which to build a community.
- Motivation to share, expertise in the problem to be solved, and an understanding of the software structure will anchor an open source community. The project founder is the starting point for what will hopefully become a community.
- The project needs to have the structural issues of license, forge, and governance sorted, even if governance becomes an evolving discussion in a growing community.
- The community needs to build a solid onramp for users, and a second onramp for contributors. The sooner this happens in a project's life, the faster it can build a community.
One can choose to publish software under an open source license and never build a community. The software isn't "lost", but neither is it hardened or evolved. It may be useful to someone that discovers it, but the dynamic aspects of software development are lost to it. Taking the steps to encourage and build a community around the open source project sets the dynamic software engine in motion and allows the economics of collaborative development and sharing to work at its best.
How do APIs and IoT relate? The answer is not as simple as merely adding an API on top of a dumb device, but rather about understanding the architectural patterns for implementing an IoT fabric. There are typically two or three trends: Exposing the device to a management framework Exposing that management framework to a business centric logic Exposing that business layer and data to end users. This last trend is the IoT stack, which involves a new shift in the separation of what stuff happens, where data lives and where the interface lies. For instance, it's a mix of architectural styles ...
May. 27, 2015 03:00 AM EDT Reads: 5,907
We certainly live in interesting technological times. And no more interesting than the current competing IoT standards for connectivity. Various standards bodies, approaches, and ecosystems are vying for mindshare and positioning for a competitive edge. It is clear that when the dust settles, we will have new protocols, evolved protocols, that will change the way we interact with devices and infrastructure. We will also have evolved web protocols, like HTTP/2, that will be changing the very core of our infrastructures. At the same time, we have old approaches made new again like micro-services...
May. 27, 2015 02:30 AM EDT Reads: 5,587
Connected devices and the Internet of Things are getting significant momentum in 2014. In his session at Internet of @ThingsExpo, Jim Hunter, Chief Scientist & Technology Evangelist at Greenwave Systems, examined three key elements that together will drive mass adoption of the IoT before the end of 2015. The first element is the recent advent of robust open source protocols (like AllJoyn and WebRTC) that facilitate M2M communication. The second is broad availability of flexible, cost-effective storage designed to handle the massive surge in back-end data in a world where timely analytics is e...
May. 27, 2015 02:00 AM EDT Reads: 6,347
Collecting data in the field and configuring multitudes of unique devices is a time-consuming, labor-intensive process that can stretch IT resources. Horan & Bird [H&B], Australia’s fifth-largest Solar Panel Installer, wanted to automate sensor data collection and monitoring from its solar panels and integrate the data with its business and marketing systems. After data was collected and structured, two major areas needed to be addressed: improving developer workflows and extending access to a business application to multiple users (multi-tenancy). Docker, a container technology, was used to ...
May. 27, 2015 01:00 AM EDT Reads: 2,517
The true value of the Internet of Things (IoT) lies not just in the data, but through the services that protect the data, perform the analysis and present findings in a usable way. With many IoT elements rooted in traditional IT components, Big Data and IoT isn’t just a play for enterprise. In fact, the IoT presents SMBs with the prospect of launching entirely new activities and exploring innovative areas. CompTIA research identifies several areas where IoT is expected to have the greatest impact.
May. 26, 2015 09:00 PM EDT Reads: 5,230
The 4th International Internet of @ThingsExpo, co-located with the 17th International Cloud Expo - to be held November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA - announces that its Call for Papers is open. The Internet of Things (IoT) is the biggest idea since the creation of the Worldwide Web more than 20 years ago.
May. 26, 2015 06:00 PM EDT Reads: 2,277
The Industrial Internet revolution is now underway, enabled by connected machines and billions of devices that communicate and collaborate. The massive amounts of Big Data requiring real-time analysis is flooding legacy IT systems and giving way to cloud environments that can handle the unpredictable workloads. Yet many barriers remain until we can fully realize the opportunities and benefits from the convergence of machines and devices with Big Data and the cloud, including interoperability, data security and privacy.
May. 26, 2015 05:00 PM EDT Reads: 4,961
The Internet of Things is tied together with a thin strand that is known as time. Coincidentally, at the core of nearly all data analytics is a timestamp. When working with time series data there are a few core principles that everyone should consider, especially across datasets where time is the common boundary. In his session at Internet of @ThingsExpo, Jim Scott, Director of Enterprise Strategy & Architecture at MapR Technologies, discussed single-value, geo-spatial, and log time series data. By focusing on enterprise applications and the data center, he will use OpenTSDB as an example t...
May. 26, 2015 02:00 PM EDT Reads: 6,698
The Internet of Things is not only adding billions of sensors and billions of terabytes to the Internet. It is also forcing a fundamental change in the way we envision Information Technology. For the first time, more data is being created by devices at the edge of the Internet rather than from centralized systems. What does this mean for today's IT professional? In this Power Panel at @ThingsExpo, moderated by Conference Chair Roger Strukhoff, panelists will addresses this very serious issue of profound change in the industry.
May. 26, 2015 01:15 PM EDT Reads: 1,160
Scott Jenson leads a project called The Physical Web within the Chrome team at Google. Project members are working to take the scalability and openness of the web and use it to talk to the exponentially exploding range of smart devices. Nearly every company today working on the IoT comes up with the same basic solution: use my server and you'll be fine. But if we really believe there will be trillions of these devices, that just can't scale. We need a system that is open a scalable and by using the URL as a basic building block, we open this up and get the same resilience that the web enjoys.
May. 26, 2015 01:00 PM EDT Reads: 7,311
We are reaching the end of the beginning with WebRTC, and real systems using this technology have begun to appear. One challenge that faces every WebRTC deployment (in some form or another) is identity management. For example, if you have an existing service – possibly built on a variety of different PaaS/SaaS offerings – and you want to add real-time communications you are faced with a challenge relating to user management, authentication, authorization, and validation. Service providers will want to use their existing identities, but these will have credentials already that are (hopefully) i...
May. 26, 2015 01:00 PM EDT Reads: 4,608
All major researchers estimate there will be tens of billions devices - computers, smartphones, tablets, and sensors - connected to the Internet by 2020. This number will continue to grow at a rapid pace for the next several decades. With major technology companies and startups seriously embracing IoT strategies, now is the perfect time to attend @ThingsExpo, June 9-11, 2015, at the Javits Center in New York City. Learn what is going on, contribute to the discussions, and ensure that your enterprise is as "IoT-Ready" as it can be
May. 26, 2015 12:15 PM EDT Reads: 2,711
Container frameworks, such as Docker, provide a variety of benefits, including density of deployment across infrastructure, convenience for application developers to push updates with low operational hand-holding, and a fairly well-defined deployment workflow that can be orchestrated. Container frameworks also enable a DevOps approach to application development by cleanly separating concerns between operations and development teams. But running multi-container, multi-server apps with containers is very hard. You have to learn five new and different technologies and best practices (libswarm, sy...
May. 26, 2015 12:00 PM EDT Reads: 2,474
SYS-CON Events announced today that DragonGlass, an enterprise search platform, will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. After eleven years of designing and building custom applications, OpenCrowd has launched DragonGlass, a cloud-based platform that enables the development of search-based applications. These are a new breed of applications that utilize a search index as their backbone for data retrieval. They can easily adapt to new data sets and provide access to both structured and unstruc...
May. 26, 2015 12:00 PM EDT Reads: 2,250
An entirely new security model is needed for the Internet of Things, or is it? Can we save some old and tested controls for this new and different environment? In his session at @ThingsExpo, New York's at the Javits Center, Davi Ottenheimer, EMC Senior Director of Trust, reviewed hands-on lessons with IoT devices and reveal a new risk balance you might not expect. Davi Ottenheimer, EMC Senior Director of Trust, has more than nineteen years' experience managing global security operations and assessments, including a decade of leading incident response and digital forensics. He is co-author of t...
May. 26, 2015 11:00 AM EDT Reads: 5,969
The Internet of Things is a misnomer. That implies that everything is on the Internet, and that simply should not be - especially for things that are blurring the line between medical devices that stimulate like a pacemaker and quantified self-sensors like a pedometer or pulse tracker. The mesh of things that we manage must be segmented into zones of trust for sensing data, transmitting data, receiving command and control administrative changes, and peer-to-peer mesh messaging. In his session at @ThingsExpo, Ryan Bagnulo, Solution Architect / Software Engineer at SOA Software, focused on desi...
May. 26, 2015 11:00 AM EDT Reads: 4,270
SYS-CON Events announced today that MetraTech, now part of Ericsson, has been named “Silver Sponsor” of SYS-CON's 16th International Cloud Expo®, which will take place on June 9–11, 2015, at the Javits Center in New York, NY. Ericsson is the driving force behind the Networked Society- a world leader in communications infrastructure, software and services. Some 40% of the world’s mobile traffic runs through networks Ericsson has supplied, serving more than 2.5 billion subscribers.
May. 26, 2015 10:15 AM EDT Reads: 1,929
Buzzword alert: Microservices and IoT at a DevOps conference? What could possibly go wrong? In this Power Panel at DevOps Summit, moderated by Jason Bloomberg, the leading expert on architecting agility for the enterprise and president of Intellyx, panelists will peel away the buzz and discuss the important architectural principles behind implementing IoT solutions for the enterprise. As remote IoT devices and sensors become increasingly intelligent, they become part of our distributed cloud environment, and we must architect and code accordingly. At the very least, you'll have no problem fil...
May. 26, 2015 10:00 AM EDT Reads: 2,182
While great strides have been made relative to the video aspects of remote collaboration, audio technology has basically stagnated. Typically all audio is mixed to a single monaural stream and emanates from a single point, such as a speakerphone or a speaker associated with a video monitor. This leads to confusion and lack of understanding among participants especially regarding who is actually speaking. Spatial teleconferencing introduces the concept of acoustic spatial separation between conference participants in three dimensional space. This has been shown to significantly improve comprehe...
May. 26, 2015 10:00 AM EDT Reads: 3,388
The Domain Name Service (DNS) is one of the most important components in networking infrastructure, enabling users and services to access applications by translating URLs (names) into IP addresses (numbers). Because every icon and URL and all embedded content on a website requires a DNS lookup loading complex sites necessitates hundreds of DNS queries. In addition, as more internet-enabled ‘Things' get connected, people will rely on DNS to name and find their fridges, toasters and toilets. According to a recent IDG Research Services Survey this rate of traffic will only grow. What's driving t...
May. 26, 2015 09:00 AM EDT Reads: 5,437