|By Kaazing Blog||
|January 29, 2013 01:29 PM EST||
After presenting to a partner of Kaazing last week I got asked what impact the emerging WebSocket standard would have on the Web, assuming we continue down the path that has already been laid out.
The impact could be the same, or even more profound, as when we were first introduced to HTTP as a means to share static documents. The difference is that this time the targeted market is already defined – it is called the Web. I have over the past several years, half jokingly and half seriously, compared the current static Web with a push to talk radio (aka Walkie Talkie) and the new living Web with a cell phone. You can get by with the WT and solve most of your problems; after all it’s been around for a while and it works. If you want to communicate with a friend in “real-time” you can solve it by getting two WTs, one to talk and one to listen. With a new Web standard, WebSocket, entering the market, Web developers now have access to the equivalent of a cellphone – one channel for “talk” and “listen”.
What would you choose if both push-to-talk and cell phone were available to you? What would developers choose if both technologies were readily available to them (e.g.: browser support)?
Well, if you are uncertain and feel like WebSocket is an unproven standard you might want to relate to this; remember the first time your friends started pushing you to buy a cellphone although you already had a stationary phone at home and one at the office that worked perfectly? Now several years later we all have at least one cell phone, each, and we can’t (at least I can’t) live without it.
This is exactly the same impact the following line of code will have on the Web in comparison with the current HTTP communication we are so used and accustomed too.
var mySocket = new WebSocket("ws://websocket.org/");
If you do understand the profound impact this one line of code will have you are in good shape and most likely are already using, extending, or pushing this new standard solution from W3C and IETF. If you are not, then let me take a short moment to explain why it is so important:
- HTTP was designed to deliver static documents, not to deliver transactional, dynamic, and real-time data updates.
- HTTP is by design stateless, so session state needs to be artificially maintained. Traditionally this is done by a legacy Web-tier solution such as an application server like Oracle WebLogic Server or IBM WebSpere.
- In every environment developers have access to a “socket” interface, which enables them to communicate using any format (read protocol) over a full-duplex connection. Not on the Web.
- Not having access to a standard, Web-friendly, socket API forces us to create transformation layers when sending data from a Web client, using HTTP, to a backend system relying on a different full-duplex TCP protocol e.g. XMPP, STOMP, AMQP.
- The above line of code opens the floodgates to use any TCP-based communication format, which in turn enables developers to freely innovate and create new types of Web applications that previously have not been feasible over the existing HTTP infrastructure.
- WebSocket offers a far better use of bandwidth by getting rid of unnecessary HTTP headers when information is shared. The improvement is at a ratio of up to 1000x.
- The latency to deliver data is greatly improved by eliminating the round trip of the HTTP request-response model, and by using the bandwidth more efficiently.
With the explosive growth of Web-enabled devices (yes, I’m thinking about the iPhone, iPad, Android, Galaxy, etc…) and the demand for more and live information, communication and distribution of data over the Web is growing exponentially. At this rate the growth of data distributed over the Web will out pace the performance principals of Moore’s Law, which we depend on to ensure that our hardware can keep up with our needs.
WebSocket traffic vs. HTTP traffic
For individuals this may not be too much of a concern, but for companies providing online services it will be, and already is, a huge and costly issue since it requires a tremendous amount of resources to deliver on the increasing demand for live information over the Web (read about Google’s move and Facebook’s move).
For example, when a user enters a single character ‘a’ in a search engine, a drop down list appears automatically showing possible search results starting with letter ‘a’. Behind the scenes an HTTP request has been issued asking the server for the information displayed in the drop down list. For every new character entered a new HTTP request is issued to the server to request for more information. The same HTTP characteristics you can find in collaborative online documents such as Google Docs, where each character entered generates a POST to ensure that users editing or looking at the same document can see each other changes in real-time.
Now, what was sent, what was received, and what was really needed?
There is a great article on websocket.org, called a “Quantum Leap in Scalability for the Web” that is outlining the difference between HTTP and WebSocket in terms of bandwidth utilization. In this article the sample application is a simple trading solution, but the math can be applied to any HTTP-based dynamic and transactional Web application.
In the article we have 0,665Gbps in header traffic to respond to 100,000 users per request.
What is the impact of using WebSocket technology? There are no sizable headers involved passing information between a client and a Websocket Gateway. Let’s apply the above math example to WebSocket technology as described by the article:
100,000 visitors receiving an update every second. (WS wireframe = 2 byte) * 100,000 * 8 = 1,600,000 bps (0.001526Gbps).
Results from this easy math:
HTTP:// = 0,665Gbps versus WS:// = 0.001526Gbps. In the above sample Websocket communication is 436 times more efficient. 436 times! We are talking about a gigantic leap of improvement, and that assuming that your cookies are not adding more data than this sample.
WebSocket is not a better Ajax!
Not only is the new standard improving bandwidth utilization it also gives us the ability to use any TCP-based high level communication format for our Web applications. This part of the HTML5 WebSocket standard has still yet to be fully appreciated. Right now most solutions and developers tinkering with the WebSocket APIs are looking at the new standard as merely a better replacement of XHR, or Ajax, when in fact it is a quantum leap forward in communicating over the Web that cannot be compared to XHR. With WebSocket we can now build client libraries in any Web technology supporting any TCP-based protocols. A simple example would be to extend the now widely used chat protocol XMPP to the Web (here is a demo site that lets you log in to Google Talk using XMPP over WebSocket) by providing a client-side implementation on top of Websocket APIs, or an advanced example would be to extend Java Message Service (JMS) over WebSocket such as the Kaazing WebSocket Gateway.
Scaling a WebSocket Solution
Web developers have been trying to work around the limitations of HTTP since the early days using techniques such as Comet, Reverse Ajax, or HTTP Streaming. With a move to persistent connections, or a stateful Web, server scalability of concurrent connections has been, and still is, a serious concern. Holding on to a thread on the server while the thread is not in use, combined with an Web-tier and infrastructure that was not designed for this, is not necessarily a scalable combination. Now, great strides have been made to ensure better scalability across technology stacks such as the use of NIO in Java.
At Kaazing we have always taken scalability and performance extremely serious and focused on making sure that our software is not in the way of scale or performance. As a matter of fact, we did a benchmark over the new year 2008/2009, to prove that scaling a WebSocket solution with persistent connections was not an issue, so we brought in a Java performance expert – Kirk Pepperdine – to help us and by the first weeks of Jan 09 we were running 1,000,000 concurrent connections on one single server. Now, is this practical? A more realistic scenario is running 1,000,000 users on a single rack or half a rack. This would enable us to have failover and high-availability, while still providing great performance and scale. So, last year we ran new tests together with DELL and Tibco to ensure not only great scale but also outstanding performance: DELL, Tibco, and Kaazing enable ‘The Fastest Million’ to revolutionize real-time data delivery over the Web.
The simplest design ideas are often the innovations with the most impact. WebSocket as an idea and design is extremely “simple” and its impact on our industry will be profound. Of course, with simple ideas you also get the “doubters”. I remember one time when my co-founder John Fallows and I met with a renowned VC in Silicon Valley and he asked us:
“If this is such a great idea why has no one come up with this idea before?”
I guess you could ask humanity a similar question about why it took several thousands of years to invent the wheel – after all it’s so obvious and simple.
What is important to understand is that we now have at our disposal a very powerful tool that will enable us to communicate securely with anything over the Web, and that it is only our own imagination that will limit our ability to fully exploit the WebSocket standard to its full potential.
If you are having performance and scalability issues with your current Web solution then it is time to look at an enterprise WebSocket platform, such as the one Kaazing provides. To round off I’m just going to ask you one short question:
If you had a choice between building a Web application using HTTP and Websocket, and both were readily available to you, which one would you choose?
Major trends and emerging technologies – from virtual reality and IoT, to Big Data and algorithms – are helping organizations innovate in the digital era. However, to create real business value, IT must think beyond the ‘what’ of digital transformation to the ‘how’ to harness emerging trends, innovation and disruption. Architecture is the key that underpins and ties all these efforts together. In the digital age, it’s important to invest in architecture, extend the enterprise footprint to the cl...
Oct. 23, 2016 01:00 PM EDT Reads: 1,773
SYS-CON Events announced today that MathFreeOn will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. MathFreeOn is Software as a Service (SaaS) used in Engineering and Math education. Write scripts and solve math problems online. MathFreeOn provides online courses for beginners or amateurs who have difficulties in writing scripts. In accordance with various mathematical topics, there are more tha...
Oct. 23, 2016 12:45 PM EDT Reads: 964
The best way to leverage your Cloud Expo presence as a sponsor and exhibitor is to plan your news announcements around our events. The press covering Cloud Expo and @ThingsExpo will have access to these releases and will amplify your news announcements. More than two dozen Cloud companies either set deals at our shows or have announced their mergers and acquisitions at Cloud Expo. Product announcements during our show provide your company with the most reach through our targeted audiences.
Oct. 23, 2016 12:30 PM EDT Reads: 4,482
@ThingsExpo has been named the Top 5 Most Influential Internet of Things Brand by Onalytica in the ‘The Internet of Things Landscape 2015: Top 100 Individuals and Brands.' Onalytica analyzed Twitter conversations around the #IoT debate to uncover the most influential brands and individuals driving the conversation. Onalytica captured data from 56,224 users. The PageRank based methodology they use to extract influencers on a particular topic (tweets mentioning #InternetofThings or #IoT in this ...
Oct. 23, 2016 12:00 PM EDT Reads: 8,341
@ThingsExpo has been named the Top 5 Most Influential M2M Brand by Onalytica in the ‘Machine to Machine: Top 100 Influencers and Brands.' Onalytica analyzed the online debate on M2M by looking at over 85,000 tweets to provide the most influential individuals and brands that drive the discussion. According to Onalytica the "analysis showed a very engaged community with a lot of interactive tweets. The M2M discussion seems to be more fragmented and driven by some of the major brands present in the...
Oct. 23, 2016 11:30 AM EDT Reads: 11,317
In the next forty months – just over three years – businesses will undergo extraordinary changes. The exponential growth of digitization and machine learning will see a step function change in how businesses create value, satisfy customers, and outperform their competition. In the next forty months companies will take the actions that will see them get to the next level of the game called Capitalism. Or they won’t – game over. The winners of today and tomorrow think differently, follow different...
Oct. 23, 2016 11:00 AM EDT Reads: 931
In an era of historic innovation fueled by unprecedented access to data and technology, the low cost and risk of entering new markets has leveled the playing field for business. Today, any ambitious innovator can easily introduce a new application or product that can reinvent business models and transform the client experience. In their Day 2 Keynote at 19th Cloud Expo, Mercer Rowe, IBM Vice President of Strategic Alliances, and Raejeanne Skillern, Intel Vice President of Data Center Group and ...
Oct. 23, 2016 10:30 AM EDT Reads: 1,452
The Internet of Things (IoT), in all its myriad manifestations, has great potential. Much of that potential comes from the evolving data management and analytic (DMA) technologies and processes that allow us to gain insight from all of the IoT data that can be generated and gathered. This potential may never be met as those data sets are tied to specific industry verticals and single markets, with no clear way to use IoT data and sensor analytics to fulfill the hype being given the IoT today.
Oct. 23, 2016 09:45 AM EDT Reads: 2,505
More and more brands have jumped on the IoT bandwagon. We have an excess of wearables – activity trackers, smartwatches, smart glasses and sneakers, and more that track seemingly endless datapoints. However, most consumers have no idea what “IoT” means. Creating more wearables that track data shouldn't be the aim of brands; delivering meaningful, tangible relevance to their users should be. We're in a period in which the IoT pendulum is still swinging. Initially, it swung toward "smart for smar...
Oct. 23, 2016 09:45 AM EDT Reads: 790
Virgil consists of an open-source encryption library, which implements Cryptographic Message Syntax (CMS) and Elliptic Curve Integrated Encryption Scheme (ECIES) (including RSA schema), a Key Management API, and a cloud-based Key Management Service (Virgil Keys). The Virgil Keys Service consists of a public key service and a private key escrow service.
Oct. 23, 2016 09:45 AM EDT Reads: 1,020
Data is the fuel that drives the machine learning algorithmic engines and ultimately provides the business value. In his session at Cloud Expo, Ed Featherston, a director and senior enterprise architect at Collaborative Consulting, will discuss the key considerations around quality, volume, timeliness, and pedigree that must be dealt with in order to properly fuel that engine.
Oct. 23, 2016 09:30 AM EDT Reads: 3,823
What happens when the different parts of a vehicle become smarter than the vehicle itself? As we move toward the era of smart everything, hundreds of entities in a vehicle that communicate with each other, the vehicle and external systems create a need for identity orchestration so that all entities work as a conglomerate. Much like an orchestra without a conductor, without the ability to secure, control, and connect the link between a vehicle’s head unit, devices, and systems and to manage the ...
Oct. 23, 2016 08:30 AM EDT Reads: 1,342
Machine Learning helps make complex systems more efficient. By applying advanced Machine Learning techniques such as Cognitive Fingerprinting, wind project operators can utilize these tools to learn from collected data, detect regular patterns, and optimize their own operations. In his session at 18th Cloud Expo, Stuart Gillen, Director of Business Development at SparkCognition, discussed how research has demonstrated the value of Machine Learning in delivering next generation analytics to impr...
Oct. 23, 2016 08:00 AM EDT Reads: 5,687
For basic one-to-one voice or video calling solutions, WebRTC has proven to be a very powerful technology. Although WebRTC’s core functionality is to provide secure, real-time p2p media streaming, leveraging native platform features and server-side components brings up new communication capabilities for web and native mobile applications, allowing for advanced multi-user use cases such as video broadcasting, conferencing, and media recording.
Oct. 23, 2016 07:00 AM EDT Reads: 4,110
Amazon has gradually rolled out parts of its IoT offerings, but these are just the tip of the iceberg. In addition to optimizing their backend AWS offerings, Amazon is laying the ground work to be a major force in IoT - especially in the connected home and office. In his session at @ThingsExpo, Chris Kocher, founder and managing director of Grey Heron, explained how Amazon is extending its reach to become a major force in IoT by building on its dominant cloud IoT platform, its Dash Button strat...
Oct. 23, 2016 06:30 AM EDT Reads: 4,753
SYS-CON Events announced today that SoftNet Solutions will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. SoftNet Solutions specializes in Enterprise Solutions for Hadoop and Big Data. It offers customers the most open, robust, and value-conscious portfolio of solutions, services, and tools for the shortest route to success with Big Data. The unique differentiator is the ability to architect and ...
Oct. 23, 2016 06:00 AM EDT Reads: 712
A critical component of any IoT project is what to do with all the data being generated. This data needs to be captured, processed, structured, and stored in a way to facilitate different kinds of queries. Traditional data warehouse and analytical systems are mature technologies that can be used to handle certain kinds of queries, but they are not always well suited to many problems, particularly when there is a need for real-time insights.
Oct. 23, 2016 05:30 AM EDT Reads: 3,930
DevOps is being widely accepted (if not fully adopted) as essential in enterprise IT. But as Enterprise DevOps gains maturity, expands scope, and increases velocity, the need for data-driven decisions across teams becomes more acute. DevOps teams in any modern business must wrangle the ‘digital exhaust’ from the delivery toolchain, "pervasive" and "cognitive" computing, APIs and services, mobile devices and applications, the Internet of Things, and now even blockchain. In this power panel at @...
Oct. 23, 2016 05:15 AM EDT Reads: 1,845
One of biggest questions about Big Data is “How do we harness all that information for business use quickly and effectively?” Geographic Information Systems (GIS) or spatial technology is about more than making maps, but adding critical context and meaning to data of all types, coming from all different channels – even sensors. In his session at @ThingsExpo, William (Bill) Meehan, director of utility solutions for Esri, will take a closer look at the current state of spatial technology and ar...
Oct. 23, 2016 03:45 AM EDT Reads: 1,703
Everyone knows that truly innovative companies learn as they go along, pushing boundaries in response to market changes and demands. What's more of a mystery is how to balance innovation on a fresh platform built from scratch with the legacy tech stack, product suite and customers that continue to serve as the business' foundation. In his General Session at 19th Cloud Expo, Michael Chambliss, Head of Engineering at ReadyTalk, will discuss why and how ReadyTalk diverted from healthy revenue an...
Oct. 23, 2016 03:30 AM EDT Reads: 2,957