language-icon Old Web
English
Sign In

Streaming data

Streaming data is data that is continuously generated by different sources. Such data should be processed incrementally using Stream Processing techniques without having access to all of the data. In addition, it should be considered that concept drift may happen in the data which means that the properties of the stream may change over time. Streaming data is data that is continuously generated by different sources. Such data should be processed incrementally using Stream Processing techniques without having access to all of the data. In addition, it should be considered that concept drift may happen in the data which means that the properties of the stream may change over time. It is usually used in the context of big data in which it is generated by many different sources at high speed. Data streaming can also be explained as a technology used to deliver content to devices over the internet, and it allows users to access the content immediately, rather than having to wait for it to be downloaded. Big data is forcing many organizations to focus on storage costs, which brings interest to data lakes and data streams. A data lake refers to the storage of a large amount of unstructured and semi data, and is useful due to the increase of big data as it can be stored in such a way that firms can dive into the data lake and pull out what they need at the moment they need it. Whereas a data stream can perform real-time analysis on streaming data, and it differs from data lakes in speed and continuous nature of analysis, without having to store the data first. In digital innovation management theories, five characteristics of digital innovative technologies are mentioned; homogenization and decoupling, modularity, connectivity, digital traces and programmability. Before these characteristics are explained and further elaborated with different examples of data streaming, it is important to understand the difference between digitalization and digitizing. The latter describes encoding from analog information to a digital format, such as light that enters the lens of a camera and transforms to a digital format/image (Yoo et al. 2012;. Where digitalization refers to a more socio-technical process, where digitized techniques are applied to broader social and institutional contexts, meaning the process of converting information into a digital format, readable by some sort of computer. Within the context of data streaming this means that media for example, (information) has been digitized since the early 90’s, however the digitalization of ‘information/media’ has started to pick up since the beginning of this century. Now, firstly homogenization and decoupling. “Because all digital information assumes the same form, it can, at least in principle, be processed by the same technologies. Consequently, digitizing has the potential to remove the tight couplings between information types and their storage, transmission, and processing technologies”. Within the context of data streaming, this means in theory that one can stream data now from any digital device. It also leads to a reduction in demand and use of music and films on CDs for example. One of the consequences of homogenization & decoupling is the declining of marginal costs. The marginal cost of data streaming because it solely uses digital information, which can be transmitted, stored, and computed in fast and low-cost ways. An example of an industry that has low marginal costs due to data streaming is the music industry. Producers can now digitize songs and upload them on Spotify, instead of paying for the creation of the physical Albums and distributing these. Another consequence is convergent user experience, meaning that previously separated experiences are now brought together in one product. Data streaming is also modular, because systems components may be separated and recombined mainly for flexibility and variety. Data streaming works in different application versions and systems such as IOS. It is also possible to change the speed of data streaming. A consequence of modularity is the creation of platforms. Data streaming platforms bring together analysis of information, but more importantly, they are able to integrate data between different sources (Myers, 2016). IBM streams for example is a analytics platform that enables the applications developed by users to gather, analyze and correlate information that comes to them from a variety of sources (IBM). The third characteristic, connectivity, describes that a digital technology not only connects applications, devices and users but also connects customers and firms. Streaming services for example connects a vast collection of music and films of ‘producers’ with their consumers, so how music on Spotify can easily reach a vast group of consumers. Another example would be data of transport vehicles that can also be connected to firms with streaming applications, via vehicle-to-roadside communications. UPS does this for example to ‘calculate’ the optimal delivery routes by streaming real time big data and thereby reducing time to deliver packages. Interoperability, which is the ability of a product or system to work with other products or systems, is a consequence of connectivity. For instance, the music industry is interoperable, because some music platforms have integrated social media platforms. Another of connectivity is network externality. This means that the value of a good to a user increases with the number of other users (installed base) of the same or similar good. Data streaming technology can utilize network externalities, because it brings together supply and demand of large networks of creators and consumers. This is very much the case at popcorn time, a service where people can stream latest movies on demand. These streams work better when people have used their content. The latter has to do with the fact that if one streams content he/she automatically also down/uploads content. While a streaming service is being used it leaves Digital Traces, which simply describes the fact that all digital technologies leave a digital trace from the user. In the past, when media was sold, the seller/provider only had information about the transaction itself. With data streaming it has become possible to actually track the behaviour of the users because it occurs in real time, directly from the distributor/providers. Morris and Powers describe this as opening the ‘black box’ of consumption. Providers of streaming services, for example, are now able to track detailed consuming behavior of the user, which in turn, they use to influence the user' decision-making process by creating algorithms to further develop a service. This kind of streaming has changed the way people consume media, which in time offered new possibilities for new ideas. These are also referred to as wakes of innovation and occur in places one wouldn’t initially expect. For instance, data streaming has enabled the development of sensors, for example that are used in a lot of sectors for different purposes. In the manufacturing sector data streaming is used for real-time analysis to improve operations. In healthcare sector sensors are being used for connected medical devices to create hubs of patients and healthcare providers, that can trigger alerts when a patient has a medical emergency.

[ "Computer network", "Real-time computing", "Data mining", "Artificial intelligence" ]
Parent Topic
Child Topic
    No Parent Topic