Big Data Analytics and Its Applications in IoT

Big Data Analytics and Its Applications in IoT

Shaila S. G., Bhuvana D. S., Monish L.
DOI: 10.4018/978-1-7998-3111-2.ch009
OnDemand:
(Individual Chapters)
Available
$37.50
No Current Special Offers
TOTAL SAVINGS: $37.50

Abstract

Big data and the internet of things (IoT) are two major ruling domains in today's world. It is observed that there are 2.5 quintillion bytes of data created each day. Big data defines a very huge amount of data in terms of both structured and unstructured formats. Business intelligence and other application domains that have high information density use big data analytics to make predictions and better decisions to improve the business. Big data analytics is used to analyze a high range of data at a time. In general, big data and IoT were built on different technologies; however, over a period of time, both of them are interlinked to build a better world. Companies are not able to achieve maximum benefit, just because the data produced by the applications are not utilized and analyzed effectively as there is a shortage of big data analysts. For real-time IoT applications, synchronization among hardware, programming, and interfacing is needed to the greater extent. The chapter discusses about IoT and big data, relation between them, importance of big data analytics in IoT applications.
Chapter Preview
Top

Overview Of Big Data:

In general, the Big Data is defined using 3Vs such as Velocity, Volume and Variety. Velocity is the rate at which the data grows and how fast the data is gathered for analysis. Volume refers to the enormous data being generated, whereas Variety refers to the different types of data being generated like structured, semi- structured and unstructured data. There is a fourth V to describe the big data, it is referred as veracity, which includes availability and accountability. In general, data generated by sensors in IoT applications will be considered as raw data. This raw data need to be fine-tuned before it is used by the decision makers. In general, during data collection, it is noticed that data missing, data redundant, data in the wrong format, etc. Hence, preprocessing is needed to get the relevant data in required format to avoid erroneous and misleading outputs that reduce the efficiency. According to the B2B report of data quality index test, it was shown that every data repository has got 40% bad data in which 15% are duplicate, 8% are missing, 11% are invalid and 6% comes from malicious or unauthorized users. These data weaken the organization's marketing and financial automation, increase the resource consumption and cost and leads to lower customer satisfaction and invalid reports. Thus, there is a great need of data preprocessing. Apache Hadoop, Apache Spark, etc are used for this purpose. Hence, we use techniques such as data mining, analytical tools and machine learning to extract the useful information from the big data.

Complete Chapter List

Search this Book:
Reset