Seventh International Conference on Advances in Computing, Electronics and Communication-ACEC2018
Author(s) : MUHAMMAD KALEEM ULLAH, SYED KHURAM SHAHZAD
A term Big Data explain innovative technologies and techniques to store, manage, capture, analyze and distribute large size or petabyte data sets with high acceleration and dynamic structures. Big Data is categorized as semi-structured, unstructured, and structured, which results in an inability of traditional data managerial techniques. Data is produced from numerous resources and reached the system at different rates. This immense amount of data is processed in an efficient, and inexpensive manner, a technique of parallelism is practiced. Big Data parameters which include diversity, scale, and complexity required new techniques, architectures, analytics, and algorithms for the purpose of management of data and the knowledge hidden in it. Hadoop is a famous software platform to make data useful for the purpose of analytics to solve problems and structured the Big Data. Distributed processing is enabled by the help of Hadoop for huge datasets across the bunch of servers. It is specially designed to scale from one to thousands of high computing machines, with high fault tolerance degree.