Hit Count Analysis Using BigData Hadoop

Page 1

International Research Journal of Engineering and Technology (IRJET) Volume: 03 Issue: 11 | Nov -2016

www.irjet.net

e-ISSN: 2395 -0056 p-ISSN: 2395-0072

Hit Count Analysis Using BigData Hadoop Navonil Sarkar Department of Computer Engineering DPCOE Pune, India navonilsrkr34@gmail.com Jitin Michael Department of Computer Engineering DPCOE Pune, India g3jitin87@gmail.com Vivek Kumar Department of Computer Engineering DPCOE Pune, India vivekumar1995@gmail.com Rahul Chaurasia Department of Computer Engineering DPCOE Pune, India rahulchaurasia_3895@yahoo.com ---------------------------------------------------------------------***---------------------------------------------------------------------

Abstract - Log files are generated every second and are

Key Words: Bigdata , Hadoop, Map, Reduce, Pig, Hive,

HDFS

increasing at an alarming rate. Every time we access anything from internet we generate a log file at the corresponding server. These logs contain various information about the user and the website. These logs are stored and analyzed to extract various useful information about user and are used for various reasons such business decisions, security, performance analysis etc. In the past few years the number of internet users has increased drastically. As a result the log files are also generated in huge amount regularly. Terabytes and Petabytes of log files are generated every day and storing such huge amount of data is a challenge. Traditional tools are unable to store and process such huge amount of data. To overcome this problem Big Data is the solution. The main purpose of this paper is to overview the storing and processing of log files in Big data environment. To store the log files, HDFS is used. Hadoop clusters are used for efficient storage of log files. Hadoop framework is used to analyze the log files and summarize the same using some front end tools. Different Hadoop based algorithms are used to extract the necessary information from the log files and present them for various clients

Š 2016, IRJET

|

Impact Factor value: 4.45

1. INTRODUCTION In today's time, the use of internet is increasing day by day. Almost everyone today is using internet for a variety of purposes like social networking, gaming, gaining knowledge, spreading knowledge, entertainment etc. etc. With this spread of usage of internet millions of Terabytes and Petabytes of data is being generated almost every minute around the world. When someone likes a photo on Facebook data gets generated, someone uploads something on Facebook data gets generated, someone makes a Tweet data gets generated, Someone sends an email data gets generated. Now this data needs to be stored somewhere and as well as managed. Hadoop Distributed File System (HDFS) can be used to store and manage this data efficiently since Hadoop is an open-source software framework used for distributed storage and processing of very large data sets. It consists of computer clusters built from commodity hardware. It can incorporate large amount of clusters. A number of machines can be organized in the form of clusters and Millions of Terabytes of data can be stored in these clusters and managed. Hadoop is the core platform for structuring Big Data, and solves the problem of formatting it for subsequent

|

ISO 9001:2008 Certified Journal

|

Page 592


Turn static files into dynamic content formats.

Create a flipbook
Issuu converts static files into: digital portfolios, online yearbooks, online catalogs, digital photo albums and more. Sign up and create your flipbook.