You are on page 1of 20

BIG DATA CHALLENGES AND OPPORTUNITIES FOR

HADOOP
Fikru Megersa Roro
Graduate Student, Department of Informatics, College of Engineering and Technology,
P.B.No. 395, Wollega University, Nekemte, Ethiopia.

Dr. Satyanaryana Gaddada 1*

Associate Professor, Department of Electrical and Computer Engineering, College of


Engineering and Technology, P.B.No. 395, Wollega University, Nekemte, Ethiopia.

Abstract:
New cohorts are going into a "big data" epoch. Because of the bottlenecks, for
example, poor versatility, establishment, support challenges, adaptation to
internal failure and low execution, in customary data system framework it is
expected to influence the distributed computing procedures and answers for
manage huge information issues. Distributed computing and huge information
are corresponding to one another and have characteristic association of
argumentative solidarity. The leap forward of enormous information methods
won't just resolve the present circumstance additionally advance the wide use of
distributed computing and the web of things procedures. This paper concentrates
on talking about the improvement and the significant procedures of enormous
information and giving a thorough portrayal of big data from a few
perspectives including the advancement of big data the current information
burst situation the relationship between big data and distributed computing,
cloud computing and the big data techniques.
Keywords: Internet of things,
computing, stream computing.

HBase,

distributed

computing,

real-time

Introduction:
These days, information technology opens the entryway which makes the human
stride into the keen society, prompted the advancement of present day
administrations, for example, Internet e-business, cutting edge logistics and emoney, and advanced the improvement of rising commercial ventures. Present
day information technology is turning into the motor of operation and
advancement of varying backgrounds. Be that as it may, the motor is confronting
an enormous test of huge information [1]. Different business information is
softening out up the type of geometric arrangement. Issues, for example,
accumulation, stockpiling, recovery, examination, application etc, can never
again be understood by the conventional data preparing innovation, has brought
awesome deterrents for human accomplishing advanced, system and insightful
society. Starting in 2009, the "enormous information" has turned into a popular
expression of Internet data innovation industry, most utilizations of huge
information toward the starting were in the Internet business, the information on
the Internet expanded by half every year, multiplying like clockwork, The
worldwide Internet organizations know about the approach of "big data" time and
the colossal importance of information. 2011 May, McKinsey Global Institute
distributed a report entitled "big data: The following boondocks for development,
rivalry and profitability" [2], since the report was discharged, "big data" has
turned into a hot idea in the PC business. Toward the start of 2012, all the salary
of extensive information related programming, equipment and administrations
1

was just about $5 billion [3, 4]. However, as organizations bit by bit understand
that the huge information and related investigation will frame another separated
upper hand and enhance operational proficiency, big data related methods and
administrations will get the significant improvement.
At present, the industry does not have a brought together meaning of big data;
big data is characterized as takes after usually: "Big Data alludes to datasets
whose size is past the capacity of commonplace database programming
instruments to catch, store, oversee and dissect." - McKinsey. "Big Data more
often than excludes information sets with sizes past the capacity of usually
utilized programming devices to catch, clergyman, oversee and handle the
information inside of a bearable passed time." Wikipedia "Big Data is high
volume, high speed, and/or high assortment data resources that require new
types of preparing to empower improved choice making, knowledge revelation
and procedure advancement." - Gartner Big Data has four qualities: Volume,
Velocity, Variety and Value [5] (alluded to as "4V", which implies a tremendous
measure of information volume, quick preparing pace, different information sort
and low esteem thickness).
Volume: Means a lot of information of big data. The size of information set
continue expanding and from GB to TB, then to PB level yet, even checked by EB
and ZB. For example, video screens of a medium-sized city can create several TB
information consistently.
Variety: Indicates the sorts of big data are unpredictable. Previously, the
information sorts that produced or handled are more straightforward and a large
portion of the information are organized. In any case, now, with the developing of
new channels and innovations, for example, person to person communication,
Internet of Things, versatile processing , web publicizing, a lot of semi-organized
or unstructured information were delivered, for example, XML, email, blog , text,
and so forth. Result in a surge of new information sorts. Organizations need to
incorporate and break down information that from complex conventional or noncustomary wellsprings of data, including organizations' interior and outside
information. With the hazardous development of sensors, shrewd gadgets and
social synergistic innovations, the sorts of information are uncountable,
including: content, miniaturized scale online journals, sensors' information,
sound, video, snap streams, log documents etc.
Velocity: The speed of information era, preparing and investigation keep on
quickening, there are three reasons, Data Creation's inclination of ongoing, the
interest of consolidating spilling information to business procedures and choice
making procedures. The speed of information preparing is high, handling limit
shifts from bunch preparing to stream handling. The business gave the handling
limit of big data a tile "one second run the show". It demonstrates big data's
handling ability enough and the crucial contrast with customary data mining.
Value: Because of the amplifying scale, big data's quality thickness of per unit
information is continually lessening, in any case, the general estimation of the
information is expanding. Some person even compare enormous information with
the gold and oil, shows huge information contains boundless business esteem. As
indicated by an expectation from IDC exploration reports, big data innovation
and administrations business sector will ascend from $3.2 billion in 2010 to
$16.9 billion in 2015 and accomplish a yearly development rate of 40%, it will be
seven times the development rate of the entire IT and interchanges industry. By
2

preparing big data, figuring out its potential business esteem, it can make
gigantic business benefits. In particular applications, huge information handling
innovation can give specialized and stage backing to the national column
endeavors, examination, process and digging information for undertakings,
separate essential data and learning and afterward change them into helpful
models and apply to the procedure of exploration, generation, operation and
deal. In the meantime, the state unequivocally advocate development of "keen
city", in the connection of urbanization and data joining, concentrating on
enhancing individuals' occupation, improving the intensity of endeavors and
advancing reasonable improvement of urban communities, use Internet of
Things, distributed computing and other data innovation devices extensively,
consolidate the city's current data base, coordinate propelled administration idea
of urban operation, set up a generally secured and profoundly connected data
Network, see numerous elements of city completely, for instance, assets,
environment, bases, industry etc, fabricate a synergistic and shared urban data
stage, to prepare and use data shrewdly, so that give canny reaction control to
city's operations and allotment of assets, give social administration of
government and open administrations with smart premise for choice making and
routines, offer clever data assets and open data utilize stage's incorporated
provincial data advancement procedure to ventures and people.
Information are without a doubt the foundation of the new IT administration and
experimental exploration and big data handling innovation turn into the hot pot
of today's data innovation advancement normally, the twist of the big data
preparing innovation has additionally proclaimed the landing of another IT upset.
Then again, with the extending of national monetary rebuilding and modern
updating, the part of data preparing innovations will turn out to be progressively
conspicuous and big data handling innovation will turn into the best leap forward
to accomplish the center innovation's overwhelming around the bend, taking
after the advancement, application break through and diminishing hijacking in
data development of the mainstays of the national economy [6].
Big data Problems:
Big data is turned into an undetectable "gold mine" for the potential worth it
contains. With the gathering and developing information of creations, operations,
administration, observing, deals, client administrations and other perspectives'
information, and additionally the ascent of the quantity of clients, by examining
the relationship examples and patterns from extensive measure of information, it
is conceivable to accomplish proficient administration, exactness promoting and
this can turn into a key to open this "gold mine". Notwithstanding, the
conventional IT foundation and the techniques for information administration and
investigation can't adjust to the quick development of big dat.
Table 1: Classification of Big data problem
Classification of Big
Description of Big data problems
data problems
Import and export problem
Statistical analysis problems
Speed
Query and retrieval problems
real-time response problems
Type and structure
Multi-source problem
Heterogeneous problems
3

Volume and flexibility

Cost

Value mining
Storage and security
Connectivity and data
sharing

The original systems infrastructure problems


Linear scaling problems
Dynamic scheduling problems
Cost Comparison between Mainframe and
small server
The control of costs of the original systems
modification
Data analysis and mining
The actual efficiency after data mining
Structured and non-structured
Data security
Privacy security
Data standards and interfaces
Shared protocols
Access permissions

Issues of Speed
Traditional relational database management systems (RDBMS) for the most part
utilize concentrated capacity and handling strategies, without utilizing a
dispersed structural planning, in numerous huge undertakings, arrangements are
regularly in view of IOE (IBM Server, Oracle Database, EMC capacity). In this
normal design, a solitary server's arrangement is generally high, there can be
many CPU centers, memory can achieve several GB, either; Databases are put
away in rapid and expansive limit plate exhibit and storage room can up to TB
level. This setup can take care of the demand of traditional Management
Information System (MIS), yet when confront steadily developing information
volumes and element information utilization situation and this brought together
approach was turning into the bottleneck, particularly for its constrained pace of
reaction. At the point when face importing and sending out a lot of information,
factual examination, recovery and inquiry, in light of its reliance of brought
together information stockpiling and file, its execution will decay pointedly as
information volume develop, not to mention the measurements and question
situations that require ongoing reaction. For example, in Internet of Things, the
information of the sensor can be up to billions of things; these information
require ongoing capacity, question and examination, so conventional RDBMS is
no more suitable for application necessities.
Type and architecture problems:
RDMBS has framed moderately develop store, query, statistical and processing
methodologies for information that are organized and have fixed patterns. With
the rapid improvement of Internet of Things, Internet and versatile
correspondence arranges, the organizations and sorts of information is
continually changing and creating. In clever transportation field, the information
included may contain content, logs, pictures, recordings, vector maps and other
various types of information that from distinctive observing sources. The
organizations of these information are normally not settled; it will be hard to
react to changing needs in the event that we embrace organized capacity
modes. So we have to utilize different methods of information preparing and
stockpiling, incorporate organized and unstructured information stockpiling to
handle these information whose sorts, sources and structures are different. The
general information administration mode and construction modeling additionally
requires new sorts of circulated record frameworks and dispersed NoSQL
4

database building design to adjust to extensive measure of information and


shifty structures.
Volume and adaptability Related issues:
At the point when the measure of information increments and the measure of
simultaneous read and compose get to be bigger and bigger, centralized file
system or single database will turn into the fatal performance bottleneck, all
things considered, a solitary machine can just withstand constrained weight. This
can appropriate the weight to numerous machines by embracing the structures
and routines for direct versatility to achieve a level that machines can hold up
under, so we can powerfully expand or diminish the measure of records or
database servers as indicated by the measure of information and the amount of
simultaneous to accomplish linear adaptability.
As far as information storage, it needs to embrace a distributed and scalable
architecture, for example, surely understood Hadoop record framework and
HBase database [7]. In the interim, in the admiration of handling the information,
it likewise needs to receive an appropriated distributed architecture, doling out
the information preparing undertakings to numerous figuring hub, and need to
consider the relationship between the information storage nodes and the
computer nodes. In the computing field, the assignment of assets and
undertakings is really a mission booking issue. Its primary errand is making the
best match in the middle of assets and employments or among undertakings on
the premise of assets including CPU, memory, storage and network resources,
and so forth. Of every individual node of current group and the operation service
quality solicitations of every client. Since the different client necessities of
operation service quality and the changing condition of assets, discovering the
proper assets for conveyed information handling is a dynamic scheduling issue.
Cost Related Problems:
For centralized information storage and processing, while selecting software and
hardware, the essential methodology is utilizing high configuration mainframe
server or minicomputer server and high access speed disk array exhibits with
high security to ensure the data processing execution. These hardware gadgets
are extremely costly and every now and again up to a few million dollars and in
programming, regularly received the results of large foreign software vendors,
for example, Oracle, IBM, SAP, Microsoft, the support of servers and databases
additionally requires proficient specialized personnel and investment and
operation costs are high. Despite the difficulties of massive information
processing, these organizations have additionally presented an "AIO" solution in
the shape as a fiddle of beast, for example, Oracle's Exadata, SAP's Hana, and so
forth, that by stacking the multi-server, huge memory, flash memory, High-speed
networks and other hardware, to calm the weight of information,
notwithstanding, the hardware cost is essentially higher, general enterprises are
difficult to manage it.
The new distributed storage building architecture and distributed databases, for
example, HDFS, HBase, Cassandra[8], and so forth. Try not to have the
bottlenecks of information centralized handling and summary as they utilize a
decentralized and huge parallel processing MPP architectural design with linear
scalability, can manage the issues of capacity and handling of big data
successfully. In the software architecture, they additionally accomplished some
self-overseeing and self-mending mechanisms to confront the occasional failure
in massive nodes and secure the strength of general framework of the system,
5

so the hardware setup of every node needn't to be high, this can even utilize a
general PC as a server, so the cost of server can be incredibly lessened and as
far as software, open source software likewise give an expansive value
advantage.
Worth Mining Related Problems:
Because of the gigantic and developing volume, the worth density per unit of big
data is always diminishing, while the general estimation of big data is
relentlessly expanding, big data is practically equivalent to the oil and gold, so
can mine its immense business value[9]. On the off chance that we need to
extricate the concealed examples from expansive measure of information, we
require profound data mining and investigation. Big data is additionally very not
the same as conventional data mining model: traditional data mining spotlights
on moderate greater part of information by and large. Its algorithm is generally
mind boggling and the meeting pace is moderate. While in big data zone the
amount of information is huge and for the procedures of information storage,
information cleaning, ETL (extraction, transformation, loading), we have to
manage the needs and difficulties of huge information, which recommends the
utilization of conveyed parallel processing strategy. For instance, on account of
Google and Microsoft's web search engines, it needs hundreds or even a great
many servers working synchronously to perform the file storage of clients' hunt
logs which produced via look practices of billions of overall clients. In the interim,
while mining the information, we additionally need to adjust traditional data
mining algorithms and its hidden processing architecture. To help the massive
information registering and investigation; it is promising to present the parallel
processing mechanism. Apache's Mahout[10] task gives a progression of parallel
implementation of data mining algorithms. In numerous application situations,
even the ongoing criticisms of results is required, which displays the framework a
colossal test: the data mining calculations generally take quite a while,
particularly when the measure of information is immense. For this situation,
perhaps just a mix of constant estimation and substantial amounts disconnected
from the net handling can take care of the demand.
The genuine addition of data mining is an issue that should be precisely
evaluated before mining big data's quality. Not the greater part of the data
mining projects will obtain the fancied results. Firstly, we have to ensure the
credibility and fulfillment of the information. Furthermore, we likewise need to
consider the expenses and advantages of the mining. On the off chance that the
investments of manpower, hardware and software platform are expensive and
the task cycle is long, however data removed is not exceptionally important for
big enterprises production creation choices, cost-viability and different aspects,
then the data mining is additionally illogical.

Capacity and Security Related Problems:


In big data's capacity and security ensure angles, its variable arrangements,
huge volume have additionally brought a great deal of difficulties. For centralized
data storage, relational database management system (RDBMS) including
capacity, access, security and reinforcement control system is somewhat
develop following quite a while of advancement. The huge volume of big data
additionally has sway on the traditional RDBMS, as said, incorporated information
storage and preparing are moving to conveyed parallel processing. As a rule, big
data are unstructured data, so it determined a great deal of dispersed document
6

storage system and distributed NoSQL databases to manage this kind of


information. Be that as it may, these developing systems need to culminate their
client administrations, information access benefit, backup mechanism, security
control and different viewpoints. So, firstly it is important to avoid information
lost, and to give sensible backup and redundancy mechanism for the massive
structure and unstructured information so that the information won't to be lost
under any circumstances. Besides, we ought to shield the information from
unapproved access. Just the clients that have the power can get to the
information. Because of the a lot of unstructured information may require
different storage and access mechanisms, the arrangement of bound together
security access control mechanisms that emphasis on multi-source and different
information sorts is a major issues to be solved. Since big data gathered more
touchy information together, so it's more attractive to potential attackers; an
attacker will have the capacity to get more data on the off chance that he/she
submit an attack effectively "cost performance" is higher. These make it less
demanding for big data to wind up the objective of attack. In 2012, LinkedIn was
accused that it released 6.5 million client account passwords; Yahoo! have faced
network attacks, bringing about 450,000 client ID leak. 2011 December, CSDN's
security system was hacked, 6000000 client's login name, password and email
were leaked.
Privacy issues likewise nearly connected with big data. Because of the fast
improvement of Internet innovation and the Internet of Things innovation, a wide
range of data identified with our works and lives have been gathered and stored.
Interoperability and Data Sharing Issues:
Systems and information between distinctive commercial enterprises have no
crossing points. The same business, for example, transportation and social
security system's inside, additionally are partitioned and developed by
administrative areas; data exchange and collaboration effort crosswise over
areas are exceptionally troublesome. More genuine, even inside of the same unit,
for example, the development of a percentage of the hospitals data systems,
medical record management, beds information, medicines management and
other different subsystems are built discretely and there is no data sharing and
interoperability. "Smart City" is the accentuation of China's Twelfth Five years
plan of information construction development. The fundamentality of the "Smart
City" is to accomplish the interoperability and sharing of data, to accomplish
astute e-government, social management and improving individuals' livelihood in
light of information mix. So on the establishment of Digital city, it additionally
needs to accomplish interconnection, open the information interface of walks of
life to accomplish interoperability thus it can accomplish the knowledge. At
present, the information sharing platform built by US government, www.data.gov,
the information resources Network (www.bjdata.gov.cn) of Beijing Municipal
Government and different platforms are compelling endeavors for information
open and information sharing.
To accomplish the cross-industry information integration, it needs to make a
uniform information principles, exchange interfaces and additionally sharing
protocols, with the goal that we can get to, exchange and share the information
from distinctive commercial industries, different departments, and different
formats in light of a uniform basis. For information access, it likewise needs to
make definite access power to characterize what clients can access to which sort
of information in which circumstances. In the big data and cloud computing era,
information from distinctive industries, enterprises may be put away in a unified
7

platform and data centers, which needs to secure some sensitive information, for
example, the information identified with the enterprises mysteries of the
undertakings and transaction data, in spite of the fact that its procedure depend
on the stages, other than their own authorized persons, it ought to guarantee the
platforms administration and different organizations can't access such
information.
Challenging Relationships between the Cloud Computing and Big Data
Cloud computing has a quick advancement since 2007. Its core model is large
scale distributed computing, providing computing, storage, systems network
administration and different assets of an excess of clients in administration
mode, the clients use them when they require [11]. Cloud computing offer
enterprises and clients uses the high versatility, high accessibility and high
unwavering quality, productive utilization of assets, can enhance asset
administration proficiency and decrease the expenses of business data
development, investment and maintenance costs. As the U.S. Amazon, Google,
and Microsoft's open cloud administrations turn out to be more develop and
more immaculate, more organizations are relocating toward the cloud computing
platform.
Cloud computing and big data are reciprocal, persuasive relationship. Cloud
computing and Internet of things' broad use is our vision and the flare-up of big
data is a thorny problem that experienced in the development; The previous is
the dream of human's quest for civilization, the latter is the bottleneck to be
unraveled of social advancement; Cloud computing is a propensity of technology
development, big data is an unavoidable phenomenon of the quick improvement
of modern information society. To take care of big data issue, it needs present
day implies. The achievement of big data innovation can take care of genuine
issues, as well as can let innovation of cloud computing and Internet of things hit
the ground and be advanced and connected.
Big data Technology
Big data brings open doors as well as difficulties. Traditional information handling
means has been not able meet the massive real-time demand of huge
information. It needs the new era of data innovation to manage the outbreak of
big data. We summarize the big data innovation into five Classifications, as
appeared in Table
Classification
of
big
data
big data technology and tools
technology
Cloud computing platform
Cloud Storage
Infrastructure Supports
Virtualization technologies
Network technology
Resource Monitoring Technology
Data Bus
Data acquisition
ETL tools
Distributed File System
Relational database
NoSQL technology
The
integration
of
Relational
Data Storage
databases
and
non-relational
database
In-Memory Database
8

Data computing

Display and Interaction

Data query, statistics and analysis


Data Mining and Prediction
spectrum process
BI (Business Intelligence)
Graphics and reports
Visualization Tools
Augmented Reality Technology

Infrastructure Supports:
This mostly incorporates infrastructure administration centers that to support big
data processing, cloud computing platforms, cloud storage equipment and
technology, network technology, resources monitoring technology. Big data
processing needs the support from cloud data centers that have extensive scale
physical resources and cloud computing platforms
that have proficient
scheduling and administration function[10][11].
Information Acquisition Technology:
Information acquisition technology is an essential for information processing;
firstly it require the method for information procurement, gathering information
and can apply the top information processing technology to them. Other than the
different types of sensor and other hardware and software equipment,
information obtaining includes the ETL( acquisition, conversion, load) process of
information, can pre-process the information, such as washing, filtering, checking
and converting, changing over the legitimate information into suitable
arrangements and types. In the meantime, to bolster multi-source and
heterogeneous information acquisition and storage access, it additionally needs
to design a data bus of organizations, to facilitate the information exchange and
sharing between the different endeavor applications and administrations.
Information Storage Technology
After gathering and converting, the information should be storied documented.
Facing the large amount of information, by and large utilizing dispersed file
storage systems and distributed databases to distribute the information to
different storage nodes furthermore need to give numerous systems, for
example, backup, security, access interfaces and protocols.
The measure of information increment quickly consistently, alongside the current
authentic data, it has conveyed extraordinary open doors and difficulties to
information storage and information processing industry. Keeping in mind the
end goal to take care of the storage demand that developing quickly, cloud
storage requires high scalability, high reliability, high availability, minimal cost,
automatic fault tolerance, decentralization and different characteristics. Basic
types of cloud storage can be separated into appropriated distributed file system
and distributed database, the distributed file system utilize a large scale
distributed storage nodes to address the issues of storage a lot of files and
distributed database NoSQL support the processing and analysis of mass
unstructured information.
At the point when Google faced the issues of storing and analysis massive web
pages right on time, as a pioneer, it built up the Google File System GFS [12] and
the Map Reduce distributed processing analysis model[13, 14] in light of GFS. As
a feature of applications need to manage a substantial number of organized and
semi-designed information, Google has built a large scale database system
9

named Big Table which has weak consistency requests for, and is capable for
indexing, querying and analyzing enormous information. This series of Google
products has open the way to mass information storage, query and processing in
cloud computing era and turn into the accepted standard in this field, has
remained the pioneer in related method.
Google's innovation is not open, so Yahoo and open source group has developed
Hadoop system framework cooperatively, which is an open source
implementation of Map Reduce and GFS. The design principles of its principal file
system HDFS is totally reliable with GFS and it likewise accomplished an open
source execution of Big table, a distributed database system named HBase.
Since their dispatch, Hadoop and HBase has been generally applied everywhere
throughout the world, they are overseen by the Apache Foundation now, Yahoo's
own search system keeps running on Hadoop groups of million units.
Google has considered the vicious environment that confronted by running
disseminated file system in a large-scale data cluster:
1) Take full record of the issues that large number of nodes may experience
failure and need to coordinate the fault tolerance and automatic recovery
functions into the system and its framework.
2) Construct exceptional file system parameters, files are typically measure in GB
and contains a large number of small files.
3) Consider the physiognomies of the applications, support document append
operations, optimize sequential read and write speeds.
4) Some particular operations of file system are no more transparent and need
the aids of utilization application programs.

10

Fig 1 Architecture of the Google File System


Figure 1 delineates the architecture of the Google File System, to be specific a
GFS cluster contains a primary server (GFS Master) and a few blocks servers
(GFS chunkserver) and accessed by different clients (GFS Client). Large files are
split into blocks with fixed size, block server store the blocks on the local hard
drive as though they are Linux documents, read and write block information as
indicated by the predetermined block handle and byte range. With a specific end
goal to ensure the unwavering quality, every block has three backups by default.
Primary server deals with the greater part of the metadata of file system,
including namespaces, access control, mapping of files to blocks, physical
location of part and other important data. By the joint design of server and client,
GFS give application supports that have ideal execution and accessibility. GFS
was designed for Google applications themselves; there are many deployments
of GFS cluster in it. A few groups have more than a thousands of storage nodes,
storage space that over PB and visited by thousands of clients from different
machines continuously and frequently.
Keeping in mind the end goal to manage massive information challenges, some
commercial database system attempts to join traditional RDBMS innovations with
distributed, parallel computing technologies to process the needs of big data.
Many systems accelerate information processing on the hardware level.
NoSQL database, by definition is to break the paradigm constraints of traditional
relational databases. From the information storage point of view, many NoSQL
databases are not relational database, but the hash database that have keyvalue information format. Because of the abandonment of powerful SQL query
language and transactional consistency and paradigm constraints of relational
databases, NoSQL database can solve many challenges faced by traditional
relational database to a great extent. In the design, they are concerned about
the high simultaneous read and write of information and huge data storage, and
so forth. Contrasted and the relational databases, they have an extraordinary
point of interest in scalability, concurrency and fault tolerance to internal failure.
The standard NoSQL databases incorporate Google Big Table, an open source
implementation like Big Table named HBase, and Facebook Cassandra, and so
forth.
As a feature of Google applications need to process a large number of formatted
and semi-formatted information, Google constructed large-scale database
systems that need frail consistency requirements named Big Table. Big Table
applications incorporate log maps, Orkut online group, RSS reader etc.

11

Fig 2 Data model in Big Table


Figure 2 depicts the information model of applications in the Big Table model. The
information model incorporates rows, columns and corresponding timestamps, all
the information are put away in the table cells. Big Table substance are isolated
by rows, it coordinates a few columns to shape a small table and save to a single
server node. This small table is called Tablet.
Like the previous systems, Big Table is also a joint configuration of client and
server, making the performance can meet the needs of the applications farthest.
Big Table systems depends on the hidden structure of the cluster framework, a
distributed cluster task scheduler, the Google file system, as well as a distributed
lock administration Chubby. Chubby is an exceptionally vigorous coarse-grained
lock, which Big Table use to save the pointer of root information, hence clients
can acquire root server's area from the Chubby lock firstly, and after that get to
the information. Big Table utilize one server as the primary server, to store and
control metadata. Other than metadata administration, the primary server is
additionally responsible for remote administration and load deployment of tablet
server (the general feeling of the information server). Client utilize the
programming interfaces for metadata communications with the main server,
information communications with tablet server [15].
Information Computing
The information queries, measurements, investigation, determining, mining,
range process, BI business knowledge and other applicable innovation are all
things considered alluded to as information Computing innovation. Information
computing innovation cover all parts of information processing and are the core
technique of big data innovation either. Big Data is partitioned into three
sections, to be specific: offline batch computing, real-time interactive computing,
and stream computing.
12

Offline Batch Computing:


With the extensive variety of uses and improvement of cloud computing
strategy, Hadoop dispersed distributed storage systems and Map Reduce
information processing mode analysis systems in view of open source have
additionally been generally utilized. Hadoop can bolster PB level of distributed
information storage through information partitioning and self-recovery
mechanism, and also break down and handle these data based on Map Reduce
distributed processing model. Map Reduce programming model can make several
general information batch processing tasks and operations parallel on a largescale cluster and have automated failover capacity. Driven by open source
software programming, for example, Hadoop, Map Reduce programming model
has been broadly received and is connected to Web search, fraud detection and
different assortments of practical applications.
Hadoop is a software framework that can accomplish a lot of information's
distributed processing and process by a reliable, effective and adaptable way,
depending on horizontal extension, enhancing the computing and storing limit by
increasing low-cost commodity servers. Clients can easily create and run
applications that for managing gigantic measures of information easily. Hadoop
has the accompanying advantages:
High Reliability: the capacity to store and process information bit deserving of
the trust;
High Scalability: Allocate the information and complete computing errands in
accessible PC clusters, these clusters can be extended to the size of a thousands
of nodes easily.
High Efficiency: can dynamically move information in between nodes and
guarantee ensure dynamic balance of every node, the processing speed is very
fast.
High Fault-resilience: can spare multiple copies of information automatically
and consequently and reassign failed tasks automatically.

13

Fig 3: The Hadoop Ecosystem


The big data processing platform innovation [16] that the Hadoop platform
represents incorporate Map Reduce, HDFS, HBase has formed a Hadoop
environment, as appeared in Figure 3.
(1) Map Reduce programming model is the heart of Hadoop and utilized for
parallel calculation of massive information clusters. It is this programming model
that has accomplished massive scalability that crosswise over hundreds or a
great many servers of a Hadoop cluster.
(2) Distributed File System HDFS gives mass information storage based on
Hadoop processing platform, NameNode gives metadata administrations,
DataNode is utilized to store the file blocks file system.
(3) HBase is based on HDFS, is utilized to give a database system that has high
dependability, superior, column storage, adaptability and continuous read and
compose, can store unstructured and semi-structured loose information.
(4) There is a large data warehouse taking into account Hadoop that can be
utilized for information extraction, transformation and loading (ETL), storage,
query and large-scale information that store in Hadoop.
(5) There is likewise a large-scale information analysis platform taking into
account Hadoop, can change SQL-like information analysis requests into a series
of Map Reduce operations that were optimized, provides a simple operation and
programming Interface for complex massive information parallel computing.
(6) There is a proficient and solid synergistic frameworks, it that is utilized to
facilitate an assortment of administrations on distributed applications that can
utilize Zookeeper to assemble a coordination service that can counteract single
purpose of failures and manage load balancing effectively.
(7) As the middleware of parallel and superior, Avro gives information
serialization capabilities and RPC services between Hadoop platforms.

14

Hadoop platform is mainly for offline batch applications, typical application is to


operate static data by scheduling batch tasks, the computing procedure is
moderately moderate, to get results, a few queries may take hours or much
more, so it's weak when face applications and services that require high
constant. Map Reduce is a good cluster parallel programming model and can
address the needs of most applications. In spite of the fact that Map Reduce is a
decent dynamic of distributed/parallel computing, is not as a matter of course
suitable for taking care of any issue of processing. Map Reduce can't give
successful medicines to these ongoing applications, because of the fact that the
processing of these application logics requires multiple operations, or splitting
the information into a modest molecule size. Map Reduce model has the
accompanying confinements:
(1) The intermediate information transfer is hard to be completely improved.
(2) Restart of individual tasks is expensive.
(3) Big intermediate information storage spending.
(4) The master node can easily become a bottleneck.
(5) The unified file fragment size backing just, hard to manage complex
accumulation of
Documents that have assortment of sizes.
(6) Difficult to store and get to organized information storage directly.
Real-Time Interactive Computing
These days, the real-time computing are by and large for gigantic information,
notwithstanding meeting a percentage of the prerequisites of non-real-time
computing, that is exact results, the most important requirement of real-time
computing is reacting computed results progressively, millisecond level by and
large. Real-time computing can be categorized into the following two application
situations:
(1) The amount of information is enormous and the results can't be computed in
advance, while the response time of clients must be real-time.
Mostly utilized for information analysis and processing in specific events. At the
point when the amount of information is large, and we have discovered that
listing all the query combinations of possible conditions is impossible, or the
comprehensive condition combinations is futile, then real-time computing can
assume a part, it defer the computing procedure to the query phase, however
need to give clients real-time response. For this situation, it can handle a part of
the information in advance and combine real-time computing results to enhance
the processing efficiency.
(2) Data sources is real-time and continuous or uninterrupted, requires a realtime user response time.
The information sources are real-time, as such, gushing information. Purported
gushing information means viewing in the information as an information stream
to process. Information stream is the collection of a progression of information
records that are unlimited in time distribution and number; Data records are the
smallest units of information streams. For Real-time information computing and
analysis can analyze and count the information dynamically and in real-time, this
has important practical significances to the monitoring of systems state and
scheduling management.
15

Fig 4 Process of real-time calculation


Real-Time Data Acquisition
It needs to guarantee that can gather all the information integrally in function
and provide real-time information to real-time applications. Accordingly, need to
guarantee real time and low latency. Configuration is straightforward and simple
to convey. The system is constant and reliable [17, 18].
Real-Time Data Computing
The traditional information manipulations, gather the information and store them
in a database management system (DBMS) firstly, then connect with DBMS by
query and get the answer that user need. In the entire process, the users are
dynamic, while the DBMS system is aloof. Be that as it may, there are a lot of
real-time information now, such information have strong real-timeliness, huge
information volume and various information formats, traditionally relational
database pattern is not appropriate. The new real-time computing architectures
generally adopt the distributed structural planning of massive parallel
processing(MPP), the information storage and handling will be relegated to largescale nodes to meet the real-time requirements, on the information storage, they
utilize large-scale distributed file system, such as Hadoop's HDFS file system, or
the new NoSQL distributed databases.
Real-Time Query Service
Its usage can be partitioned into three ways:
1) Full Memory: give information read services directly, dump to disks or
databases for ingenuity consistently.
2) Semi-Memory: Use Red is, Mem cache, Mongo DB, Berkeley DB and different
databases to give Real-time Polling Service, completing ingenuity operations by
these systems.
3) Full Disk: use NoSQL databases that based on distributed file system (HDFS),
for example, HBase, with respect to key-value engine, the key is to design the
distribution of the key. [19]
Streaming Computing
In some real-time application scenarios, for example, real-time trading systems,
real-time fraud analysis and real-time ad push [20], real-time monitoring, realtime analysis of social networks, and so forth huge measure of existing
information, high real-time requirements and the information source is
persistent. New information processing must be processed immediately or
subsequent data will pile up and the processing will never end. It regularly needs
a sub-second or even sub-millisecond response time, which requires an
exceedingly adaptable streaming computing solution.
16

Stream Computing [21, 22] is for the real-time and ceaseless types of
information. Analysis progressively amid the development that the stream
information are changing, capture information that may be helpful to the clients
and sends the result out. All the while, the information analysis and processing
system is dynamic, the clients are in a latent state of gathering as demonstrated
as follows.

Figure 5 Process of Streaming computing


Traditional streaming computing systems are generally based on the event
mechanism and the amount of information processed by them is small. The new
stream processing strategy, for example, Yahoo's S4 [23][24], for the most part
to streaming processing issue that have a high information rate and a lot of
data.S4 is a general-purpose, distributed, scalable, partially fault-tolerant,
pluggable platform. Developing Engineers can without much of a stretch create
applications for unbounded, uninterrupted streaming information processing on
it. Information events are steered to processing Elements (PEs), PEs devour these
events and manage as takes after:
(1)Send out one or more events that may be processed by other PE.
(2)Publish results.
S4's outline is principally determined by information acquisitions and machine
discovering that utilized as a part of the production environments on a large
scale. Its primary features are:
(1)Provides a straightforward programming interface to handle the
information streaming.
(2)Design a high-accessibility cluster that are versatile on general
hardware.
(3)Use local memory of each processing node to stay away from disk I/O
bottleneck and minimize the inactivity.
(4)Use a de-centered, peer-to-peer architecture; all nodes give the same
functions and responsibilities. There is not a central node that takes a
special responsibility. This significantly rearranges the organization and
support.
(5)Use a pluggable architecture to make design general and adjustable as
much as conceivable.
(6)Friendly design concept, simple to program and with adaptability.
There are numerous characteristics between S4's design and IBM's stream
processing center (SPC) middleware [25]. Both architectures are intended for a
lot of information. Both of them have ability to utilize user-defined operations to
gather data in nonstop information streams. The fundamental difference is in the
17

structure design, SPC's system design is from Publish/Subscribe mode, however


S4's system structure originate from the mix of Map Reduce and Actor model.
Yahoo! Believe that on account of its equal structure, S4's system design has
accomplished a high level of simplicity. All nodes in the group are
indistinguishable, there is no central control.
SPC is a distributed stream processing middleware to support applications that
extract information from large-scale information stream. SPC contains
programming models and advancement environments to accomplish distributed,
dynamic, adaptable applications, its programming models incorporates API for
declaring and creating processing element (PE), and additionally a toolset that
for gathering, testing, troubleshooting, and applications deployments. Not at all
like the other stream processing middleware, notwithstanding supporting
relational operations, it likewise support non-relational operators and client
defined functions.
Storm [26] is a real-time data processing structure and open source by Twitter
and it's like Hadoop, this type of streaming computing solutions that with high
adaptability and can process high-recurrence information, large scale information
will be connected to real-time search, high-recurrence exchanging and social
networks. Storm has three activity scopes:
Stream Processing: Storm can be utilized to process new information
continuously and redesign the databases, have both adaptation to internal failure
and versatility.
Consistent Computation: Storm can do nonstop query and criticism the results
to the clients, for example, sending the hotly debated issue of Twitter to the
clients.
Appropriated RPC: Storm can be utilized to process escalated queries
simultaneously, Storm's topology is a distributed function that waiting for
summon messages, when gets a conjuring message, and it will compute the
queries and returns the results.
Information's Display and Interaction
Information's appear and cooperation is additionally vital in big data strategy,
since the information will in the long run be used by individuals to give choice
backings to preparations, operations and arranging. Picking a suitable, vivid and
visual method for presentation can present to us a better understanding of the
information, its meanings and affiliation relationship and can likewise offer us
some assistance with interpreting and utilize the information all the more
adequately, to add to their worth. In the methods for appear, notwithstanding
traditional reporting, graphics, it can likewise consolidate modern visualization
tools and human-computer interactions; even utilize the Augmented Reality
Technology, for example, Google Glasses, to accomplish a consistent interface
between the information and reality.
Conclusion
Big Data is a hot boondocks of today's information innovation advancement.
Internet of Things(IOT) , Internet and the rapid advancement of mobile portable
communication networking systems has produced the big data issue and have
brought problems of a different viewpoints, for example, speed's, structure's,
volume's, cost's, value's, security privacy's, interoperability's. Traditional IT
18

processing methods are frail when face the big data problem, for their lack of
adaptability and effectiveness. Big data issue needs distributed computing
technique to be solved, while big data additionally can advance the real landing
and execution of the cloud computing technique. There is a correlative
relationship between them. It concentrate on infrastructure, information
obtaining, information storage, information processing, information display and
association and different perspectives to depict a few kinds of system techniques
by big data, portray the difficulties and chances of big data strategy from
another angle for the researchers from related fields and give reference
classification techniques for big data technology. Big data technology is
continually developing with the surge of information amount and processing
requirements, influencing human life propensities and styles.
References:
1. Zikopoulos PC, Eaton C, DeRoos D, Deutsch T, Lapis G. Understanding big
data[J]. New York et al: McGraw-Hill, 2012.
2. Manyika J, Chui M, Brown B, Bughin J, Dobbs R, Roxburgh C, Byers AH. Big
data: The next frontier for innovation competition and productivity. May
2011[J]. MacKinsey Global Institute, 2011.
3. Big
Data
Research
and
Development
Initiative,
http://www.whitehouse.gov/sites/default/files/microsites/ostp/big_data_press_
release_final_2.pdf
4. http://wikibon.org/wiki/v/Big_Data_Market_Size_and_Vendor_Revenues
5. Foster I, Zhao Y, Raicu I, Shiyong L. Cloud computing and grid computing
360-degree compared[C]//Grid Computing Environments Workshop, 2008.
GCE'08. IEEE, 2008: 1-10.
6. OpenNebula, http://www.opennebula.org.
7. Taylor RC. An overview of the Hadoop/MapReduce/HBase framework and its
current applications in bioinformatics [J]. BMC bioinformatics, 2010, 11(Suppl
12): S1.
8. Openstack, http://www.openstack.org.
9. Keahey K, Freeman T. Contextualization: Providing one-click virtual
clusters[C]//eScience, 2008. EScience08. IEEE Fourth International
Conference on. IEEE, 2008: 301-308.
10. Ghemawat S, Gobioff H, Leung ST. The Google file system. In: Proc. of the
19th ACM Symp. On Operating Systems Principles. New York: ACM Press,
2003. 2943.
11. Chang F, Dean J, Ghemawat S, Hsieh WC, Wallach DA, Burrows M, Chandra T,
Fikes A, Gruber RE. Bigtable: A distributed storage system for structured
data. In: Proc. of the 7th USENIX Symp. On Operating Systems Design and
Implementation. Berkeley: USENIX Association, 2006. 205218.
12. Zheng QL, Fang M, Wang S, Wang XQ, Wu XW, Wang H. Scientific Parallel
Computing Based on MapReduce Model. Micro Electronics & Computer, 2009,
26(8):13-17 (in Chinese with English abstract).
13. Li GJ, Cheng XQ. Research Status and Scientific Thinking of Big Data [J].
Bulletin of Chinese Academy of Sciences, 2012, 27(6): 647-657 (in Chinese
with English abstract).
14. Dean J, Ghemawat S. MapReduce: simplified data processing on large
clusters [J]. Communications of the ACM, 2008, 51(1): 107-113.
15. Malkin J, Schroedl S, Nair A, Neumeyer L. Tuning Hyperparameters on Live
Traffic with S4. In TechPulse 2010: Internal Yahoo! Conference, 2010.
16. Schroedl S, Kesari A, Neumeyer L. Personalized ad placement in web
search[C]//Proceedings of the 4th Annual International Workshop on Data

19

17.
18.
19.
20.
21.
22.
23.
24.
25.

26.

Mining and Audience Intelligence for Online Advertising (AdKDD), Washington


USA. 2010.
Stonebraker M, etintemel U, Zdonik S. The 8 requirements of real-time
stream processing [J]. ACM SIGMOD Record, 2005, 34(4): 42-47.
Apache Hadoop. http://hadoop.apache.org/.
Neumeyer L, Robbins B, Nair A, Kesari A. S4: Distributed stream computing
platform[C]//Data Mining Workshops (ICDMW), 2010 IEEE International
Conference on. IEEE, 2010: 170-177.
Khetrapal A, Ganesh V. HBase and Hypertable for large scale distributed
storage systems [J]. Dept. of Computer Science, Purdue University, 2006.
http://cassandra.apache.org/
http://www.mongodb.org/
http://mahout.apache.org/
Li YL, Dong J. Study and Improvement of MapReduce based on Hadoop.
Computer Engineering and Design. 2012, 33(8):3110-3116 (in Chinese with
English abstract).
Baker J, Bond C, Corbett JC, Furman JJ, Khorlin A, Larson J, Leon JM, Li YW,
Lloyd A, Yushprakh V. Megastore: Providing Scalable, Highly Available Storage
for Interactive Services[C]//CIDR. 2011, 11: 223-234.
Wang S, Wang HJ, Qin XP, Zhou X. Architecting Big Data: Challenges, Studies
and Forecasts. Chinese Journal of Computers, 2011, 34(10): 1741-1752 (in
Chinese with English abstract).

20

You might also like