22,99 €
Find the right big data solution for your business or organization Big data management is one of the major challenges facing business, industry, and not-for-profit organizations. Data sets such as customer transactions for a mega-retailer, weather patterns monitored by meteorologists, or social network activity can quickly outpace the capacity of traditional data management tools. If you need to develop or manage big data solutions, you'll appreciate how these four experts define, explain, and guide you through this new and often confusing concept. You'll learn what it is, why it matters, and how to choose and implement solutions that work. * Effectively managing big data is an issue of growing importance to businesses, not-for-profit organizations, government, and IT professionals * Authors are experts in information management, big data, and a variety of solutions * Explains big data in detail and discusses how to select and implement a solution, security concerns to consider, data storage and presentation issues, analytics, and much more * Provides essential information in a no-nonsense, easy-to-understand style that is empowering Big Data For Dummies cuts through the confusion and helps you take charge of big data solutions for your organization.
Sie lesen das E-Book in den Legimi-Apps auf:
Seitenzahl: 487
Veröffentlichungsjahr: 2013
Big Data For Dummies®
Published byJohn Wiley & Sons, Inc.111 River StreetHoboken, NJ 07030-5774
www.wiley.com
Copyright © 2013 by John Wiley & Sons, Inc., Hoboken, New Jersey
Published simultaneously in Canada
No part of this publication may be reproduced, stored in a retrieval system or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, scanning or otherwise, except as permitted under Sections 107 or 108 of the 1976 United States Copyright Act, without either the prior written permission of the Publisher, or authorization through payment of the appropriate per-copy fee to the Copyright Clearance Center, 222 Rosewood Drive, Danvers, MA 01923, (978) 750-8400, fax (978) 646-8600. Requests to the Publisher for permission should be addressed to the Permissions Department, John Wiley & Sons, Inc., 111 River Street, Hoboken, NJ 07030, (201) 748-6011, fax (201) 748-6008, or online at http://www.wiley.com/go/permissions.
Trademarks: Wiley, the Wiley logo, For Dummies, the Dummies Man logo, A Reference for the Rest of Us!, The Dummies Way, Dummies Daily, The Fun and Easy Way, Dummies.com, Making Everything Easier, and related trade dress are trademarks or registered trademarks of John Wiley & Sons, Inc. and/or its affiliates in the United States and other countries, and may not be used without written permission. All other trademarks are the property of their respective owners. John Wiley & Sons, Inc. is not associated with any product or vendor mentioned in this book.
Limit of Liability/Disclaimer of Warranty: The publisher and the author make no representations or warranties with respect to the accuracy or completeness of the contents of this work and specifically disclaim all warranties, including without limitation warranties of fitness for a particular purpose. No warranty may be created or extended by sales or promotional materials. The advice and strategies contained herein may not be suitable for every situation. This work is sold with the understanding that the publisher is not engaged in rendering legal, accounting, or other professional services. If professional assistance is required, the services of a competent professional person should be sought. Neither the publisher nor the author shall be liable for damages arising herefrom. The fact that an organization or Website is referred to in this work as a citation and/or a potential source of further information does not mean that the author or the publisher endorses the information the organization or Website may provide or recommendations it may make. Further, readers should be aware that Internet Websites listed in this work may have changed or disappeared between when this work was written and when it is read.
For general information on our other products and services, please contact our Customer Care Department within the U.S. at 877-762-2974, outside the U.S. at 317-572-3993, or fax 317-572-4002.
For technical support, please visit www.wiley.com/techsupport.
Wiley publishes in a variety of print and electronic formats and by print-on-demand. Some material included with standard print versions of this book may not be included in e-books or in print-on-demand. If this book refers to media such as a CD or DVD that is not included in the version you purchased, you may download this material at http://booksupport.wiley.com. For more information about Wiley products, visit www.wiley.com.
Library of Congress Control Number: 2013933950
ISBN: 978-1-118-50422-2 (pbk); ISBN 978-1-118-64417-1 (ebk); ISBN 978-1-118-64396-9 (ebk); ISBN 978-1-118-64401-0 (ebk)
Manufactured in the United States of America
10 9 8 7 6 5 4 3 2 1
About the Authors
Judith S. Hurwitz is President and CEO of Hurwitz & Associates, a research and consulting firm focused on emerging technology, including cloud computing, big data, analytics, software development, service management, and security and governance. She is a technology strategist, thought leader, and author. A pioneer in anticipating technology innovation and adoption, she has served as a trusted advisor to many industry leaders over the years. Judith has helped these companies make the transition to a new business model focused on the business value of emerging platforms. She was the founder of Hurwitz Group. She has worked in various corporations, including Apollo Computer and John Hancock. She has written extensively about all aspects of distributed software. In 2011 she authored Smart or Lucky? How Technology Leaders Turn Chance into Success (Jossey Bass, 2011). Judith is a co-author on five retail For Dummies titles including Hybrid Cloud For Dummies (John Wiley & Sons, Inc., 2012), Cloud Computing For Dummies (John Wiley & Sons, Inc., 2010), Service Management For Dummies, and Service Oriented Architecture For Dummies, 2nd Edition (both John Wiley & Sons, Inc., 2009). She is also a co-author on many custom published For Dummies titles including Platform as a Service For Dummies, CloudBees Special Edition (John Wiley & Sons, Inc., 2012), Cloud For Dummies, IBM Midsize Company Limited Edition (John Wiley & Sons, Inc., 2011), Private Cloud For Dummies, IBM Limited Edition (2011), and Information on Demand For Dummies, IBM Limited Edition (2008) (both John Wiley & Sons, Inc.).
Judith holds BS and MS degrees from Boston University, serves on several advisory boards of emerging companies, and was named a distinguished alumnus of Boston University’s College of Arts & Sciences in 2005. She serves on Boston University’s Alumni Council. She is also a recipient of the 2005 Massachusetts Technology Leadership Council award.
Alan F. Nugent is a Principal Consultant with Hurwitz & Associates. Al is an experienced technology leader and industry veteran of more than three decades. Most recently, he was the Chief Executive and Chief Technology Officer at Mzinga, Inc., a leader in the development and delivery of cloud-based solutions for big data, real-time analytics, social intelligence, and community management. Prior to Mzinga, he was executive vice president and Chief Technology Officer at CA, Inc. where he was responsible for setting the strategic technology direction for the company. He joined CA as senior vice president and general manager of CA’s Enterprise Systems Management (ESM) business unit and managed the product portfolio for infrastructure and data management. Prior to joining CA in April of 2005, Al was senior vice president and CTO of Novell, where he was the innovator behind the company’s moves into open source and identity-driven solutions. As consulting CTO for BellSouth he led the corporate initiative to consolidate and transform all of BellSouth’s disparate customer and operational data into a single data instance.
Al is the independent member of the Board of Directors of Adaptive Computing in Provo, UT, chairman of the advisory board of SpaceCurve in Seattle, WA, and a member of the advisory board of N-of-one in Waltham, MA. He is a frequent writer on business and technology topics and has shared his thoughts and expertise at many industry events throughout the years.
He is an instrument rated private pilot and has played professional poker for the past three decades. In his sparse spare time he enjoys rebuilding older American muscle cars and motorcycles, collecting antiquarian books, epicurean cooking, and has passion for cellaring American and Italian wines.
Fern Halper, PhD, is a Fellow with Hurwitz & Associates and Director of TDWI Research for Advanced Analytics. She has more than 20 years of experience in data analysis, business analysis, and strategy development. Fern has published numerous articles on data analysis and advanced analytics. She has done extensive research, writing, and speaking on the topic of predictive analytics and text analytics. Fern publishes a regular technology blog. She has held key positions at AT&T Bell Laboratories and Lucent Technologies, where she was responsible for developing innovative data analysis systems as well as developing strategy and product-line plans for Internet businesses. Fern has taught courses in information technology at several universities. She received her BA from Colgate University and her PhD from Texas A&M University.
Fern is a co-author on four retail For Dummies titles including Hybrid Cloud For Dummies (John Wiley & Sons, Inc., 2012), Cloud Computing For Dummies (John Wiley & Sons, Inc., 2010), Service Oriented Architecture For Dummies, 2nd Edition, and Service Management For Dummies (both John Wiley & Sons, Inc., 2009). She is also a co-author on many custom published For Dummies titles including Cloud For Dummies, IBM Midsize Company Limited Edition (John Wiley & Sons, Inc., 2011), Platform as a Service For Dummies, CloudBees Special Edition (John Wiley & Sons, Inc., 2012), and Information on Demand For Dummies, IBM Limited Edition (John Wiley & Sons, Inc., 2008).
Marcia A. Kaufman is a founding Partner and COO of Hurwitz & Associates, a research and consulting firm focused on emerging technology, including cloud computing, big data, analytics, software development, service management, and security and governance. She has written extensively on the business value of virtualization and cloud computing, with an emphasis on evolving cloud infrastructure and business models, data-encryption and end-point security, and online transaction processing in cloud environments. Marcia has more than 20 years of experience in business strategy, industry research, distributed software, software quality, information management, and analytics. Marcia has worked within the financial services, manufacturing, and services industries. During her tenure at Data Resources, Inc. (DRI), she developed sophisticated industry models and forecasts. She holds an AB from Connecticut College in mathematics and economics and an MBA from Boston University.
Marcia is a co-author on five retail For Dummies titles including Hybrid Cloud For Dummies (John Wiley & Sons, Inc., 2012), Cloud Computing For Dummies (John Wiley & Sons, Inc., 2010), Service Oriented Architecture For Dummies, 2nd Edition, and Service Management For Dummies (both John Wiley & Sons, Inc., 2009). She is also a co-author on many custom published For Dummies titles including Platform as a Service For Dummies, CloudBees Special Edition (John Wiley & Sons, Inc., 2012), Cloud For Dummies, IBM Midsize Company Limited Edition (John Wiley & Sons, Inc., 2011), Private Cloud For Dummies, IBM Limited Edition (2011), and Information on Demand For Dummies (2008) (both John Wiley & Sons, Inc.).
Dedication
Judith dedicates this book to her husband, Warren, her children, Sara and David, and her mother, Elaine. She also dedicates this book in memory of her father, David.
Alan dedicates this book to his wife Jane for all her love and support; his three children Chris, Jeff, and Greg; and the memory of his parents who started him on this journey.
Fern dedicates this book to her husband, Clay, daughters, Katie and Lindsay, and her sister Adrienne.
Marcia dedicates this book to her husband, Matthew, her children, Sara and Emily, and her parents, Gloria and Larry.
Authors’ Acknowledgments
We heartily thank our friends at Wiley, most especially our editor, Nicole Sholly. In addition, we would like to thank our technical editor, Brenda Michelson, for her insightful contributions.
The authors would like to acknowledge the contribution of the following technology industry thought leaders who graciously offered their time to share their technical and business knowledge on a wide range of issues related to hybrid cloud. Their assistance was provided in many ways, including technology briefings, sharing of research, case study examples, and reviewing content. We thank the following people and their organizations for their valuable assistance:
Context Relevant: Forrest Carman
Dell: Matt Walken
Epsilon: Bob Zurek
IBM: Rick Clements, David Corrigan, Phil Francisco, Stephen Gold, Glen Hintze, Jeff Jones, Nancy Kop, Dave Lindquist, Angel Luis Diaz, Bill Mathews, Kim Minor, Tracey Mustacchio, Bob Palmer, Craig Rhinehart, Jan Shauer, Brian Vile, Glen Zimmerman
Kognitio: Michael Hiskey, Steve Millard
Opera Solutions: Jacob Spoelstra
RainStor: Ramon Chen, Deidre Mahon
SAS Institute: Malcom Alexander, Michael Ames
VMware: Chris Keene
Xtremedata: Michael Lamble
Publisher’s Acknowledgments
We’re proud of this book; please send us your comments at http://dummies.custhelp.com. For other comments, please contact our Customer Care Department within the U.S. at 877-762-2974, outside the U.S. at 317-572-3993, or fax 317-572-4002.
Some of the people who helped bring this book to market include the following:
Acquisitions, Editorial
Senior Project Editor: Nicole Sholly
Project Editor: Dean Miller
Acquisitions Editor: Constance Santisteban
Copy Editor: John Edwards
Technical Editor: Brenda Michelson
Editorial Manager: Kevin Kirschner
Editorial Assistant: Anne Sullivan
Sr. Editorial Assistant: Cherie Case
Cover Photo: © Baris Simsek / iStockphoto
Composition Services
Project Coordinator: Sheree Montgomery
Layout and Graphics: Jennifer Creasey, Joyce Haughey
Proofreaders: Debbye Butler, Lauren Mandelbaum
Indexer: Valerie Haynes Perry
Publishing and Editorial for Technology Dummies
Richard Swadley, Vice President and Executive Group Publisher
Andy Cummings, Vice President and Publisher
Mary Bednarek, Executive Acquisitions Director
Mary C. Corder, Editorial Director
Publishing for Consumer Dummies
Kathleen Nebenhaus, Vice President and Executive Publisher
Composition Services
Debbie Stailey, Director of Composition Services
Visit www.dummies.com/cheatsheet/bigdata to view this book's cheat sheet.
Table of Contents
Introduction
About This Book
Foolish Assumptions
How This Book Is Organized
Part I: Getting Started with Big Data
Part II: Technology Foundations for Big Data
Part III: Big Data Management
Part IV: Analytics and Big Data
Part V: Big Data Implementation
Part VI: Big Data Solutions in the Real World
Part VII: The Part of Tens
Glossary
Icons Used in This Book
Where to Go from Here
Part I: Getting Started with Big Data
Chapter 1: Grasping the Fundamentals of Big Data
The Evolution of Data Management
Understanding the Waves of Managing Data
Wave 1: Creating manageable data structures
Wave 2: Web and content management
Wave 3: Managing big data
Defining Big Data
Building a Successful Big Data Management Architecture
Beginning with capture, organize, integrate, analyze, and act
Setting the architectural foundation
Performance matters
Traditional and advanced analytics
The Big Data Journey
Chapter 2: Examining Big Data Types
Defining Structured Data
Exploring sources of big structured data
Understanding the role of relational databases in big data
Defining Unstructured Data
Exploring sources of unstructured data
Understanding the role of a CMS in big data management
Looking at Real-Time and Non-Real-Time Requirements
Putting Big Data Together
Managing different data types
Integrating data types into a big data environment
Chapter 3: Old Meets New: Distributed Computing
A Brief History of Distributed Computing
Giving thanks to DARPA
The value of a consistent model
Understanding the Basics of Distributed Computing
Why we need distributed computing for big data
The changing economics of computing
The problem with latency
Demand meets solutions
Getting Performance Right
Part II: Technology Foundations for Big Data
Chapter 4: Digging into Big Data Technology Components
Exploring the Big Data Stack
Layer 0: Redundant Physical Infrastructure
Physical redundant networks
Managing hardware: Storage and servers
Infrastructure operations
Layer 1: Security Infrastructure
Interfaces and Feeds to and from Applications and the Internet
Layer 2: Operational Databases
Layer 3: Organizing Data Services and Tools
Layer 4: Analytical Data Warehouses
Big Data Analytics
Big Data Applications
Chapter 5: Virtualization and How It Supports Distributed Computing
Understanding the Basics of Virtualization
The importance of virtualization to big data
Server virtualization
Application virtualization
Network virtualization
Processor and memory virtualization
Data and storage virtualization
Managing Virtualization with the Hypervisor
Abstraction and Virtualization
Implementing Virtualization to Work with Big Data
Chapter 6: Examining the Cloud and Big Data
Defining the Cloud in the Context of Big Data
Understanding Cloud Deployment and Delivery Models
Cloud deployment models
Cloud delivery models
The Cloud as an Imperative for Big Data
Making Use of the Cloud for Big Data
Providers in the Big Data Cloud Market
Amazon’s Public Elastic Compute Cloud
Google big data services
Microsoft Azure
OpenStack
Where to be careful when using cloud services
Part III: Big Data Management
Chapter 7: Operational Databases
RDBMSs Are Important in a Big Data Environment
PostgreSQL relational database
Nonrelational Databases
Key-Value Pair Databases
Riak key-value database
Document Databases
MongoDB
CouchDB
Columnar Databases
HBase columnar database
Graph Databases
Neo4J graph database
Spatial Databases
PostGIS/OpenGEO Suite
Polyglot Persistence
Chapter 8: MapReduce Fundamentals
Tracing the Origins of MapReduce
Understanding the map Function
Adding the reduce Function
Putting map and reduce Together
Optimizing MapReduce Tasks
Hardware/network topology
Synchronization
File system
Chapter 9: Exploring the World of Hadoop
Explaining Hadoop
Understanding the Hadoop Distributed File System (HDFS)
NameNodes
Data nodes
Under the covers of HDFS
Hadoop MapReduce
Getting the data ready
Let the mapping begin
Reduce and combine
Chapter 10: The Hadoop Foundation and Ecosystem
Building a Big Data Foundation with the Hadoop Ecosystem
Managing Resources and Applications with Hadoop YARN
Storing Big Data with HBase
Mining Big Data with Hive
Interacting with the Hadoop Ecosystem
Pig and Pig Latin
Sqoop
Zookeeper
Chapter 11: Appliances and Big Data Warehouses
Integrating Big Data with the Traditional Data Warehouse
Optimizing the data warehouse
Differentiating big data structures from data warehouse data
Examining a hybrid process case study
Big Data Analysis and the Data Warehouse
The integration lynchpin
Rethinking extraction, transformation, and loading
Changing the Role of the Data Warehouse
Changing Deployment Models in the Big Data Era
The appliance model
The cloud model
Examining the Future of Data Warehouses
Part IV: Analytics and Big Data
Chapter 12: Defining Big Data Analytics
Using Big Data to Get Results
Basic analytics
Advanced analytics
Operationalized analytics
Monetizing analytics
Modifying Business Intelligence Products to Handle Big Data
Data
Analytical algorithms
Infrastructure support
Studying Big Data Analytics Examples
Orbitz
Nokia
NASA
Big Data Analytics Solutions
Chapter 13: Understanding Text Analytics and Big Data
Exploring Unstructured Data
Understanding Text Analytics
The difference between text analytics and search
Analysis and Extraction Techniques
Understanding the extracted information
Taxonomies
Putting Your Results Together with Structured Data
Putting Big Data to Use
Voice of the customer
Social media analytics
Text Analytics Tools for Big Data
Attensity
Clarabridge
IBM
OpenText
SAS
Chapter 14: Customized Approaches for Analysis of Big Data
Building New Models and Approaches to Support Big Data
Characteristics of big data analysis
Understanding Different Approaches to Big Data Analysis
Custom applications for big data analysis
Semi-custom applications for big data analysis
Characteristics of a Big Data Analysis Framework
Big to Small: A Big Data Paradox
Part V: Big Data Implementation
Chapter 15: Integrating Data Sources
Identifying the Data You Need
Exploratory stage
Codifying stage
Integration and incorporation stage
Understanding the Fundamentals of Big Data Integration
Defining Traditional ETL
Data transformation
Understanding ELT — Extract, Load, and Transform
Prioritizing Big Data Quality
Using Hadoop as ETL
Best Practices for Data Integration in a Big Data World
Chapter 16: Dealing with Real-Time Data Streams and Complex Event Processing
Explaining Streaming Data and Complex Event Processing
Using Streaming Data
Data streaming
The need for metadata in streams
Using Complex Event Processing
Differentiating CEP from Streams
Understanding the Impact of Streaming Data and CEP on Business
Chapter 17: Operationalizing Big Data
Making Big Data a Part of Your Operational Process
Integrating big data
Incorporating big data into the diagnosis of diseases
Understanding Big Data Workflows
Workload in context to the business problem
Ensuring the Validity, Veracity, and Volatility of Big Data
Data validity
Data volatility
Chapter 18: Applying Big Data within Your Organization
Figuring the Economics of Big Data
Identification of data types and sources
Business process modifications or new process creation
The technology impact of big data workflows
Finding the talent to support big data projects
Calculating the return on investment (ROI) from big data investments
Enterprise Data Management and Big Data
Defining Enterprise Data Management
Creating a Big Data Implementation Road Map
Understanding business urgency
Projecting the right amount of capacity
Selecting the right software development methodology
Balancing budgets and skill sets
Determining your appetite for risk
Starting Your Big Data Road Map
Chapter 19: Security and Governance for Big Data Environments
Security in Context with Big Data
Assessing the risk for the business
Risks lurking inside big data
Understanding Data Protection Options
The Data Governance Challenge
Auditing your big data process
Identifying the key stakeholders
Putting the Right Organizational Structure in Place
Preparing for stewardship and management of risk
Setting the right governance and quality policies
Developing a Well-Governed and Secure Big Data Environment
Part VI: Big Data Solutions in the Real World
Chapter 20: The Importance of Big Data to Business
Big Data as a Business Planning Tool
Stage 1: Planning with data
Stage 2: Doing the analysis
Stage 3: Checking the results
Stage 4: Acting on the plan
Adding New Dimensions to the Planning Cycle
Stage 5: Monitoring in real time
Stage 6: Adjusting the impact
Stage 7: Enabling experimentation
Keeping Data Analytics in Perspective
Getting Started with the Right Foundation
Getting your big data strategy started
Planning for Big Data
Transforming Business Processes with Big Data
Chapter 21: Analyzing Data in Motion: A Real-World View
Understanding Companies’ Needs for Data in Motion
The value of streaming data
Streaming Data with an Environmental Impact
Using sensors to provide real-time information about rivers and oceans
The benefits of real-time data
Streaming Data with a Public Policy Impact
Streaming Data in the Healthcare Industry
Capturing the data stream
Streaming Data in the Energy Industry
Using streaming data to increase energy efficiency
Using streaming data to advance the production of alternative sources of energy
Connecting Streaming Data to Historical and Other Real-Time Data Sources
Chapter 22: Improving Business Processes with Big Data Analytics: A Real-World View
Understanding Companies’ Needs for Big Data Analytics
Improving the Customer Experience with Text Analytics
The business value to the big data analytics implementation
Using Big Data Analytics to Determine Next Best Action
Preventing Fraud with Big Data Analytics
The Business Benefit of Integrating New Sources of Data
Part VII: The Part of Tens
Chapter 23: Ten Big Data Best Practices
Understand Your Goals
Establish a Road Map
Discover Your Data
Figure Out What Data You Don’t Have
Understand the Technology Options
Plan for Security in Context with Big Data
Plan a Data Governance Strategy
Plan for Data Stewardship
Continually Test Your Assumptions
Study Best Practices and Leverage Patterns
Chapter 24: Ten Great Big Data Resources
Hurwitz & Associates
Standards Organizations
The Open Data Foundation
The Cloud Security Alliance
National Institute of Standards and Technology
Apache Software Foundation
OASIS
Vendor Sites
Online Collaborative Sites
Big Data Conferences
Chapter 25: Ten Big Data Do’s and Don’ts
Do Involve All Business Units in Your Big Data Strategy
Do Evaluate All Delivery Models for Big Data
Do Think about Your Traditional Data Sources as Part of Your Big Data Strategy
Do Plan for Consistent Metadata
Do Distribute Your Data
Don’t Rely on a Single Approach to Big Data Analytics
Don’t Go Big Before You Are Ready
Don’t Overlook the Need to Integrate Data
Don’t Forget to Manage Data Securely
Don’t Overlook the Need to Manage the Performance of Your Data
Introduction
Welcome to Big Data For Dummies. Big data is becoming one of the most important technology trends that has the potential for dramatically changing the way organizations use information to enhance the customer experience and transform their business models. How does a company go about using data to the best advantage? What does it mean to transform massive amounts of data into knowledge? In this book, we provide you with insights into how technology transitions in software, hardware, and delivery models are changing the way that data can be used in new ways.
Big data is not a single market. Rather, it is a combination of data-management technologies that have evolved over time. Big data enables organizations to store, manage, and manipulate vast amounts of data at the right speed and at the right time to gain the right insights. The key to understanding big data is that data has to be managed so that it can meet the business requirement a given solution is designed to support. Most companies are at an early stage with their big data journey. Many companies are experimenting with techniques that allow them to collect massive amounts of data to determine whether hidden patterns exist within that data that might be an early indication of an important change. Some data may indicate that customer buying patterns are changing or that new elements are in the business that need to be addressed before it is too late.
As companies begin to evaluate new types of big data solutions, many new opportunities will unfold. For example, manufacturing companies may be able to monitor data coming from machine sensors to determine how processes need to be modified before a catastrophic event happens. It will be possible for retailers to monitor data in real time to upsell customers related products as they are executing a transaction. Big data solutions can be used in healthcare to determine the cause of an illness and provide a physician with guidance on treatment options.
Big data is not an isolated solution, however. Implementing a big data solution requires that the infrastructure be in place to support the scalability, distribution, and management of that data. Therefore, it is important to put both a business and technical strategy in place to make use of this important technology trend.
For many important reasons, we think that it is important for you to understand big data technologies and know the ways that companies are using emerging technologies such as Hadoop, MapReduce, and new database engines to transform the value of their data. We wrote this book to provide a perspective on what big data is and how it’s changing the way that organizations can leverage more data than was possible in the past. We think that this book will give you the context to make informed decisions.
About This Book
Big data is new to many people, so it requires some investigation and understanding of both the technical and business requirements. Many different people need knowledge about big data. Some of you want to delve into the technical details, while others want to understand the economic implications of making use of big data technologies. Other executives need to know enough to be able to understand how big data can affect business decisions. Implementing a big data environment requires both an architectural and a business approach — and lots of planning.
No matter what your goal is in reading this book, we address the following issues to help you understand big data and the impact it can have on your business:
What is the architecture for big data? How can you manage huge volumes of data without causing major disruptions in your data center?
When should you integrate the outcome of your big data analysis with your data warehouse?
What are the implications of security and governance on the use of big data? How can you keep your company safe?
What is the value of different data technologies, and when should you consider them as part of your big data strategy?
What types of data sources can you take advantage of with big data analytics? How can you apply different types of analytics to business problems?
Foolish Assumptions
Try as we might to be all things to all people, when it came to writing this book, we had to pick who we thought would be most interested in Big Data For Dummies. Here’s who we think you are:
You’re smart. You’re no dummy, yet the topic of big data gives you an uneasy feeling. You can’t quite get your head around it, and if you’re pressed for a definition, you might try to change the subject.
You’re a businessperson who wants little or nothing to do with technology. But you live in the 21st century, so you can’t escape it. People are saying, “It’s all about big data,” so you think that you better find out what they’re talking about.
You’re an IT person who knows a heck of a lot about technology. The thing is, you’re new to big data. Everybody says it’s something different. Once and for all, you want the whole picture.
Whoever you are, welcome. We’re here to help.
How This Book Is Organized
We divided our book into seven parts for easy reading. Feel free to skip about.
Part I: Getting Started with Big Data
In this part, we explain the basic concepts you need for a full understanding of big data, from both a technical and a business perspective. We also introduce you to the major concepts and components so that you can hold your own in any meaningful conversation about big data.
Part II: Technology Foundations for Big Data
Part II is for both technical and business professionals who need to understand the different types of big data components and the underlying technology concepts that support big data. In this section, we give you an understanding about the type of infrastructure that will make big data more practical.
Part III: Big Data Management
Part III is for both technical and business professionals, but it gets into a lot more of the details of different database options and emerging technologies such as MapReduce and Hadoop. Understanding these underlying technologies can help you understand what is behind this important trend.
Part IV: Analytics and Big Data
How do you analyze the massive amounts of data that become part of your big data infrastructure? In this part of the book, we go deeper into the different types of analytics that are helpful in getting real meaning from your data. This part helps you think about ways that you can turn big data into action for your business.
Part V: Big Data Implementation
This part gets to the details of what it means to actually manage data, including issues such as operationalizing your data and protecting the security and privacy of that data. This section gives you plenty to think about in this critical area.
Part VI: Big Data Solutions in the Real World
In this section, you get an understanding of how companies are beginning to use big data to transform their business operations. If you want to get a peek into the future at what you might be able to do with data, this section is for you.
Part VII: The Part of Tens
If you’re new to the For Dummies treasure-trove, you’re no doubt unfamiliar with The Part of Tens. In this section, Wiley editors torture For Dummies authors into creating useful bits of information that are easily accessible in lists containing ten (or so) elucidating elements. We started these chapters kicking and screaming but are ultimately very glad that they’re here. After you read through the big data best practices, and the do’s and don’ts we provide in The Part of Tens, we think you’ll be glad, too.
Glossary
We include a glossary of terms frequently used when people discuss big data. Although we strive to define terms as we introduce them in this book, we think you’ll find the glossary a useful resource.
Icons Used in This Book
Pay attention. The bother you save may be your own.
You may be sorry if this little tidbit slips your mind.
With this icon, we mark particularly useful points to pay attention to.
Here you find tidbits for the more technically inclined.
Where to Go from Here
We’ve created an overview of big data and introduced you to all its significant components. We recommend that you read the first four chapters to give you the context for what big data is about and what technologies are in place to make implementations a reality. The next two chapters introduce you to some of the underlying infrastructure issues that are important to understand. The following eight chapters get into a lot more detail about the different types of data structures that are foundational to big data.
You can read the book from cover to cover, but if you’re not that kind of person, we’ve tried to adhere to the For Dummies style of keeping chapters self-contained so that you can go straight to the topics that interest you most. Wherever you start, we wish you well.
Many of these chapters could be expanded into full-length books of their own. Big data and the emerging technology landscape are a big focus for us at Hurwitz & Associates, and we invite you to visit our website and read our blogs and insights at www.hurwitz.com.
Occasionally, John Wiley & Sons, Inc., has updates to its technology books. If this book has technical updates, they will be posted at www.dummies.com/go/bigdatafdupdates.
Part I
Visit www.dummies.com for more great Dummies content online.
In this part . . .
Trace the evolution of data management.
Define big data and its technology components.
Understand the different types of big data.
Integrate structured and unstructured data.
Understand the difference between real-time and non-real-time data.
Scale your big data operation with distributed computing.
Chapter 1
Grasping the Fundamentals of Big Data
In This Chapter
Looking at a history of data management
Understanding why big data matters to business
Applying big data to business effectiveness
Defining the foundational elements of big data
Examining big data’s role in the future
Managing and analyzing data have always offered the greatest benefits and the greatest challenges for organizations of all sizes and across all industries. Businesses have long struggled with finding a pragmatic approach to capturing information about their customers, products, and services. When a company only had a handful of customers who all bought the same product in the same way, things were pretty straightforward and simple. But over time, companies and the markets they participate in have grown more complicated. To survive or gain a competitive advantage with customers, these companies added more product lines and diversified how they deliver their product. Data struggles are not limited to business. Research and development (R&D) organizations, for example, have struggled to get enough computing power to run sophisticated models or to process images and other sources of scientific data.
Indeed, we are dealing with a lot of complexity when it comes to data. Some data is structured and stored in a traditional relational database, while other data, including documents, customer service records, and even pictures and videos, is unstructured. Companies also have to consider new sources of data generated by machines such as sensors. Other new information sources are human generated, such as data from social media and the click-stream data generated from website interactions. In addition, the availability and adoption of newer, more powerful mobile devices, coupled with ubiquitous access to global networks will drive the creation of new sources for data.
Although each data source can be independently managed and searched, the challenge today is how companies can make sense of the intersection of all these different types of data. When you are dealing with so much information in so many different forms, it is impossible to think about data management in traditional ways. Although we have always had a lot of data, the difference today is that significantly more of it exists, and it varies in type and timeliness. Organizations are also finding more ways to make use of this information than ever before. Therefore, you have to think about managing data differently. That is the opportunity and challenge of big data. In this chapter, we provide you a context for what the evolution of the movement to big data is all about and what it means to your organization.
The Evolution of Data Management
It would be nice to think that each new innovation in data management is a fresh start and disconnected from the past. However, whether revolutionary or incremental, most new stages or waves of data management build on their predecessors. Although data management is typically viewed through a software lens, it actually has to be viewed from a holistic perspective. Data management has to include technology advances in hardware, storage, networking, and computing models such as virtualization and cloud computing. The convergence of emerging technologies and reduction in costs for everything from storage to compute cycles have transformed the data landscape and made new opportunities possible.
As all these technology factors converge, it is transforming the way we manage and leverage data. Big data is the latest trend to emerge because of these factors. So, what is big data and why is it so important? Later in the book, we provide a more comprehensive definition. To get you started, big data is defined as any kind of data source that has at least three shared characteristics:
Extremely large Volumes of data
Extremely high Velocity of data
Extremely wide Variety of data
Big data is important because it enables organizations to gather, store, manage, and manipulate vast amounts data at the right speed, at the right time, to gain the right insights. But before we delve into the details of big data, it is important to look at the evolution of data management and how it has led to big data. Big data is not a stand-alone technology; rather, it is a combination of the last 50 years of technology evolution.
Organizations today are at a tipping point in data management. We have moved from the era where the technology was designed to support a specific business need, such as determining how many items were sold to how many customers, to a time when organizations have more data from more sources than ever before. All this data looks like a potential gold mine, but like a gold mine, you only have a little gold and lot more of everything else. The technology challenges are “How do you make sense of that data when you can’t easily recognize the patterns that are the most meaningful for your business decisions? How does your organization deal with massive amounts of data in a meaningful way?” Before we get into the options, we take a look at the evolution of data management and see how these waves are connected.
Understanding the Waves of Managing Data
Each data management wave is born out of the necessity to try and solve a specific type of data management problem. Each of these waves or phases evolved because of cause and effect. When a new technology solution came to market, it required the discovery of new approaches. When the relational database came to market, it needed a set of tools to allow managers to study the relationship between data elements. When companies started storing unstructured data, analysts needed new capabilities such as natural language–based analysis tools to gain insights that would be useful to business. If you were a search engine company leader, you began to realize that you had access to immense amounts of data that could be monetized. To gain value from that data required new innovative tools and approaches.
The data management waves over the past five decades have culminated in where we are today: the initiation of the big data era. So, to understand big data, you have to understand the underpinning of these previous waves. You also need to understand that as we move from one wave to another, we don’t throw away the tools and technology and practices that we have been using to address a different set of problems.
Wave 1: Creating manageable data structures
As computing moved into the commercial market in the late 1960s, data was stored in flat files that imposed no structure. When companies needed to get to a level of detailed understanding about customers, they had to apply brute-force methods, including very detailed programming models to create some value. Later in the 1970s, things changed with the invention of the relational data model and the relational database management system (RDBMS) that imposed structure and a method for improving performance. Most importantly, the relational model added a level of abstraction (the structured query language [SQL], report generators, and data management tools) so that it was easier for programmers to satisfy the growing business demands to extract value from data.
The relational model offered an ecosystem of tools from a large number of emerging software companies. It filled a growing need to help companies better organize their data and be able to compare transactions from one geography to another. In addition, it helped business managers who wanted to be able to examine information such as inventory and compare it to customer order information for decision-making purposes. But a problem emerged from this exploding demand for answers: Storing this growing volume of data was expensive and accessing it was slow. Making matters worse, lots of data duplication existed, and the actual business value of that data was hard to measure.
At this stage, an urgent need existed to find a new set of technologies to support the relational model. The Entity-Relationship (ER) model emerged, which added additional abstraction to increase the usability of the data. In this model, each item was defined independently of its use. Therefore, developers could create new relationships between data sources without complex programming. It was a huge advance at the time, and it enabled developers to push the boundaries of the technology and create more complex models requiring complex techniques for joining entities together. The market for relational databases exploded and remains vibrant today. It is especially important for transactional data management of highly structured data.
When the volume of data that organizations needed to manage grew out of control, the data warehouse provided a solution. The data warehouse enabled the IT organization to select a subset of the data being stored so that it would be easier for the business to try to gain insights. The data warehouse was intended to help companies deal with increasingly large amounts of structured data that they needed to be able to analyze by reducing the volume of the data to something smaller and more focused on a particular area of the business. It filled the need to separate operational decision support processing and decision support — for performance reasons. In addition, warehouses often store data from prior years for understanding organizational performance, identifying trends, and helping to expose patterns of behavior. It also provided an integrated source of information from across various data sources that could be used for analysis. Data warehouses were commercialized in the 1990s, and today, both content management systems and data warehouses are able to take advantage of improvements in scalability of hardware, virtualization technologies, and the ability to create integrated hardware and software systems, also known as appliances.
Sometimes these data warehouses themselves were too complex and large and didn’t offer the speed and agility that the business required. The answer was a further refinement of the data being managed through data marts. These data marts were focused on specific business issues and were much more streamlined and supported the business need for speedy queries than the more massive data warehouses. Like any wave of data management, the warehouse has evolved to support emerging technologies such as integrated systems and data appliances.
Data warehouses and data marts solved many problems for companies needing a consistent way to manage massive transactional data. But when it came to managing huge volumes of unstructured or semi-structured data, the warehouse was not able to evolve enough to meet changing demands. To complicate matters, data warehouses are typically fed in batch intervals, usually weekly or daily. This is fine for planning, financial reporting, and traditional marketing campaigns, but is too slow for increasingly real-time business and consumer environments.
How would companies be able to transform their traditional data management approaches to handle the expanding volume of unstructured data elements? The solution did not emerge overnight. As companies began to store unstructured data, vendors began to add capabilities such as BLOBs (binary large objects). In essence, an unstructured data element would be stored in a relational database as one contiguous chunk of data. This object could be labeled (that is, a customer inquiry) but you couldn’t see what was inside that object. Clearly, this wasn’t going to solve changing customer or business needs.
Enter the object database management system (ODBMS). The object database stored the BLOB as an addressable set of pieces so that we could see what was in there. Unlike the BLOB, which was an independent unit appended to a traditional relational database, the object database provided a unified approach for dealing with unstructured data. Object databases include a programming language and a structure for the data elements so that it is easier to manipulate various data objects without programming and complex joins. The object databases introduced a new level of innovation that helped lead to the second wave of data management.
Wave 2: Web and content management
It’s no secret that most data available in the world today is unstructured. Paradoxically, companies have focused their investments in the systems with structured data that were most closely associated with revenue: line-of-business transactional systems. Enterprise Content Management systems evolved in the 1980s to provide businesses with the capability to better manage unstructured data, mostly documents. In the 1990s with the rise of the web, organizations wanted to move beyond documents and store and manage web content, images, audio, and video.
The market evolved from a set of disconnected solutions to a more unified model that brought together these elements into a platform that incorporated business process management, version control, information recognition, text management, and collaboration. This new generation of systems added metadata (information about the organization and characteristics of the stored information). These solutions remain incredibly important for companies needing to manage all this data in a logical manner. But at the same time, a new generation of requirements has begun to emerge that drive us to the next wave. These new requirements have been driven, in large part, by a convergence of factors including the web, virtualization, and cloud computing. In this new wave, organizations are beginning to understand that they need to manage a new generation of data sources with an unprecedented amount and variety of data that needs to be processed at an unheard-of speed.
Wave 3: Managing big data
Is big data really new or is it an evolution in the data management journey? The answer is yes — it is actually both. As with other waves in data management, big data is built on top of the evolution of data management practices over the past five decades. What is new is that for the first time, the cost of computing cycles and storage has reached a tipping point. Why is this important? Only a few years ago, organizations typically would compromise by storing snapshots or subsets of important information because the cost of storage and processing limitations prohibited them from storing everything they wanted to analyze.
In many situations, this compromise worked fine. For example, a manufacturing company might have collected machine data every two minutes to determine the health of systems. However, there could be situations where the snapshot would not contain information about a new type of defect and that might go unnoticed for months.
With big data, it is now possible to virtualize data so that it can be stored efficiently and, utilizing cloud-based storage, more cost-effectively as well. In addition, improvements in network speed and reliability have removed other physical limitations of being able to manage massive amounts of data at an acceptable pace. Add to this the impact of changes in the price and sophistication of computer memory. With all these technology transitions, it is now possible to imagine ways that companies can leverage data that would have been inconceivable only five years ago.
But no technology transition happens in isolation; it happens when an important need exists that can be met by the availability and maturation of technology. Many of the technologies at the heart of big data, such as virtualization, parallel processing, distributed file systems, and in-memory databases, have been around for decades. Advanced analytics have also been around for decades, although they have not always been practical. Other technologies such as Hadoop and MapReduce have been on the scene for only a few years. This combination of technology advances can now address significant business problems. Businesses want to be able to gain insights and actionable results from many different kinds of data at the right speed — no matter how much data is involved.
If companies can analyze petabytes of data (equivalent to 20 million four-drawer file cabinets filled with text files or 13.3 years of HDTV content) with acceptable performance to discern patterns and anomalies, businesses can begin to make sense of data in new ways. The move to big data is not just about businesses. Science, research, and government activities have also helped to drive it forward. Just think about analyzing the human genome or dealing with all the astronomical data collected at observatories to advance our understanding of the world around us. Consider the amount of data the government collects in its antiterrorist activities as well, and you get the idea that big data is not just about business.
Different approaches to handling data exist based on whether it is data in motion or data at rest. Here’s a quick example of each. Data in motion would be used if a company is able to analyze the quality of its products during the manufacturing process to avoid costly errors. Data at rest would be used by a business analyst to better understand customers’ current buying patterns based on all aspects of the customer relationship, including sales, social media data, and customer service interactions.
Keep in mind that we are still at an early stage of leveraging huge volumes of data to gain a 360-degree view of the business and anticipate shifts and changes in customer expectations. The technologies required to get the answers the business needs are still isolated from each other. To get to the desired end state, the technologies from all three waves will have to come together. As you will see as you read this book, big data is not simply about one tool or one technology. It is about how all these technologies come together to give the right insights, at the right time, based on the right data — whether it is generated by people, machines, or the web.
Defining Big Data
Big data is not a single technology but a combination of old and new technologies that helps companies gain actionable insight. Therefore, big data is the capability to manage a huge volume of disparate data, at the right speed, and within the right time frame to allow real-time analysis and reaction. As we note earlier in this chapter, big data is typically broken down by three characteristics:
Volume: How much data
Velocity: How fast that data is processed
Variety: The various types of data
Although it’s convenient to simplify big data into the three Vs, it can be misleading and overly simplistic. For example, you may be managing a relatively small amount of very disparate, complex data or you may be processing a huge volume of very simple data. That simple data may be all structured or all unstructured. Even more important is the fourth V: veracity. How accurate is that data in predicting business value? Do the results of a big data analysis actually make sense?
It is critical that you don’t underestimate the task at hand. Data must be able to be verified based on both accuracy and context. An innovative business may want to be able to analyze massive amounts of data in real time to quickly assess the value of that customer and the potential to provide additional offers to that customer. It is necessary to identify the right amount and types of data that can be analyzed to impact business outcomes. Big data incorporates all data, including structured data and unstructured data from e-mail, social media, text streams, and more. This kind of data management requires that companies leverage both their structured and unstructured data.
Building a Successful Big Data Management Architecture
We have moved from an era where an organization could implement a database to meet a specific project need and be done. But as data has become the fuel of growth and innovation, it is more important than ever to have an underlying architecture to support growing requirements.
Beginning with capture, organize, integrate, analyze, and act
Before we delve into the architecture, it is important to take into account the functional requirements for big data. Figure 1-1 illustrates that data must first be captured, and then organized and integrated. After this phase is successfully implemented, data can be analyzed based on the problem being addressed. Finally, management takes action based on the outcome of that analysis. For example, Amazon.com might recommend a book based on a past purchase or a customer might receive a coupon for a discount for a future purchase of a related product to one that was just purchased.
Figure 1-1: The cycle of big data management.
Although this sounds straightforward, certain nuances of these functions are complicated. Validation is a particularly important issue. If your organization is combining data sources, it is critical that you have the ability to validate that these sources make sense when combined. Also, certain data sources may contain sensitive information, so you must implement sufficient levels of security and governance. We cover data management in more detail in Chapter 7.
Of course, any foray into big data first needs to start with the problem you’re trying to solve. That will dictate the kind of data that you need and what the architecture might look like.
Setting the architectural foundation
In addition to supporting the functional requirements, it is important to support the required performance. Your needs will depend on the nature of the analysis you are supporting. You will need the right amount of computational power and speed. While some of the analysis you will do will be performed in real time, you will inevitably be storing some amount of data as well. Your architecture also has to have the right amount of redundancy so that you are protected from unanticipated latency and downtime.
Your organization and its needs will determine how much attention you have to pay to these performance issues. So, start out by asking yourself the following questions:
How much data will my organization need to manage today and in the future?
How often will my organization need to manage data in real time or near real time?
How much risk can my organization afford? Is my industry subject to strict security, compliance, and governance requirements?
How important is speed to my need to manage data?
How certain or precise does the data need to be?
To understand big data, it helps to lay out the components of the architecture. A big data management architecture must include a variety of services that enable companies to make use of myriad data sources in a fast and effective manner. To help you make sense of this, we put the components into a diagram (see Figure 1-2) that will help you see what’s there and the relationship between the components. In the next section, we explain each component and describe how these components are related to each other.
Figure 1-2: The big data architecture.
Interfaces and feeds
Before we get into the nitty-gritty of the big data technology stack itself, we’d like you to notice that on either side of the diagram are indications of interfaces and feeds into and out of both internally managed data and data feeds from external sources. To understand how big data works in the real world, it is important to start by understanding this necessity. In fact, what makes big data big is the fact that it relies on picking up lots of data from lots of sources. Therefore, open application programming interfaces (APIs) will be core to any big data architecture. In addition, keep in mind that interfaces exist at every level and between every layer of the stack. Without integration services, big data can’t happen.
Redundant physical infrastructure
The supporting physical infrastructure is fundamental to the operation and scalability of a big data architecture. In fact, without the availability of robust physical infrastructures, big data would probably not have emerged as such an important trend. To support an unanticipated or unpredictable volume of data, a physical infrastructure for big data has to be different than that for traditional data. The physical infrastructure is based on a distributed computing model. This means that data may be physically stored in many different locations and can be linked together through networks, the use of a distributed file system, and various big data analytic tools and applications.
Redundancy is important because we are dealing with so much data from so many different sources. Redundancy comes in many forms. If your company has created a private cloud, you will want to have redundancy built within the private environment so that it can scale out to support changing workloads. If your company wants to contain internal IT growth, it may use external cloud services to augment its internal resources. In some cases, this redundancy may come in the form of a Software as a Service (SaaS) offering that allows companies to do sophisticated data analysis as a service. The SaaS approach offers lower costs, quicker startup, and seamless evolution of the underlying technology.
Security infrastructure
The more important big data analysis becomes to companies, the more important it will be to secure that data. For example, if you are a healthcare company, you will probably want to use big data applications to determine changes in demographics or shifts in patient needs. This data about your constituents needs to be protected both to meet compliance requirements and to protect the patients’ privacy. You will need to take into account who is allowed to see the data and under what circumstances they are allowed to do so. You will need to be able to verify the identity of users as well as protect the identity of patients. These types of security requirements need to be part of the big data fabric from the outset and not an afterthought.
Operational data sources
When you think about big data, it is important to understand that you have to incorporate all the data sources that will give you a complete picture of your business and see how the data impacts the way you operate your business. Traditionally, an operational data source consisted of highly structured data managed by the line of business in a relational database. But as the world changes, it is important to understand that operational data now has to encompass a broader set of data sources, including unstructured sources such as customer and social media data in all its forms.
You find new emerging approaches to data management in the big data world, including document, graph, columnar, and geospatial database architectures. Collectively, these are referred to as NoSQL, or not only SQL, databases. In essence, you need to map the data architectures to the types of transactions. Doing so will help to ensure the right data is available when you need it. You also need data architectures that support complex unstructured content. You need to include both relational databases and nonrelational databases in your approach to harnessing big data. It is also necessary to include unstructured data sources, such as content management systems, so that you can get closer to that 360-degree business view.
All these operational data sources have several characteristics in common: