最新章節
- Index
- Sensitive data masking and encryption using Hadoop
- Web log analytics
- Call Data Record analytics
- Introduction
- Chapter 10. Hadoop Use Cases
品牌:中圖公司
上架時間:2021-07-09 18:22:03
出版社:Packt Publishing
本書數字版權由中圖公司提供,并由其授權上海閱文信息技術有限公司制作發行
- Index 更新時間:2021-07-09 20:03:08
- Sensitive data masking and encryption using Hadoop
- Web log analytics
- Call Data Record analytics
- Introduction
- Chapter 10. Hadoop Use Cases
- Conducting predictive analytics using Spark MLib
- Processing graphs using Graph X
- Analyzing JSON data using Spark
- Analyzing Parquet files using Spark
- Twitter trending topics using Spark streaming
- Creating Twitter trending topics using Spark Streaming
- Performing Olympics Athletes analytics using the Spark Shell
- Running Spark on YARN
- Running Spark standalone
- Introduction
- Chapter 9. Integration with Apache Spark
- Performing Predictive Analytics using R
- Twitter Sentiment Analytics using R
- Population Data Analytics using R
- Text data clustering using K-Means using Mahout
- Predictive analytics on Bank Data using Mahout
- Creating a user-based recommendation engine using Mahout
- Creating an item-based recommendation engine using Mahout
- Setting up the Mahout development environment
- Introduction
- Chapter 8. Machine Learning and Predictive Analytics Using Mahout and R
- Scheduling a job in Oozie
- Executing parallel jobs using Oozie (fork)
- Implementing an e-mail action job using Oozie
- Implementing a Pig action job using Oozie
- Implementing a Hive action job using Oozie
- Implementing a Java action job using Oozie
- Implementing a Map Reduce action job using Oozie
- Implementing a Sqoop action job using Oozie
- Introduction
- Chapter 7. Automation of Hadoop Tasks Using Oozie
- Importing web logs data into HDFS using Flume
- Importing data from Kafka into HDFS using Flume
- Importing Twitter data into HDFS using Flume
- Importing data from RDBMS to Hbase using Sqoop
- Creating and executing Sqoop job
- Incremental import using Sqoop
- Importing data into HDFS from Mainframes
- Importing data into Hive tables using Sqoop
- Performing Atomic export using Sqoop
- Importing data using Sqoop in compressed format
- Using query operator in Sqoop import
- Exporting data from HDFS to RDBMS
- Importing data from RDMBS to HDFS using Sqoop
- Introduction
- Chapter 6. Data Import/Export Using Sqoop and Flume
- Multiple table inserting using Hive
- Implementing Change Data Capture using Hive
- Twitter sentiment analysis using Hive
- Call Data Record Analytics using Hive
- Performing context Ngram in Hive
- Executing map side joins in Hive
- Performing table joins in Hive
- Writing a user-defined function in Hive
- Processing Hive data in the Avro format
- Processing XML data in Hive using XML SerDe
- Processing JSON data in Hive using JSON SerDe
- Introduction
- Chapter 5. Advanced Data Analysis Using Hive
- Executing the MapReduce programming with an Hbase Table
- Performing Hbase operations in Java
- Performing the Hbase operation in CLI
- Analyzing web log data using Pig
- Writing a user-defined function in Pig
- Performing JOINS in Pig
- Performing Order By queries in Pig
- Performing Group By queries in Pig
- Performing FILTER By queries in Pig
- Storing and processing Hive data in the Parquet file format
- Storing and processing Hive data in the ORC file format
- Storing and processing Hive data in the RC file format
- Storing and processing Hive data in a sequential file format
- Introduction
- Chapter 4. Data Analysis Using Hive Pig and Hbase
- Unit testing the Map Reduce code using MRUnit
- Performing Reduce side Joins using Map Reduce
- Writing Map Reduce results to multiple output files
- Map Reduce program to partition data using a custom partitioner
- Map Reduce program to find distinct values
- Map Reduce program to find the top X
- Implementing a user-defined counter in a Map Reduce program
- Adding support for a new writable data type in Hadoop
- Executing the Map Reduce program in a Hadoop cluster
- Writing the Map Reduce program in Java to analyze web log data
- Introduction
- Chapter 3. Mastering Map Reduce Programs
- Saving compressed data in HDFS
- Recycling deleted data from trash to HDFS
- Importing data from another Hadoop cluster
- Enabling transparent encryption for HDFS
- Setting the HDFS block size for a specific file in a cluster
- Setting the HDFS block size for all the files in a cluster
- Changing the replication factor of an existing file in HDFS
- Exporting HDFS data to a local machine
- Loading data from a local machine to HDFS
- Introduction
- Chapter 2. Exploring HDFS
- Performing benchmarking on a Hadoop cluster
- Decommissioning DataNodes
- Entering and exiting from the safe mode in a Hadoop cluster
- Executing the balancer command for uniform data distribution
- Adding new nodes to existing Hadoop clusters
- Installing a multi-node Hadoop cluster
- Installing a single-node Hadoop Cluster
- Introduction
- Chapter 1. Getting Started with Hadoop 2.X
- Preface
- www.PacktPub.com
- About the Reviewer
- Acknowledgements
- About the Author
- Credits
- 版權信息
- 封面
- 封面
- 版權信息
- Credits
- About the Author
- Acknowledgements
- About the Reviewer
- www.PacktPub.com
- Preface
- Chapter 1. Getting Started with Hadoop 2.X
- Introduction
- Installing a single-node Hadoop Cluster
- Installing a multi-node Hadoop cluster
- Adding new nodes to existing Hadoop clusters
- Executing the balancer command for uniform data distribution
- Entering and exiting from the safe mode in a Hadoop cluster
- Decommissioning DataNodes
- Performing benchmarking on a Hadoop cluster
- Chapter 2. Exploring HDFS
- Introduction
- Loading data from a local machine to HDFS
- Exporting HDFS data to a local machine
- Changing the replication factor of an existing file in HDFS
- Setting the HDFS block size for all the files in a cluster
- Setting the HDFS block size for a specific file in a cluster
- Enabling transparent encryption for HDFS
- Importing data from another Hadoop cluster
- Recycling deleted data from trash to HDFS
- Saving compressed data in HDFS
- Chapter 3. Mastering Map Reduce Programs
- Introduction
- Writing the Map Reduce program in Java to analyze web log data
- Executing the Map Reduce program in a Hadoop cluster
- Adding support for a new writable data type in Hadoop
- Implementing a user-defined counter in a Map Reduce program
- Map Reduce program to find the top X
- Map Reduce program to find distinct values
- Map Reduce program to partition data using a custom partitioner
- Writing Map Reduce results to multiple output files
- Performing Reduce side Joins using Map Reduce
- Unit testing the Map Reduce code using MRUnit
- Chapter 4. Data Analysis Using Hive Pig and Hbase
- Introduction
- Storing and processing Hive data in a sequential file format
- Storing and processing Hive data in the RC file format
- Storing and processing Hive data in the ORC file format
- Storing and processing Hive data in the Parquet file format
- Performing FILTER By queries in Pig
- Performing Group By queries in Pig
- Performing Order By queries in Pig
- Performing JOINS in Pig
- Writing a user-defined function in Pig
- Analyzing web log data using Pig
- Performing the Hbase operation in CLI
- Performing Hbase operations in Java
- Executing the MapReduce programming with an Hbase Table
- Chapter 5. Advanced Data Analysis Using Hive
- Introduction
- Processing JSON data in Hive using JSON SerDe
- Processing XML data in Hive using XML SerDe
- Processing Hive data in the Avro format
- Writing a user-defined function in Hive
- Performing table joins in Hive
- Executing map side joins in Hive
- Performing context Ngram in Hive
- Call Data Record Analytics using Hive
- Twitter sentiment analysis using Hive
- Implementing Change Data Capture using Hive
- Multiple table inserting using Hive
- Chapter 6. Data Import/Export Using Sqoop and Flume
- Introduction
- Importing data from RDMBS to HDFS using Sqoop
- Exporting data from HDFS to RDBMS
- Using query operator in Sqoop import
- Importing data using Sqoop in compressed format
- Performing Atomic export using Sqoop
- Importing data into Hive tables using Sqoop
- Importing data into HDFS from Mainframes
- Incremental import using Sqoop
- Creating and executing Sqoop job
- Importing data from RDBMS to Hbase using Sqoop
- Importing Twitter data into HDFS using Flume
- Importing data from Kafka into HDFS using Flume
- Importing web logs data into HDFS using Flume
- Chapter 7. Automation of Hadoop Tasks Using Oozie
- Introduction
- Implementing a Sqoop action job using Oozie
- Implementing a Map Reduce action job using Oozie
- Implementing a Java action job using Oozie
- Implementing a Hive action job using Oozie
- Implementing a Pig action job using Oozie
- Implementing an e-mail action job using Oozie
- Executing parallel jobs using Oozie (fork)
- Scheduling a job in Oozie
- Chapter 8. Machine Learning and Predictive Analytics Using Mahout and R
- Introduction
- Setting up the Mahout development environment
- Creating an item-based recommendation engine using Mahout
- Creating a user-based recommendation engine using Mahout
- Predictive analytics on Bank Data using Mahout
- Text data clustering using K-Means using Mahout
- Population Data Analytics using R
- Twitter Sentiment Analytics using R
- Performing Predictive Analytics using R
- Chapter 9. Integration with Apache Spark
- Introduction
- Running Spark standalone
- Running Spark on YARN
- Performing Olympics Athletes analytics using the Spark Shell
- Creating Twitter trending topics using Spark Streaming
- Twitter trending topics using Spark streaming
- Analyzing Parquet files using Spark
- Analyzing JSON data using Spark
- Processing graphs using Graph X
- Conducting predictive analytics using Spark MLib
- Chapter 10. Hadoop Use Cases
- Introduction
- Call Data Record analytics
- Web log analytics
- Sensitive data masking and encryption using Hadoop
- Index 更新時間:2021-07-09 20:03:08