舉報

會員
Practical Real-time Data Processing and Analytics
最新章節:
Summary
IfyouareaJavadeveloperwhowouldliketobeequippedwithallthetoolsrequiredtodeviseanend-to-endpracticalsolutiononreal-timedatastreaming,thenthisbookisforyou.Basicknowledgeofreal-timeprocessingwouldbehelpful,andknowingthefundamentalsofMaven,Shell,andEclipsewouldbegreat.
最新章節
- Summary
- Visualization using Kibana
- Start simulator
- Deploy topology
- Generate Vehicle static value
- Load Hazelcast
品牌:中圖公司
上架時間:2021-07-08 09:22:19
出版社:Packt Publishing
本書數字版權由中圖公司提供,并由其授權上海閱文信息技術有限公司制作發行
- Summary 更新時間:2021-07-08 10:23:51
- Visualization using Kibana
- Start simulator
- Deploy topology
- Generate Vehicle static value
- Load Hazelcast
- Running the case study
- Complete Topology
- Elasticsearch Bolt
- Generate alert Bolt
- Check distance and alert bolt
- Parser bolt
- Building Storm topology
- Hazelcast loader
- Building the data simulator
- Implementing the case study
- Setting up the infrastructure
- Tools and frameworks
- Data modeling
- Introduction
- Case Study
- Summary
- DIY
- Graph transformations
- Graph creation
- Graph representation
- Gelly API
- Gelly
- Example
- Selecting from patterns
- Detecting pattern
- Pattern API
- FlinkCEP
- Running example
- Integration with Cassandra
- Flink persistence
- DataSet API
- DataStream API
- Flink processing and computation
- Running example
- Integration with RabbitMQ
- Example
- Integration with Apache Kafka
- Integration of source stream to Flink
- Flink basic components and processes
- Flink architecture and execution engine
- Working with Apache Flink
- Summary
- Connecting Kafka to Spark Streaming
- Spark Streaming operations
- Spark Streaming APIs
- Packaging structure of Spark Streaming
- Spark Streaming - introduction and architecture
- Spark Streaming concepts
- Spark Streaming
- Summary
- Accumulators
- Broadcast variables
- Shared variables – broadcast variables and accumulators
- Actions
- Transformations
- RDD pragmatic exploration
- Spark – packaging and API
- Working with Spark Operations
- Summary
- Spark 2.x – advent of data frames and datasets
- RDD – the name says it all
- Spark pragmatic concepts
- Spark architecture - working inside the engine
- Spark – use cases
- When to avoid using Spark
- Distinct advantages of Spark
- Spark framework and schedulers
- Spark overview
- Working with Spark
- Summary
- Do It Yourself
- DRPC
- Merge and joins
- Grouping
- Aggregator
- Reducer aggregator
- Combiner aggregator
- Persistence aggregate
- Partition aggregate
- Aggregate
- Aggregation
- Sliding window
- Tumbling window
- Windowing
- Filters
- peek
- map and flatMap
- Functions
- Trident operations
- Trident internals
- Basic Storm Trident topology
- Opaque transactional Spout
- Transactional spout
- State retention and the need for Trident
- Storm Trident
- Summary
- Do It Yourself
- Visualizing the output on Grafana
- Executing code
- Writing code
- Adding the Elasticsearch datasource in Grafana
- Running Grafana
- Installing the Elasticsearch plugin in Grafana
- Configuring Grafana
- Downloading Grafana
- Setting up Grafana with the Elasticsearch plugin
- Integrating the presentation layer with Storm
- Storm and IMDB integration for dimensional data
- Storm and Cassandra topology
- Configuring Cassandra
- Setting up Cassandra
- Setting up and configuring Cassandra
- From Storm to Sink
- Summary
- String together Storm-RMQ-PubNub sensor data topology
- PubNub data stream publisher
- AMQPSpout
- RabbitMQ – integration with Storm
- RabbitMQ — publish and subscribe
- RabbitMQ setup
- Headers exchanges
- Topic exchanges
- Fanout exchanges
- Direct exchanges
- RabbitMQ exchanges
- RabbitMQ – messaging that works
- Integrating Storm with a Data Source
- Summary
- Balancing in Apache Beam
- MinimalWordCount example walk through
- Running example
- Beam model
- Setting up and a quick execution of Apache Beam
- Running example
- Download Flink
- Build Flink source
- Setting up and a quick execution of Flink
- Running an example
- Downloading Spark
- Building from source
- Setting up and a quick execution of Spark
- Configuring Apache Spark and Flink
- Summary
- Cluster
- Local
- Running job
- Real-time processing job on Storm
- Running
- Configuring
- Installing
- Setting up Apache Storm
- Running
- Cluster
- Standalone
- Configuring
- Installing
- Setting up Zookeeper
- Setting up and configuring Storm
- Stream grouping
- Components
- Characteristics
- Storm architecture and its components
- Overview of Storm
- Setting up the Infrastructure for Storm
- Summary
- Setting up Elasticsearch
- Do it yourself
- Comparing and choosing what works best for your use case
- Taping data from source to the processor - expectations and caveats
- Flume
- Fluentd
- Logstash
- Apache NiFi
- Apache Kafka
- Setting up infrastructure for data ingestion
- Understanding data streams
- Understanding and Tailing Data Streams
- Summary
- Storage
- Transformation and processing
- Broker
- Collection
- Event producer
- NRT – technology view
- NRT – high-level system view
- Analytical layer – serve it to the end user
- Stream processing
- Data collection
- The NRT system and its building blocks
- Real Time Applications – The Basic Ingredients
- Summary
- Cloud – considerations for NRT and IOT
- Edge analytics
- IOT – thoughts and possibilities
- Lambda architecture – analytics possibilities
- NRT – The Spark solution
- NRT – The Storm solution
- Near real–time solution – an architecture that works
- Real–time analytics – the myth and the reality
- Big data infrastructure
- What is big data?
- Introducing Real-Time Analytics
- Questions
- Piracy
- Errata
- Downloading the example code
- Customer support
- Reader feedback
- Conventions
- Who this book is for
- What you need for this book
- What this book covers
- Preface
- Customer Feedback
- Why subscribe?
- www.PacktPub.com
- About the Reviewers
- About the Authors
- Credits
- Practical Real-Time Data Processing and Analytics
- Copyright
- Title Page
- coverpage
- coverpage
- Title Page
- Copyright
- Practical Real-Time Data Processing and Analytics
- Credits
- About the Authors
- About the Reviewers
- www.PacktPub.com
- Why subscribe?
- Customer Feedback
- Preface
- What this book covers
- What you need for this book
- Who this book is for
- Conventions
- Reader feedback
- Customer support
- Downloading the example code
- Errata
- Piracy
- Questions
- Introducing Real-Time Analytics
- What is big data?
- Big data infrastructure
- Real–time analytics – the myth and the reality
- Near real–time solution – an architecture that works
- NRT – The Storm solution
- NRT – The Spark solution
- Lambda architecture – analytics possibilities
- IOT – thoughts and possibilities
- Edge analytics
- Cloud – considerations for NRT and IOT
- Summary
- Real Time Applications – The Basic Ingredients
- The NRT system and its building blocks
- Data collection
- Stream processing
- Analytical layer – serve it to the end user
- NRT – high-level system view
- NRT – technology view
- Event producer
- Collection
- Broker
- Transformation and processing
- Storage
- Summary
- Understanding and Tailing Data Streams
- Understanding data streams
- Setting up infrastructure for data ingestion
- Apache Kafka
- Apache NiFi
- Logstash
- Fluentd
- Flume
- Taping data from source to the processor - expectations and caveats
- Comparing and choosing what works best for your use case
- Do it yourself
- Setting up Elasticsearch
- Summary
- Setting up the Infrastructure for Storm
- Overview of Storm
- Storm architecture and its components
- Characteristics
- Components
- Stream grouping
- Setting up and configuring Storm
- Setting up Zookeeper
- Installing
- Configuring
- Standalone
- Cluster
- Running
- Setting up Apache Storm
- Installing
- Configuring
- Running
- Real-time processing job on Storm
- Running job
- Local
- Cluster
- Summary
- Configuring Apache Spark and Flink
- Setting up and a quick execution of Spark
- Building from source
- Downloading Spark
- Running an example
- Setting up and a quick execution of Flink
- Build Flink source
- Download Flink
- Running example
- Setting up and a quick execution of Apache Beam
- Beam model
- Running example
- MinimalWordCount example walk through
- Balancing in Apache Beam
- Summary
- Integrating Storm with a Data Source
- RabbitMQ – messaging that works
- RabbitMQ exchanges
- Direct exchanges
- Fanout exchanges
- Topic exchanges
- Headers exchanges
- RabbitMQ setup
- RabbitMQ — publish and subscribe
- RabbitMQ – integration with Storm
- AMQPSpout
- PubNub data stream publisher
- String together Storm-RMQ-PubNub sensor data topology
- Summary
- From Storm to Sink
- Setting up and configuring Cassandra
- Setting up Cassandra
- Configuring Cassandra
- Storm and Cassandra topology
- Storm and IMDB integration for dimensional data
- Integrating the presentation layer with Storm
- Setting up Grafana with the Elasticsearch plugin
- Downloading Grafana
- Configuring Grafana
- Installing the Elasticsearch plugin in Grafana
- Running Grafana
- Adding the Elasticsearch datasource in Grafana
- Writing code
- Executing code
- Visualizing the output on Grafana
- Do It Yourself
- Summary
- Storm Trident
- State retention and the need for Trident
- Transactional spout
- Opaque transactional Spout
- Basic Storm Trident topology
- Trident internals
- Trident operations
- Functions
- map and flatMap
- peek
- Filters
- Windowing
- Tumbling window
- Sliding window
- Aggregation
- Aggregate
- Partition aggregate
- Persistence aggregate
- Combiner aggregator
- Reducer aggregator
- Aggregator
- Grouping
- Merge and joins
- DRPC
- Do It Yourself
- Summary
- Working with Spark
- Spark overview
- Spark framework and schedulers
- Distinct advantages of Spark
- When to avoid using Spark
- Spark – use cases
- Spark architecture - working inside the engine
- Spark pragmatic concepts
- RDD – the name says it all
- Spark 2.x – advent of data frames and datasets
- Summary
- Working with Spark Operations
- Spark – packaging and API
- RDD pragmatic exploration
- Transformations
- Actions
- Shared variables – broadcast variables and accumulators
- Broadcast variables
- Accumulators
- Summary
- Spark Streaming
- Spark Streaming concepts
- Spark Streaming - introduction and architecture
- Packaging structure of Spark Streaming
- Spark Streaming APIs
- Spark Streaming operations
- Connecting Kafka to Spark Streaming
- Summary
- Working with Apache Flink
- Flink architecture and execution engine
- Flink basic components and processes
- Integration of source stream to Flink
- Integration with Apache Kafka
- Example
- Integration with RabbitMQ
- Running example
- Flink processing and computation
- DataStream API
- DataSet API
- Flink persistence
- Integration with Cassandra
- Running example
- FlinkCEP
- Pattern API
- Detecting pattern
- Selecting from patterns
- Example
- Gelly
- Gelly API
- Graph representation
- Graph creation
- Graph transformations
- DIY
- Summary
- Case Study
- Introduction
- Data modeling
- Tools and frameworks
- Setting up the infrastructure
- Implementing the case study
- Building the data simulator
- Hazelcast loader
- Building Storm topology
- Parser bolt
- Check distance and alert bolt
- Generate alert Bolt
- Elasticsearch Bolt
- Complete Topology
- Running the case study
- Load Hazelcast
- Generate Vehicle static value
- Deploy topology
- Start simulator
- Visualization using Kibana
- Summary 更新時間:2021-07-08 10:23:51