- Apache Spark Machine Learning Blueprints
- Alex Liu
- 184字
- 2021-07-16 10:39:48
Spark computing for machine learning
With its innovations on RDD and in-memory processing, Apache Spark has truly made distributed computing easily accessible to data scientists and machine learning professionals. According to the Apache Spark team, Apache Spark runs on the Mesos cluster manager, letting it share resources with Hadoop and other applications. Therefore, Apache Spark can read from any Hadoop input source like HDFS.

For the above, the Apache Spark computing model is very suitable to distributed computing for machine learning. Especially for rapid interactive machine learning, parallel computing, and complicated modelling at scale, Apache Spark should definitely be utilized.
According to the Spark development team, Spark's philosophy is to make life easy and productive for data scientists and machine learning professionals. Due to this, Apache Spark has:
- Well documented, expressive API's
- Powerful domain specific libraries
- Easy integration with storage systems
- Caching to avoid data movement
Per the introduction by Patrick Wendell, co-founder of Databricks, Spark is especially made for large scale data processing. Apache Spark supports agile data science to iterate rapidly, and Spark can be integrated with IBM and other solutions easily.
- 協作機器人技術及應用
- Hands-On Cybersecurity with Blockchain
- 機器學習流水線實戰
- 工業機器人應用案例集錦
- Google SketchUp for Game Design:Beginner's Guide
- C++程序設計基礎(上)
- Mastering Geospatial Analysis with Python
- Mastering Ansible(Second Edition)
- 機器學習案例分析(基于Python語言)
- 基于Proteus的PIC單片機C語言程序設計與仿真
- 案例解說Delphi典型控制應用
- 大話數據科學:大數據與機器學習實戰(基于R語言)
- 我的IT世界
- ARM? Cortex? M4 Cookbook
- Azure Serverless Computing Cookbook