Optimizing the Storing and Processing of Data for Machine Learning Problems
All of the preceding uses for artificial intelligence rely heavily on optimized data storage and processing. Optimization is necessary for machine learning because the data size can be huge, as seen in the following examples:
- A single X-ray file can be many gigabytes in size.
- Translation corpora (large collections of texts) can reach billions of sentences.
- YouTube's stored data is measured in exabytes.
- Financial data might seem like just a few numbers; these are generated in such large quantities per second that the New York Stock Exchange generates 1 TB of data daily.
While every machine learning system is unique, in many systems, data touches the same components. In a hypothetical machine learning system, data might be dealt with as follows:

Figure 1.4: Hardware used in a hypothetical machine learning system
Each of these is a highly specialized piece of hardware, and although not all of them store data for long periods in the way traditional hard disks or tape backups do, it is important to know how data storage can be optimized at each stage. Let's pe into a text classification AI project to see how optimizations can be applied at some stages.
- Windows phone 7.5 application development with F#
- 圖解西門子S7-200系列PLC入門
- 辦公通信設備維修
- 硬件產品經理手冊:手把手構建智能硬件產品
- AMD FPGA設計優化寶典:面向Vivado/SystemVerilog
- Mastering Adobe Photoshop Elements
- 單片機系統設計與開發教程
- Istio服務網格技術解析與實踐
- 筆記本電腦芯片級維修從入門到精通(圖解版)
- 觸摸屏應用技術從入門到精通
- 微服務實戰(Dubbox +Spring Boot+Docker)
- Drupal Rules How-to
- 計算機應用基礎案例教程(Windows 7+Office 2010)
- Service Mesh微服務架構設計
- CPU設計實戰:LoongArch版