Summary
In this chapter, we discussed the non-functional requirements for data storage solutions. It has become clear that a data lake, which is an evolution of a data warehouse, consists of multiple layers that have their own requirements and thus technology. We have discussed the key requirements for a raw data store where primarily flat files need to be stored in a robust way, for a historical database where temporal information is saved, and for analytics data stores where fast querying is necessary. Furthermore, we have explained the requirements for a streaming data engine and for a model development environment. In all cases, requirements management is an ongoing process in an AI project. Rather than setting all the requirements in stone at the start of the project, architects and developers should be agile, revisiting and revising the requirements after every iteration.
In the next chapter, we will connect the layers of the architecture we have explored in this chapter by creating a data processing pipeline that transforms data from the raw data layer to the historical data layer and to the analytics layer. We will do this to ensure that all the data has been prepared for use in machine learning models. We will also cover data preparation for streaming data scenarios.
- 24小時學會電腦組裝與維護
- 龍芯應用開發標準教程
- INSTANT Wijmo Widgets How-to
- Creating Flat Design Websites
- Machine Learning with Go Quick Start Guide
- Wireframing Essentials
- FL Studio Cookbook
- FreeSWITCH Cookbook
- Mastering Machine Learning on AWS
- 單片機項目設計教程
- Zabbix 4 Network Monitoring
- 微服務架構實戰:基于Spring Boot、Spring Cloud、Docker
- 計算機組裝與維護立體化教程(微課版)
- ActionScript Graphing Cookbook
- Nagios系統監控實踐(原書第2版)