- Statistics for Data Science
- James D. Miller
- 269字
- 2021-07-02 14:58:55
Assessment
When a data scientist evaluates a model or data science process for performance, this is referred to as assessment. Performance can be defined in several ways, including the model's growth of learning or the model's ability to improve (with) learning (to obtain a better score) with additional experience (for example, more rounds of training with additional samples of data) or accuracy of its results.
One popular method of assessing a model or processes performance is called bootstrap sampling. This method examines performance on certain subsets of data, repeatedly generating results that can be used to calculate an estimate of accuracy (performance).
The bootstrap sampling method takes a random sample of data, splits it into three files--a training file, a testing file, and a validation file. The model or process logic is developed based on the data in the training file and then evaluated (or tested) using the testing file. This tune and then test process is repeated until the data scientist is comfortable with the results of the tests. At that point, the model or process is again tested, this time using the validation file, and the results should provide a true indication of how it will perform.
- OpenStack for Architects
- Mastering D3.js
- 電腦上網(wǎng)直通車
- Hands-On Cybersecurity with Blockchain
- 大數(shù)據(jù)平臺異常檢測分析系統(tǒng)的若干關鍵技術研究
- AWS Administration Cookbook
- 人工智能與人工生命
- RPA(機器人流程自動化)快速入門:基于Blue Prism
- 新手學電腦快速入門
- 多媒體制作與應用
- 網(wǎng)絡管理工具實用詳解
- Mastering MongoDB 3.x
- 漢字錄入技能訓練
- 智能+:制造業(yè)的智能化轉(zhuǎn)型
- D3.js Quick Start Guide