- Hands-On GPU:Accelerated Computer Vision with OpenCV and CUDA
- Bhaumik Vaidya
- 96字
- 2021-08-13 15:48:26
Cache memory
On the latest GPUs, there is an L1 cache per multiprocessor and an L2 cache, which is shared between all multiprocessors. Both global and local memories use these caches. As L1 is near to thread execution, it is very fast. As shown in the diagram for memory architecture earlier, the L1 cache and shared memory use the same 64 KB. Both can be configured for how many bytes they will use out of the 64 KB. All global memory access goes through an L2 cache. Texture memory and constant memory have their separate caches.
推薦閱讀
- Advanced Splunk
- 數(shù)據(jù)庫原理及應(yīng)用(Access版)第3版
- PaaS程序設(shè)計
- PostgreSQL技術(shù)內(nèi)幕:事務(wù)處理深度探索
- Java持續(xù)交付
- Backbone.js Blueprints
- Unity Shader入門精要
- Learning Apache Kafka(Second Edition)
- 精通Python自動化編程
- 第一行代碼 C語言(視頻講解版)
- C++20高級編程
- Struts 2.x權(quán)威指南
- Hacking Android
- Xamarin Cross-Platform Development Cookbook
- Mastering Machine Learning with R