File tree Expand file tree Collapse file tree 4 files changed +86
-0
lines changed
Expand file tree Collapse file tree 4 files changed +86
-0
lines changed Original file line number Diff line number Diff line change @@ -14,3 +14,13 @@ This chapter has the following learning objectives:
1414
15153 . Understand the typical techniques used to optimize the performance
1616 of accelerators.
17+
18+ ``` toc
19+ :maxdepth: 2
20+
21+ Overview
22+ Components_of_Hardware_Accelerators
23+ Programming_Methods
24+ Performance_Optimization_Methods
25+ Chapter_Summary
26+ ```
Original file line number Diff line number Diff line change 1+ # Hardware Accelerator
2+
3+ In the field of AI frameworks, hardware accelerators play a vital role
4+ in enabling efficient neural network computations. This chapter delves
5+ into the design of modern hardware accelerators, their programming
6+ techniques, and the typical approaches to optimize accelerator
7+ performance.
8+
9+ This chapter has the following learning objectives:
10+
11+ 1 . Understand the architecture of a modern hardware accelerator.
12+
13+ 2 . Understand the methods of programming hardware accelerators.
14+
15+ 3 . Understand the typical techniques used to optimize the performance
16+ of accelerators.
17+
18+ ``` toc
19+ :maxdepth: 2
20+
21+ Overview
22+ Components_of_Hardware_Accelerators
23+ Programming_Methods
24+ Performance_Optimization_Methods
25+ Chapter_Summary
26+ ```
Original file line number Diff line number Diff line change @@ -20,3 +20,17 @@ The key aspects this chapter explores are as follows:
2020 optimization.
2121
22224 . Common methods for model security protection.
23+
24+ ``` toc
25+ :maxdepth: 2
26+
27+ Overview
28+ Conversion_to_Inference_Model_and_Model_Optimization
29+ Model_Compression
30+ Advanced_Efficient_Techniques
31+ Model_Inference
32+ Security_Protection_of_Models
33+ Chapter_Summary
34+ Further_Reading
35+ ```
36+
Original file line number Diff line number Diff line change 1+ # Model Deployment {#ch: deploy }
2+
3+ In earlier chapters, we discussed the basic components of the machine
4+ learning model training system. In this chapter, we look at the basics
5+ of model deployment, a process whereby a trained model is deployed in a
6+ runtime environment for inference. We explore the conversion from a
7+ training model into an inference model, model compression methods that
8+ adapt to hardware restrictions, model inference and performance
9+ optimization, and model security protection.
10+
11+ The key aspects this chapter explores are as follows:
12+
13+ 1 . Conversion and optimization from a training model to an inference
14+ model.
15+
16+ 2 . Common methods for model compression: quantization, sparsification,
17+ and knowledge distillation.
18+
19+ 3 . Model inference process and common methods for performance
20+ optimization.
21+
22+ 4 . Common methods for model security protection.
23+
24+ ``` toc
25+ :maxdepth: 2
26+
27+ Overview
28+ Conversion_to_Inference_Model_and_Model_Optimization
29+ Model_Compression
30+ Advanced_Efficient_Techniques
31+ Model_Inference
32+ Security_Protection_of_Models
33+ Chapter_Summary
34+ Further_Reading
35+ ```
36+
You can’t perform that action at this time.
0 commit comments