Skip to content

LinkedInLearning/advanced-quantization-techniques-for-large-language-models-4003000

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Advanced Quantization Techniques for Large Language Models

This is the repository for the LinkedIn Learning course Advanced Quantization Techniques for Large Language Models. The full course is available from LinkedIn Learning.

lil-thumbnail-url

Course Description

Discover cutting-edge quantization techniques for large language models, focusing on the algorithms and optimization strategies that deliver the best performance. Instructor Nayan Saxena begins by covering mathematical foundations, before progressing through advanced methods including GPTQ, AWQ, and SmoothQuant with hands-on examples in Google Colab. Along the way, gather quick tips to master critical concepts such as precision formats, calibration strategies, and evaluation methodologies. Leveraging both theoretical principles and practical applications, this course equips you with in-demand skills to significantly reduce model size and accelerate inference while maintaining performance quality.

Learning objectives:

  • Analyze the mathematical foundations of quantization and their impact on transformer architectures.
  • Apply state-of-the-art quantization techniques including GPTQ, AWQ, and SmoothQuant to LLMs.
  • Evaluate the trade-offs between different quantization approaches using appropriate metrics.
  • Optimize quantization results through advanced calibration strategies.
  • Compare and select quantization methods based on model architecture and use case requirements.

Instructor

Nayan Saxena

Deep Learning Expert

Check out my other courses on LinkedIn Learning.

About

This repo is for the LinkedIn Learning course: Advanced Quantization Techniques for Large Language Models

Resources

License

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •