|
591 | 591 | "outputs": [], |
592 | 592 | "source": [ |
593 | 593 | "# 1. Basic Post-Training Quantization for PyTorch\n", |
594 | | - "flag, quantized_model = PTQ_Pytorch(float_model)" |
| 594 | + "flag, quantized_mode_ptq = PTQ_Pytorch(float_model)" |
595 | 595 | ] |
596 | 596 | }, |
597 | 597 | { |
|
602 | 602 | "outputs": [], |
603 | 603 | "source": [ |
604 | 604 | "# 2. PTQ with Mixed Precision (optimized size/accuracy trade-off for PyTorch)\n", |
605 | | - "flag, quantized_model2 = PTQ_Pytorch_mixed_precision(float_model)" |
| 605 | + "flag, quantized_model_ptq_mixed_precision = PTQ_Pytorch_mixed_precision(float_model)" |
606 | 606 | ] |
607 | 607 | }, |
608 | 608 | { |
|
613 | 613 | "outputs": [], |
614 | 614 | "source": [ |
615 | 615 | "# 3. Gradient-based PTQ (improved accuracy through fine-tuning for PyTorch)\n", |
616 | | - "flag, quantized_model3 = GPTQ_Pytorch(float_model)" |
| 616 | + "flag, quantized_model_gptq = GPTQ_Pytorch(float_model)" |
617 | 617 | ] |
618 | 618 | }, |
619 | 619 | { |
|
624 | 624 | "outputs": [], |
625 | 625 | "source": [ |
626 | 626 | "# 4. GPTQ with Mixed Precision (best accuracy with optimal compression for PyTorch)\n", |
627 | | - "flag, quantized_model4 = GPTQ_Pytorch_mixed_precision(float_model)" |
| 627 | + "flag, quantized_model_gptq_mixed_precision = GPTQ_Pytorch_mixed_precision(float_model)" |
628 | 628 | ] |
629 | 629 | }, |
630 | 630 | { |
|
673 | 673 | "source": [ |
674 | 674 | "# Evaluate PTQ quantized PyTorch model accuracy\n", |
675 | 675 | "print(\"\\n=== PyTorch PTQ Model Evaluation ===\")\n", |
676 | | - "evaluate(quantized_model, val_dataloader, 'PTQ_Pytorch')" |
| 676 | + "evaluate(quantized_mode_ptq, val_dataloader, 'PTQ_Pytorch')" |
677 | 677 | ] |
678 | 678 | }, |
679 | 679 | { |
|
685 | 685 | "source": [ |
686 | 686 | "# Evaluate PTQ + Mixed Precision PyTorch model accuracy\n", |
687 | 687 | "print(\"\\n=== PyTorch PTQ + Mixed Precision Model Evaluation ===\")\n", |
688 | | - "evaluate(quantized_model2, val_dataloader, 'PTQ_Pytorch_mixed_precision')" |
| 688 | + "evaluate(quantized_model_ptq_mixed_precision, val_dataloader, 'PTQ_Pytorch_mixed_precision')" |
689 | 689 | ] |
690 | 690 | }, |
691 | 691 | { |
|
697 | 697 | "source": [ |
698 | 698 | "# Evaluate GPTQ quantized PyTorch model accuracy\n", |
699 | 699 | "print(\"\\n=== PyTorch GPTQ Model Evaluation ===\")\n", |
700 | | - "evaluate(quantized_model3, val_dataloader, 'GPTQ_Pytorch')" |
| 700 | + "evaluate(quantized_model_gptq, val_dataloader, 'GPTQ_Pytorch')" |
701 | 701 | ] |
702 | 702 | }, |
703 | 703 | { |
|
709 | 709 | "source": [ |
710 | 710 | "# Evaluate GPTQ + Mixed Precision PyTorch model accuracy\n", |
711 | 711 | "print(\"\\n=== PyTorch GPTQ + Mixed Precision Model Evaluation ===\")\n", |
712 | | - "evaluate(quantized_model4, val_dataloader, 'GPTQ_Pytorch_mixed_precision')" |
| 712 | + "evaluate(quantized_model_gptq_mixed_precision, val_dataloader, 'GPTQ_Pytorch_mixed_precision')" |
713 | 713 | ] |
714 | 714 | }, |
715 | 715 | { |
|
0 commit comments