Skip to content
GitLab
Explore
Sign in
b1075
75923754
·
Better perplexity for 2- and 3-bit quantization for LLaMA-v2-70B (#2807)
·
Aug 26, 2023