-
horseee authored
[ADD] SKIM: Any-bit Quantization Pushing The Limits of Post-Training Quantization & CPTQuant -- A Novel Mixed Precision Post-Training Quantization Techniques for Large Language Models & Efficient LLM Inference using Dynamic Input Pruning and Cache-Aware Masking
589c1433horseee authored[ADD] SKIM: Any-bit Quantization Pushing The Limits of Post-Training Quantization & CPTQuant -- A Novel Mixed Precision Post-Training Quantization Techniques for Large Language Models & Efficient LLM Inference using Dynamic Input Pruning and Cache-Aware Masking
Loading