EfficientLLM: Pruning-Aware Pretraining Collection This is the models of our paper "EfficientLLM: Scalable Pruning-Aware Pretraining for Architecture-Agnostic Edge Language Models". • 3 items • Updated 14 days ago
BiPFT: Binary Pre-trained Foundation Transformer with Low-rank Estimation of Binarization Residual Polynomials Paper • 2312.08937 • Published Dec 14, 2023
LM-Cocktail: Resilient Tuning of Language Models via Model Merging Paper • 2311.13534 • Published Nov 22, 2023 • 3