Segui
Minsoo Kim
Minsoo Kim
Email verificata su hanyang.ac.kr - Home page
Titolo
Citata da
Citata da
Anno
NN-LUT: Neural approximation of non-linear operations for efficient transformer inference
J Yu, J Park, S Park, M Kim, S Lee, DH Lee, J Choi
Proceedings of the 59th ACM/IEEE Design Automation Conference, 577-582, 2022
242022
Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders
M Kim, S Lee, S Hong, DS Chang, J Choi
Proceedings of the 2022 Conference on Empirical Methods in Natural Language …, 2022
82022
Token-scaled logit distillation for ternary weight generative language models
M Kim, S Lee, J Lee, S Hong, DS Chang, W Sung, J Choi
Advances in Neural Information Processing Systems 36, 2023
62023
Enhancing computation efficiency in large language models through weight and activation quantization
J Lee, M Kim, S Baek, SJ Hwang, W Sung, J Choi
Proceedings of the 2023 Conference on Empirical Methods in Natural Language …, 2023
32023
Teacher Intervention: Improving Convergence of Quantization Aware Training for Ultra-Low Precision Transformers
M Kim, K Shim, S Park, W Sung, J Choi
Proceedings of the 17th Conference of the European Chapter of the …, 2023
12023
Il sistema al momento non può eseguire l'operazione. Riprova più tardi.
Articoli 1–5