| 0 |
(example) Neural Image Compression with Text-guided Encoding for both Pixel-level and Perceptual Fidelity |
Hagyeong Lee |
| 1 |
Is Bigger Edit Batch Size Always Better? – An Empirical Study on Model Editing with Llama-3 |
Jin Hyun, Gyuhyun Jung |
| 2 |
Spectrally Pruned Gaussian Fields with Neural Compensation |
Donggeon Lee, Chiho yoon |
| 3 |
Unit Scaling: Out-of-the-Box Low-Precision Training |
SeongRok Moon, Changyoung Ju |
| 4 |
Better & Faster Large Language Models via Multi-token Prediction |
Jinoh Cho, Seonghyeon Park |
| 5 |
Lossless Self-Speculative Decoding via Double Early Exiting |
Nayoung Kwon, Jiwoong Im |
| 6 |
XC-Cache: Cross-Attending to Cached Context for Efficient LLM Inference |
Hyundong Kim, Sangil Han |
| 7 |
VeRA: Vector-based Random Matrix Adaptation |
Kyumin Cho, Sejin Park |
| 8 |
Mixture of LoRA Experts |
Jegwang Ryu, Sangbeom Ha |
| 9 |
MobileNetV4 – Universal Models for the Mobile Ecosystem |
JoonSeok Kim, DongGyu Kim |
| 10 |
Megalodon: Efficient LLM Pretraining and Inference with Unlimited Context Length |
Hyunho Kook |
| 11 |
Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention |
Younghyun Cho, Sangjun Lee |
| 12 |
Scaling (Down) CLIP: A Comprehensive Analysis of Data, Architecture, and Training Strategies |
Junkyeong Park, Harit Keawmuang |
| 13 |
A Large-Scale Exploration of ΞΌ-Transfer |
Jeonghyun Choi, Minhye Choo |
| 14 |
BinaryDM: Towards Accurate Binarization of Diffusion Model |
Junhyuk So, Juncheol Shin |
| 15 |
Training LLMs over Neurally Compressed Text |
Seonghyun Park, Jiwoo Kim |
| 16 |
Mixture-of-Depths: Dynamically allocating compute in transformer-based language models |
Minjae Park, Inkwan Hwang |
| 17 |
QuaRot: Outlier-Free 4-Bit Inference in Rotated LLMs |
MyeongJi Yun, Jung Gyu Min |
| 18 |
ViTAR: Vision Transformer with Any Resolution |
Jungwon Lee, Minsang Seok |
| 19 |
LISA: Layerwise Importance Sampling for Memory-Efficient Large Language Model Fine-Tuning |
Sungbin Shin, Dongyeop Lee |
| 20 |
Evolutionary Optimization of Model Merging Recipes |
Youngkil Song, Jaehyeon Park |
| 21 |
A Unified Framework for Model Editing |
Jonghyun Chae, Donggeun An |
| 22 |
Larimar: Large Language Models with Episodic Memory Control |
Sunggyu Jang, Hyeonwoo Park |
| 23 |
Beyond Language Models: Byte Models are Digital World Simulators |
Dohun Kim, Yeongwoo Kim |
| 24 |
LLMLingua-2: Data Distillation for Efficient and Faithful Task-Agnostic Prompt Compression |
Seungjoo Shin, Sua Choi |
| 25 |
Merging Text Transformer Models from Different Initializations |
Minwoo Kim, Kyungtae Kim |