peng
superpeng
·
AI & ML interests
None yet
Organizations
None yet
LLM Pretrain
-
How to Train Data-Efficient LLMs
Paper • 2402.09668 • Published • 42 -
Adapting Large Language Models via Reading Comprehension
Paper • 2309.09530 • Published • 81 -
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Paper • 2403.03507 • Published • 189 -
MathScale: Scaling Instruction Tuning for Mathematical Reasoning
Paper • 2403.02884 • Published • 17
reward
LLM Pretrain
-
How to Train Data-Efficient LLMs
Paper • 2402.09668 • Published • 42 -
Adapting Large Language Models via Reading Comprehension
Paper • 2309.09530 • Published • 81 -
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Paper • 2403.03507 • Published • 189 -
MathScale: Scaling Instruction Tuning for Mathematical Reasoning
Paper • 2403.02884 • Published • 17
models
0
None public yet
datasets
0
None public yet