Simplify-This: A Comparative Analysis of Prompt-Based and Fine-Tuned LLMs
Abstract
A comparative analysis of fine-tuning and prompt engineering approaches for text simplification using encoder-decoder large language models shows that fine-tuned models provide better structural simplification while prompting achieves higher semantic similarity but with tendency to copy inputs.
Large language models (LLMs) enable strong text generation, and in general there is a practical tradeoff between fine-tuning and prompt engineering. We introduce Simplify-This, a comparative study evaluating both paradigms for text simplification with encoder-decoder LLMs across multiple benchmarks, using a range of evaluation metrics. Fine-tuned models consistently deliver stronger structural simplification, whereas prompting often attains higher semantic similarity scores yet tends to copy inputs. A human evaluation favors fine-tuned outputs overall. We release code, a cleaned derivative dataset used in our study, checkpoints of fine-tuned models, and prompt templates to facilitate reproducibility and future work.
Models citing this paper 9
Browse 9 models citing this paperDatasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper