Efficient Unsupervised Sentence Compression by Fine-tuning Transformers with Reinforcement Learning
We design loss functions for unsupervised text-compression that use auxiliary signals for text compression quality, such PLM-derived fluency and consistency with source inputs. The models outperform existing approaches that use discrete-search and are also very efficient at inference time due to a policy-based reinforcement learning training setup, which distills the ensemble of training targets into a single classification decision.
Code