Corpus Conscious Coaching (CAT) leverages helpful corpus metadata throughout coaching by injecting corpus info into every coaching instance, and has been discovered efficient within the literature, generally often called the “tagging” strategy. Fashions educated with CAT inherently be taught the standard, area and nuance between corpora straight from knowledge, and may simply swap to totally different inference conduct. To attain the perfect analysis, CAT fashions pre-define a gaggle of top of the range knowledge earlier than coaching begins which may be error-prone and inefficient. On this work, we suggest Optimum Corpus Conscious Coaching (OCAT), which fine-tunes a CAT pre-trained mannequin by freezing a lot of the mannequin parameters and solely tuning small set of corpus-related parameters. We present that OCAT is light-weight, resilient to overfitting, and efficient in boosting mannequin accuracy. We use WMT23 English to Chinese language and English to German translation duties as our take a look at floor and present +3.6 and +1.8 chrF enchancment, respectively, over vanilla coaching. Moreover, our strategy is on-par or barely higher than different state-of-the-art fine-tuning methods whereas being much less delicate to hyperparameter settings.







