Goto

Collaborating Authors

 linguistics





e4d2b6e6fdeca3e60e0f1a62fee3d9dd-Paper.pdf

Neural Information Processing Systems

AwidevarietyofNLPapplications, suchasmachinetranslation, summarization, and dialog, involve text generation. One major challenge for these applications is how to evaluate whether such generated texts are actually fluent, accurate, or effective. In this work, we conceptualize theevaluation of generated text as a text generation problem, modeled using pre-trained sequence-to-sequence models. The general idea is that models trained to convert the generated text to/from a reference output or the source text will achieve higher scores when the generated text is better.





BMoreExperimentalSetups

Neural Information Processing Systems

Example Reweightingdirectly assigns an importance weight to the standard CE training loss, accordingtothebiasdegreeβ: Lreweight = (1 β)y logpm (3) Confidence Regularizationis based on knowledge distillation [9]. It involves a teacher model trainedwiththestandardCEloss. Specifically, we calculate the weighted average of the F1 score of each class. The splits used for evaluation are highlightedwithredcolor. To address this problem, we select the best checkpoint after0.7 tmax of training, butstill according to the performance on the ID devset.