Leveraging Natural Supervision: Tailoring Textual Resources for Evaluation Tasks

Written by textmodels | Published 2024/06/01
Tech Story Tags: llm-natural-supervision | llm-self-supervision | llm-language-pretraining | llm-data-to-text-generation | llm-text-summarization | llm-story-generation | llm-story-generation-datasets | llm-summarization-datasets

TLDRIn this study, researchers build evaluation tasks from naturally-occurring textual resources.via the TL;DR App

Author:

(1) Mingda Chen.

Table of Links

CHAPTER 6 - TAILORING TEXTUAL RESOURCES FOR EVALUATION TASKS

This chapter describes our contributions to building evaluation tasks from naturally-occurring textual resources. In Section 6.1, we cast generating arbitrary Wikipedia sections as a data-to-text generation problem. We leverage different data sources to create tabular data for a given section text. In Section 6.2 and Section 6.3, we use fan-contributed websites to create summarization and story generation datasets. Due to the rich information provided on these websites, the resulting datasets offer unique challenges in their respective task settings.

The material in this chapter is adapted from Chen et al. (2022a), Chen et al. (2021), and Chen and Gimpel (2021).

This paper is available on arxiv under CC 4.0 license.


Written by textmodels | We publish the best academic papers on rule-based techniques, LLMs, & the generation of text that resembles human text.
Published by HackerNoon on 2024/06/01