NoticIA-it / README.md
Iker's picture
Update README.md
335a647 verified
metadata
language:
  - es
size_categories:
  - n<1K
task_categories:
  - summarization
pretty_name: Resumen Noticias Clickbait
dataset_info:
  features:
    - name: id
      dtype: int64
    - name: titular
      dtype: string
    - name: respuesta
      dtype: string
    - name: pregunta
      dtype: string
    - name: texto
      dtype: string
    - name: idioma
      dtype: string
    - name: periodo
      dtype: string
    - name: tarea
      dtype: string
    - name: registro
      dtype: string
    - name: dominio
      dtype: string
    - name: país_origen
      dtype: string
  splits:
    - name: train
      num_bytes: 5440051
      num_examples: 700
    - name: validation
      num_bytes: 462364
      num_examples: 50
    - name: test
      num_bytes: 782440
      num_examples: 100
  download_size: 3417692
  dataset_size: 6684855
configs:
  - config_name: default
    data_files:
      - split: train
        path: data/train-*
      - split: validation
        path: data/validation-*
      - split: test
        path: data/test-*
tags:
  - summarization
  - clickbait
  - news

NoticIA: A Clickbait Article Summarization Dataset in Spanish.

We present NoticIA, a dataset consisting of 850 Spanish news articles featuring prominent clickbait headlines, each paired with high-quality, single-sentence generative summarizations written by humans.

Dataset Details

Dataset Description

We define a clickbait article as one that seeks to attract the reader's attention through curiosity. For this purpose, the headline poses a question or an incomplete, sensationalist, exaggerated, or misleading statement. The answer to the question raised in the headline usually does not appear until the end of the article, preceded by a large amount of irrelevant content. The goal is for the user to enter the website through the headline and then scroll to the end of the article, viewing as much advertising as possible. Clickbait articles tend to be of low quality and provide no value to the reader beyond the initial curiosity. This phenomenon undermines public trust in news sources and negatively affects the advertising revenue of legitimate content creators, who could see their web traffic reduced.

We introduce NoticIA, a dataset consisting of 850 Spanish news articles with clickbait headlines, each paired with high-quality, single-sentence generative summaries written by humans. This task demands advanced skills in text comprehension and summarization, challenging the ability of models to infer and connect various pieces of information to satisfy the user's informational curiosity generated by the clickbait headline.

The project is inspired by the X/Twitter account @ahorrandoclick1. @ahorrandoclick1 has 300,000 followers, demonstrating the great value of summarizing clickbait news articles. However, creating these summaries manually is a labor-intensive task, and the number of clickbait news articles published greatly exceeds the number of summaries one person can perform. Therefore, there is a need for automatic summarization of clickbait news articles. Additionally, as mentioned earlier, this is an ideal task for analyzing the text comprehension capabilities of a language model in Spanish.

The following Figure illustrates examples of clickbait headlines from our dataset, together with the human-written summaries.

Dataset Sources

Uses

This dataset has been compiled for use in scientific research. Specifically, for use in the evaluation of language models in Spanish. Commercial use of this dataset is subject to the licenses of each news and media outlet. If you want to make commercial use of the dataset you will need to have the express permission of the media from which the news has been obtained.

Direct Use

  • 📈 Evaluation of Language Models in Spanish.
  • 🤖 Instruction-Tuning of Spanish Language Models
  • 📚 Develop new datasets on top of our data
  • 🎓 Any other academic research purpose.

Out-of-Scope Use

We expressly prohibit the use of these data for two use cases that we consider to be that may be harmful: The training of models that generate sensational headlines or clickbait, and the training of models that generate articles or news automatically.

Dataset Structure

The dataset is ready to be used to evaluate language models. For this aim, we have developed a prompt that makes use of the news headline and text. The prompt is as follows:

def clickbait_prompt(
    headline: str,
    body: str,
) -> str:
    """
    Generate the prompt for the model.
    Args:
        headline (`str`):
            The headline of the article.
        body (`str`):
            The body of the article.
    Returns:
        `str`: The formatted prompt.
    """
    return (
        f"Ahora eres una Inteligencia Artificial experta en desmontar titulares sensacionalistas o clickbait. "
        f"Tu tarea consiste en analizar noticias con titulares sensacionalistas y "
        f"generar un resumen de una sola frase que revele la verdad detrás del titular.\n"
        f"Este es el titular de la noticia: {headline}\n"
        f"El titular plantea una pregunta o proporciona información incompleta. "
        f"Debes buscar en el cuerpo de la noticia una frase que responda lo que se sugiere en el título. "
        f"Responde siempre que puedas parafraseando el texto original. "
        f"Usa siempre las mínimas palabras posibles. "
        f"Recuerda responder siempre en Español.\n"
        f"Este es el cuerpo de la noticia:\n"
        f"{body}\n"
    )

The expected output of the model is the summary. Below is an example of how to evaluate gemma-2b in our dataset:

from transformers import pipeline
from datasets import load_dataset

generator = pipeline(model="google/gemma-2b-it",device_map="auto")
dataset = load_dataset("somosnlp/NoticIA-it",split="test")

outputs = generator(dataset[0]["prompt"], return_full_text=False,max_length=4096)
print(outputs)

The dataset includes the following fields:

  • ID: id of the example
  • Titular (headline): headline of the article
  • Respuesta (response): Summary written by a human being
  • Pregunta (question): Prompt ready to be used as input to a language model.
  • Texto (text): Text of the article, obtained from the HTML.
  • idioma (language): ISO code of the language. In the case of Spanish, it also includes the geographic variant ("Mexican Spanish" = es_mx, "Ecuadorian Spanish" = es_ec, ...).
  • Tarea (task) Task of the example. Every example has the task resumen (summary)
  • Registro (Language Register): coloquial, medio o culto (colloquial, medium or educated)
  • Dominio (Domain): The domain (prensa, press) and the subdomain.
  • País de origen (Country of origin): Country of origin of the data.

The Idioma (language), Registro (Language Register), Dominio (Domain) and País de origen (Country of origin) labels have been automatically generated using GPT3.5-Turbo.

Dataset Creation

Curation Rationale

NoticIA offers an ideal scenario to test the ability of language models to understand Spanish texts. This task is complex, involving discerning the hidden question in a clickbait headline or identifying the information that the user is actually seeking. This challenge involves filtering large volumes of superfluous content to find and succinctly summarize the relevant information accurately.

In addition, by making our data and models public, we aim to exert pressure against the use of deceptive tactics by online news providers to increase advertising revenue,

Source Data

Data Collection and Processing

We have compiled clickbait news using the timeline of the X/Twitter user @ahorrandoclick1. To do this, we extracted the URLs of the news mentioned by the user. Additionally, we have added about 100 clickbait news articles chosen by us. The following image shows the source of the news in the dataset.

We have classified each of the news articles based on the category to which they belong. As can be seen, our dataset includes a wide variety of categories.

Annotation process

Although @ahorrandoclick1 provides summaries of clickbait news, these summaries do not follow any guidelines, and in many cases, their summaries do not refer to the text, but are rather of the style "This is advertising", "They still haven't realized that...". Therefore, we have manually generated the summaries for the 850 news articles. To do this, we have defined strict annotation guidelines, available at the following link: https://huggingface.co/spaces/Iker/ClickbaitAnnotation/blob/main/guidelines.py. The dataset has been annotated by Iker García-Ferrero and Begoña Altuna, and this process has taken approximately 40 hours.

Dataset Statistics

We have divided the dataset into three splits, which facilitates the training of models. As can be seen in the following table, the summaries of the news are extremely concise. They respond to the clickbait headline using the fewest words possible.

Train Validation Test Total
Number of articles 700 50 100 850
Average number of words in headlines 16 17 17 17
Average number of words in news text 544 663 549 552
Average number of words in summaries 12 11 11 12
[More Information Needed]

Who are the annotators?

  • Iker García-Ferrero: PhD Student HiTZ, the Basque center for language technology
  • Begoña Altuna: Postdoctoral research fellow at HiTZ, the Basque center for language technology

Annotation Validation

To validate the dataset, the 100 summaries from the Test set were annotated by two annotators. This data is available here: https://huggingface.co/datasets/Iker/NoticIA_Human_Validation The overall agreement between the annotators was high, as they provided exactly the same answer in 26% of the cases and provided responses that partially shared information in 48% of the cases (same response but with some variation in the words used). This demonstrates that it was easy for humans to find the information referred to by the headline. We also identified a list of cases where the annotators provided different but equally valid responses, which accounts for 18% of the cases. Lastly, we identified 8 cases of disagreement. In 3 cases, one of the annotators made an incorrect summary, likely due to fatigue after annotating multiple examples. In the remaining 5 cases, the disagreement was due to contradictory information in the article and different interpretations of this information. In these cases, determining the correct summary is subject to the reader's interpretation.

Regarding the evaluation of the guidelines, overall, they were not ambiguous, although the request to select the minimum number of words to generate a valid summary is sometimes interpreted differently by the annotators: For example, the minimum length could be understood as focusing on the question in the headline or a minimum well-formed phrase.

Massive Evaluation of Language Models

As is customary in summary tasks, we use the ROUGE scoring metric to automatically evaluate the summaries produced by models. Our main metric is ROUGE-1, which considers whole words as basic units. To calculate the ROUGE score, we lowercase both summaries and remove punctuation marks. In addition to the ROUGE score, we also consider the average length of the summaries. For our task, we aim for the summaries to be concise, an aspect that the ROUGE score does not evaluate. Therefore, when evaluating models, we consider both the ROUGE-1 score and the average length of the summaries. Our goal is to find a model that achieves the highest possible ROUGE score with the shortest possible summary length, balancing quality and brevity.

We have evaluated the best current instruction-following language models. We used the previously defined prompt. The prompt is converted into the specific chat template of each model.

The code to reproduce the results is available at the following link: https://github.com/ikergarcia1996/NoticIA

Bias, Risks, and Limitations

The dataset contains a small number of articles from Latin America; however, the vast majority of the articles are from Spanish news sources. Therefore, this dataset will evaluate the proficiency of language models in Spanish from Spain.

Although explicitly prohibited, a bad actor could use our data to train models that can generate clickbait articles automatically, contributing to polluting the internet with low-quality content. In any case, we consider the advantages of having a text comprehension dataset to evaluate language models in Spanish to be superior to the possible risks.

License

We release our annotations under the Apache 2.0 license. However, commercial use of this dataset is subject to the licenses of each news and media outlet.

Citation

If you use this dataset, please cite our paper: NoticIA: A Clickbait Article Summarization Dataset in Spanish

BibTeX:

@misc{garcíaferrero2024noticia,
      title={NoticIA: A Clickbait Article Summarization Dataset in Spanish}, 
      author={Iker García-Ferrero and Begoña Altuna},
      year={2024},
      eprint={2404.07611},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

More Information

This project was developed during the Hackathon #Somos600M organized by SomosNLP. Demo endpoints were sponsored by HuggingFace.

Team:

Contact: {iker.garciaf,begona.altuna}@ehu.eus

This dataset was created by Iker García-Ferrero and Begoña Altuna. We are researchers in NLP at the University of the Basque Country, within the IXA research group, and we are part of HiTZ, the Basque Language Technology Center.