Skip to main content

MARB

Standard reference Information

Tom Södahl Bladsjö and Ricardo Muñoz Sánchez. 2025. Introducing MARB — A Dataset for Studying the Social Dimensions of Reporting Bias in Language Models. In Proceedings of the 6th Workshop on Gender Bias in Natural Language Processing (GeBNLP), pages 59–74, Vienna, Austria. Association for Computational Linguistics.

Data citation Information

Södahl Bladsjö, Tom, & Muñoz Sánchez, Ricardo (2025). MARB (updated: 2025-06-05). [Data set]. Språkbanken Text. https://doi.org/10.23695/v3wp-6c64
BibTeX Additional ways to cite the dataset.
A dataset for studying Marked Attribute Reporting Bias

Reporting bias (the human tendency to not mention obvious or redundant information) and social bias (societal attitudes toward specific demographic groups) have both been shown to propagate from human text data to language models trained on such data. However, the two phenomena have not previously been studied in combination. The MARB dataset was developed to begin to fill this gap by studying the interaction between social biases and reporting bias in language models. Unlike many existing benchmark datasets, MARB does not rely on artificially constructed templates or crowdworkers to create contrasting examples. Instead, the templates used in MARB are based on naturally occurring written language from the 2021 version of the enTenTen corpus (Jakubíček et al., 2013).

Annotation

The dataset consists of nearly 28.5K template sequences. 9.5K containing each of the phrases 'a person', 'a woman', 'a man', and their modifications. The templates are based on naturally occurring written language from the 2021 version of the enTenTen corpus (Jakubíček et al., 2013).

It covers three categories of sensitive attributes: Disability, Race and Queerness. Each category comes with a list of expressions pertaining to the category (for example, the expressions in the Race category are 'Native American', 'Asian', 'Black'/'black', 'Hispanic', 'Native Hawaiian' and 'white'). Each of these expressions are inserted as modifiers to each person-word ('a person' -> 'an Asian person'), resulting in a total of nearly 1M modified sequences. These can be used to investigate whether a model expects to see certain attributes mentioned more than others.

Caveats

Ethical considerations

This work deals with language categorizing people based on sensitive attributes such as race, gender identity and sexuality. This is a sensitive topic, and care should be taken not to oversimplify complex real-world power structures or to confuse real-life demographic groups with the words used to describe them.

Things to watch out for

There are often many ways to refer to a specific social group, and they carry different connotations and underlying assumptions. For example, both terminology and ontological definitions relating to disability are contested, and there is great variation in the language used both by in-group and out group members. Additionally, in many cases there is a complete lack of established terms describing normative attributes, such as not having a disability. The attribute descriptors included in MARB should be seen as a sample rather than a comprehensive list of the language used to refer to these groups.

Furthermore, most common metrics for how well a model predicts a sequence (such as perplexity) are affected by factors such as sequence length and model vocabulary. While steps have been taken to mitigate these effects (especially that of sequence length), they will still be present to some extent, particularly in cases where an attribute descriptor is divided into many tokens. When using the dataset for evaluation, take care to analyse your results with this in mind.

Finally, while MARB can be used to identify certain types of bias in a model, a lack of visible bias using MARB is no guarantee that the model is not biased.

Intended uses

  • Key applications: Identify and study reporting bias with regards to human attributes in masked and autoregressive language models
  • Intended task(s)/usage(s): While MARB can be used to investigate a range of different research questions, it is mainly intended as an intrinsic diagnostic for out-of-the-box language models: How likely is the model to predict the sets of sequences containing different attribute descriptors and person-words? Are there systematic differences depending on the attribute mentioned?
  • Recommended evaluation measures: Perplexity, Rank biserial correlation (r)
  • Dataset function(s): Diagnostics
  • Recommended split(s): Test data only

References

  • Tom Södahl Bladsjö. 2024. Don't Mention the Norm. Masters Thesis, Gothenburg, Sweden. University of Gothenburg.

Accessible through

Access Platform Licence
MIT License

Download

File Size Modified Licence
12.42 MB 2025-09-08 MIT License
15.99 KB 2025-09-08 MIT License

Type

  • Corpus

Language

English

Size

Sentences: 940,500

Keywords

  • bias and fairness
  • reporting bias
  • marked attributes
  • language model evaluation
  • llm evaluation
  • lgbtq+
  • disability
  • race
  • gender
  • pseudo-perplexity
  • diagnostic dataset
  • language resources

Creators

  • Södahl Bladsjö, Tom
  • Muñoz Sánchez, Ricardo

Created

2024-05-27

Updated

2025-06-05

Contact

sb-info@svenska.gu.se