Diversity-Aware NLP Intelligent Systems (DANIS)

Independent Research Group

Reflecting Intelligent Systems for Diversity, Demography, and Democracy (IRIS3D)

Diversity-Aware NLP Intelligent Systems (DANIS)

Project focus

The immense influence of NLP systems on human lives raises increasing concerns about the possible harm these tools can cause. Harmful behaviors of such systems are regarded as symptoms of their bias, i.e., the systematic preference or discrimination against certain groups of users. NLP tools are commonly trained on textual corpora that display such biases already at the level of their authors. For example, Wikipedia, which is one of the most commonly used sources of training data, is created by a predominantly white and male group of editors. Such a lack of diversity among authors can lessen the impact of data from minorities and, as a consequence, result in NLP models that reflect the underlying demographic imbalances. DANIS contributes to the discourse of fairness in AI by facilitating the design of NLP intelligent systems that can recognize inputs from underrepresented groups of users and strengthen their role in the training processes.


January 2023 - December 2026




The project is funded by the Ministry of Science, Research and the Arts of the State of Baden-Württemberg.

IRIS3D Research Group DANIS

News related to the DANIS Group

To the top of the page