Machine learning models/Production/Language-agnostic revert risk
This model card page currently has a draft status. It is a piece of model documentation that is in the process of being written. Once the model card is completed, this template should be removed. |
How can we help editors to identify revisions that need to be “patrolled”? The goal of this model is to detect revisions that might be reverted, independently of whether they were made in good faith or with the intention of creating damage.
Model card | |
---|---|
This page is an on-wiki machine learning model card. | |
Model Information Hub | |
This model uses revision content and metadata to predict the risk of being reverted. | |
Patrolling content in more than 250+ Wikipedia projects is a difficult task. The amount of revisions plus the different languages involved requires a complex human effort. The aim of this model is to help patrollers quickly identify potential problems and revert damaging edits when needed.
Previous models had tried to solve this by creating language-specific solutions, however, that approach is difficult to escalate and maintain, because it requires as many models as languages used on the Wikimedia projects. Moreover, complex-language models are just available in certain languages, leaving out smaller Wikipedia editions. Therefore, this model is based on Language Agnostic features, making it possible to use it for any existing Wikipedia, and for new language projects that can appear in the future.
This model was trained using the two tables from the Wikimedia Data Lake: The MediaWiki History , and the Wikitext History one. Meta-data was extracted for the former, and other features such as number of references, images and wikilinks, were extracted from the latter.
This model is deployed on LiftWing. Right now, it is available for internal usage. Technical details can be found here. This model can be used to detect revisions that might need to be reverted. A high “revert probability” output (over .9) would provide good precision, while lower threshold (0.5) would provide recall. This model should be used just for Wikipedia Articles (namespace 0), its features won't work outside Wikipedia.
Motivation
editKnowledge Integrity is one of the strategic programs of Wikimedia Research with the goal of identifying and addressing threats to content on Wikipedia, increasing the capabilities of patrollers, and providing mechanisms for assessing the reliability of sources[1]. The main goal of the project is to create a new generation of patrolling models, improving accuracy, fairness, and maintainability compared to previous state-of-the-art ORES[2].
The current model is completely language agnostic and can run in any Wikipedia language edition.
Supported languages |
---|
|
Users and uses
edit- Automatically find revisions that requires patrolling.
- Vandalism detection.
- Create bots for assisting admins and patrollers to remove vandalism or non good-faith edits.
- As ground-truth for training other models
- Making predictions on projects other than Wikipedia language editions
- Making predictions on the first revision of a page or a revision that is the only one for a page.
- Research
- Our products, including via Wikimedia Enterprise APIs.
Ethical considerations, caveats, and recommendations
editThe model is built using meta features that take into account user characteristics. However, it may exhibit bias against edits from new users, temporary accounts, or IP edits due to past experiences. To address this issue, we have developed an alternative Revert Risk Multilingual Model (RRML) that specifically mitigates such biases, however that model requires more processing power and might be slow. Therefore, for anonymous edits - on the 47 languages covered - we recommend using the RRML. For the reminding edits (non-anonymous or not covered by RRML), we recommend to use this model.
Model
editThis model uses this set of features:
- Article features:
- We used the features developed for the Language Agnostic Article Quality Model.
- We computed the article quality features for the current and parent revision.
- We measured the quality differences between these revisions.
- User features:
- Account "age" (difference between revision timestamp and the user creation date)
- Number of previous revisions made.
- Number of users groups.
Performance
editImplementation
editThe model is build using the XgBoost library.
The detailed model training procedure and configuration can be found in this repository.{
model_name: <model name string>
model_version: <model version string>
wiki_db: <wiki code string>,
revision_id: <revision_id string>,
output: {
prediction: <boolean decision result>
probabilities: {
true: <probability of being reverted>,
false: <probability of being NOT reverted>
}
}
Example input:
curl https://api.wikimedia.org/service/lw/inference/v1/models/revertrisk-language-agnostic:predict -X POST -d '{"rev_id": 1215916855, "lang":"en"}'
Example output:
{
"model_name":"revertrisk-language-agnostic",
"model_version":"2",
"wiki_db":"enwiki",
"revision_id":123855516,
"output":{
"prediction":true,
"probabilities":{
"true":0.6868777275085449,
"false":0.3131222724914551
}
}
}
Data
editThe model was trained on a dataset collected using the two tables from the Wikimedia Data Lake. We used the MediaWiki History table, and the Wikitext History one. Snapshot dated 2023-05 was used with the observation period from 2022-01-01 to 2023-01-01 (12 months). We filtered out revisions created by bots. We used the 70% of the data for training, and 30% testing, using a random split.
The data collection process can by found on this repository.Licenses
edit- Code: Apache 2.0 License
- Model: Apache 2.0 License
Citation
editCite this model as:
@inproceedings{trokhymovych2023fair,
title={Fair multilingual vandalism detection system for Wikipedia},
author={Trokhymovych, Mykola and Aslam, Muniza and Chou, Ai-Jou and Baeza-Yates, Ricardo and Saez-Trumper, Diego},
booktitle={Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining},
pages={4981--4990},
year={2023}
}
Notice this paper also covers the Multilingual Revert Risk Model.
References
edit- ↑ Zia, Leila and Johnson, Isaac and Mansurov, Bahodir and Morgan, Jonathan and Redi, Miriam and Saez-Trumper, Diego and Taraborelli, Dario. 2019. Knowledge Integrity. https://doi.org/10.6084/m9.figshare.7704626
- ↑ https://www.mediawiki.org/wiki/ORES