Machine learning models/Production/French Wikipedia article quality
Model card | |
---|---|
This page is an on-wiki machine learning model card. | |
Model Information Hub | |
Model creator(s) | Aaron Halfaker (User:EpochFail) and Amir Sarabadani |
Model owner(s) | WMF Machine Learning Team (ml@wikimediafoundation.org) |
Model interface | Ores homepage |
Code | ORES Github, ORES training data, and ORES model binaries |
Uses PII | No |
In production? | Yes |
Which projects? | French Wikipedia |
This model uses data about a revision to predict the likelihood that the article is of a certain quality level. | |
Motivation
editThis model card describes a model for predicting the quality of Wikipedia articles. It uses structural features extracted from the article to label Wikipedia articles with a probability score for each article quality class.
Wikipedia articles range in quality from rich, well-illustrated, fully-referenced articles that fully cover their topic and are easy to read to single sentence stubs that define the topic of the article but do not offer much more information. It is very useful to be able to reliably distinguish between these extremes and the various stages of quality along this spectrum. Wikipedia editors have developed rich rubrics for how to evaluate the quality of Wikipedia articles and are constantly assessing article quality to assist in coordinating work on the wikis. Editors use these quality scores to evaluate and prioritize their work. Researchers use these quality scores to understand content dynamics. Developers use these quality scores as filters when building recommender systems or other tools.
Wikipedia is always changing, which makes it time-consuming (and largely impossible) for editors to keep these quality assessments complete and up-to-date. An automatic quality model can help fill these gaps by evaluating the quality for articles that are unassessed or have changed substantially since they were last assessed. In doing so, it can provide researchers and tool developers with more consistent data and even potentially help editors identify articles that would benefit from a human assessment.
Users and uses
edit- high-level analyses of article quality trends
- filtering / ranking articles in tools – e.g. only show low-quality articles in a recommender system
- identifying potential ways to improve articles – e.g. using the lowest-value feature from the model as a recommendation
- projects outside of French Wikipedia
- namespaces outside of 0, disambiguation pages, and redirects
- directly comparing article quality across language editions – the scores are for a given project so e.g., an article that received a 0.5 score on English Wikipedia would get a much higher score if it had been on Simple English Wikipedia instead (because high-quality articles on English Wikipedia generally have more content than high-quality articles on Simple English Wikipedia)
This model is a part of ORES, and generally accessible via API. It is used for high-level analysis of Wikipedia, platform research, and other on-wiki tasks.
Example API call:https://ores.wikimedia.org/v3/scores/frwiki/1234/articlequality
Ethical considerations, caveats, and recommendations
edit- The source data for this model is several years old — data drift may skew current outputs relative to the training data.
- The model does not currently take into account the quality of the specific writing, so a long article with many fake words would register as high quality. It does take into account structure though, so a long article would be penalized if it did not have many sections or was poorly referenced.
- Different wikis have different labeling schemes — do not use this model in conjunction with other models to conduct an interwiki analysis.
Model
editPerformance
editTest data confusion matrix:
Label | n | ~e | ~bd | ~b | ~a | ~ba | ~adq |
---|---|---|---|---|---|---|---|
e | 1491 | 1083 | 368 | 37 | 3 | 0 | 0 |
bd | 1477 | 253 | 878 | 310 | 18 | 14 | 4 |
b | 1487 | 63 | 286 | 746 | 119 | 179 | 94 |
a | 1494 | 37 | 105 | 269 | 227 | 475 | 381 |
ba | 1490 | 10 | 16 | 92 | 139 | 819 | 414 |
adq | 1491 | 0 | 1 | 59 | 78 | 337 | 1016 |
Test data sample rates:
e | bd | b | a | ba | adq | |
---|---|---|---|---|---|---|
sample | 0.167 | 0.165 | 0.167 | 0.167 | 0.167 | 0.167 |
population | 0.731 | 0.231 | 0.03 | 0.003 | 0.002 | 0.001 |
Test data performance:
Statistic | e | bd | b | a | ba | adq |
---|---|---|---|---|---|---|
match_rate | 0.544 | 0.218 | 0.115 | 0.048 | 0.136 | 0.121 |
filter_rate | 0.456 | 0.782 | 0.885 | 0.952 | 0.864 | 0.879 |
recall | 0.726 | 0.594 | 0.502 | 0.152 | 0.55 | 0.681 |
precision | 0.976 | 0.632 | 0.132 | 0.009 | 0.01 | 0.008 |
f1 | 0.833 | 0.613 | 0.209 | 0.017 | 0.019 | 0.016 |
accuracy | 0.787 | 0.826 | 0.885 | 0.95 | 0.864 | 0.88 |
fpr | 0.049 | 0.104 | 0.103 | 0.048 | 0.135 | 0.12 |
roc_auc | 0.954 | 0.89 | 0.824 | 0.754 | 0.845 | 0.902 |
pr_auc | 0.979 | 0.689 | 0.137 | 0.018 | 0.01 | 0.021 |
Implementation
edit{
"type": "GradientBoosting",
"params": {
"presort": "deprecated",
"min_impurity_decrease": 0.0,
"n_estimators": 100,
"verbose": 0,
"criterion": "friedman_mse",
"max_depth": 7,
"min_samples_leaf": 1,
"ccp_alpha": 0.0,
"random_state": null,
"labels": [
"e",
"bd",
"b",
"a",
"ba",
"adq"
],
"multilabel": false,
"warm_start": false,
"loss": "deviance",
"population_rates": null,
"validation_fraction": 0.1,
"center": true,
"init": null,
"label_weights": null,
"subsample": 1.0,
"max_features": "log2",
"n_iter_no_change": null,
"scale": true,
"min_impurity_split": null,
"min_samples_split": 2,
"tol": 0.0001,
"max_leaf_nodes": null,
"learning_rate": 0.01,
"min_weight_fraction_leaf": 0.0
}
}
{
"properties": {
"prediction": {
"type": "string",
"description": "The most likely label predicted by the estimator"
},
"probability": {
"properties": {
"e": {
"type": "number"
},
"bd": {
"type": "number"
},
"a": {
"type": "number"
},
"b": {
"type": "number"
},
"ba": {
"type": "number"
},
"adq": {
"type": "number"
}
},
"type": "object",
"description": "A mapping of probabilities onto each of the potential output labels"
}
},
"title": "Scikit learn-based classifier score with probability",
"type": "object"
}
https://ores.wikimedia.org/v3/scores/frwiki/1234/articlequality
Output:
{
"frwiki": {
"models": {
"articlequality": {
"version": "0.8.0"
}
},
"scores": {
"1234": {
"articlequality": {
"score": {
"prediction": "e",
"probability": {
"a": 0.09291616777238719,
"adq": 0.061338631094345654,
"b": 0.20091211688282035,
"ba": 0.07865775129072539,
"bd": 0.2582831067626765,
"e": 0.3078922261970448
}
}
}
}
}
}
}
Data
editLicenses
edit- Code: MIT license
- Model: MIT license
Citation
editCite this model card as:
@misc{
Triedman_Bazira_2023_French_Wikipedia_article_quality,
title={ French Wikipedia article quality model card },
author={ Triedman, Harold and Bazira, Kevin },
year={ 2023 },
url={ https://meta.wikimedia.org/wiki/Machine_learning_models/Production/French_Wikipedia_article_quality }
}