Research:Topical coverage of Edit Wars: Difference between revisions
mNo edit summary |
Diego (WMF) (talk | contribs) Adding contextual information about related projects |
||
Line 60: | Line 60: | ||
}} <!-- End of Research project template --> |
}} <!-- End of Research project template --> |
||
This project is run by the [[mw:Wikimedia Research|Research Team] as part of the [[Community health initiative]]. Other (parallel) research studies focus on the usage of toxic language <ref>https://blog.wikimedia.org/2017/02/07/scaling-understanding-of-harassment </ref> or specific behavior such as wiki-hounding <ref> https://meta.wikimedia.org/wiki/Research:Wikihounding_and_Machine_Learning_Analysis</ref>, mainly working with the content of user interactions. |
|||
Here we adopt a complementary language agnostic approach, that focus on edit wars, trying to differentiate topic-centered from person-centered conflicts. |
|||
== Introduction == |
== Introduction == |
||
[[en:Wikipedia:Edit_warring |Edit wars ]] in Wikipedia has been largely studied. An edit war is usually consider to be the consequence of different opinions about an specific topic |
[[en:Wikipedia:Edit_warring |Edit wars ]] in Wikipedia has been largely studied. An edit war is usually consider to be the consequence of different opinions about an specific topic between two or more users. Naturally, users with different political views might have different opinions on many articles related with politics, and these differences can scale in a multi-article edit war. These actions can be consider toxic, but is not necessarily a stalking behavior. However, in the case (if exist) that edit wars start happening across multiple topics, this can be an indicator of a person-centered attack (instead of topic-centered), that might be categorized as wikihounding. |
||
Taking the advantage that edit wars can be detected in content agnostic approach (without analyzing the text), we propose to study the topical span of those wars, characterizing usual and unusual (potentially toxic) behaviors. |
Taking the advantage that edit wars can be detected in content agnostic approach (without analyzing the text), we propose to study the topical span of those wars, characterizing usual and unusual (potentially toxic) behaviors. |
||
Line 100: | Line 103: | ||
We use two metrics to characterize the topical focus/coverage of a given user, one focused on the probability of editing on a giving topic, and other focused on the ''user topical stability'': |
We use two metrics to characterize the topical focus/coverage of a given user, one focused on the probability of editing on a giving topic, and other focused on the ''user topical stability'': |
||
*'''Probability of U on editing on topic ''T'' (u<sub>p<sub>T</sub></sub> ''': The probability of user ''U'' editing topic ''T'', is the her total number of editions on ''T'' divided by total number of editions in the set. Example: in our dataset T= 40,833,278 (total number of revisions), consider a user ''U'', with 3 editions on topic Sports, his probability of editing on topic sports p<sub>sports</sub> is 3/T = 7.3e-08. |
*'''Probability of U on editing on topic ''T'' (u<sub>p<sub>T</sub></sub> ''': The probability of user ''U'' editing topic ''T'', is the her total number of editions on ''T'' divided by total number of editions in the set. Example: in our dataset T= 40,833,278 (total number of revisions), consider a user ''U'', with 3 editions on topic Sports, his probability of editing on topic sports p<sub>sports</sub> is 3/T = 7.3e-08. |
||
*'''Topical stability (u<sub>s</sub>)''': For each user U, we obtain the '''distance for his next edition '''. For example, given a user ''U'', doing three editions, the first one in the topic 'Sports', the second one in the same topic, and the third one in the topic biology (with distance 4 from Sports), the probability of user U to edit with topical distance 0, is 2/3, with distance 4 is 1/3, and 0 for the other distances. This metric gives an idea of user stability in terms of topics. |
*'''Topical stability (u<sub>s</sub>)''': For each user U, we obtain the '''distance for his next edition '''. For example, given a user ''U'', doing three editions, the first one in the topic 'Sports', the second one in the same topic, and the third one in the topic biology (with distance 4 from Sports), the probability of user U to edit with topical distance 0, is 2/3, with distance 4 is 1/3, and 0 for the other distances. This metric gives an idea of user stability in terms of topics. |
||
Line 172: | Line 175: | ||
== Future Work == |
== Future Work == |
||
* Improve mapping system from pages to Wikiprojects: The aforementioned [query https://quarry.wmflabs.org/query/23214] returns around 22% of pages matching with no wikiproject. However, manually reviewing we found cases where the Wikiproject is not correctly (?) assigned as category in the Talk pages (example [[en:Classon_Avenue_(IND_Crosstown_Line) ]] belongs to Wikiproject Trains, but that project is not listed as category) |
* Improve mapping system from pages to Wikiprojects: The aforementioned [query https://quarry.wmflabs.org/query/23214] returns around 22% of pages matching with no wikiproject. However, manually reviewing we found cases where the Wikiproject is not correctly (?) assigned as category in the Talk pages (example [[en:Classon_Avenue_(IND_Crosstown_Line) ]] belongs to Wikiproject Trains, but that project is not listed as category) |
||
== Timeline == |
== Timeline == |
||
Q1, Q2 |
Q1, Q2 |
Revision as of 05:18, 2 January 2018
This page documents a research project in progress.
Information may be incomplete and change as the project progresses.
Please contact the project lead before formally citing or reusing results from this page.
This project is run by the [[mw:Wikimedia Research|Research Team] as part of the Community health initiative. Other (parallel) research studies focus on the usage of toxic language [1] or specific behavior such as wiki-hounding [2], mainly working with the content of user interactions.
Here we adopt a complementary language agnostic approach, that focus on edit wars, trying to differentiate topic-centered from person-centered conflicts.
Introduction
Edit wars in Wikipedia has been largely studied. An edit war is usually consider to be the consequence of different opinions about an specific topic between two or more users. Naturally, users with different political views might have different opinions on many articles related with politics, and these differences can scale in a multi-article edit war. These actions can be consider toxic, but is not necessarily a stalking behavior. However, in the case (if exist) that edit wars start happening across multiple topics, this can be an indicator of a person-centered attack (instead of topic-centered), that might be categorized as wikihounding.
Taking the advantage that edit wars can be detected in content agnostic approach (without analyzing the text), we propose to study the topical span of those wars, characterizing usual and unusual (potentially toxic) behaviors.
The main tasks to develop such model are:
- Generate a representative dataset of edit war in Wikipedia.
- Detect pairs or groups of users involved in more than X (define X part of the study) controversies.
- Define and implement a robust topic model.
- Define a distance metric for topics (eg: Geography is N steps far from Politics, and M steps far from Sports.)
- Apply an outlier detection mechanism to find potential cases of harassment .
- Our approach focus in obtain good recall [3].
Methodology
- Define a topic model, that allows to measure topical distance between Wikipedia pages.
- Model user behavior according to the topics that she/he covers (edit on) and the amount of reverts that she/he does.
- Compute the probability of pair of users of co-edit a page, and the probability that this co-edition is a revert.
- Based on the aforementioned co-editions' probability, identify anomalous behavior that are potentially related with stalking or wikihounding behavior, and do manual evaluation.
Topic Model
Pages to Topic
- We use wikiprojects as proxy for topic assignment.
- Mapping from pages to Wikiprojects can be obtained with this query: https://quarry.wmflabs.org/query/23214
- Therefore, we consider each wikiproject as one topic, however, usually each page belongs two more than one topic.
- Each page can belong to 0, 1 or several Wikiprojects. For example the page about en:Diego_Maradona is the interest of en:Wikipedia:WikiProject_Biography/Sports_and_games, en:Wikipedia:WikiProject_Football and en:Wikipedia:WikiProject_Argentina
Topic Distance
- Wikiprojects can be represented as a graph (we use this library https://github.com/wiki-ai/drafttopic/tree/master/drafttopic/utilities).
- Then, giving that each page can belong to more than one category, we define the distance between two pages as the minimum shortest path [4] among all pairs of nodes on the Wikiprojects graph.
- Example: Given a page X and Y, with X in Wikiprojects a and b and Y in Wikprojects c, and d. We compute the lenght of shortest path between (a,c), (a,d), (b,c) and (b,d), and return the minimum value among these results.
User Behavior
Topical Coverage
We use two metrics to characterize the topical focus/coverage of a given user, one focused on the probability of editing on a giving topic, and other focused on the user topical stability:
- Probability of U on editing on topic T (upT) : The probability of user U editing topic T, is the her total number of editions on T divided by total number of editions in the set. Example: in our dataset T= 40,833,278 (total number of revisions), consider a user U, with 3 editions on topic Sports, his probability of editing on topic sports psports is 3/T = 7.3e-08.
- Topical stability (us): For each user U, we obtain the distance for his next edition . For example, given a user U, doing three editions, the first one in the topic 'Sports', the second one in the same topic, and the third one in the topic biology (with distance 4 from Sports), the probability of user U to edit with topical distance 0, is 2/3, with distance 4 is 1/3, and 0 for the other distances. This metric gives an idea of user stability in terms of topics.
Reverting Behavior
- For user we compute portion of reverts (within the dataset) compared with his/her total amount editions. Considering that our dataset just contains editions done by users 10 or more editions, we are just considering the reverts among these users. Later, in the results section we discuss the implications of such constrain.
Probability of Reverting Other User
- TODO
Dataset, Tools and Resources
Revisions
- We consider all the users with more than 10 revisions in the selected period (1/Jan/2017 to 16/Nov/2017)
- All the revisions from those users can be obtained from this query: https://quarry.wmflabs.org/query/23424
- Pages to Wikiprojects: https://quarry.wmflabs.org/query/23214
Wikiprojects Graph
- Using the Wikiproject graph generated from this script[5]
Code
- Find all the code used in this study here:
Terminology / Conventions
- We use edition as synonym of a Wikipedia revision.
- We use Wikiproject as proxy for Topics
Results
(Work in progress... )
Characterization on Topical Distance in Multipage editwars
In order to have notion of how frequent are reverts and editwars across multiple topics, we consider all pair of users with U, V, where U has reverted V more than 2 times, and compute the topical distance between all the pairs of pages reverted, next we compute the mode (most frequent value) for each pairs of users, and report the frequency of those values. As expected, most of wars (71%) focus on one page, 22% in the same topic but different page. The remaining 7% are a cross-topic reverts, reinforcing our intuition that cross-topic edir wars are rare.
Distance | % |
---|---|
-1 | 0.7107 |
0 | 0.2297 |
4 | 0.0143 |
5 | 0.0121 |
2 | 0.0103 |
3 | 0.0074 |
6 | 0.0056 |
7 | 0.0034 |
8 | 0.0027 |
1 | 0.0016 |
9 | 0.0012 |
10 | 0.0007 |
11 | 0.0002 |
12 | 0.0001 |
Future Work
- Improve mapping system from pages to Wikiprojects: The aforementioned [query https://quarry.wmflabs.org/query/23214] returns around 22% of pages matching with no wikiproject. However, manually reviewing we found cases where the Wikiproject is not correctly (?) assigned as category in the Talk pages (example en:Classon_Avenue_(IND_Crosstown_Line) belongs to Wikiproject Trains, but that project is not listed as category)
Timeline
Q1, Q2