DS4UX (Spring 2016)/Panama Papers: Difference between revisions

From CommunityData
(from http://wiki.communitydata.cc/Wikipedia_(CDSW))
 
 
(56 intermediate revisions by the same user not shown)
Line 1: Line 1:
[[File:Wikipedia.png|right|250px]]
[[File:Wikipedia.png|right|250px]]


In this project, we will explore a few ways to gather data using the Wikipedia API. Once we've done that, we will extend this to code to create our own datasets of Wikipedia edits or other data that we might be able to use to ask and answer questions in the final session.
In this project, we will explore a few ways to gather data using two Wikipedia APIs: one provides data related to edits, and the other provides data related to pageviews. Once we've done that, we will extend this to code to create our own datasets of Wikipedia edits or other data that you use as the basis of your Final Project.


== Goals ==
This project is adapted from material being developed for the [[CDSW|Community Data Science Workshops]] by Ben Lewis and Morten Wang ([https://github.com/nettrom/wikipedia-session GitHub repo]).


* Get set up to build datasets with the Wikipedia API
== Overview ==
In this project we will look at the viewing and editing history of a recently created Wikipedia article about a breaking news event—''[[w:Panama_Papers| Panama Papers]]''. When events of global significance occur, Wikipedia is often among the first places that people look for information about these events. By examining both the editing and viewing history of this article, we can learn a lot about how people create ''and'' consume information on Wikipedia.
 
The process by which 'breaking news' articles are created on Wikipedia is [http://dgergle.soc.northwestern.edu/resources/KeeganGergleContractor_StayingInTheLoop_WikiSym2012.pdf a fascinating area of research] for data scientists who are study how humans work together. For more links to interesting research on Wikipedia, see the [[DS4UX_(Spring_2016)/Wikipedia_API#Research_using_Wikipedia_data|Resources section]] of this page.
 
=== Goals ===
 
* Understand the basic anatomy of an API request
* Have fun collecting different types of data from Wikipedia
* Have fun collecting different types of data from Wikipedia
* Practice reading API documentation
* Pracice testing API queries in an [https://en.wikipedia.org/wiki/Special:ApiSandbox API Sandbox]
* Practice reading and extending other people's code
* Practice reading and extending other people's code
* Create a few collections of different types of data from Wikipedia that you can do research with in the final section


== Download and test the Wikipedia API project ==
=== Download and test the Panama Papers project ===
 
# Right click the following file, click "Save Target as..." or "Save link as...", and save it to your Desktop directory: https://jtmorgan.net/ds4ux/week6/panama-papers.zip
# Find <code>panama-papers.zip</code> on your Desktop and double-click on it to "unzip" it. That will create a folder called <code>panama-papers</code> containing several files.
# In PowerShell or Terminal, navigate to the <code>panama-papers</code> directory and type:
 
python wikipedia-1.py
 
at the command prompt to execute the <code>wikipedia-1.py</code> Python script. Wait a little while while your computer connects to Wikipedia. You should see the result of a query from Wikipedia API on your screen. If you don't, let Jonathan or Ray know .
 
=== About the APIs ===
 
;Wikipedia Edit API: An API that provides data on edits to Wikipedia pages. Use the [https://en.wikipedia.org/wiki/Special:ApiSandbox API sandbox] to build and test queries with this API.
* Also known as "the MediaWiki API", because it is the standard API for all sites that use [https://en.wikipedia.org/wiki/MediaWiki MediaWiki].
* You can use this API to edit Wikipedia, and do all sorts of other things. We will just be using [https://www.mediawiki.org/wiki/API:Query the Query module of the API] to gather data about individual pages and editors.
* Example request for Panama Papers (current content of the page): https://en.wikipedia.org/w/api.php?format=json&action=query&titles=Panama_Papers&prop=revisions&rvprop=content
 
 
;Wikipedia Page View API: Use [https://wikimedia.org/api/rest_v1/?doc#!/Pageviews_data the (experimental!) Wikipedia page view API] to find out how many views a given Wikipedia article gets on a daily basis.
* This API is a little different than the Wikipedia edit API because it uses relative paths (<code>/.../.../</code>) instead of parameters (<code>somekey=someval&otherkey=otherval</code>).
* This API is case-sensitive, and space-sensitive (replace spaces in page titles with underscores). It also requires you to [http://www.w3schools.com/tags/ref_urlencode.asp URL encode] any special characters (like !, $, # and %) that might appear in a page title.
* Example request for Panama Papers (page views 4/01 - 4/20): https://wikimedia.org/api/rest_v1/metrics/pageviews/per-article/en.wikipedia/all-access/user/Panama_Papers/daily/20160401/20160420
 
== Exercises ==
=== Building queries in the Wikipedia editing API Sandbox ===
 
;When was the article about the Panama Papers created?
 
* [https://en.wikipedia.org/wiki/Special:ApiSandbox#action=query&format=json&prop=revisions&list=&meta=&titles=Panama_Papers&rvprop=ids%7Ctimestamp%7Cflags%7Ccomment%7Cuser&rvlimit=1&rvdir=newer View query in sandbox]
* [https://en.wikipedia.org/w/api.php?action=query&format=json&prop=revisions&list=&meta=&titles=Panama_Papers&rvprop=ids%7Ctimestamp%7Cflags%7Ccomment%7Cuser&rvlimit=1&rvdir=newer View result in browser]
* [https://www.mediawiki.org/wiki/API:Revisions View API:Revisions documentation]
 
 
;When was the most recent edit to the Panama Papers article?
 
* [https://en.wikipedia.org/wiki/Special:ApiSandbox#action=query&format=json&prop=revisions&list=&meta=&titles=Panama_Papers&rvprop=ids%7Ctimestamp%7Cflags%7Ccomment%7Cuser&rvlimit=1&rvdir=older View query in sandbox]
* [https://en.wikipedia.org/w/api.php?action=query&format=json&prop=revisions&list=&meta=&titles=Panama_Papers&rvprop=ids%7Ctimestamp%7Cflags%7Ccomment%7Cuser&rvlimit=1&rvdir=newer View result in browser]
* [https://www.mediawiki.org/wiki/API:Revisions View API:Revisions documentation]
 
 
; How many edits has the creator of Panama Papers made to Wikipedia?
 
* [https://en.wikipedia.org/wiki/Special:ApiSandbox#action=query&format=json&list=users&usprop=editcount%7Cregistration&ususers=Czar View query in sandbox]
* [https://en.wikipedia.org/w/api.php?action=query&format=json&list=users&usprop=editcount%7Cregistration&ususers=Czar View result in browser]
* [https://www.mediawiki.org/wiki/API:Users View API:Users documentation]
 
 
;What was the text of the Panama Papers article 24 hours after it was created?


If you are confused by these steps, go back and refresh your memory with the [[Community Data Science Workshops (Fall 2015)/Day 0 setup and tutorial|Day 0 setup and tutorial]] and [[Community Data Science Workshops (Fall 2015)/Day 0 tutorial|Day 0 tutorial]]
* [https://en.wikipedia.org/wiki/Special:ApiSandbox#action=query&format=json&prop=revisions&titles=Panama+Papers&rvprop=ids%7Ctimestamp%7Ccontent&rvstart=2016-04-04T17%3A58%3A00.000Z&rvend=2016-04-04T17%3A59%3A05.000Z&rvdir=newer View query in sandbox]
* [https://en.wikipedia.org/w/api.php?action=query&format=json&prop=revisions&titles=Panama+Papers&rvprop=ids%7Ctimestamp%7Ccontent&rvstart=2016-04-04T17%3A58%3A00.000Z&rvend=2016-04-04T17%3A59%3A05.000Z&rvdir=newer View result in browser]
* [https://www.mediawiki.org/wiki/API:Revisions View API:Revisions documentation]
* [https://en.wikipedia.org/w/index.php?title=Panama_Papers&oldid=713548359 View the text of this revision on Wikipedia]
* [https://en.wikipedia.org/w/index.php?title=Panama_Papers&diff=713548359&oldid=713548357 View the "diff" version of revision] (shows what was changed between this edit and the previous one)


(Estimated time: 10 minutes)


===Download the Wikipedia API project===
;Who has edited Panama Papers?
* [https://www.mediawiki.org/wiki/API:Contributors View API:Contributors documentation]
* [https://en.wikipedia.org/wiki/Special:ApiSandbox#action=query&format=json&prop=contributors&list=&titles=Panama+Papers&pclimit=500 View query in sandbox]


* Right click the following file, click "Save Target as..." or "Save link as...", and save it to your Desktop directory:
https://communitydata.cc/~groceryheist/wikipedia-cdsw-answers.zip


* The ".zip" extension on the above file indicates that it is a compressed Zip archive. We need to "extract" its contents. To do this, click on "Start", then "Computer" on Windows or open Finder and navigate to your Desktop directory if you are a Mac. Find <code>wikipedia-cdsw.zip</code> on your Desktop and double-click on it to "unzip" it. That will create a folder called <code>wikipedia-cdsw</code> containing several files.
;What categories is Panama Papers in?
* [https://www.mediawiki.org/wiki/API:Categories View API:Categories documentation]
* [https://en.wikipedia.org/wiki/Special:ApiSandbox#action=query&format=json&prop=categories&list=&titles=Panama+Papers&clprop=timestamp&cllimit=500 View query in API sandbox]


===Test the Wikipedia API code===


<div style="background-color:#CEE7DA; width:80%; padding:1.2em;">
;What other articles are in the ''Category'' Panama Papers?
'''On Windows'''
* [https://www.mediawiki.org/wiki/API:Categorymembers View API:Categorymembers documentation]
* [https://en.wikipedia.org/wiki/Special:ApiSandbox#action=query&format=json&prop=&list=categorymembers&titles=Panama+Papers&cmtitle=Category%3APanama_Papers&cmprop=ids%7Ctitle%7Ctimestamp&cmtype=page&cmlimit=50 View query in sandbox]


Start up PowerShell and navigate to the <code>Desktop\wikipedia-cdsw</code> directory where the Wikipedia API code lives. For example, if the Wikipedia API project is at <code>C:\Users\'''YOURUSERNAME'''\Desktop\wikipedia-cdsw</code>,


cd C:\Users\'''YOURUSERNAME'''\Desktop\wikipedia-cdsw
;What other articles does Panama Papers link to?
* [https://www.mediawiki.org/wiki/API:Links View API:Links documentation]
* [https://en.wikipedia.org/wiki/Special:ApiSandbox#action=query&format=json&prop=links&list=&titles=Panama+Papers&plnamespace=0&pllimit=500 View query in API sandbox]


</div>
<div style="background-color:#D8E8FF; width:80%; padding:1.2em;">
'''On Mac'''


Start a command prompt and navigate to the Desktop/wikipedia-data-examples directory where the Wikipedia API code lives. For example, if the Wikipedia API project is at <code>~/Desktop/wikipedia-cdsw</code>,
<br/>


  cd ~/Desktop/wikipedia-cdsw
=== Building queries in the Wikipedia page view API Sandbox ===
   
;How many views did Panama Papers have during its first week?


</div>
*[https://wikimedia.org/api/rest_v1/metrics/pageviews/per-article/en.wikipedia.org/all-access/all-agents/Panama_Papers/daily/20160403/20160409 View results in browser]


This will change you into the Wikipedia example code directory. Running <code>ls</code> will show you the source code files in that directory. One of the files is <code>wikipedia-1.py</code>, which has a <code>.py</code> extension indicating that it is a Python script. Type:
;How many views did Panama papers have during its first week from mobile devices?


python wikipedia-1.py
* [https://wikimedia.org/api/rest_v1/metrics/pageviews/per-article/en.wikipedia.org/mobile-web/all-agents/Panama_Papers/daily/20160403/20160409 View results for mobile website in browser]
* [https://wikimedia.org/api/rest_v1/metrics/pageviews/per-article/en.wikipedia.org/mobile-app/all-agents/Panama_Papers/daily/20160403/20160409 View results for mobile app in browser]
<br/>
 
=== Building queries with Python <code>requests</code> ===
Now that we're comfortable building API queries in the sandbox, we will focus on how we can access these APIs with Python. If you would like to review the steps involved in building an API query in Python, check out the resources listed below.
* [https://github.com/makoshark/wikipedia-cdsw/blob/master/building-a-query.md Querying APIs from Python] a written lecture by Ben Lewis that walks you step-by-step through the process of building and executing an API query in Python. The 'companion script' <code>building_a_query_code.py</code> in the project directory executes all of the code shown in this lecture step-by-step. If you want to just execute some of the code in the lecture, comment out all the stuff below the blocks of code you want to execute it before you run the script.
* <code>wikipedia-1.py</code> — This is the script you were asked to execute to 'test' your code when you downloaded the project. It's also a valid API request that gathers metadata about the first revision to Panama Papers, and prints it to your terminal. The JSON that this query returns can be seen here: https://en.wikipedia.org/w/api.php?action=query&prop=revisions&titles=Panama_Papers&rvdir=newer&rvlimit=1&format=jsonfm
* <code>introduce_while.py</code> — (in project directory) this script uses a while loop to roll two 'virtual dice' until they both come up 6's. This example doesn't make API calls or use Wikipedia data—the point is to help you understand that sometimes you will need to loop through an operation (like an API request) an indeterminate number of times. In these situations, a 'while' loop is more appropriate than a 'for' loop.
* <code>introduce_continue.py</code> — (in project directory) this script shows you two ways to use the value of the 'continue' key that is embedded inside the JSON returned by your API request. Each API request returns a chunk of data, but there may be more data available! By passing the value of 'continue' back in subsequent requests, you can pick up where the last request left off.
 
 
;1. How many edits did Panama Papers receive in its first 24 hours?


at the command prompt to execute the <code>wikipedia-1.py</code> Python script. Wait a little while while your computer connects to Wikipedia. You should see the result of a query from Wikipedia API on your screen. If you don't, let a staff member know.
Completing this exercise requires you to first query the Wikipedia API to get revisions to the article that have a timestamp between 2016/4/3 17:59:05 and 2016/4/4 17:59:05, and then use Python to count the number of revisions in the resulting JSON.


== Example topics we might cover in the session ==


=== Main Wikipedia API ===
;2. How many edits has Panama Papers receive from mobile devices since it was created?  
* explain [http://www.mediawiki.org/wiki/API:Main_page MediaWiki], exists on other wikis
* navigate to [http://en.wikipedia.org/w/api.php api page] and show the documentation, point out examples
* introduce the [https://en.wikipedia.org/wiki/Special:ApiSandbox API sandbox] as a tool for building queries
* looking at the images within a page http://en.wikipedia.org/w/api.php?action=query&titles=Seattle&prop=images&imlimit=20&format=jsonfm
* change the city with a custom URL
* edit count http://en.wikipedia.org/w/api.php?action=query&list=users&ususers=Benjamin_Mako_Hill|Jtmorgan|Sj|Koavf&usprop=editcount&format=jsonfm
* get the content of the main page http://en.wikipedia.org/w/api.php?format=json&action=query&titles=Main%20Page&prop=revisions&rvprop=content


=== Page View API ===
Completing this exercise requires you to perform two queries with the Wikipedia pageview API, because there are ''two'' types of mobile device counts—<code>mobile-app</code> and <code>mobile-web</code>—and you can only query them one at a time.
* Use [https://wikimedia.org/api/rest_v1/?doc#!/Pageviews_data the experimental API]
* Explain that this API is a little different because it uses relative paths instead of parameters.
* Also note that this API is case-sensitive.
* Request for Panama Papers: https://wikimedia.org/api/rest_v1/metrics/pageviews/per-article/en.wikipedia/all-access/user/Panama_Papers/daily/20160401/20160420


== Questions to answer ==


Warm up questions:
;3. How many times was Panama Papers viewed in the first week? What proportion of those views came from mobile devices?


<ol start="1">
Completing this exercise also requires two API requests: one to gather pageview data for ALL devices, and then performing a request that only gathers data about devices that viewed the page using the [https://en.m.wikipedia.org/wiki/Main_Page Wikipedia mobile website].
<li>When was the article about the Panama Papers created?</li>
<li>When was the most recent edit to the Panama Papers article?</li>
<li>Think of one or two articles that interest you. Which ones were created first?</li>
<li>Which have been edited most recently?</li>
</ol>


How many views did Panama_Papers have…


<ol start="5">
;4. How many other articles has [https://en.wikipedia.org/wiki/User_talk:Czar User:Czar] edited on Wikipedia since they created Panama Papers?
<li>the day it was created?</li>
* [https://www.mediawiki.org/wiki/API:Usercontribs View documentation for API:Usercontribs]
<li>the first week? </li>
* [https://en.wikipedia.org/wiki/Special:ApiSandbox#action=query&format=json&prop=&list=usercontribs&titles=&uclimit=500&ucstart=2016-04-04T17%3A59%3A00.000Z&ucend=2016-05-01T00%3A22%3A24.000Z&ucuser=Czar&ucdir=newer&ucnamespace=0&ucprop=ids%7Ctitle%7Ctimestamp%7Ccomment%7Csize View example query in API sandbox]
<li>How does this compare to the articles that interest you? </li>
</ol>


How many edits did it get in…
The sample query above returns a list of edits by Czar between 4/03 and 5/01, that includes the ''title'' of each article that was edited. How would you iterate over this JSON data in Python to come up with a list that contained the title of every article Czar edited, with no duplicates?
<ol start="9">
<br/>
<li> the first 24 hours?  </li>
<li> the first week? </li>
<li> How many edits did the articles that interest you get? </li>
</ol>


More difficult questions:
== Coding challenges ==


<ol start="12">
Once you have worked through these exercises and feel confident that you can gather data from the Wikipedia edit and pageview APIs using Python, you can get started on [[DS4UX_(Spring_2016)/Day_6_coding_challenge|this week's coding challenges]]!
<li>Who made the total most edits to the article?</li>
<br/>
<li>What’s the number of edits per day in the first two weeks of the article?</li>
<li>What’s the peak number of edits per hour of the article? When did it occur?</li>
<li>Who were the top editors during that hour?</li>
<li>What day did it have the most views, and how many views did it have?</li>
<li>How many views did it have per day?</li>
<li>How many views did it have per day on German Wikipedia?</li>
<li>Who’s the person named in the Panama papers with the most views to their Wikipedia page?</li>
</ol>


== Resources ==
== Resources ==
Line 108: Line 148:
* [https://en.wikipedia.org/w/api.php?action=help&modules=query API documentation for the query module]
* [https://en.wikipedia.org/w/api.php?action=help&modules=query API documentation for the query module]
* [https://en.wikipedia.org/wiki/Special:ApiSandbox API Sandbox]
* [https://en.wikipedia.org/wiki/Special:ApiSandbox API Sandbox]
* [[Sample Wikipedia API queries]]
* [[Sample Wikipedia API queries|More sample Wikipedia API queries]]
* [https://github.com/ben-zen/wikipedia-session The session lecture notes (in Markdown) and python sources.]
 
* [[Sample Wikipedia API questions]]
 
=== Research using Wikipedia data ===
* ''[https://www.research.ibm.com/visual/projects/history_flow/ HistoryFlow]'' — A colorful visualization of the development of Wikipedia articles over time.
* [http://www.brianckeegan.com/papers/CSCW_2015.pdf ‘Is’ to ‘Was’: Coordination and Commemoration on Posthumous Wikipedia Biographies] — an exploration of editing patterns around Wikipedia articles about people who have recently died.
* [http://www.brianckeegan.com/papers/ICS_2015.pdf WikiWorthy: Judging a Candidate’s Notability in the Community] — A study that uses the editing activity on Wikipedia articles about political candidates as a predictor of election success.
 
 
=== Websites that use the MediaWiki API ===
* [http://listen.hatnote.com/ Listen to Wikipedia] — a dynamic, audiovisual experience based on what is being edited on Wikipedia right now.
* [http://histography.io/ HistoryGraph] — an interactive timeline of world events based on Wikipedia articles.
* [https://www.google.com/intl/es419/insidesearch/features/search/knowledge.html Google's knowledge graph]


[[Category:DS4UX (Spring 2016)]]
[[Category:DS4UX (Spring 2016)]]

Latest revision as of 00:05, 3 May 2016

In this project, we will explore a few ways to gather data using two Wikipedia APIs: one provides data related to edits, and the other provides data related to pageviews. Once we've done that, we will extend this to code to create our own datasets of Wikipedia edits or other data that you use as the basis of your Final Project.

This project is adapted from material being developed for the Community Data Science Workshops by Ben Lewis and Morten Wang (GitHub repo).

Overview[edit]

In this project we will look at the viewing and editing history of a recently created Wikipedia article about a breaking news event— Panama Papers. When events of global significance occur, Wikipedia is often among the first places that people look for information about these events. By examining both the editing and viewing history of this article, we can learn a lot about how people create and consume information on Wikipedia.

The process by which 'breaking news' articles are created on Wikipedia is a fascinating area of research for data scientists who are study how humans work together. For more links to interesting research on Wikipedia, see the Resources section of this page.

Goals[edit]

  • Understand the basic anatomy of an API request
  • Have fun collecting different types of data from Wikipedia
  • Practice reading API documentation
  • Pracice testing API queries in an API Sandbox
  • Practice reading and extending other people's code

Download and test the Panama Papers project[edit]

  1. Right click the following file, click "Save Target as..." or "Save link as...", and save it to your Desktop directory: https://jtmorgan.net/ds4ux/week6/panama-papers.zip
  2. Find panama-papers.zip on your Desktop and double-click on it to "unzip" it. That will create a folder called panama-papers containing several files.
  3. In PowerShell or Terminal, navigate to the panama-papers directory and type:
python wikipedia-1.py

at the command prompt to execute the wikipedia-1.py Python script. Wait a little while while your computer connects to Wikipedia. You should see the result of a query from Wikipedia API on your screen. If you don't, let Jonathan or Ray know .

About the APIs[edit]

Wikipedia Edit API
An API that provides data on edits to Wikipedia pages. Use the API sandbox to build and test queries with this API.


Wikipedia Page View API
Use the (experimental!) Wikipedia page view API to find out how many views a given Wikipedia article gets on a daily basis.

Exercises[edit]

Building queries in the Wikipedia editing API Sandbox[edit]

When was the article about the Panama Papers created?


When was the most recent edit to the Panama Papers article?


How many edits has the creator of Panama Papers made to Wikipedia?


What was the text of the Panama Papers article 24 hours after it was created?


Who has edited Panama Papers?


What categories is Panama Papers in?


What other articles are in the Category Panama Papers?


What other articles does Panama Papers link to?



Building queries in the Wikipedia page view API Sandbox[edit]

How many views did Panama Papers have during its first week?
How many views did Panama papers have during its first week from mobile devices?


Building queries with Python requests[edit]

Now that we're comfortable building API queries in the sandbox, we will focus on how we can access these APIs with Python. If you would like to review the steps involved in building an API query in Python, check out the resources listed below.

  • Querying APIs from Python a written lecture by Ben Lewis that walks you step-by-step through the process of building and executing an API query in Python. The 'companion script' building_a_query_code.py in the project directory executes all of the code shown in this lecture step-by-step. If you want to just execute some of the code in the lecture, comment out all the stuff below the blocks of code you want to execute it before you run the script.
  • wikipedia-1.py — This is the script you were asked to execute to 'test' your code when you downloaded the project. It's also a valid API request that gathers metadata about the first revision to Panama Papers, and prints it to your terminal. The JSON that this query returns can be seen here: https://en.wikipedia.org/w/api.php?action=query&prop=revisions&titles=Panama_Papers&rvdir=newer&rvlimit=1&format=jsonfm
  • introduce_while.py — (in project directory) this script uses a while loop to roll two 'virtual dice' until they both come up 6's. This example doesn't make API calls or use Wikipedia data—the point is to help you understand that sometimes you will need to loop through an operation (like an API request) an indeterminate number of times. In these situations, a 'while' loop is more appropriate than a 'for' loop.
  • introduce_continue.py — (in project directory) this script shows you two ways to use the value of the 'continue' key that is embedded inside the JSON returned by your API request. Each API request returns a chunk of data, but there may be more data available! By passing the value of 'continue' back in subsequent requests, you can pick up where the last request left off.


1. How many edits did Panama Papers receive in its first 24 hours?

Completing this exercise requires you to first query the Wikipedia API to get revisions to the article that have a timestamp between 2016/4/3 17:59:05 and 2016/4/4 17:59:05, and then use Python to count the number of revisions in the resulting JSON.


2. How many edits has Panama Papers receive from mobile devices since it was created?

Completing this exercise requires you to perform two queries with the Wikipedia pageview API, because there are two types of mobile device counts—mobile-app and mobile-web—and you can only query them one at a time.


3. How many times was Panama Papers viewed in the first week? What proportion of those views came from mobile devices?

Completing this exercise also requires two API requests: one to gather pageview data for ALL devices, and then performing a request that only gathers data about devices that viewed the page using the Wikipedia mobile website.


4. How many other articles has User:Czar edited on Wikipedia since they created Panama Papers?

The sample query above returns a list of edits by Czar between 4/03 and 5/01, that includes the title of each article that was edited. How would you iterate over this JSON data in Python to come up with a list that contained the title of every article Czar edited, with no duplicates?

Coding challenges[edit]

Once you have worked through these exercises and feel confident that you can gather data from the Wikipedia edit and pageview APIs using Python, you can get started on this week's coding challenges!

Resources[edit]


Research using Wikipedia data[edit]


Websites that use the MediaWiki API[edit]