I need to get the number of daily pageviews of the English Wikipedia article on "Dollar" and "Euro" from 06/2012-06/2016.
Raw dumps (*.bz2) are available at: https://dumps.wikimedia.org/other/pagecounts-ez/merged/
For example, https://dumps.wikimedia.org/other/pagecounts-ez/merged/pagecounts-2014-01-views-ge-5-totals.bz2 provides hourly/daily data for January 2014.
Problem: The unzipped files are too big to be opened in any text editor.
Desired solution: A Python script (?) that reads each of the .bz2 files, searches for the en wikipedia "Dollar" / "Euro" entry only and puts the daily pageviews into a Data Frame.
Hint: Using the Pageviews API (https://wikitech.wikimedia.org/wiki/Pageviews_API) won't be helpful as I'll need consistent data before 2015. stats.grok data (http://stats.grok.se/) is neither an option, as the generated data is different and incompatible.