Cristian Consonni bio photo

Cristian Consonni

Ph.D. in Computer Science, free software activist, physicist and storyteller

Email Twitter Facebook LinkedIn Github Stackoverflow keybase

Datasets: Wikipedia pagecounts-all-sites sorted by page (years 2014 – 2016)

This dataset consists of hourly pagecounts for Wikipedia pages sorted by article, ordered by (project, page, timestamp). It has been created by processing Wikimedia’s pagecounts-all-sites dataset.

The original dataset holds output from September, 2014 to August, 2016 that mimics pagecounts-raw files, but gets generated from Hadoop data using Hive. The original dataset holds the desktop sites’ pageview data with hourly granularity for all Wikipedia editions. More info about the original dataset are available on Wikitech. Note that these are not unique visits. Note that this dataset files have a one hour shift later than any other dataset handled by the analytics-team (particularly webrequest, pageview-hourly, projectview-hourly). For instance for data between 2018-09-27T13:00:00 and 2018-09-27T14:00:00, pagecounts-all-sites uses 2018-09-27T14:00:00 while other dataset uses 2018-09-27T13:00:00.

The CSV uses spaces as delimiter, without any form of escaping because it is not needed. It has 5 columns:

  • project: the project name
  • page: the page requested, url-escaped
  • timestamp: the timestamp of the hour (format: %Y%m%d-%H%M%S)
  • count: the number of times the page has been requested (in that hour)
  • bytes: the number of bytes transferred (in that hour)

The original dataset has been normalized in the following ways:

  • the project column has been converted to lowercase
  • the page columns has been unquoted and then re-quoted according to RFC 1308, using the following equivalent Python 3 code:
import urllib.parse
page_unquoted = urllib.parse.unquote(page, encoding='utf-8',
page_requoted = urllib.parse.quote(page_unquoted)
  • if two lines are now equal because of this normalization, their count and bytes columns are summed up.

This dataset is split in many gzip’d files, each of them containing 1,000,000 records.

An index folder is included: for every month there is a file describing the the first record of every partial file.

Here’s an excerpt of the file pagecounts-20140923-010000.gz (lines 1612915–1612924):

en Albert_Einstein 505 44692476
en Albert_Einstein%27s_brain 13 259793
en Albert_Einstein%27s_political_views 2 60335
en Albert_Einstein%27s_religious_views 3 101366
en Albert_Einstein's_brain 12 159387
en Albert_Einstein:_The_Practical_Bohemian 1 30429
en Albert_Einstein_(album) 2 37854
en Albert_Einstein_(disambiguation) 1 9232
en Albert_Einstein_ATV 2 142976
en Albert_Einstein_Award 1 14635

and in the processed dataset in file 2014-09/part-00082.gz (lines 9079271–9079280) you can find:.

en Albert_Einstein 20140923-090000 514 50705784
en Albert_Einstein 20140923-230000 493 47997224
en Albert_Einstein 20140925-000000 477 43672460
en Albert_Einstein 20140925-020000 522 47596725
en Albert_Einstein 20140925-190000 840 74566033
en Albert_Einstein 20140926-070000 420 40327383
en Albert_Einstein 20140926-220000 427 38073829
en Albert_Einstein 20140927-140000 577 51050092
en Albert_Einstein 20140928-060000 457 38843362
en Albert_Einstein 20140928-190000 727 61539311


This dataset can be downloaded in two different ways:


You can find the dataset on:

You can use the scripts at pagecounts-download-tools on GitHub.

How-to download a month worth of data

  1. clone the repository:
        ╭─ ~
        ╰─$ git clone
  2. go to the `sizes` directory and execute the download sizes:
        ╭─ ~/pagecounts-download-tools/sizes
        ╰─$ ./
  3. go to the `downloadlists` directory and execute the download sizes:
        ╭─ ~/pagecounts-download-tools/downloadlists
        ╰─$ ./ ../sizes/2014-09.txt
  4. from the repository base directory and dowload files:
        ╭─ ~/pagecounts-download-tools
        ╰─$ ./ ./ -d 2014 9

dat (experimental)

You can download the dataset using dat, the dataset is available at

Once you have installed dat, you can download the dataset with:

dat clone dat://d4ac75cda06e991b3181abb7365a1761581c2d54e962f14015f52ed5c8e9f6b2 ~/dat-wikipedia-pagecounts-all-sites-sorted




The original dataset was published in the Public Domain (Public Domain Mark 1.0).

You can reuse this dataset under the same license.

How to cite

Please cite this dataset as:

Cristian Consonni, Alberto Montresor. Wikipedia pagecounts-raw sorted by article. doi coming soon

This dataset superseedes the previous version with just the data from 2014: doi:10.6084/m9.figshare.2085643.v1, see also datasets/wikipedia-pagecounts-sorted-by-page-year-2014.


What is the total size of the dataset, the number of files and the largest file in the dataset?

The total dataset size is 1.1TB, and it contains ~15,800 files. The average size is 35 MB and the largest file is 1.22GB.

How are files organized?

Files are divided in directories, one for each month, like this:

├── 2014-09
│   ├── part-00000.gz
│   ...
│   └── part-00190.gz
├── 2014-10
│   ├── part-00000.gz
│   ...
│   └── part-00743.gz
├── 2014-11
└── 2016-08
    ├── part-00000.gz
    └── part-00108.gz

The average directory size is ~45GB, and on average each contains 657 part files.

Who produced this dataset and why?

The dataset has been produced by Cristian Consonni and Alberto Montresor, from the Department of Information Engineering and Computer Science (DISI), University of Trento, Trento, Italy.

This research has been supported by Microsoft Azure Research Award CRM:0518942 as part of the “Azure for Research Award: Data Science” program.

This dataset has also been utilized in the research related to the ENGINEROOM project, in collaboration with David Laniado of Eurecat - Centre Tecnològic de Catalunya, Barcelona, Spain. EU ENGINEROOM has received funding from the European Union’s Horizon 2020 research and innovation programme under the Grant Agreement no 780643.

Is this dataset used in currently in-review or future papers that cite it?

This dataset has been used for multiple papers that are currently in-review or in preparation.

This dataset is published as part of the ENGINEROOM project and it will be cited in the related scientific publications.


For further info send me an e-mail.