[10:19:36] Hi, I am iterning with the abstract wikipedia team and I wanted to gather pageview informations. [10:19:36] I used the php API (https://www.mediawiki.org/wiki/API:Main_page) and saw we only have pageviews for the last 60 days. [10:19:36] Later I found this page (https://pageviews.toolforge.org/) which uses a REST API (https://wikimedia.org/api/rest_v1) to generate info. I queried the REST API and I could get information from last year! Although I couldnt find the source code to this API or any doc on how many days of data they hold on to. [10:19:36] I was wondering what can be the best way to get pageview information about pages and where in the databases are these stored anyways? (i dont see pageviews in databases). [10:19:36] I running things from toolforge and have used databses and php API to gather other data for our needs. [10:20:29] Before I start using secondary resources, I'd like to know how they work or fetch the data from source myself? What can be a good approach? [10:26:20] tanny411: raw pageview data is retained for something like 90 day. Aggregate data is retained indefinitely. [10:26:47] docs are at https://wikitech.wikimedia.org/wiki/Analytics/AQS/Pageviews [10:28:10] The rest API is the source of truth, everything else uses that. The various toolforge tools provide more convenient access. [18:10:37] tanny411: the PHP API is mainly convenience since there is already familiarity with and tooling for that API. Also it has heavier caching which is sometimes useful when the software makes use of the data internally.