There are millions of projects on GitHub. Every day, people from around the world are working to make these projects better. Opening issues, pushing code, submitting Pull Requests, discussing project details — GitHub activity is a papertrail of progress. Have you ever wondered what all that data looks like? There are millions of stories to tell; you just have to look.

Last year we held our first data challenge.
We saw incredible visualizations, interesting timelines and compelling analysis.

What stories will be told this year? It’s up to you!

To Enter

Send a link to a GitHub repository or gist with your graph(s) along with a description to before midnight, May 8th, 2013 PST.

Data access

The GitHub public timeline is a featured public dataset available on Google BigQuery. The “timeline” table has over a year’s worth of public activity and is approaching 100M rows. You can find even more data available in JSON format on The GitHub Archive project.

You are free to use any tools you like. If you choose to use BigQuery, running queries against the GitHub dataset is free for the first 100GB of query processing. Pricing information for additional query processing is available here. After signing up for BigQuery, add the project name “githubarchive”.



GitHub staff will be voting on our favorite visualizations and there will be prizes for the top three spots:

We will also feature the three winning entries on the GitHub blog. Winners will be announced the week of May 20th.

Analyzing Millions of GitHub Commits

Last year Ilya Grigorik and I spoke at the Strata conference, showcasing some of the interesting analysis and visualizations from the data challenge including which programming language results in most frustration (VimL), amusement (Ruby) and surprise (Perl).