Swiss Snow Report Data Analysis

By clicking on an active square a histogram of the fresh snowfall over the past 15 years for a given coordinate pops up.
1The median describes the "middle" amount of snowfall in a given time period (Wiki).
2"Major Dump Probability" describes the ratio of years with over 30cm snowfall within 48h in a given time period.
3This displays the rank of the median snowfall (1 is the most snowfall).

Caution

The data presented herein is based on the daily snow reports from the swiss avalanche institute (SLF). Because raw snow data is not available I first scraped the available archive of snow reports from the last 15 years in the form of 2'614 pictures. The pictures depict estimated fresh snow models from the last 24 hours, a service from SLF. Unfortunately pictures from different seasons had slightly different formats, so I had to detect the coordinates of the borders on the map with the help of Open CV.

Categories in cm
The maps were then split into a grid of 34 x 22 squares, for which the most likely fresh snow category was defined via histogram correlation. To sum up the daily fresh snow data for the months December to April, the average value of each range category was assumed (e.g. "1-10cm" = 5 cm; "10-25cm" = 17.5 cm), which leads to major inaccuracies. A major dump (more than 30cm of snowfall within 48h) could therefore be due to a single "25-50cm" snow report (average is 37.5), or two successive "10-25cm" reports. All code and data is available on Github.

This is an ongoing leisure project, so don't hold me accountable ;)

PS:

I'm currently looking for a Data-Science Internship. Know something or somebody? Let me know via helloadrianoesch[at]gmail.com or LinkedIn.

Feedback | About Me | Twitter