Patch 2.9
Of all games, about 4.9% was not available to collect
Match Id recovered: 151061
Match metadata recovered: 101921
Ranked games: 40435 matches 1
Match metadata still missing: 45124
| Region | Play Rate | N |
|---|---|---|
| Bilgewater | 2.92% | 4707 |
| Demacia | 5.37% | 8668 |
| Freljord | 11.3% | 18240 |
| Ionia | 11.49% | 18559 |
| MtTargon | 7.55% | 12197 |
| Noxus | 11.89% | 19204 |
| PnZ | 9.22% | 14884 |
| ShadowIsles | 14.61% | 23584 |
| Shurima | 25.65% | 41415 |
| Region | Play Rate | America | Asia | Europe |
|---|---|---|---|---|
| Bilgewater | 2.96% | 3.04% | 3.80% | 2.20% |
| Demacia | 5.75% | 5.76% | 6.12% | 5.44% |
| Freljord | 12.02% | 12.45% | 11.13% | 12.25% |
| Ionia | 14.17% | 13.10% | 13.37% | 15.94% |
| MtTargon | 7.97% | 7.91% | 8.99% | 7.24% |
| Noxus | 11.26% | 10.29% | 11.99% | 11.71% |
| PnZ | 9.69% | 9.68% | 10.25% | 9.25% |
| ShadowIsles | 16.27% | 17.11% | 15.25% | 16.19% |
| Shurima | 19.91% | 20.66% | 19.10% | 19.77% |
In this section I provide the play rate of which combinations of champions (plus the regions) are used in a deck. The champions showed right before a game starts for example. Right now it’s a simple approximation of the archetypes played in the ladder as such information is not restrictive enough.
With the previous reports showing that there’s a huge cluster of decks all with a very low win rate the following graph will show the play rate day by day only for the most played decks.
Tie games are excluded
Win rates of the most played combination of champions, against all decks, this week. This time I left all cases with more than 1000 games
Top Win rates of the least played combination of champions. Min 300 games and a play rate of less than 2% play rate (in this sample)
The win rates on the grid are among the 10 most played champion combination. Match-ups with less than 300 games are not included
While this is one of the most interesting data for many the results are still heavily affected by the small sample size. They may be the 10 most played combination of champions but it’s still a 100 cell grid. Also, while my approximation for archetypes usually works, this weeks more than usual, two decks are not well reported: Dragons and Azir/Noxus. In these two cases the inclusion of J4 / Garen / Zoe and Draven / Darius create different values for in the end it’s the same deck with just a couple of different card. I already have an idea and part of the code how to solve this but will use it for next week. I could use Dr.LoR approach with bayesian statistics but I want to be consistent with my metholody in order to have a better comparison for my results. Of course I can still make changes in the future if the quality vastly improve with them.
No filter this times because of the small sample size
The LMI is an Index I developed to measure the performance of decks in the metagame. For those who are familiar with basic statistical concept I wrote a document to explain the theory behind it: LMI - Early Theory, it’s very similar to vicioussyndicate (vS) Meta Score from their data reaper report.
The score of each deck is not just their “strength”, it takes in consideration both play rates and win rates that’s why I prefer to say it measure the “performance”.
The values range from 0 to 100 and the higher the value, the higher is the performance.
It’s just an early concept and will surely be expanded in the future. 2
The following table are the values I used:
This Meta Report was created under Riot Games’ “Legal Jibber Jabber” policy using assets owned by Riot Games. Riot Games does not endorse or sponsor this project.