Data

Patch 2.9

Regions’ Frequencies

By number of times a region is chosen

Region Play Rate N
Bilgewater 2.92% 4707
Demacia 5.37% 8668
Freljord 11.3% 18240
Ionia 11.49% 18559
MtTargon 7.55% 12197
Noxus 11.89% 19204
PnZ 9.22% 14884
ShadowIsles 14.61% 23584
Shurima 25.65% 41415

By number of times a card within a region is chosen

Region Play Rate America Asia Europe
Bilgewater 2.96% 3.04% 3.80% 2.20%
Demacia 5.75% 5.76% 6.12% 5.44%
Freljord 12.02% 12.45% 11.13% 12.25%
Ionia 14.17% 13.10% 13.37% 15.94%
MtTargon 7.97% 7.91% 8.99% 7.24%
Noxus 11.26% 10.29% 11.99% 11.71%
PnZ 9.69% 9.68% 10.25% 9.25%
ShadowIsles 16.27% 17.11% 15.25% 16.19%
Shurima 19.91% 20.66% 19.10% 19.77%

Champions’ combinations

In this section I provide the play rate of which combinations of champions (plus the regions) are used in a deck. The champions showed right before a game starts for example. Right now it’s a simple approximation of the archetypes played in the ladder as such information is not restrictive enough.

Date by date

With the previous reports showing that there’s a huge cluster of decks all with a very low win rate the following graph will show the play rate day by day only for the most played decks.

Win rates

Tie games are excluded

Win rates of the most played combination of champions, against all decks, this week. This time I left all cases with more than 1000 games

Underdog(?) win rates

Top Win rates of the least played combination of champions. Min 300 games and a play rate of less than 2% play rate (in this sample)

Most played combination

The win rates on the grid are among the 10 most played champion combination. Match-ups with less than 300 games are not included

Match-up grid

While this is one of the most interesting data for many the results are still heavily affected by the small sample size. They may be the 10 most played combination of champions but it’s still a 100 cell grid. Also, while my approximation for archetypes usually works, this weeks more than usual, two decks are not well reported: Dragons and Azir/Noxus. In these two cases the inclusion of J4 / Garen / Zoe and Draven / Darius create different values for in the end it’s the same deck with just a couple of different card. I already have an idea and part of the code how to solve this but will use it for next week. I could use Dr.LoR approach with bayesian statistics but I want to be consistent with my metholody in order to have a better comparison for my results. Of course I can still make changes in the future if the quality vastly improve with them.

Match-up table

No filter this times because of the small sample size

LoR-Meta Index (LMI)

The LMI is an Index I developed to measure the performance of decks in the metagame. For those who are familiar with basic statistical concept I wrote a document to explain the theory behind it: LMI - Early Theory, it’s very similar to vicioussyndicate (vS) Meta Score from their data reaper report.
The score of each deck is not just their “strength”, it takes in consideration both play rates and win rates that’s why I prefer to say it measure the “performance”.

The values range from 0 to 100 and the higher the value, the higher is the performance.

It’s just an early concept and will surely be expanded in the future. 2

The following table are the values I used:

Cards’ presence

Most played cards

Top 3 most used cards within each Region


    • Ranked games / Patch 2.8 / Master players

    • Last update: 2021-06-08 17:30 (UTC)

    ↩︎
  1. Judging some of the early feedback, it’s very likely I’m going to change the aggregation method.↩︎