1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
|
import json
from itertools import combinations
from os import path
import __main__
from f1elo.model import *
class Elo:
def __init__(self, session):
self.session = session
self.config = json.load(
open(path.join(path.dirname(__main__.__file__), 'config', 'elo.json')))
def get_ranking(self, driver, rank_date=None):
rank = driver.get_ranking(rank_date)
if rank:
return rank.ranking
return self.config['initial_ranking']
def get_entry_ranking(self, entry, date=None):
return sum([self.get_ranking(d, date) for d in entry.drivers]) / len(entry.drivers)
def rank_race(self, race):
entries = race.entries
entries_to_compare = []
rankings = {}
new_rankings = {}
for e in entries:
rankings[e] = self.get_entry_ranking(e, race.date)
new_rankings[e] = 0.0
if e.result_group:
entries_to_compare.append(e)
for c in combinations(entries_to_compare, 2):
score = self.get_score(
rankings[c[0]] - rankings[c[1]],
self.get_outcome(c),
self.get_importance(race,
[rankings[c[0]],
rankings[c[1]]]))
new_rankings[c[0]] += score
new_rankings[c[1]] -= score
return new_rankings
def get_importance(self, race, rankings):
base_importance = self.config['importance'][race.type.code]
min_rank = min(rankings)
if min_rank < min(self.config['importance_threshold']):
return base_importance
if min_rank <= max(self.config['importance_threshold']):
return base_importance * 0.75
return base_importance / 2
def get_outcome(self, entries):
if entries[0].result_group < entries[1].result_group:
return 1
elif entries[0].result_group > entries[1].result_group:
return 0
return 0.5
def get_score(self, difference, outcome, importance):
return importance * (outcome - 1 / (1 + (10 ** (-difference / self.config['disparity']))))
|