import __main__ from os import path from f1elo.model import * import json from itertools import combinations class Elo: def __init__(self, session): self.session = session self.config = json.load(open(path.dirname(__main__.__file__) + '/config/elo.json')) def get_ranking(self, driver, rank_date=None): rank = driver.get_ranking(rank_date) if rank: return rank.ranking return self.config['initial_ranking'] def get_entry_ranking(self, entry, date=None): return sum([self.get_ranking(d, date) for d in entry.drivers]) / len(entry.drivers) def rank_race(self, race): entries = race.entries entries_to_compare = [] rankings = {} new_rankings = {} for e in entries: rankings[e] = self.get_entry_ranking(e, race.date) new_rankings[e] = 0.0 if e.result_group: entries_to_compare.append(e) for c in combinations(entries_to_compare, 2): score = self.get_score(rankings[c[0]] - rankings[c[1]], self.get_outcome(c), self.get_importance(race, [rankings[c[0]], rankings[c[1]]])) #print c[0], '@', rankings[c[0]] #print 'against' #print c[1], '@', rankings[c[1]] #print 'score: ', score #print new_rankings[c[0]] += score new_rankings[c[1]] -= score return new_rankings def get_importance(self, race, rankings): base_importance = self.config['importance'][race.type.code] min_rank = min(rankings) if min_rank < 2100: return base_importance if min_rank <= 2400: return base_importance * 0.75 return base_importance / 2 def get_outcome(self, entries): if entries[0].result_group < entries[1].result_group: return 1 elif entries[0].result_group > entries[1].result_group: return 0 return 0.5 def get_score(self, difference, outcome, importance): return importance * (outcome - 1 / (1 + (10 ** (-difference / self.config['disparity']))))