When initializing with priors - convergence rescales the average rating far away from the baseline priors.
import trueskillthroughtime as ttt
player = set_of_player_names
priors = {player: ttt.Player(ttt.Gaussian(mu=25.0, sigma=8.3333)) for matchup in composition for team in matchup for player in team}
h = ttt.History(
composition=composition,
priors=priors,
)
# Get pre convergence values
latest_rating_values = [rating_array[-1][1].mu for rating_array in h.learning_curves().values()]
avg_rating_value = sum(latest_rating_values)/len(latest_rating_values)
h.convergence()
converged_latest_rating_values = [rating_array[-1][1].mu for rating_array in h.learning_curves().values()]
avg_converged_value = sum(converged_latest_rating_values)/len(converged_latest_rating_values)
print(f"{avg_rating_value} -> {avg_converged_value}")
This results in: 23.776132491952705 -> 5.261152040525114
Is there a good reason for this?
When initializing with priors - convergence rescales the average rating far away from the baseline priors.
This results in: 23.776132491952705 -> 5.261152040525114
Is there a good reason for this?