|
|
Line 1: |
Line 1: |
| [[File:ArpadElo.jpg|right|thumb|[[Arpad Elo]], the inventor of the Elo rating system]]
| | Dolomiti Stars is a dedicated tourism consortium that has set off to promote several areas in the Italy such as the [http://www.givefreeachance.com/article.php?id=559200 Dolomites]. Because there are more and more people who visit Italy every year, it was high time that these tourists discovered others areas such as the Arabba or the Marmolada mountain. This part of Italy is worth visiting in the summer, as well as in the winter, as there are plenty of activities that can be done in both periods. Dolomites Stars can inform clients about this topic. Since the majority of clients might be confused regarding the actual meaning of a tourism consortium, it is best to clarify this aspect from the beginning. The idea of creating such a company is based on the need for promotion. Although individuals passionate about winter sports are familiar to Arrabba ski resorts for instance, it might be a good idea to attract other tourists, perhaps those that love trekking routes or walks in wood. These are all activities that can be done in the [http://www.givefreeachance.com/article.php?id=559200 Dolomites Italy] area. By making use of the services offered by Dolomiti Stars, all clients are awarded with exactly what they expect, a greatly organized vacation in the outdoors. <br><br>As mentioned in the beginning, the focus of this tourism consortium is that of promoting traveling locations. This is performed by means of information. Dolomiti Stars offers interested visitors all the details they might be in need of, before deciding to travel to any of the locations promoted on the website. The online platform is adequately structured, providing all clients with revealing pieces of information on topics such as accommodation, things to do or see, events, restaurants, ski resorts and the list can go on. Each one of the categories mentioned above is adequately adapted according to the user. For instance, if the client is looking for a specific type of accommodation, he or she may rest assured that the Dolomoti Stars system will offer a true diversity in options and alternatives, as well as a search bar that will direct the client to the best choice for his profile, of course after certain details are provided. Also, since there is no better promotion than by means of deals and packages, know that these can be found on the official platform. Indeed, Dolomiti Stars can provide the clients with several such options that are suitable for different budgets. All the services included in these packages are of the highest quality. <br><br>In order to gain access to these offers, as well as to the pieces of information regarding the entire Dolomites area, clients have to visit the official website. If there should be further questions regarding the packages found on the website or the apartments or hotels listed, clients should feel free to contact the team working there. The necessary details will be found on the online platform. It is worth mentioning that the staff working at Dolomiti Stars is highly professional, dedicated to its work and ready to answer all the questions and inquiries clients might have. |
| The '''Elo rating system''' is a method for calculating the relative skill levels of players in such competitor-versus-competitor games as [[chess]]. It is named after its creator [[Arpad Elo]], a [[Hungary|Hungarian]]-born [[United States|American]] [[physics]] professor.
| |
| | |
| The Elo system was invented as an improved [[chess rating system]] and is also used in many other games. It is also used as a rating system for multiplayer competition in a number of [[video game]]s,<ref>{{cite web |url=http://www.enemydown.eu/news/502 |title=Enemydown uses Elo in its Counterstrike:Source Multilplayer Ladders |last=Wood |first=Ben |publisher=TWNA Group}}</ref> and has been adapted to team sports including [[association football]], American college football, basketball, [[Major League Baseball]], and [[eSports]].
| |
| | |
| The difference in the ratings between two players serves as a predictor of the outcome of a match. Two players with equal ratings who play against each other are expected to score an equal number of wins. A player whose rating is 100 points greater than his or her opponent's is expected to score 64%; if the difference is 200 points, then the expected score for the stronger player is 76%.<ref>In chess, a win counts 1 point and a [[draw (chess)|draw]] counts ½ point. As an example, suppose that in an eight-game match a player wins three games, draws three, and loses two. The player gains 4½ points out of 8, a 56.25% score.</ref>
| |
| | |
| A number represents a player's Elo rating and increases or decreases based upon the outcome of games between rated players. After every game, the winning player takes points from the losing one. The difference between the ratings of the winner and loser determines the total number of points gained or lost after a game. In a series of games between a high-rated player and a low-rated player, the high-rated player is expected to score more wins. If the high-rated player wins, then only a few rating points will be taken from the low-rated player whereas if the lower rated player scores an upset win, then many rating points will be transferred. The lower rated player will also gain a few points from the higher rated player in the event of a draw. The rating system thus is self-correcting. A player whose rating is too low should, in the long run, do better than the rating system predicts and thereby gain rating points until the rating reflects his or her true playing strength.
| |
| | |
| ==History==
| |
| Arpad Elo was a master-level chess player and an active participant in the [[United States Chess Federation]] (USCF) from its founding in 1939.<ref>{{cite web |url=http://www.uschess.org/ratings/RedmanremembersRichard.pdf |title=Remembering Richard, Part II |last=Redman |first=Tim |publisher=United States Chess Federation}}</ref> The USCF used a numerical ratings system, devised by [[Kenneth Harkness]], to allow members to track their individual progress in terms other than tournament wins and losses. The Harkness system was reasonably fair, but in some circumstances gave rise to ratings which many observers considered inaccurate. On behalf of the USCF, Elo devised a new system with a more sound [[Statistics|statistical]] basis.
| |
| | |
| Elo's system replaced earlier systems of competitive rewards with a system based on statistical estimation. Rating systems for many sports award points in accordance with subjective evaluations of the 'greatness' of certain achievements. For example, winning an important [[golf]] tournament might be worth an arbitrarily chosen five times as many points as winning a lesser tournament.
| |
| | |
| A statistical endeavor, by contrast, uses a model that relates the game results to underlying variables representing the ability of each player.
| |
| | |
| Elo's central assumption was that the chess performance of each player in each game is a [[normal distribution|normally distributed]] [[random variable]]. Although a player might perform significantly better or worse from one game to the next, Elo assumed that the mean value of the performances of any given player changes only slowly over time. Elo thought of a player's true skill as the mean of that player's performance random variable.
| |
| | |
| A further assumption is necessary, because chess performance in the above sense is still not measurable. One cannot look at a sequence of moves and say, "That performance is 2039." Performance can only be inferred from wins, draws and losses. Therefore, if a player wins a game, he is assumed to have performed at a higher level than his opponent for that game. Conversely if he loses, he is assumed to have performed at a lower level. If the game is a draw, the two players are assumed to have performed at nearly the same level.
| |
| | |
| Elo did not specify exactly how close two performances ought to be to result in a draw as opposed to a win or loss. And while he thought it is likely that each player might have a different [[standard deviation]] to his performance, he made a simplifying assumption to the contrary.
| |
| | |
| To simplify computation even further, Elo proposed a straightforward method of estimating the variables in his model (''i.e.'' the true skill of each player). One could calculate relatively easily, from tables, how many games a player would be expected to win based on a comparison of his rating to the ratings of his opponents. If a player won more games than expected, his rating would be adjusted upward, while if he won fewer than expected his rating would be adjusted downward. Moreover, that adjustment was to be in exact linear proportion to the number of wins by which the player had exceeded or fallen short of his expected number.
| |
| | |
| From a modern perspective, Elo's simplifying assumptions are not necessary because computing power is inexpensive and widely available. Moreover, even within the simplified model, more efficient estimation techniques are well known. Several people, most notably [[Mark Glickman]], have proposed using more sophisticated statistical machinery to estimate the same variables. On the other hand, the computational simplicity of the Elo system has proven to be one of its greatest assets. With the aid of a pocket calculator, an informed chess competitor can calculate to within one point what his next officially published rating will be, which helps promote a perception that the ratings are fair.
| |
| | |
| === Implementing Elo's scheme ===
| |
| The [[United States Chess Federation|USCF]] implemented Elo's suggestions in 1960,<ref name="aboutUSCF">{{cite web |url=http://www.uschess.org/about/about.php |title=About the USCF |publisher=United States Chess Federation |accessdate=2008-11-10}}</ref> and the system quickly gained recognition as being both more fair and accurate than the [[Chess rating systems#Harkness system|Harkness rating system]]. Elo's system was adopted by the [[Fédération Internationale des Échecs|World Chess Federation]] (FIDE) in 1970. Elo described his work in some detail in the book ''The Rating of Chessplayers, Past and Present'', published in 1978.
| |
| | |
| Subsequent statistical tests have suggested that chess performance is almost certainly not distributed as a [[normal distribution]], as weaker players have significantly (but not highly significantly) greater winning chances than Elo's model predicts.{{citation needed|date=June 2012}} Therefore, the USCF and some chess sites use a formula based on the [[logistic distribution]]. Significant statistical anomalies have also been found when using the logistic distribution in chess.<ref>{{cite web | url = http://blog.kaggle.com/2011/04/24/the-deloittefide-chess-competition-play-by-play/ | title = Deloitte Chess Rating Competition }}</ref> FIDE continues to use the normal distribution. The normal and logistic distribution points are, in a way, arbitrary points in a spectrum of distributions which would work well. In practice, both of these distributions work very well for a number of different games.
| |
| | |
| == Different ratings systems ==
| |
| The phrase "Elo rating" is often used to mean a player's chess rating as calculated by FIDE. However, this usage is confusing and misleading, because Elo's general ideas have been adopted by many organizations, including the USCF (before FIDE), the [[Internet Chess Club]] (ICC), [[Free Internet Chess Server]] (FICS), [[Yahoo!]] Games, and the now-defunct [[Professional Chess Association]] (PCA). Each organization has a unique implementation, and none of them follows Elo's original suggestions precisely. It would be more accurate to refer to all of the above ratings as Elo ratings, and none of them as ''the'' Elo rating.
| |
| | |
| Instead one may refer to the organization granting the rating, e.g. "As of August 2002, [[Gregory Kaidanov]] had a FIDE rating of 2638 and a USCF rating of 2742." It should be noted that the Elo ratings of these various organizations are not always directly comparable. For example, someone with a FIDE rating of 2500 will generally have a USCF rating near 2600 and an ICC rating in the range of 2500 to 3100. On FICS, the command "tell surveybot ratings fide estimate" gives an FIDE rating estimate based on the user's FICS rating.
| |
| | |
| ===FIDE ratings===
| |
| {{See also|FIDE World Rankings}}
| |
| For top players, the most important rating is their [[FIDE]] rating. Since July 2012, FIDE issues a ratings list once every month.
| |
| | |
| The following analysis of the November 2012 FIDE rating list gives a rough impression of what a given FIDE rating means:
| |
| * 5839 players had an active rating in the range 2200 to 2299, which is usually associated with the [[Candidate Master]] title.
| |
| * 2998 players had an active rating in the range 2300 to 2399, which is usually associated with the [[FIDE Master]] title.
| |
| * 1382 players had an active rating between 2400 and 2499, most of whom had either the [[International Master]] or the [[International Grandmaster]] title.
| |
| * 587 players had an active rating between 2500 and 2599, most of whom had the [[International Grandmaster]] title.
| |
| * 178 players had an active rating between 2600 and 2699, all but one had the [[International Grandmaster]] title.
| |
| * 42 players had an active rating between 2700 and 2799.
| |
| * 4 active players had a rating over 2800: [[Magnus Carlsen]], [[Vladimir Kramnik]], [[Veselin Topalov]] and [[Levon Aronian]].
| |
| | |
| November 2011 marked the first time four players had a rating of 2800 or higher. The highest ever FIDE rating was 2872, which [[Magnus Carlsen]] had on the January 2013 list. A list of highest ever rated players is at [[Comparison of top chess players throughout history]].
| |
| | |
| ====Performance rating====
| |
| {| class="wikitable" style="width:100px; float:right; text-align:center; margin-top:0; margin-left:10px;"
| |
| |-
| |
| ! style="width:50%;"|<math>p</math>
| |
| ! style="width:50%;"|<math>d_p</math>
| |
| |-
| |
| | 1.00 || +800
| |
| |-
| |
| | 0.99 || +677
| |
| |-
| |
| | 0.9 || +366
| |
| |-
| |
| | 0.8 || +240
| |
| |-
| |
| | 0.7 || +149
| |
| |-
| |
| | 0.6 || +72
| |
| |-
| |
| | 0.5 || 0
| |
| |-
| |
| | 0.4 || -72
| |
| |-
| |
| | 0.3 || -149
| |
| |-
| |
| | 0.2 || -240
| |
| |-
| |
| | 0.1 || -366
| |
| |-
| |
| | 0.01 || -677
| |
| |-
| |
| | 0.00 || -800
| |
| |}
| |
| Performance rating is a hypothetical rating that would result from the games of a single event only. Some chess organizations use the "algorithm of 400" to calculate performance rating. According to this algorithm, performance rating for an event is calculated by taking (1) the rating of each player beaten and adding 400, (2) the rating of each player lost to and subtracting 400, (3) the rating of each player drawn, and (4) summing these figures and dividing by the number of games played.<br />
| |
| Example: 2 Wins, 2 Losses<br />
| |
| <math>\displaystyle{\frac{\left(w+400+x+400+y-400+z-400\right)}{4}}</math><br />
| |
| <math>\displaystyle{\frac{\left[w+x+y+z+400\left(2\right)-400\left(2\right)\right]}{4}}</math><br />
| |
| This can be expressed by the following formula:<br />
| |
| <math>
| |
| \text{Performance rating} = \frac{\text{Total of opponents' ratings } + 400 \times (\text{Wins} - \text{Losses})}{\text{Games}}</math><br />
| |
| Example: If you beat a player with an Elo rating of 1000,<br />
| |
| <math>
| |
| \displaystyle
| |
| \text{Performance rating} = \frac{1000 + 400 \times (1)}{1} = 1400</math><br />
| |
| If you beat two players with Elo ratings of 1000,<br />
| |
| <math>
| |
| \displaystyle
| |
| \text{Performance rating} = \frac{2000 + 400 \times (2)}{2} = 1400</math><br />
| |
| If you draw,<br />
| |
| <math>
| |
| \displaystyle
| |
| \text{Performance rating} = \frac{1000 + 400 \times (0)}{1} = 1000</math><br />
| |
| | |
| This is a simplification, because it doesn't take account of K-factors (this factor is explained further below), but it offers an easy way to get an estimate of PR (performance rating).
| |
| | |
| [[FIDE]], however, calculates performance rating by means of the formula: Opponents' Rating Average + Rating Difference. Rating Difference <math>d_p</math> is based on a player's tournament percentage score <math>p</math>, which is then used as the key in a lookup table where <math>p</math> is simply the number of points scored divided by the number of games played. Note that, in case of a perfect or no score <math>d_p</math> is 800. The full table can be found in the [http://www.fide.com/fide/handbook.html?id=163&view=article FIDE handbook, B. Permanent Commissions 1.0. Requirements for the titles designated in 0.31, 1.48 ] online. A simplified version of this table is on the right.
| |
| {{-}}
| |
| | |
| ====FIDE tournament categories<!-- "Category (chess)" redirects here. If renaming the cat, fix that redirect please. -->====
| |
| {| class="wikitable" style="float:right; text-align:center; margin-top:0; margin-left:10px"
| |
| |-
| |
| ! rowspan="2" | Category
| |
| ! colspan="2" | Average rating
| |
| |-
| |
| ! Minimum
| |
| ! Maximum
| |
| |-
| |
| |14||2576 || 2600
| |
| |-
| |
| |15||2601 || 2625
| |
| |-
| |
| |16||2626 || 2650
| |
| |-
| |
| |17||2651 || 2675
| |
| |-
| |
| |18||2676 || 2700
| |
| |-
| |
| |19||2701 || 2725
| |
| |-
| |
| |20||2726 || 2750
| |
| |-
| |
| |21||2751 || 2775
| |
| |-
| |
| |22||2776 || 2800
| |
| |-
| |
| |23||2801 || 2825
| |
| |}
| |
| | |
| [[FIDE]] classifies tournaments into categories according to the average rating of the players. Each category is 25 rating points wide. Category 1 is for an average rating of 2251 to 2275, category 2 is 2276 to 2300, etc. For women's tournaments, the categories are 200 rating points lower, so a Category 1 is an average rating of 2051 to 2075, etc.<ref>[http://www.fide.com/component/handbook/?id=57&view=article FIDE Handbook, Section B.0.0], [[FIDE]] web site</ref> The highest rated tournament has been category 23, with an average from 2801 to 2825.<!-- (Zurich Chess Challenge 2014) --> The top categories are in the table.
| |
| {{-}}
| |
| | |
| ===Live ratings===
| |
| [[FIDE]] updates its ratings list at the beginning of each month. In contrast, the unofficial "Live ratings" calculate the change in players' ratings after every game. These Live ratings are based on the previously published FIDE ratings, so a player's Live rating is intended to correspond to what the FIDE rating would be if FIDE were to issue a new list that day.
| |
| | |
| Although Live ratings are unofficial, interest arose in Live ratings in August/September 2008 when five different players took the "Live" No. 1 ranking.<ref>Anand lost No. 1 to Morozevich ([http://www.chessbase.com/newsdetail.asp?newsid=4860 Chessbase, August 24 2008]), then regained it, then Carlsen took No. 1 ([http://www.chessbase.com/newsdetail.asp?newsid=4892 Chessbase, September 5 2008]), then Ivanchuk ([http://www.chessbase.com/newsdetail.asp?newsid=4901 Chessbase, September 11 2008]), and finally Topalov ([http://www.chessbase.com/newsdetail.asp?newsid=4908 Chessbase, September 13 2008])</ref>
| |
| | |
| The unofficial live ratings were published and maintained by Hans Arild Runde at [http://chess.liverating.org the Live Rating website] until August 2011. Another website [http://www.2700chess.com www.2700chess.com] has been maintained since May 2011 by Artiom Tsepotan. Both websites cover players over 2700 only.
| |
| | |
| Currently, the No. 1 spot in both the official FIDE rating list and the live rating list is taken by [[Magnus Carlsen]].
| |
| | |
| === United States Chess Federation ratings ===
| |
| The [[United States Chess Federation]] (USCF) uses its own classification of players:<ref>[http://www.uschess.org/ratings/ratedist.php US Chess Federation:]{{dead link|date=February 2012}}</ref>
| |
| *2400 and above: Senior Master
| |
| *2200–2399 plus 300 games above 2200: Original <!--The title has been changed, though I am too lazy to find a source--> Life Master<ref>[http://main.uschess.org/content/view/7327#Master USCF Glossary Quote:"a player who competes in over 300 games with a rating over 2200"] from The United States Chess Federation</ref>
| |
| *2200–2399: National Master
| |
| *2000–2199: Expert
| |
| *1800–1999: Class A
| |
| *1600–1799: Class B
| |
| *1400–1599: Class C
| |
| *1200–1399: Class D
| |
| *1000–1199: Class E
| |
| *800–999: Class F
| |
| *600–799: Class G
| |
| *400–599: Class H
| |
| *200–399: Class I
| |
| *100–199: Class J
| |
| | |
| In general, a beginner is around 800, a mid-level player is around 1600, and a professional, around 2400. In 2007, the median rating of all USCF members was 657.<ref>[http://www.eddins.net/steve/chess/2008/06/06/131 USCF Ratings Distribution 2008]</ref>
| |
| | |
| ====The K-factor used by the USCF====
| |
| The ''K-factor'', in the USCF rating system, can be estimated by dividing 800 by the effective number of games a player's rating is based on (''N<sub>e</sub>'') plus the number of games the player completed in a tournament ('''m''').<ref>[http://math.bu.edu/people/mg/ratings/approx/approx.html "Approximating Formulas for the USCF Rating System"], [[United States Chess Federation]], Mark Glickman, Department of Mathematics and Statistics at Boston University, 22 February 2001</ref>
| |
| : <math>K = 800 / (N_e + m)\, </math>
| |
| | |
| ====Rating floors====
| |
| The USCF maintains an absolute rating floor of 100 for all ratings. Thus, no member can have a rating below 100, no matter their performance at USCF sanctioned events. However, players can have higher individual absolute rating floors, calculated using the following formula:
| |
| :<math>AF = \operatorname{min}\{100+4N_W+2N_D+N_R , 150\}</math>
| |
| | |
| where <math>N_W</math> is the number of rated games won, <math>N_D</math> is the number of rated games drawn, and <math>N_R</math> is the number of events in which the player completed three or more rated games.
| |
| | |
| Higher rating floors exist for experienced players who have achieved significant ratings. Such higher rating floors exist, starting at ratings of 1200 in 100 point increments up to 2100 (1200, 1300, 1400, ..., 2100). A player's rating floor is calculated by taking their peak rating, subtracting 200 points, and then rounding down to the nearest rating floor. For example, a player who has reached a peak rating of 1464 would have a rating floor of 1464 − 200 = 1264, which would be rounded down to 1200. Under this scheme, only Class C players and above are capable of having a higher rating floor than their absolute player rating. All other players would have a floor of at most 150.
| |
| | |
| There are two ways to achieve higher rating floors other than under the standard scheme presented above. If a player has achieved the rating of Original Life Master, his or her rating floor is set at 2200. The achievement of this title is unique in that no other recognized USCF title will result in a new floor. For players with ratings below 2000, winning a cash prize of $2,000 or more raises that player's rating floor to the closest 100-point level that would have disqualified the player for participation in the tournament. For example, if a player won $4,000 in a 1750 and under tournament, the player would now have a rating floor of 1800.
| |
| | |
| ===Ratings of computers===
| |
| Since 2005–06, [[human-computer chess matches]] have demonstrated that [[chess computers]] are capable of defeating even the strongest human players ([[Deep Blue versus Garry Kasparov]]). However, ratings of computers are difficult to quantify. There have been too few games under tournament conditions to give computers or software engines an accurate rating.<ref>For instance, see the comments at [http://www.chessbase.com/newsdetail.asp?newsid=2476 ChessBase.com - Chess News - Adams vs Hydra: Man 0.5 – Machine 5.5<!-- Bot generated title -->]</ref> Also, for [[chess engine]]s, the rating is dependent on the machine a program runs on.
| |
| | |
| For some ratings estimates, see [[Chess engine#Ratings|here]].
| |
| | |
| ==Theory==
| |
| | |
| === Mathematical details ===
| |
| Performance isn't measured absolutely; it is inferred from wins, losses, and draws against other players. A player's rating depends on the ratings of his or her opponents, and the results scored against them. The difference in rating between two players determines an estimate for the expected score between them. Both the average and the spread of ratings can be arbitrarily chosen. Elo suggested scaling ratings so that a difference of 200 rating points in chess would mean that the stronger player has an ''expected score'' (which basically is an expected average score) of approximately 0.75, and the USCF initially aimed for an average club player to have a rating of 1500.
| |
| | |
| A player's ''expected score'' is his probability of winning plus half his probability of drawing. Thus an expected score of 0.75 could represent a 75% chance of winning, 25% chance of losing, and 0% chance of drawing. On the other extreme it could represent a 50% chance of winning, 0% chance of losing, and 50% chance of drawing. The probability of drawing, as opposed to having a decisive result, is not specified in the Elo system. Instead a draw is considered half a win and half a loss.
| |
| | |
| If Player A has a rating of <math>R_A</math> and Player B a rating of <math>R_B</math>, the exact formula (using the [[logistic curve]]) for the expected score of Player A is
| |
| | |
| :<math>E_A = \frac 1 {1 + 10^{(R_B - R_A)/400}}.</math>
| |
| | |
| Similarly the expected score for Player B is
| |
| | |
| :<math>E_B = \frac 1 {1 + 10^{(R_A - R_B)/400}}.</math>
| |
| | |
| This could also be expressed by
| |
| | |
| :<math>E_A = \frac{Q_A}{Q_A + Q_B}</math>
| |
| | |
| and
| |
| | |
| :<math>E_B = \frac{Q_B}{Q_A + Q_B},</math>
| |
| | |
| where <math>Q_A = 10^{R_A/400}</math> and <math>Q_B = 10^{R_B/400}</math>. Note that in the latter case, the same denominator applies to both expressions. This means that by studying only the numerators, we find out that the expected score for player A is <math>Q_A/Q_B</math> times greater than the expected score for player B. It then follows that for each 400 rating points of advantage over the opponent, the chance of winning is magnified ten times in comparison to the opponent's chance of winning.
| |
| | |
| Also note that <math>E_A + E_B = 1</math>. In practice, since the true strength of each player is unknown, the expected scores are calculated using the player's current ratings.
| |
| | |
| When a player's actual tournament scores exceed his expected scores, the Elo system takes this as evidence that player's rating is too low, and needs to be adjusted upward. Similarly when a player's actual tournament scores fall short of his expected scores, that player's rating is adjusted downward. Elo's original suggestion, which is still widely used, was a simple linear adjustment proportional to the amount by which a player overperformed or underperformed his expected score. The maximum possible adjustment per game, called the K-factor, was set at ''K'' = 16 for masters and ''K'' = 32 for weaker players.
| |
| | |
| Supposing Player A was expected to score <math>E_A</math> points but actually scored <math>S_A</math> points. The formula for updating his rating is
| |
| | |
| :<math>R_A^\prime = R_A + K(S_A - E_A).</math>
| |
| | |
| This update can be performed after each game or each tournament, or after any suitable rating period. An example may help clarify. Suppose Player A has a rating of 1613, and plays in a five-round tournament. He loses to a player rated 1609, draws with a player rated 1477, defeats a player rated 1388, defeats a player rated 1586, and loses to a player rated 1720. His actual score is (0 + 0.5 + 1 + 1 + 0) = 2.5. His expected score, calculated according to the formula above, was (0.506 + 0.686 + 0.785 + 0.539 + 0.351) = 2.867. Therefore his new rating is (1613 + 32×(2.5 − 2.867)) = 1601, assuming that a K-factor of 32 is used.
| |
| | |
| Note that while two wins, two losses, and one draw may seem like a par score, it is worse than expected for Player A because his opponents were lower rated on average. Therefore he is slightly penalized. If he had scored two wins, one loss, and two draws, for a total score of three points, that would have been slightly better than expected, and his new rating would have been (1613 + 32×(3 − 2.867)) = 1617.
| |
| | |
| This updating procedure is at the core of the ratings used by [[FIDE]], [[United States Chess Federation|USCF]], [[Yahoo! Games]], the [[Internet Chess Club]] (ICC) and the [[Free Internet Chess Server]] (FICS). However, each organization has taken a different route to deal with the uncertainty inherent in the ratings, particularly the ratings of newcomers, and to deal with the problem of ratings inflation/deflation. New players are assigned provisional ratings, which are adjusted more drastically than established ratings.
| |
| | |
| The principles used in these rating systems can be used for rating other competitions—for instance, international [[association football|football]] matches.
| |
| | |
| Elo ratings have also been applied to games without the possibility of [[Draw (chess)|draw]]s, and to games in which the result can also have a quantity (small/big margin) in addition to the quality (win/loss). See [[Go ranks and ratings#Elo Ratings as used in Go|go rating with Elo]] for more.
| |
| | |
| {{see also|Hubbert curve}}
| |
| | |
| === Mathematical issues ===
| |
| There are three main mathematical concerns relating to the original work of Elo, namely the correct curve, the correct K-factor, and the provisional period crude calculations.{{cn|date=January 2013}} A 2012 study lead by a team of Swedish psychologists at Lund University has spawned arguments for the inclusion of a fourth concern, albeit theoretical; intangibles.<ref>Lund University; Social Cognition: Theoretical intangibles of the ELO system study, 2012 http://www.psy.lu.se/o.o.i.s/15545</ref>
| |
| | |
| ==== Most accurate distribution model ====
| |
| The first mathematical concern addressed by the USCF was the use of the [[normal distribution]]. They found that this did not accurately represent the actual results achieved, particularly by the lower rated players. Instead they switched to a [[logistic distribution]] model, which the USCF found provided a better fit for the actual results achieved. FIDE still uses the [[normal distribution]] as the basis for rating calculations as suggested by Elo himself.<ref>{{cite web|url=http://www.chessbase.com/newsdetail.asp?newsid=4326 |title=Arpad Elo and the Elo Rating System |publisher=Chessbase article |date=2007-12-16 |accessdate=2012-02-19| author=Daniel Ross}}</ref>
| |
| | |
| ==== Most accurate K-factor ====
| |
| The second major concern is the correct "K-factor" used. The chess statistician [[Jeff Sonas]] believes that the original K=10 value (for players rated above 2400) is inaccurate in Elo's work. If the K-factor coefficient is set too large, there will be too much sensitivity to just a few, recent events, in terms of a large number of points exchanged in each game. Too low a K-value, and the sensitivity will be minimal, and the system will not respond quickly enough to changes in a player's actual level of performance.
| |
| | |
| Elo's original K-factor estimation was made without the benefit of huge databases and statistical evidence. Sonas indicates that a K-factor of 24 (for players rated above 2400) may be more accurate both as a predictive tool of future performance, and also more sensitive to performance.<ref>A key Sonas article is [http://www.chessbase.com/newsdetail.asp?newsid=562 Jeff Sonas: ''The Sonas Rating Formula — Better than Elo?'']</ref>
| |
| | |
| Certain Internet chess sites seem to avoid a three-level K-factor staggering based on rating range. For example the ICC seems to adopt a global K=32 except when playing against provisionally rated players. The USCF (which makes use of a [[logistic distribution]] as opposed to a [[normal distribution]]) has staggered the K-factor according to three main rating ranges of:
| |
| | |
| * Players below 2100: K-factor of 32 used
| |
| * Players between 2100 and 2400: K-factor of 24 used
| |
| * Players above 2400: K-factor of 16 used.
| |
| | |
| FIDE uses the following ranges:<ref name="FideRules">{{cite web|url=http://www.fide.com/fide/handbook?id=124&view=article |title=FIDE Online. FIDE Handbook: Chess Rules |publisher=Fide.com |date=2009-07-01 |accessdate=2012-02-19}}</ref>
| |
| | |
| * K = 30 (was 25), for a player new to the rating list until the completion of events with a total of 30 games.<ref name="Changes to FIDE Rating Regulations">{{cite web|url=http://www.fide.com/component/content/article/1-fide-news/5421-changes-to-rating-regulations.html |title=FIDE Online. Changes to Rating Regulations news release |publisher=Fide.com |date=2011-07-21 |accessdate=2012-02-19}}</ref>
| |
| * K = 15, for players with a rating always under 2400.
| |
| * K = 10, for players with any published rating of at least 2400 and at least 30 games played in previous events. Thereafter it remains permanently at 10.
| |
| | |
| <!-- I find the remainder of this section incomprehensible. Help!! --~~~~ -->
| |
| <!-- I think -- but am not certain -- that it is saying a high K-factor helps high rated players exploit other flaws in the rating system in situations where they can choose their opponents. -->
| |
| The gradation of the K-factor reduces ratings changes at the top end of the rating spectrum, reducing the possibility for rapid ratings inflation or deflation for those with a low K factor. This might in theory apply equally to an online chess site or over-the-board players, since it is more difficult for players to get much higher ratings when their K-factor is reduced. When playing online, it may simply be the selection of highly-rated opponents that enables 2800+ players to further increase their rating,<ref>{{cite web|url=http://www.chessclub.com/help/k-factor |title=ICC Help: k-factor |publisher=Chessclub.com |date=2002-10-18 |accessdate=2012-02-19}}</ref> since a grandmaster on the ICC playing site can play a string of different opponents who are all rated over 3100. In over-the-board events, it would only be in very high level all-play-all events that a player would be able to engage that number of 2700+ opponents, while in a normal open Swiss-paired chess tournament, frequently there would be many opponents rated less than 2500, reducing the ratings gains possible from a single contest.
| |
| | |
| == Practical issues ==
| |
| | |
| === Game activity versus protecting one's rating ===
| |
| In some cases the rating system can discourage game activity for players who wish to protect their rating.<ref>[http://www.chesscafe.com/text/skittles176.pdf A Parent's Guide to Chess] ''Skittles'', Don Heisman, Chesscafe.com, August 4, 2002</ref> In order to discourage players from sitting on a high rating, a recent proposal by British Grandmaster [[John Nunn]] for choosing qualifiers to the chess world championship included an activity bonus, to be combined with the rating.<ref>{{cite web|url=http://www.chessbase.com/newsdetail.asp?newsid=2440 |title=Chess News - The Nunn Plan for the World Chess Championship |publisher=ChessBase.com |date= |accessdate=2012-02-19}}</ref>
| |
| | |
| Beyond the chess world, concerns over players avoiding competitive play to protect their ratings caused [[Wizards of the Coast]] to abandon the Elo system for [[Magic: the Gathering]] tournaments in favour of a system of their own devising called "Planeswalker Points".<ref name=Planeswalkerpointsarticle>{{cite web|url=http://www.wizards.com/Magic/Magazine/Article.aspx?x=mtg/daily/feature/159b|title=Introducing Planeswalker Points|date=September 6, 2011|accessdate=September 9, 2011}}</ref><ref name=planeswalkerpointsarticle2>{{cite web|url=http://www.wizards.com/Magic/Magazine/Article.aspx?x=mtg/daily/twtw/159|title=Getting to the Points|date=September 9, 2011|accessdate=September 9, 2011}}</ref>
| |
| | |
| === Selective pairing ===
| |
| A more subtle issue is related to pairing. When players can choose their own opponents, they can choose opponents with minimal risk of losing, and maximum reward for winning. Such a luxury of being able to hand-pick your opponents is not present in Over-the-Board Elo type calculations, and therefore this may account strongly for the ratings on the ICC using Elo which are well over 2800.
| |
| | |
| Particular examples of 2800+ rated players choosing opponents with minimal risk and maximum possibility of rating gain include: choosing computers that they know they can beat with a certain strategy; choosing opponents that they think are overrated; or avoiding playing strong players who are rated several hundred points below them, but may hold chess titles such as IM or GM. In the category of choosing overrated opponents, new-entrants to the rating system who have played fewer than 50 games are in theory a convenient target as they may be overrated in their provisional rating. The ICC compensates for this issue by assigning a lower K-factor to the established player if they do win against a new rating entrant. The K-factor is actually a function of the number of rated games played by the new entrant.
| |
| | |
| Therefore, Elo ratings online still provide a useful mechanism for providing a rating based on the opponent's rating. Its overall credibility, however, needs to be seen in the context of at least the above two major issues described — engine abuse, and selective pairing of opponents.
| |
| | |
| The ICC has also recently introduced "auto-pairing" ratings which are based on random pairings, but with each win in a row ensuring a statistically much harder opponent who has also won x games in a row. With potentially hundreds of players involved, this creates some of the challenges of a major large Swiss event which is being fiercely contested, with round winners meeting round winners. This approach to pairing certainly maximizes the rating risk of the higher-rated participants, who may face very stiff opposition from players below 3000, for example. This is a separate rating in itself, and is under "1-minute" and "5-minute" rating categories. Maximum ratings achieved over 2500 are exceptionally rare.
| |
| | |
| === Ratings inflation and deflation ===
| |
| An increase or decrease in the average rating over all players in the rating system is often referred to as ''rating inflation'' or ''rating deflation'' respectively. For example, if there is inflation, a modern rating of 2500 means less than a historical rating of 2500, while the reverse is true if there is deflation. Using ratings to compare players between different eras is made more difficult when inflation and deflation is present. (See also [[Comparison of top chess players throughout history]].)
| |
| | |
| It is commonly believed that, at least at the top level, modern ratings are inflated. For instance [[Nigel Short]] said in September 2009, "The recent ChessBase article on rating inflation by Jeff Sonas would suggest that my rating in the late 1980s would be approximately equivalent to 2750 in today's much debauched currency".<ref>[http://www.chessbase.com/newsdetail.asp?newsid=5736 Nigel Short on being number one in Britain again], [[Chessbase]], 4 September 2009</ref> (Short's highest rating in the 1980s was 2665 in July 1988, which was equal third in the world. When he made this comment, 2665 would have ranked him 65th, while 2750 would have ranked him equal 10th. In the September 2012 FIDE rating list, 2665 would have ranked him equal 86th, while 2750 would have ranked him 13th.)
| |
| | |
| It has been suggested that an overall increase in ratings reflects greater skill. The advent of strong chess computers allows a somewhat objective evaluation of the absolute playing skill of past chess masters, based on their recorded games, but this is also a measure of how computerlike the players' moves are, not merely a measure of how strongly they have played.<ref name="inflation">{{cite web|url=http://www.chessbase.com/newsdetail.asp?newsid=5608|title=Rating Inflation - Its Causes and Possible Cures|last=Sonas|first=Jeff|date=July 27, 2009|publisher=[[ChessBase]]|accessdate=2009-08-23}}</ref>
| |
| | |
| The number of people with ratings over 2700 has increased. Around 1979 there was only one active player ([[Anatoly Karpov]]) with a rating this high. In 1992 [[Viswanathan Anand]] was only the 8th player in chess history to reach the 2700 mark at that point of time.<ref name="Viswanathan Anand">{{cite web|url=http://www.chessgames.com/perl/chessplayer?pid=12088 |title=Viswanathan Anand |publisher=Chessgames.com}}</ref> This increased to 15 players by 1994. 33 players had a 2700+ rating in 2009 and 44 as of September 2012. The current benchmark for elite players lies beyond 2800.
| |
| | |
| One possible cause for this inflation was the rating floor, which for a long time was at 2200, and if a player dropped below this they were stricken from the rating list. As a consequence, players at a skill level just below the floor would only be on the rating list if they were overrated, and this would cause them to feed points into the rating pool.<ref name="inflation"/> In July 2000 the average rating of the top 100 was 2644. By July 2012 it had increased to 2703.<ref name="Viswanathan Anand"/>
| |
| | |
| In a pure Elo system, each game ends in an equal transaction of rating points. If the winner gains N rating points, the loser will drop by N rating points. This prevents points from entering or leaving the system when games are played and rated. However, players tend to enter the system as novices with a low rating and retire from the system as experienced players with a high rating. Therefore, in the long run a system with strictly equal transactions tends to result in rating deflation.<ref>{{cite web|url=http://www.sjakk.no/nsf/elosystem_main.html|title=ELO-SYSTEMET|last=Bergersen|first=Per A|publisher=Norwegian Chess Federation|language=Norwegian|accessdate=21 October 2013}}</ref>
| |
| | |
| In 1995, the USCF acknowledged that several young scholastic players were improving faster than what the rating system was able to track. As a result, established players with stable ratings started to lose rating points to the young and underrated players. Several of the older established players were frustrated over what they considered an unfair rating decline, and some even quit chess over it.<ref name="glickman">A conversation with Mark Glickman [http://www.glicko.net/ratings/cl-article.pdf], Published in ''Chess Life'' October 2006 issue</ref>
| |
| | |
| ====Combatting deflation====
| |
| Because of the significant difference in timing of when inflation and deflation occur, and in order to combat deflation, most implementations of Elo ratings have a mechanism for injecting points into the system in order to maintain relative ratings over time. FIDE has two inflationary mechanisms. First, performances below a "ratings floor" are not tracked, so a player with true skill below the floor can only be unrated or overrated, never correctly rated. Second, established and higher-rated players have a lower K-factor. New players have a K=30, which drops to K=15 after 30 played games, and to K=10 when the player reaches 2400.<ref>{{cite web|url=http://www.fide.com/fide/handbook?id=73&view=article|title=FIDE Handbook 8.0. The working of the FIDE Rating System|publisher=[[FIDE]]|accessdate=2009-08-23}}</ref>
| |
| | |
| The current system in the United States includes a bonus point scheme which feeds rating points into the system in order to track improving players, and different K-values for different players.<ref name="glickman"/> Some methods, used in Norway for example, differentiate between juniors and seniors, and use a larger K factor for the young players, even boosting the rating progress by 100% for when they score well above their predicted performance.<ref>[http://www.sjakk.no/nsf/elosystem_index.html ]{{dead link|date=February 2012}}</ref>
| |
| | |
| Rating floors in the USA work by guaranteeing that a player will never drop below a certain limit. This also combats deflation, but the chairman of the USCF Ratings Committee has been critical of this method because it does not feed the extra points to the improving players. A possible motive for these rating floors is to combat sandbagging, i.e. deliberate lowering of ratings to be eligible for lower rating class sections and prizes.<ref name="glickman"/>
| |
| | |
| ==Other chess rating systems==
| |
| {{Main|Chess rating system}}
| |
| * Ingo system, designed by Anton Hoesslinger, used in Germany 1948-1992 {{harvcol|Harkness|1967|pp=205–6}}.
| |
| * [[Harkness rating system#Harkness system|Harkness System]], invented by [[Kenneth Harkness]], who published it in 1956 {{harvcol|Harkness|1967|pp=185–88}}.
| |
| * [[British Chess Federation]] Rating System, published in 1958, now termed the [[ECF grading system]].
| |
| * [[ICCF U.S.A.|Correspondence Chess League of America]] Rating System (now uses Elo).
| |
| * [[Glicko rating system]]
| |
| * [[Chessmetrics]]
| |
| * In November 2005, the [[Xbox Live]] online gaming service proposed the [[TrueSkill]] ranking system that is an extension of the [[Glicko rating system]] to multi-player and multi-team games.
| |
| * [[Elo++]] of Yannis Sismanis performed best of the rating systems evaluated in the Kaggle competition of 2010.<ref>
| |
| Lasek, Jan, Zoltán Szlávik, and Sandjai Bhulai. "[http://www.few.vu.nl/~z.szlavik/papers/IJAPR.pdf The predictive power of ranking systems in association football]" (2009).</ref><ref>Sismanis, Yannis. "[http://blog.kaggle.com/wp-content/uploads/2011/02/kaggle_win.pdf How I won the "Chess Ratings: Elo vs the rest of the world" Competition]" arXiv preprint arXiv:1012.4571 v1 (2010).</ref>
| |
| | |
| == Elo ratings beyond chess ==
| |
| [[College football|American Collegiate Football]] uses the Elo method as a portion of its [[Bowl Championship Series]] rating systems. [[Jeff Sagarin]] of ''[[USA Today]]'' publishes team rankings for most American sports, including Elo system ratings for College Football. The [[NCAA]] uses his Elo ratings as part of a formula to determine the annual participants in the College Football [[NCAA Division I FBS National Football Championship|National Championship Game]].
| |
| | |
| National [[Scrabble]] organizations compute normally-distributed Elo ratings except in the [[United Kingdom]], where a different system is used. The [[North American Scrabble Players Association]] has the largest rated population of active members, numbering about 2,000 as of early 2011. [[Lexulous]] also uses the Elo system.
| |
| | |
| The popular [[First Internet Backgammon Server]] calculates ratings based on a modified Elo system. New players are assigned a rating of 1500, with the best humans and bots rating over 2000. The same formula has been adopted by several other backgammon sites, such as [[Play65]], [[DailyGammon]], [[GoldToken]] and [[VogClub]]. VogClub sets a new player's rating at 1600.
| |
| | |
| The [[European Go Federation]] adopted an Elo based rating system initially pioneered by the Czech Go Federation.
| |
| The [[Groshin's Score]] an Elo based game.
| |
| | |
| In other sports, individuals maintain rankings based on the Elo algorithm. These are usually unofficial, not endorsed by the sport's governing body. The [[World Football Elo Ratings]] rank national teams in men's [[association football|football]]. In 2006, Elo ratings were adapted for [[Major League Baseball]] teams by [[Nate Silver]] of [[Baseball Prospectus]].<ref>[[Nate Silver]], "We Are Elo?" June 28, 2006.[http://baseballprospectus.com/article.php?articleid=5247]</ref> Based on this adaptation, Baseball Prospectus also makes Elo-based [[Monte Carlo method|Monte Carlo]] simulations of the odds of whether teams will make the playoffs.<ref>{{cite web|url=http://www.baseballprospectus.com/statistics/ps_oddselo.php |title=Postseason Odds, ELO version |publisher=Baseballprospectus.com |date= |accessdate=2012-02-19}}</ref> Glenn O'Brien publishes rankings using several rating systems, including variants of the Elo rating system.<ref>[http://www.glennobrien.net/ratings.html Glenn O'Brien's ratings page]</ref> One of the few Elo-based rankings endorsed by a sport's governing body is the [[FIFA Women's World Rankings]], based on a simplified version of the Elo algorithm, which [[FIFA]] uses as its official ranking system for national teams in [[women's association football|women's football]].
| |
| | |
| Sports-reference.com uses the Elo rating system to rate the best professional players in basketball, football, baseball (batters and pitchers rated separately), and hockey (goalies and skaters rated separately).
| |
| | |
| The English [[Korfball]] Association rated teams based on Elo ratings, to determine handicaps for their cup competition for the 2011/12 season.
| |
| | |
| Rugbyleagueratings.com uses the Elo rating system to rank international and club [[rugby league]] teams.
| |
| | |
| Various online games use Elo ratings for player-versus-player rankings. Since 2005, [[Golden Tee Golf|Golden Tee Live]] has rated players based on the Elo system. New players start at 2100, with top players rating over 3000.<ref>{{cite web|url=http://www.goldenteefan.com/statistics/player-rating-handicap/ |title=Golden Tee Fan Player Rating Page |accessdate=2013-12-31}}</ref> In ''[[Guild Wars]]'', Elo ratings are used to record guild rating gained and lost through Guild versus Guild battles, which are two-team fights. The initial K-value was 30, but was changed to 5 in January 2007, then changed to 15 in July 2009.<ref>{{cite web|url=http://wiki.guildwars.com/wiki/Guild_ladder |title=Guild ladder |publisher=Wiki.guildwars.com |date= |accessdate=2012-02-19}}</ref> ''[[World of Warcraft]]'' formerly used the Elo Rating system when teaming up and comparing Arena players, but now uses a system similar to Microsoft's [[TrueSkill]].<ref>{{cite web|url=http://www.wow-europe.com/en/info/basics/arena/index.html |title=World of Warcraft Europe -> The Arena |publisher=Wow-europe.com |date=2011-12-14 |accessdate=2012-02-19}}</ref> The MOBA game [[League of Legends]] used an Elo rating system prior to the second season of competitive play.<ref>{{cite web|url=http://na.leagueoflegends.com/learn/gameplay/matchmaking |title=Matchmaking | LoL - League of Legends |publisher=Na.leagueoflegends.com |date=2010-07-06 |accessdate=2012-02-19}}</ref> The game ''[[Puzzle Pirates]]'' uses the Elo rating system to determine the standings in the various puzzles. [[Roblox]] introduced the Elo rating in 2010.<ref>{{cite web|url=http://blog.roblox.com/?p=2032|title=ROBLOX Contests|date=February 9, 2010|accessdate=9 February 2010}}</ref> [[Counter-Strike: Global Offensive]] also uses the Elo rating.<ref>{{cite web|url=http://kotaku.com/5834542/an-hour-with-counter+strike-go |title=An Hour with Counter-Strike: GO |publisher=Kotaku.com |date=2011-08-25 |accessdate=2012-02-19}}</ref> The browser game [[Quidditch Manager]] uses the Elo rating to measure a team's performance.<ref>{{cite web|url=http://beta.quidditch-manager.com/help.php |title=Quidditch Manager - Help and Rules |publisher=Quidditch-Manager.com |date=2012-08-25 |accessdate=2013-10-20}}</ref> Another recent game to start using the Elo rating system is [[AirMech]], using Elo ratings for 1v1, 2v2, and 3v3 random/team Matchmaking.
| |
| | |
| Despite questions of the appropriateness of using the Elo system to rate games in which luck is a factor, trading-card game manufacturers often use Elo ratings for their organized play efforts. The [[Duelists' Convocation International|DCI]] (formerly Duelists' Convocation International) used Elo ratings for tournaments of ''[[Magic: The Gathering]]'' and other [[Wizards of the Coast]] games. However, the DCI abandoned this system in 2012 in favour of a new cumulative system of "Planeswalker Points", chiefly because of the above-noted concern that Elo encourages highly rated players to avoid playing to "protect their rating".<ref name="Planeswalkerpointsarticle" /><ref name="planeswalkerpointsarticle2" /> [[Pokémon USA]] uses the Elo system to rank its TCG organized play competitors. Prizes for the top players in various regions included holidays and world championships invites until the 2011-2012 season, where awards were based on a system of Championship Points, their rationale being the same as the DCI's for Magic: The Gathering. Similarly, [[Decipher, Inc.]] used the Elo system for its ranked games such as ''[[Star Trek Customizable Card Game]]'' and ''[[Star Wars Customizable Card Game]]''.
| |
| | |
| Moreover, [[online judge]] sites are also using Elo rating system or its derivatives. For example, [[TopCoder]] is using a modified version based on normal distribution,<ref>{{cite web|url=http://apps.topcoder.com/wiki/display/tc/Algorithm+Competition+Rating+System|title=Algorithm Competition Rating System|date=December 23, 2009|accessdate=September 16, 2011}}</ref> while [[Codeforces]] is using another version based on logistic distribution.<ref>{{cite web|url=http://www.codeforces.com/help|title=FAQ: What are the rating and the divisions?|accessdate=September 16, 2011}}</ref><ref>{{cite web|url=http://www.codeforces.com/blog/entry/870|title=Rating Distribution|accessdate=September 16, 2011}}</ref><ref>{{cite web|url=http://www.codeforces.com/blog/entry/893|title=Regarding rating: Part 2|accessdate=September 16, 2011}}</ref>
| |
| | |
| ==Applications apart from gaming==
| |
| The Elo rating system has been used in a crowdsourcing application for determining best quality photographs. A company called Brilliant Arc has created the "Keep/Toss" game. Two photographs are presented and workers in Amazon's [[Amazon Mechanical Turk|Mechanical Turk]] simply select which photograph they believe is the better photograph. Three votes are cast for each photograph and an Elo score is determined for all photographs. The photographs with a score above a certain threshold are then selected for further processing. The results have been well received.{{cn|date=January 2014}}
| |
| | |
| Additionally, the Elo rating system has been used in [[soft biometrics]].<ref>[http://eprints.soton.ac.uk/272922/1/IJCB%20(3).pdf "Using Comparative Human Descriptions for Soft Biometrics"], D.A. Reid and M.S. Nixon, International Joint Conference on Biometrics (IJCB), 2011</ref> Soft biometrics concerns the identification of individuals using human descriptions. Comparative descriptions were utilized alongside the Elo rating system to provide robust and discriminative 'relative measurements', permitting accurate identification.
| |
| | |
| The Elo rating system has also been used in biology for assessing male dominance hierarchies.<ref>{{cite journal|last=Pörschmann|coauthors=et al.|journal=Molecular Ecology|year=2010|volume=19|url=http://onlinelibrary.wiley.com/doi/10.1111/j.1365-294X.2010.04665.x/abstract|title=Male reproductive success and its behavioural correlates in a polygynous mammal, the Galápagos sea lion (Zalophus wollebaeki)|doi=10.1111/j.1365-294X.2010.04665.x}}</ref>
| |
| | |
| ==References in the media==
| |
| The Elo rating system was featured prominently in ''[[The Social Network]]'' during the algorithm scene where [[Mark Zuckerberg]] released [[Facemash]]. In the scene [[Eduardo Saverin]] writes mathematical formulas for the Elo rating system on Zuckerberg's dorm room window. Behind the scenes, the movie claims, the Elo system is employed to rank coeds by their attractiveness. The equations driving the algorithm are shown briefly, written on the window. The correct equations are
| |
| | |
| <math>E_a = \frac{1}{1 + 10^{\tfrac{R_b - R_a}{400}}}</math> and <math>E_b = \frac{1}{1 + 10^{\tfrac{R_a - R_b}{400}}}</math>
| |
| | |
| ''E<sub>x</sub>'' is the expected probability that X will win the match, i.e. ''E<sub>a</sub>'' + ''E<sub>b</sub>'' = 1.
| |
| | |
| ''R<sub>x</sub>'' is the rating of X, which changes after every match, according to the formula (''R<sub>x</sub>'' )<sub>''n''</sub> = (''R<sub>x</sub>'' )<sub>''n''-1</sub> + 32 (''W'' - ''E<sub>x</sub>'' ) where ''W'' = 1 if X wins and ''W'' = 0 if X loses.
| |
| | |
| Everyone starts with an ''R<sub>x</sub>'' = 1400.<ref>[http://flash.sonypictures.com/video/movies/thesocialnetwork/awards/thesocialnetwork_screenplay.pdf Screenplay for "The Social Network," Sony Pictures], p. 16</ref>
| |
| | |
| ==See also==
| |
| * [[FIDE World Rankings]]
| |
| * [[Sports rating system]]
| |
| * [[World Football Elo Ratings]]
| |
| | |
| ==Notes==
| |
| {{reflist|30em}}
| |
| | |
| ==References==
| |
| * {{Cite book
| |
| |last=Elo|first=Arpad|authorlink=Arpad Elo
| |
| |title=The Rating of Chessplayers, Past and Present
| |
| |year=1978
| |
| |publisher=Arco
| |
| |isbn=0-668-04721-6}}
| |
| * {{Cite book
| |
| |last=Harkness|first=Kenneth|authorlink=Kenneth Harkness
| |
| |title=Official Chess Handbook
| |
| |publisher=McKay
| |
| |year=1967}}
| |
| | |
| == External links ==
| |
| * [http://ratings.fide.com/toplist.phtml/ Official FIDE Rating List]
| |
| * [http://chess.eusa.ed.ac.uk/Chess/Trivia/AlltimeList.html All Time Rankings] - lists the top 10 from 1970 to 1997
| |
| * [http://www.2700chess.com/ Live Chess Ratings] - an unofficial [[#Live ratings|live ratings]] list with daily updates
| |
| * [http://gobase.org/rating/elo.html GoBase.org: Introduction to the Elo Rating System]
| |
| * [http://www.natipuj.eu/en The Software Natipuj uses the ELO rating to prediction of football matches]
| |
| * [http://citeseer.nj.nec.com/context/344636/0 Citations for Elo, A.E. 1978. The Ratings of Chess Players: Past and Present.]
| |
| * [http://www.glicko.net/research.html Mark Glickman's research page, with a number of links to technical papers on chess rating systems]
| |
| * [http://www.glicko.net/ratings/rating.system.pdf PDF from Glickman's site explaining USCF ratings system]
| |
| * [http://www.benoni.de/schach/elo/index_e.html Elo query with world ranking list and historical development since 1990]
| |
| * [http://elo.divergentinformatics.com Multiplayer Elo Calculator]
| |
| *[http://www.statsilk.com/maps/chess-fide-country-and-player-rankings-interactive-world-map Interactive world map of FIDE/Elo chess country rankings]
| |
| * [https://github.com/Extrawurst/elo-rating-d ELO Rating implementation in the D programming language]
| |
| | |
| {{DEFAULTSORT:Elo Rating System}}
| |
| [[Category:Chess rating systems]]
| |
| [[Category:Sports records and statistics]]
| |