|
< Earlier Kibitzing · PAGE 19 OF 57 ·
Later Kibitzing> |
Sep-10-09
 | | alexmagnus: October 2008:
1 . Topalov, Veselin BUL : 2658
2 . Morozevich, Alexander RUS : 2654
3 . Carlsen, Magnus NOR : 2653
4 . Ivanchuk, Vassily UKR : 2653
5 . Anand, Viswanathan IND : 2650
6 . Kramnik, Vladimir RUS : 2639
7 . Aronian, Levon ARM : 2624
8 . Radjabov, Teimour AZE : 2619
9 . Leko, Peter HUN : 2614
10 . Jakovenko, Dmitry RUS : 2604
11 . Wang, Yue CHN : 2603
12 . Adams, Michael ENG : 2601 |
|
Sep-10-09
 | | alexmagnus: January 2009:
1 . Topalov, Veselin BUL : 2662
2 . Anand, Viswanathan IND : 2657
3 . Ivanchuk, Vassily UKR : 2645
4 . Carlsen, Magnus NOR : 2642
5 . Morozevich, Alexander RUS : 2637
6 . Radjabov, Teimour AZE : 2627
7 . Jakovenko, Dmitry RUS : 2626
8 . Kramnik, Vladimir RUS : 2625
9 . Leko, Peter HUN : 2617
10 . Movsesian, Sergei SVK : 2617
11 . Aronian, Levon ARM : 2616
12 . Shirov, Alexei ESP : 2611
13 . Wang, Yue CHN : 2605 |
|
Sep-10-09
 | | alexmagnus: April 2009:
1 Topalov, Veselin g BUL 2667
2 Anand, Viswanathan g IND 2648
3 Carlsen, Magnus g NOR 2635
4 Kramnik, Vladimir g RUS 2624
5 Radjabov, Teimour g AZE 2621
6 Aronian, Levon g ARM 2619
7 Jakovenko, Dmitry g RUS 2618
8 Morozevich, Alexander g RUS 2616
9 Leko, Peter g HUN 2616
10 Grischuk, Alexander g RUS 2613
11 Movsesian, Sergei g SVK 2612
12 Ivanchuk, Vassily g UKR 2611
13 Shirov, Alexei g ESP 2610
14 Wang, Yue g CHN 2603 |
|
Sep-10-09
 | | alexmagnus: July 2009:
1 Topalov, Veselin g BUL 2672
2 Anand, Viswanathan g IND 2647
3 Carlsen, Magnus g NOR 2631
4 Aronian, Levon g ARM 2627
5 Jakovenko, Dmitry g RUS 2619
6 Kramnik, Vladimir g RUS 2618
7 Leko, Peter g HUN 2615
8 Radjabov, Teimour g AZE 2615
9 Gelfand, Boris g ISR 2614
10 Morozevich, Alexander g RUS 2610 |
|
Sep-10-09
 | | alexmagnus: September 2009:
1 . Topalov, Veselin BUL : 2670
2 . Anand, Viswanathan IND : 2645
3 . Aronian, Levon ARM : 2630
4 . Carlsen, Magnus NOR : 2629
5 . Kramnik, Vladimir RUS : 2629
6 . Leko, Peter HUN : 2619
7 . Radjabov, Teimour AZE : 2614
8 . Ivanchuk, Vassily UKR : 2613
9 . Gelfand, Boris ISR : 2613
10 . Morozevich, Alexander RUS : 2607 |
|
Sep-10-09
 | | alexmagnus: This gets this top-20 by dominance:
1.Fischer 2810 Jan 1972
2.Kasparov 2764 Jan 2000
3.Karpov 2725 Jan 1978, Jan 1980, Jan 1982
4.Kramnik 2714 Oct 2001
5.Spassky 2710 Jan 1971
6.Tal 2705 Jan 1980
6.Anand 2705 Jul 1998
8.Korchnoi 2695 Jan 1979, Jan 1980
9.Topalov 2694 Jul 2006
10.Ivanchuk 2685 Jul 1991
11.Larsen 2680 Jan 1971
12.Petrosian 2670 Jan 1972
12.Polugayevsky 2670 Jan 1972
12.Portisch 2670 Jan 1972
15.Morozevich 2667 Jul 1999
16.Shirov 2665 Jul 1994
16.Kamsky 2665 Jul 1996
18.Adams 2661 Jul 2000, Oct 2000
19.Timman 2660 Jan 1982
20.Botvinnik 2655 Jan 1972
20.Ljubojevic 2655 Jan 1983
20.Gelfand 2655 Jan 1991
20.Leko 2655 Oct 2000 |
|
Sep-11-09
 | | alexmagnus: That makes 42 points the greatest distance between Delo at the point and peak Delo at the moment of peak Elo (Gelfand, who on Sept 2009 list had Elo 2756 (peak), Delo 2613 and peak Delo 2655 (on Jan 1991 list). Among the 20 top-Delo-players ever, that is. |
|
| Sep-18-09 | | whatthefat: <alexmagnus: OK I admit: this kind of adjustment is good for measuring dominance. And if we fix #100 at 2500, we actually get a good definition for "elite player" a player over 2600 on this "fixed" system (let's call it Delo, from dominance+elo).> Interesting work.
This method makes implicit assumptions about the relative strengths of #100 ranked players across time, that will artificially inflate the ratings of players from times when the playing pool was smaller, e.g., Fischer. Today, the Earth's population is almost double what it was in 1970, and rapidly growing regions such as India and China have started to partake in elite chess only relatively recently. Just look at how the number of Soviet or Russian/Former-Soviet players has reduced from 15/19 on your 1971 list, to 7/14 in 2009. Sonas' chessmetrics system makes a similar (bad) assumption, in fixing the average rating of the players ranked 3-20 (although that doesn't seem to check out if you crunch the chessmetrics numbers). I can sympathize with the problem - there's no really good way to normalize the ratings without having some measure of just how many players were active at any one time. Presumably the relative strengths of chess players at any given point in time will lie on a similar statistical distribution, and the top 100 will always be at the upper end of this distribution. One might therefore be able to make inferences based on how the top 100 ratings are distributed. |
|
Sep-18-09
 | | alexmagnus: <whatthefat> The population grew, but the size of the perceived elite never changes. You know, like in tennis: even if population gets thousandfold bigger, only those coming to Grand Slams will be more or less known - and that number is fixed... That's what the list is based on. The population doubled, but nobody perceives #20 of today as "elitaire" as #10 of 1970. Also note, the list isn't intended to measure <strength>, it is intended to measure <domination> on the elite level. On the elite level and only there. |
|
| Sep-18-09 | | whatthefat: <alexmagnus: <whatthefat> The population grew, but the size of the perceived elite never changes. You know, like in tennis: even if population gets thousandfold bigger, only those coming to Grand Slams will be more or less known - and that number is fixed...> Well, I think this is debatable. The number of top-level tournaments is also greater today. Show me a top 100 list today and I will have heard of nearly all the players. I doubt I could do the same in 1970, and I certainly couldn't if we went even further back. <That's what the list is based on. The population doubled, but nobody perceives #20 of today as "elitaire" as #10 of 1970. Also note, the list isn't intended to measure <strength>, it is intended to measure <domination> on the elite level. On the elite level and only there.> Okay, I see what you're trying to do, and I agree that this is not a measure of objective strength. After all, a rating (be it Elo, chessmetrics or whatever) doesn't compare strength between eras, it only compares relative playing strengths in the same era. But I still have a problem with the argument - let me show you why: Let us suppose for the sake of argument, that the strengths of chess players worldwide follow a bell curve with Elo rating. Now the point is, that if the playing pool becomes larger, the top 100 becomes a smaller slice of the tail of the distribution. For example, suppose the total number of players is n = 10^5 (in reality there are many more than this), the mean rating is 2000, and the standard deviation is 150. Running a simulation of this, I find: #1 - 2691
#2 - 2619
#3 - 2616
...
#100 - 2468
So the Delo rating of #1 is 2823. Running this simulation 10 times, I found an average Delo rating for the #1 player of <2817>. Now, let's increase the total number of players to n = 10^6. Running a simulation of this, I find: #1 - 2733
#2 - 2703
#3 - 2691
...
#100 - 2553
So now the Delo rating of #1 is 2781. Running this simulation 10 times, I found an average Delo rating for the #1 player of <2768>. Now suppose I increase n to 10^7. Now the average Delo rating for the #1 player is <2754>. You see what's happening - the larger the playing pool, the smaller the #1 Delo rating (all else being equal). So in practice, Fischer is being given a significant advantage by this approach. Note also that the mean of the distribution could be moving over time (inflation), and this would not matter, since we care only about rating differences. |
|
Sep-18-09
 | | alexmagnus: whatthefat: I understand it, but what made me use this approach is Sonas' observation that the distance between f.x. #100 and #1000 almost doesn't change... |
|
| Sep-18-09 | | whatthefat: <alexmagnus>
Do you know where he said that? I find it difficult to believe that that could be true. |
|
Sep-18-09
 | | alexmagnus: <whatthefat> He presented some graphs in his article on rating inflation on Chessbase (somewhere at the end on July). |
|
Sep-18-09
 | | alexmagnus: http://www.chessbase.com/newsdetail..., you'll find the mentioned graph there. |
|
| Sep-18-09 | | whatthefat: Okay thanks, I'll check it out. |
|
| Sep-18-09 | | metatron2: <alexmagnus: but what made me use this approach is Sonas' observation that the distance between f.x. #100 and #1000 almost doesn't change...> Alex, Sonas observation is just a specific case of rating distance being kept over time between two absolute ranking positions. This doesn't hold for the vast majority of the ranking positions. I wrote about this here: zarg chessforum (empty ignore list needed, and it starts from the " Jeff showed a diagram where.." paragraph). The whole idea of trying to conclude something from rating change over time of absolute ranks, when the pool size seriously increases, is completely wrong. Did you ever see anyone referring to IQ change over time of person ranked #X on the list? Absolute rank can only be relevant when the pool size doesn't change over time. <whatthefat: After all, a rating (be it Elo, chessmetrics or whatever) doesn't compare strength between eras, it only compares relative playing strengths in the same era> That's not very accurate, Fide desires that rating values will be kept over time, and they are currently studying the issue of rating inflation in order to try making this happen. Rating inflation would have been a none-issue for fide if rating was just supposed to measure temporal relative strength of each specific time. It is true that this issue is complex, and the longer the periods between two "eras", the bigger the problem, especially if there were changes in the rating system during that time, such as changes in rating floors. |
|
Sep-19-09
 | | alexmagnus: <<Then we have the gap between #100 and the closer #500-#1000 ranks. Those ranks are in the 24XX-25XX rating range. I suspect that the top 100 are "escaping" from this range so fast because the top 100 are full time and the most talented players in the world, while the other range is mostly made of semi/retired/none-professionals players, hence the top 100 are simply improving much-much faster then them, compensating for the natural tendency of the gap to decrease. (This assumption is of course in line with my assumption that rating changes do reflect <true change> in chess strength).
Seems like those 3 different regions indicate that the distribution function is not concave around their boundaries. Hence It might be more convenient to separate the rating pool distribution into 3 different distributions: "Professionals" (rating > 2600), "Semi professionals" (2400-2600), and "Amateurs" (rating < 2400).>> <metatron> I myself, as you know, support the idea of changes reflecting changes in chess strength, but do you really think that the number of professionals grew like 15-fold in the last 40 years? Isn't it more so that in the modern world one can come further as an amateur? |
|
Sep-19-09
 | | alexmagnus: Btw, metatron, there is a little flaw in your logic... I mean regarding the gaps. You talk about ranks being close to 2400 and say about possible influence of the change of K-factors at this level. But note: the gap to #100 remains constant, while #100 itself is moving, i.e. the rating of those ranks has moved too and is no longer around 2400... |
|
| Sep-19-09 | | metatron2: <I myself, as you know, support the idea of changes reflecting changes in chess strength> I know that alex, that's why I don't understand why all of the sudden u decided that rectifying rating lists to #100 = 2500-rating was a good idea. By doing that you are actually saying the opposite, namely that the
#100 ranked player's rating has been unjustifiably increasing since 1985, and hence you need to "re-adjust" it to 2500 in order to compare dominance level from different periods. You posted lists showing how wrong such an idea would be, and then for reason I didn't understand, you decided it was a good idea after all. <but do you really think that the number of professionals grew like 15-fold in the last 40 years? Isn't it more so that in the modern world one can come further as an amateur?> I don't think one comes on the expense of the other: That we have much more amateurs today thanks to a much handier accessibility and a lower rating floor, doesn't mean that we don't have more professionals as well. Having more amateurs increases the chance more of them can advance further (as you said), but that also mean that more of them would become professionals, or at least professionals to some period of time. As I see it, the main reason the rating of #100 position has been constantly increasing, is because the number of players in the pool is constantly increasing, and the average chess level is increasing as well. Hence I see no reason to rectify the top 100 ratings like that in order to make such comparison, and surely not in such brutal way, reducing the modern players rating by 130 points.. <You talk about ranks being close to 2400 and say about possible influence of the change of K-factors at this level. But note: the gap to #100 remains constant, while #100 itself is moving, i.e. the rating of those ranks has moved too and is no longer around 2400...> I talked about the entire red group in Sonas' chart, and as I recall it, that group started with players ranked higher then #10000, and so this rank did not actually reach the 2400 boundary. But this was a general assumption anyway, got to go now, I might try thinking about it in more detail later. |
|
Sep-19-09
 | | alexmagnus: <metatron> As I said, the lists are intended to compare dominance, not chess strength. The latter is IMO quite well reflected by the absolute rating (here we both agree, and both have numerous opponents:)). Still, sometimes it's interesting to compare who dominated more. Since I talk about dominance on the elite level only, it makes sense to chose a constant number of players (100) for the definition of "elite" (see my comparison to Grand Slam in tennis). I.e. I compare a totally different concept of "greatness" than chess strength, namely dominance among the elite players. Of course this comparison is unfair (the better general the level of played chess on the elite level becomes, the harder it is to dominate it), but it makes sense. To see the <greatness> of a player, one has to explore <both> factors - chess strength <and> dominance (though I don't think a simple sum Elo+Delo does it, one needs also some weightings for the time period in question). |
|
Sep-19-09
 | | alexmagnus: The average of peak Elo and peak Delo actually is interesting in the sense that it would almost equalize Fischer and Kasparov (Fischer 2797.5, Kasparov 2807.5). |
|
| Sep-19-09 | | whatthefat: <metatron2: <whatthefat: After all, a rating (be it Elo, chessmetrics or whatever) doesn't compare strength between eras, it only compares relative playing strengths in the same era> That's not very accurate, Fide desires that rating values will be kept over time, and they are currently studying the issue of rating inflation in order to try making this happen. Rating inflation would have been a none-issue for fide if rating was just supposed to measure temporal relative strength of each specific time.> Consider this thought experiment:
Tomorrow, all chess players become objectively 400 Elo points stronger. So somebody who is rated 2100 today, will tomorrow play at the level that a 2500 player today plays at. Will anybody's ratings change as a result? No, of course not. All it means to be 400 rating points better than someone else is that on average you score 10/11 against them. If everyone's ratings shift simultaneously then there is no difference, because the Elo formula is designed to measure relative performance and nothing more. Something similar has happened in the world of chess, only over a much longer timescale, and ratings formulae are in no way designed to measure that. There is no way of predicting how a player rated x will perform against a player rated y if they played in totally different eras. Only analysis of the moves themselves can do that (along the lines of what <nimh> is doing - see his profile if you haven't already checked it out). |
|
Sep-19-09
 | | alexmagnus: <whatthefat> Unlike your thought experiment, in reality chess progress is not simultaneous. Improvement in chess is a chain reaction, a continuous process. |
|
Sep-19-09
 | | alexmagnus: What nimh is doing is not good at all. The error percentage depends not only on own playing level, but also on opposition. The worse your opponent plays, the easier it is to find perfect moves. |
|
Sep-19-09
 | | alexmagnus: Also, what the term "error percentage" misses: the level of a chess move depends not only on its "errorness" but also on the amount of pressure it puts. As I already once said, one could easily find error-free games between 2100s. It just takes a stronger opponent who is able to put enough pressure to make those 2100s err. |
|
 |
 |
|
< Earlier Kibitzing · PAGE 19 OF 57 ·
Later Kibitzing> |
|
|
|