Jump to content

Rating Players


hotShot

Recommended Posts

I was just ready to give up this whole project, when I came across these forums. Hopefully, this is the correct one in which to post my dilemma (sic). I have not played bridge since '92. Before that I became an LM quite quickly, played in regional and national events. Now, even though I belonged to OK since 94 and BBO for years, never played, just kibitzed on occasion. I am now trying to return and am frustrated. Need to brush up (to put it mildly) on bidding and play; however, the radom players with whom I am paired, usually play only the most basic system and I cannot get a chance to brush up on 2/1 or conventions---I forget alot of things and need practice. I have been playing with GIB 'cause he plays 2/1; but would really like to move along to real play. I am ready to give up the whole thing here on BBO--OK's no better. Perhaps if players were more honest about how they rated themselves, that would be a big step. I have played with "advanced" and "expers" that only played Stayman. I don't think I am too picky--if anyone has a suggestion or critical comment, please let me know. Thanks for listening.
Link to comment
Share on other sites

I was just ready to give up this whole project, when I came across these forums. Hopefully, this is the correct one in which to post my dilemma (sic). I have not played bridge since '92. Before that I became an LM quite quickly, played in regional and national events. Now, even though I belonged to OK since 94 and BBO for years, never played, just kibitzed on occasion. I am now trying to return and am frustrated. Need to brush up (to put it mildly) on bidding and play; however, the radom players with whom I am paired, usually play only the most basic system and I cannot get a chance to brush up on 2/1 or conventions---I forget alot of things and need practice. I have been playing with GIB 'cause he plays 2/1; but would really like to move along to real play. I am ready to give up the whole thing here on BBO--OK's no better. Perhaps if players were more honest about how they rated themselves, that would be a big step. I have played with "advanced" and "expers" that only played Stayman. I don't think I am too picky--if anyone has a suggestion or critical comment, please let me know. Thanks for listening.

Play with people you know.

Link to comment
Share on other sites

If I knew someone, I would.

Sally,

 

Yup, it's a jungle out there. The response from TimG is correct I think. You either have to make a list of 'friends' on BBO or take your chances in the Main Bridge Club.

 

When I drop in the Main Bridge Club, I look at pard's profile first. If no information, or no system profile, or pard is -50 IMPS, I leave.

 

Good luck.

Link to comment
Share on other sites

I have a non-BBO question about rating systems.

 

Suppose that I have access to a large number of results from matchpoint events at a local club. I would like to apply some semblance of a sensible rating system to this data. I've seen a number of rating systems proposed (the Lehman system being prominent, but I'm aware of others).

 

Where can I get software which will, given the event results in some reasonable format, compute the ratings for each player according to some system? Or do I need to write this software from scratch?

Link to comment
Share on other sites

I have a non-BBO question about rating systems.

 

Suppose that I have access to a large number of results from matchpoint events at a local club. I would like to apply some semblance of a sensible rating system to this data. I've seen a number of rating systems proposed (the Lehman system being prominent, but I'm aware of others).

 

Where can I get software which will, given the event results in some reasonable format, compute the ratings for each player according to some system? Or do I need to write this software from scratch?

Hi,

 

not sure, if this answers your question, and is helpful at all, but the scoring

program ruderyv has a rating module for club tournments:

 

http://www.rudersyv.de

 

It seemed to be reasonaable, our club uses the program, but we did not activate

the rating option.

=> Other similar programs may also have such a feature.

 

With kind regards

Marlowe

Link to comment
Share on other sites

If I knew someone, I would.

Hi,

 

one possible option is to join the IAC club tournaments.

http://pigpen.org.uk/IAC/iac.php

 

In general, if you play in tournaments with a fairly well defined player set,

the niveau will be better than the niveau in the MC.

You will also have a better chance to find the same people, and if you start

a tournament you will play with your p the whole tournment (most of the time),

i.e. you will have a chance to get to know each other.

 

With kind regards

Marlowe

Link to comment
Share on other sites

  • 2 weeks later...
I have a non-BBO question about rating systems.

 

Suppose that I have access to a large number of results from matchpoint events at a local club. I would like to apply some semblance of a sensible rating system to this data. I've seen a number of rating systems proposed (the Lehman system being prominent, but I'm aware of others).

 

Where can I get software which will, given the event results in some reasonable format, compute the ratings for each player according to some system? Or do I need to write this software from scratch?

Doesn't answer your question directly but you may want to contact Chris Champion -

http://www.coloradospringsbridge.com/pr.htm

Link to comment
Share on other sites

  • 2 months later...

I don't think a rating system that is accurate and fair is possible given the number of variables in the game of bridge.

 

I do think, however, that bridge knowledge is at least an indicator of performance at the table.

 

What if a standardized test(s) were developed that could be taken online and scored automatically? Such a test could cover the basic bidding of various common systems ..... in my opinion to be rated an expert, a player should have at least a rudimentary knowledge of Standard American, SAYC, Acol, 2/1, and Precision.

 

Such a test should also include a section on play of the hand including basic leads, defensive signals and carding, how to play basic card combinations, etc.

 

Also the test should cover the more common conventions.

 

Rating would then be based on overall score acheived on the test.

 

The test should be available to be taken online and players should be allowed to retake it as often as they wish.

Link to comment
Share on other sites

  • 4 months later...

Competitive Chess has had an effective and efficient rating system for decades. It's worth a look for anyone who seriously wants to generate a rating system for Bridge.

 

If my memory is working, this is how it worked 25 years ago:

 

1) Ratings run from a low of about 1200 (beginning players) to roughly 2800 (National Masters). Above that, a player's FIDE (International) rating is more important, although it is calculated in a similar fashion.

 

2) Each recorded game contributes to a player's rating, except as noted below. Typically, a tournament player will play 3 or 4 games per day in a local tournament.

 

3) A player's rating increases by 16 points for each win, and decreases by 16 points for each loss, plus or minus 4% of the difference in the players ratings, up to a difference of 400 points. So, if you beat a player who is 400 points lower in the rankings than you are, the effect on your rating is (16 - (.04 * 400)) = 0. Similarly, losing to a player who is ranked 400 points higher than you, you don't lose any rating points. If you beat a player who is ranked 200 points higher, the effect is (16 + (.04 * 200)) = 24;

 

4) As a corollary to (3), players never lose rating points as a result of winning a game, and never gain points by losing a game. A player can never lose more than 32 points or gain more than 32 points as a result of a single game.

 

5) Ratings are provisional until a player has some number of rated games - 24 I think.

 

There are some grumblings about the chess rating system (of course, since there are grumblings about everything), but it has worked well. Bridge players who want a rating system should take a look at it.

Link to comment
Share on other sites

 

Excluding pairs that play and practice together (THE VAST VAST VAST VAST VAST VAST VAST VAST (that's a lot of VASTs) majority of who DO NOT CHEAT) is ridiculous. it randomizes the field, causes bad and high variance bridge and rewards good luck rather than good actions.

 

 

How do you know that the (VAST *8) majority of established partnerships do not cheat?

 

I wouldn't say that they DO cheat, because I have no evidence. Apparently you have some evidence that they don't cheat. What is it?

Link to comment
Share on other sites

Bridge players who want a rating system should take a look at it.

 

I don't care much for the chess rating system. Last time i played, my p led the pawn of black, later crossed the rook with the king and sacrificed the queen to the opponents' knight. Before I could do anything my p already blew the board and my rating went down... and the other pair were NOOOOBS.

  • Upvote 1
Link to comment
Share on other sites

How do you know that the (VAST *8) majority of established partnerships do not cheat?

 

I wouldn't say that they DO cheat, because I have no evidence. Apparently you have some evidence that they don't cheat. What is it?

 

 

I am not sure how you get through life thinking that unless you have some evidence to the contrary people are doing you wrong?

 

Or perhaps all the partnerships you have been involved in have cheated?

Link to comment
Share on other sites

The beautiful chess rating, go rating, golf handicap...all work really well for individual events.

 

How good is LeBron? Really, really good, we all know that. But how good is LeBron when paired with a high school centre? Or with Random Ball Hog (I'm sure we all can think of at least one example). How good is LeBron, against competition at his level, playing pickup street against a professional team allowed to practise together? And how does that affect his rankings?

 

I'm a decent flight A player in the fields I play in. Not great, decent. With a team of people at my level, we were one exhausted cardplay mistake away from qualifying for day 2 of the Sat-Sun Swiss in Reno last year. But I was playing with my regular partner, my teammates played together frequently; if I had to play pickup - even if my partner was measurably "better" than the one I had - we wouldn't have done as well, pretty much guaranteed. So, what does that do to rating?

 

The other thing that ratings-that-can-drop do is to make people not *want* to play with new people, or lesser players, because it might affect their rating in ways they can't help. So, welcome to even *more* cliques and stratifying of the game. I haven't *proven*, but it's pretty obvious, that there is no way to set a single-person rating that can't be obviously gamed (either way, really - if I want a *bad* rating, so that I can get into an event where I can clean up, I can arrange that, without "dumping").

 

Really, the only sane thing one can measure is *partnership* ratings; and that's going to leave many players, who don't just don't play "regular" partnerships, out in the cold. It certainly won't help the reason "everybody" wants a rating system - "how good is he? Am I willing to play with him?" And even those would probably have to have separate "MPs/IMPs/BAM" ratings, the same way FIDE has separate regular/speed ratings - it's a different game.

 

Attendance points suck for rating. But really, very few things are *unambiguously* better - just less worse.

  • Upvote 1
Link to comment
Share on other sites

  • 1 month later...

Competitive Chess has had an effective and efficient rating system for decades. It's worth a look for anyone who seriously wants to generate a rating system for Bridge.

 

 

I think that developing a rating system for bridge requires considerable research. Apologies for the technical language that follows, but you cannot really discuss the technical merits of a rating system without some mathematical language. I don't have anything to say about social and business questions such as whether BBO should have a rating system, whether some games should be rated and others unrated and so on.

 

The post above game the numerical calculations involved in running a rating system, but omitted the heart of the system -- that mathematical model that leads to these calculation. The basic model for chess, proposed by Arpad Elo, is the following:

 

1. We assume that every player has a current "average ability level" (this is the number that the rating system will be trying to measure). In any individual game, the player will play at an ability level which is randomly distributed around his "average" level, and the winner of the game is the player who happened to display a higher level during that game.

2. The key assumption (to be discussed below): the distribution of abilities around the average is normal (Gaussian), with a standard deviation of 200 rating points.

3. Say we have two players with true abilities r_1, r_2. We can calculate the probability that player 1 will beat player 2 (that's the probability that a random sample from N(r_1,200) is larger than a random sample from N(r_2,200)). This is the "expected result of the game". Now assume that r_1 and r_2 were only the ratings associated to the players instead of their true abilities. We can still compare the calculated probability to the actual result of the game (player 1 can score 0,0.5, or 1 point). If player 1 did better than expected, we infer that we underestimated r_1 and overestimated r_2, and make a small adjustment to the ratings accordingly (dlbalt's post above explains how this is done). Conversely if player 1 did worse than expected.

4. Mathematical fact: If assumption 2 is valid, then then o a long series of games the ratings of all players will converge to their true ability levels.

 

The main weakness of this model is the fixed standard deviation of 200 points. If all players of a given level had the same standard deviation, this would simply serve to fix the value of "1 rating point". In practice they don't quite. Another weakness is that an input to the rating update process is a measure of the "rating uncertainty" (if we are very confident that r_1 is close to player 1's ability, we want to only make a small change after the game; if we are unsure we want to make a large change), which affects the speed of convergence of the ratings to the true abilities. The original Elo system had a fixed value (later versions give players a higher uncertainty during their first games). Glicko's system improves on this in two ways. First, he modifies assumption 2 by assigning a different variance to each player (which the system will also try to measure). Second, we improve step 3 by assigning to every player a "ratings deviation" which more directly measures our uncertainty about his rating. The first change is fundamental (makes the model closer to reality) while the second is designed to improve the convergence properties).

 

How does one test the system? Take a set of players and a large collection of games. Use the first part of the sample to calculate ratings as described above, using enough games for the ratings to converge. Then examine the remaining games to see whether the statistical model really predicts the results (i.e. whether the probability of winning is really given by the larger-of-two-Gaussian-samples model, and whether our model for the variance [either 200 points or per-player] agrees with reality). In chess this works to some extent -- see for example the papers by Mark Glickman.

 

What about adapting this to bridge? Here are some thoughts:

 

1. Formally, one could simply look at each hand played separately (say looking at the points margin). Given enough time (hands), the ratings may converge. I think the general perception is that convergence here would be too slow, and that it's better to compare what player did with the same cards rather than compare NS to EW directly. Also probably different models are needed for team games (only one other table) and duplicate games (many other tables).

 

2. On the other hand, bridge has more detailed scoring information than just "win/draw/lose", which should help. It is not a-priori clear which component of the score best predicts future success.

 

This much is for rating pairs. But what we really want is rating individuals, and this raises the last question:

 

3. Can players be assigned individual ratings which effectively predict the performance of pairs? The naive answer is that established partnerships perform much better than pick-up partnerships with the same ability, but I don't know of research into how large this effect actually is and (more importantly) to what extent it limits the accuracy of individual ratings in predicting results. To get a genuine answer we have to create a statistical rating model and then test it against actual data.

 

Side thought: in trying to extract individual ability from pair data, it would be helpful to incorporate the identity of the declarer into the model (that is, probably the ability of declarer affects the result of the hand more than the ability of dummy), but since the identity of declarer is not simply a function of the cards I don't see an obvious way to do this.

 

In summary: before we discuss whether BBO should have a rating system, it may be worth doing some statistical research and actually <i>validate</i> a rating system for bridge.

Link to comment
Share on other sites

In summary: before we discuss whether BBO should have a rating system, it may be worth doing some statistical research and actually <i>validate</i> a rating system for bridge.

 

Developing an accurate rating system for a given population of players isn't particularly hard.

 

Convincing half the players that they are below average is much more daunting, as is the mind numbing tedium trying to convince numerically illiterate yahoos why this algorithm says that you suck...

 

Figure out how to deal with this and I'll invest the time/effort to develop an accurate rating system

  • Upvote 1
Link to comment
Share on other sites

So.... convincing.

 

For what its worth, I had the chance to discuss this topic (bridge ratings) with Glickman after a talk he gave a few monthes back.

 

I posited (and Glickman concurred) that the best way to approach the problem was to focus on developing an accurate rating system for pairs.

Once you have an accurate system for rating pairs, you can then try to decompose accurate ratings for individuals out of a sets of pair ratings.

 

It's entirely possible that Glickman doesn't believe any such thing and thought that agreeing with me was the best way to get me to go away.

Link to comment
Share on other sites

For what its worth, I had the chance to discuss this topic (bridge ratings) with Glickman after a talk he gave a few monthes back.

 

I posited (and Glickman concurred) that the best way to approach the problem was to focus on developing an accurate rating system for pairs.

Once you have an accurate system for rating pairs, you can then try to decompose accurate ratings for individuals out of a sets of pair ratings.

 

It's entirely possible that Glickman doesn't believe any such thing and thought that agreeing with me was the best way to get me to go away.

 

one of my partner always says this about rating pairs

 

but says also that no bridge organization will consider such a thing for pecuniary reasons

Link to comment
Share on other sites

  • 5 months later...

The ELO system is highly regarded in International Football:

 

http://www.eloratings.net/system.html

 

and "about ratings" gives details. The important thing is to have a high weight constant in the early stages reducing as the players get more experience. A percentage in any event can be converted into a rating difference, and compared with the player's expected result.

Link to comment
Share on other sites

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.

Guest
Reply to this topic...

×   Pasted as rich text.   Paste as plain text instead

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.

Loading...
×
×
  • Create New...