tnevolin, on 2016-July-16, 09:59, said:
What I meant about coefficients is that there are different ways to get them. Computer and people solve same task of calculating coefficients (feature values). They do it differently. Sometimes they converge on coefficient and people are happy to see two different approaches match in the end. That is all to it. We cannot actually judge the way computer think if coefficients do not converge as good as we would like. Our attempt to "explain" it is just a rationalization of our own model that doesn't actually prove that we are right.
This is true but these two models are not independent. I mean if you analyse a situation and you think you understand it, and then a computer analyses the situation by means of a complex model, you expect it to get a similar result, at least qualitatively, right?
I believe your model is somewhat similar to what climate scientitst do to understand the warming of the earth. Use a formula or a set of formulas and determine a multitude of coefficients from a huge set of data. So if simple physics tells us that the temperature should go up when the CO
2 content increases, and if the model would tell us that the temperature should go down instead, that would be quite spectacular. On the other hand, the influence of water vapor is too complicated to treat it with simple physics because you have opposing effects, so that's where you need the model. But I believe most of the situations in your analysis are more simple than the water vapor example. Although some might not be.
tnevolin, on 2016-July-16, 09:59, said:
Now, if we continue our speculative mind game
, keep in mind that "High card combinations in side suit with 8+ cards on line" and "Value duplication" feature are
corrective ones. You can see "Optional. Count only if known." note for each of them. That means that even if you do not count them due to lack of partner's hand knowledge the result still be correct. These two coefficients allow you to do finer tuning in case you have information to use them. That's why they go to both positive and negative sides.
Yes yes. But still, when one of these coefficients is positive, you expect a higher probability of making more tricks.
tnevolin, on 2016-July-16, 09:59, said:
So K-x = -1 doesn't say anything about king trick taking potential. It says that king trick taking potential and singleton trick taking potential clash and the result of this clash is that the combined trick taking potential of kind and singleton when they are in the same suit is 1 point less than if they were in different suits.
This is certainly a precise formulation but it does not help to understand the problem of these figures. The problem is this:
You have a hand with an average side suit, you add a king in the opposite hand in this suit, and the combined value of the hand rises by 2 points equivalent of the probability of making 2/3 of a trick.
You have a hand with a singleton in a side suit, you add a king in the opposite hand in this suit, and the combined value of the hand rises by 1 point equivalent of the probability of making 1/3 of a trick.
You have a hand with a void in a side suit, you add a king in the opposite hand in this suit, and the combined value of the hand rises by 2 points equivalent of the probability of making 2/3 of a trick.
This is strange. I can't prove it wrong, but I find it strange.
tnevolin, on 2016-July-16, 09:59, said:
You are right that some features occur more often than others. That's why I explicitly excluded unstable coefficient with insufficient statistics. Those included in the document are reliable!
I see, OK.
tnevolin, on 2016-July-16, 09:59, said:
In numbers, I excluded feature those occur less than 100-200 times overall. With 100 results statistical error for corresponding coefficient is about 10%. So if its numeric value is less than 5 then absolute error is less than 0.5 which is OK.
This is true for independent random events of equal weight on a single variable. Is it correct that you kept the other coefficients constant in this part of the test and only determined the value duplication coefficients? In that case I would accept the test. If you did it within a multi-factor analysis, working with several variables within one test, I wonder if you could have additional noise from other sources.
5 points sounds OK but not too much considering that a contract can make an overtrick or be 2 - 3 tricks down depending on play. 0.5 means still 32 % of being wrong. Well okay, that's the borderline figures.
tnevolin, on 2016-July-16, 09:59, said:
This is only for very very rare features. I can tell you void is not rare. 10 cards suit is rare.
I never had a 10-card suit.
Well with ~ 400k boards (boards are independent, observations are not), voids have ~ 4.5 %, voids by the playing party in a side suit have ~ 2 %, that is ~ 8000 observations. Voids opposite some definite combination of honors, you are down to ~ 1000 events. That seems to be on the safe side. But I started to understand why you need 400k boards
.
If your figures are correct, what could that mean? Could it mean that the opponents, when they know declarer is very short in a suit, lead their aces carelessly promoting tricks for declarer?