Because none of the claims I've made are obviously factually wrong, while Your Guy has committed significant and obvious blunders, and made factually incorrect claims, as detailed previously.AndrewGrant wrote: ↑Sun Sep 11, 2022 10:30 amMy point is to say, why should I take your analysis as true and correct, and not this guy's analysis as true and correct?
Usually, when one guy says multiple blatantly and factually wrong things, and another guy has not yet been shown to say any wrong things, you should place more credence in the claims and conclusions of the person who has not yet been shown to say any wrong things. (Or if you're not willing to do so, you should verify the information yourself.)
They take place at a significantly shorter time control, so I'd imagine most people consider it correct to exclude them from an analysis of performance at classical time controls, as I do. They also aren't FIDE rated, and rarely have meaningful prizes, so I'd imagine most people consider there to be less incentive to cheat in quick events, as I do.Is removing quick events proper?
For example, suppose that the typical method of cheating requires the participation of a confidant, which imposes a nonzero upfront cost, not to mention further exposes yourself to potentially getting caught. Probably most cheaters are less likely to spend resources (or take the chance of getting caught) in order to cheat at such events.
As I said in my original post, I was unable to locate any information contradicting the claimed broadcast status of any of the events in the original tweet. So if by "the few events pointed out marked correctly or incorrectly" you mean the ones that Your Guy claimed were wrong, then all I can say is that I have seen no evidence that the original tweet was wrong, and I verified what I could by going to each tournament's official website.Does that fact that you could not verify a few of the events have a significant impact? Were the few events pointed out marked correctly or incorrectly, can you answer that?
Your Guy gives absolutely no evidence for his corrections. His only links are to the USCF pages for the tournaments, and he provides no links to any pages that mention a live broadcast for the contested tournaments.
I can certainly find a record of the games of most of the tournaments online, for example, FollowChess has the 2019 World Open games (as noted by Your Guy), but I can find neither mention of nor links to a live broadcast of the 2019 World Open games on its official website, nor in any contemporaneous mentions of the 2019 World Open. That doesn't mean the 2019 World Open wasn't broadcast online, and that doesn't mean I didn't miss something, but I can only work with the information that I have available to me. As Your Guy actually presented no new information, and committed obvious blunders that call into question the thoroughness of his process, I see no reason to amend my analysis at this time.
I also consider it pretty unlikely that the original tweet systematically erred/lied about precisely those tournaments whose broadcast status I wasn't able to definitively confirm. While possible, it presupposes the original tweeter would somehow know exactly which information I would miss in my checks, or that he used the exact same information gathering process that I did, or that he otherwise committed the exact same errors that I did. But let's go wild and do some robustness checks, anyway.
If you flip the broadcast status of the Niemann's most significant tournament in the dataset, the 2019 World Open, broadcast status still explains 34% of the variation in his performance, and its regression coefficient is large and positive, with a p-value of 0.064:

True, that's just outside the traditional significance threshold of 0.05, but that's arbitrary to begin with. The 95% confidence interval is [-1, 29], and it's overwhelmingly likely that broadcast status is the single most important factor in determining Niemann's performance across the dataset.
If you additionally flip the broadcast status of Niemann's second most significant tournament in the dataset, the 2019 Marshall Chess Club Championship, broadcast status still explains 16% of the variation in his performance, and its regression coefficient is large and positive, with a p-value of 0.267:

Obviously, we're now starting to veer well into statistical insignificance, but the 95% confidence interval is [-7, 24], and it's still significantly more likely than not that the single most important factor in determining Niemann's performance remains broadcast status.
If you additionally flip the broadcast status of Niemann's third most significant tournament in the dataset, the 2020 Charlotte Fall GM, broadcast status still explains 6% of the variation in his performances, and its regression coefficient is large and positive, with a p-value of 0.5, and a 95% confidence interval of [-10, 20]. In this scenario, broadcast status is finally unlikely to be the single most important factor in determining Niemann's performance, at the last overtaken by the average rating of his opponents.
If you mean that you believe the difference between those types of events implies a difference in preparation, which you (without evidence) believe is very strongly and positively correlated with broadcast status, you know what to do. Run the numbers and let me know what you find.Should your analysis take into account what I mentioned before about seeded events and Swiss events?
Niemann wasn't that young, and actually had very little ratings volatility (or growth) over the period. His USCF classical rating went from 2541 in December 2018 to 2569 at the end of the dataset in November 2020. His rating never dropped below 2496 or went above 2576 in that time. That's remarkably stable given USCF's K-factor of 15. (There's conflicting data online, but I believe that's the current schedule used for non-provisional players.)Should your analysis take into account the increased volatility of ratings of young players?
I easily answered your questions.No easy answer to those questions, and so I am not inclined to take anyone's spreadsheets as proof of anything more than one's ability to find data that supports their view.
I also didn't "find data that supports [my] view", as I more or less didn't have a view, though as I previously mentioned, I was somewhat (but not strongly) disinclined to believe Niemann after his Sinquefield interviews. I was sent the original tweet, and I thought it was interesting. I also wondered if the claims were true, and how statistically significant the observed effect was. I verified as much of the information as I reasonably could, found no verifiably contradictory information, gathered a bit more detailed data, and performed a simple and standard statistical analysis on it.
Notably, the only counterargument you've offered came from someone who made blatant, factual errors that were trivially easy to debunk.
I mean, obviously there isn't and probably won't ever be any definitive proof that Niemann cheated at Sinquefield. Even the mechanical clicking caught on video just as Niemann appeared to fiddle with something behind his ear in his post round 3 interview can only ever be seen as highly suggestive. We will never know. But we can certainly use our heads and make well-informed arguments based on the information we do have.The only people who can offer compelling arguments here are 1. Carlsen, who has refused to speak, and 2. Chesscom, who has claimed to send their information to Hans. Hans has not refuted this, so we can assume that Chesscom did indeed send Hans their information, and that Hans would prefer it not be public. I'm happy to hop on the online cheating bandwagon as a result, but OTB Carlsen must speak.
My use of "random" was obviously intended to be disparaging and connote the easily falsifiable nature of the claims made by Your Guy. In that sense, I'm not "random". My claims are verifiable to the extent that I have claimed, and neither my methods nor my results have yet to be seriously challenged. I provided links directly to every single data point I used, and showed the results of the (completely standard) regressions that I ran. Repeating my analysis is as simple as clicking on the links, verifying the numbers, and typing them into Excel, or R, or whatever. If you attempt to verify the claims of Your Guy, you immediately run into the issue that he committed obvious blunders and made factually incorrect claims, for example, by failing to understand how USCF classical ratings are calculated.As far as the people outside this forum are concerned, you are also a random tweet from a random person with little to no citations.