How an “Expert” Will Dismiss the Death Data Analysis
By Steve Kirsch
I got this email from one of my readers (Jeff) who explained why “experts” won’t believe what the death data indicates.
The plan is to have a third party survey company ask people survey questions like:
- Did anyone in your family die in the last 18 months?
- When did they die (date)?
- How old were they?
- Did they ever receive a COVID vaccine prior to death?
From this we can figure out whether the vaccine is safe or not. That is all NO opinion or belief is required. This is basically pure fact regurgitation.
“Jeff” wrote me:
I passed along the survey concept to “prove” increase in all cause mortality to see how it would be received by a Ph.D. in social sciences / marketing who works near enough to vaccine proponents to stand as an indicator of likely rejection of my representation of what I think you have in mind – also expert in social stats, who I personally know, and below you will kindly find his response, which if not addressed, will likely be the kind of resistant efforts, from the pro-vaxx side of the discourse, you will likely encounter.
That is the best I can offer to you – I hope it saves time and money and further, increases the likely acceptance of your survey results by both anti-vaxxers and by those who support covid vaccinations.
His friend wrote to him:
A social survey is an inappropriate method out of the box.
Social surveys about knowing others who had X happen after Y experience is 100% hearsay, but could look and feel “legitimate”, which is an issue. Its called bias.
It won’t likely fly in either law or psychological (or opinion) measurement. You might be able to flip it, asking beliefs about cov vax efficacy and safety, and then ask if they know anyone personally whom they believe was injured by admin of the cov vax. That gives a measure of belief and the potential reason for the belief. Social surveys, e.g. opinion surveys are good at getting at beliefs and poor at predicting behavior based on beliefs, and awful at medical diagnosis.
The method itself is inappropriate, even if you could ask valid and reliable questions.
Valid and reliable questions require at least three samples (of say 500) depending on how many constructs are included, and the indicators used to measure the constructs.
In my example above you would need to test at least 3-4 questions that measure beliefs about cov vax efficacy and safety (I suspect the questions are out there already tested by public health related centers, schools or health communications studies programs. Might go quick.) The questions about personally knowing someone you believe was injured by cov vax, would take some testing to get valid/reliable phrasing. Then you have belief.
Since the effort is geared from the outset to prove rather than test, it’s biased in formulation. Another reason I wouldn’t go that route. It’s the exact same thing that the parties involved are alleging others are doing, ironically.
In other words, there will be people you will never convince no matter how hard you try.
I thought this was so absurd that I wanted to share it with you.