Alex, as we've repeatedly tried to tell you, feel free to run the numbers any way you want but don't expect your results to be held in high regard. Garbage in - garbage out, it's really that simple. Your computer normalization program works so well on weather because you are only asking it to predict an overall average for temperature or rainfall for example and you have a huge sample size obtained over more than a hundred years with calibrated instruments which results in only a few outlying results that can be disregarded. Even with this ideal setup we are more than 30 degrees below our expected average temp today, have received snow and are expecting a killing frost overnight.
I'm sure 30 degrees below normal falls well within the expected range or probability density of temps for today. Statistics are great for possible ranges based on lots of good data. They are also good for determining probability densities when even small samples are closely matched to limit noisy data and are precisely monitored - that's why drug trials use similar aged patients, with similar conditions, and base dosages on weight. The point being that there is lots of verifiable data and/or lots of controls to ensure good data. We don't have any of that with the bigfoot reports you plan on using. There is no control over who submits, no control over data accuracy and no way to clean up the data by normalizing because we have no standard like we would in a rainfall study with calibrated gauges.
I believe you mean to use no criteria for inclusion but rather rely on throwing out some percentage of the reports? Are you using sightings only or including encounters based on sounds or possible footprints? SWAG1. What if 90% or more of whatever reports you choose were mistaken or hoaxed as some proponents believe? What if it's 50%? Because you are not using calibrated instruments but will instead rely on humans who are prone to error your study will obviously be less precise but by how much? Imagine basing your rainfall data on people's quick perceptions gathered in the amount of time a sighting takes place rather a calibrated gauge. How close would you expect the data to track now? Introduce SWAG2. What percentage will you ascribe to outright lying? Introduce SWAG3. How many are sightings of the same creature at different times or places? SWAG4. How many SWAGs will there be? Do you realize that depending on how you set this up potential errors in the SWAGs can compound?
The difficulty in doing statistics is not in applying simple math to a sample but in choosing the sample and correctly determining what factors to apply. It's impossible to do if you if you can't trust the data itself due to unreliable reporting. All I can say is good luck!