Getting Good Data, Part II (of many)
Okay, so by now you’ve heard dozens and dozens of times that you guys produce really good data: your aggregated answers are 97% correct overall (see here and here and here). But we also know that not all images are equally easy. More specifically, not all species are equally easy. It’s a lot easier to identify a giraffe or zebra than it is to decide between an aardwolf and striped hyena.
The plot below shows the different error rates for each species. Note that error comes in two forms. You can have a “false negative” which means you miss a species given that it’s truly there. And then you can have a “false positive,” in which you report a species as being there when it really isn’t. Error is a proportion from 0 to 1.
We calculated this by comparing the consensus data to the gold standard dataset that Margaret collated last year. Note that at the bottom of the chart there are a handful of species that don’t have any values for false negatives. That’s because, for statistical reasons, we could only calculate false negative error rates from completely randomly sampled images, and those species are so rare that they didn’t appear in the gold standard dataset. But for false positives, we could randomly sample images from any consensus classification – so I gathered a bunch of images that had been identified as these rare species and checked them to calculate false positive rates.
Now, if a species has really low rates of false negatives and really low rates of false positives, then it’s one people are really good at identifying all the time. Note that in general, species have pretty low rates of both types of error. Furthermore, species with lower rates of false negatives have higher rates of false positives. There aren’t really any species with high rates of both types of error. Take rhinos, for example: folks often identify a rhino when it’s not actually there, but never miss a rhino if it is there.
Also: we see that rare species are just generally harder to identify correctly than common species. The plot below shows the same false negative and false positive error rates plotted against the total number of pictures for every species. Even though there is some noise, those lines reflect significant trends: in general, the more pictures of an animal, the more often folks get it right!
This makes intuitive sense. It’s really hard to get a good “search image” for something you never see. But also folks are especially excited to see something rare. You can see this if you search the talk pages for “rhino” or “zorilla.” Both of these have high false positive rates, meaning people say it’s a rhino or zorilla when it’s really not. Thus, most of the images that show up tagged as these really rare creatures aren’t.
But that’s okay for the science. Because recall that we can assess how confident we are in an answer based on the evenness score, fraction support, and fraction blanks. Because such critters are so rare, we want to be really sure that those IDs are right — but because those animals are so rare, and because you have high levels of agreement on the vast majority of images, it makes it really easy to review any “uncertain” image that’s been ID’d as a rare species.
Pretty cool, huh?
Check out Mammal March Madness!
For some reason I missed this in 2013 and 2014. Maybe it was because I was finishing up my dissertation the first time and then recovering from a cross-country move the second time. But now I am totally excited about 2015’s
Mammal March Madness
What is Mammal March Madness? I’ll let organizer Katie Hinde explain:
In honor of the NCAA College Basketball March Madness Championship Tournament, Mammals Suck is featuring *simulated* combat competition among mammals. … Battle outcome is a function of the two species’ attributes within the battle environment. Attributes considered in calculating battle outcome include temperament, weaponry, armor, body mass, fight style, and other fun facts that are relevant to the outcome.
As a spectator to Mammal March Madness, you fill out a bracket and then follow along on Twitter or on the Mammals Suck … Milk! blog. The first game is on Monday, March 9, and direct elimination games continue until the championship on March 26.
I’ll note that the 2014 winner was Hyena who defeated Orca in the championship game. This year, we’ve got some Serengeti representation as well. But with Lion, Baboon, and Vervet monkey ranked just 8th, 12th, and 13th in the ‘Sexy Beasts’ division, they’re going to need all the cheering-on they can get.
So head on over, print out a bracket, and tell me who you think will make it all the way to the top this year.
(And just to be clear, I am not involved in Mammal March Madness in any way except as a participant. But it looks fun!)