Hey all, the first update to the Computer Vision model since March of 2020 has been released! The iNaturalist website (and all network member sites) are using it as well as the Android and iOS iNaturalist apps. Seek, however, is still running on the older model as we still need to compress this new model so it will fit within the Seek app.
There is also a new
Pending - The current Computer Vision Model does not know about this taxon, so while it might be included in automated suggestions with the “Nearby” label, it will not have the “Visually Similar” label. While the requirements for model inclusion change with each model, generally inclusion is based on number of observations, so to increase the chance of this taxon getting included during the next model training, add or identify more observations of this taxon.
Yes the CV Demo page uses the same model as the website and the mobile apps. As mentioned in the blog post, the only exception is the on-device CV model that Seek uses, which we plan to update in the near future.
The taxon Osmunda spectabilis, the American royal fern, also has the “pending” label, and sure enough, it does not appear in CV suggestions. Yet there are over 8000 research grade observations for it. Why is this?
Hmm, so one downside of using an output as an input in a taxon split like that is that Osmunda regalis sensu stricto computer vision suggestions are likely based in part on many photos of Osmunda spectabilis?
I thought it might be something like that. It is suggesting O. regalis if I “include suggestions not seen nearby.”
This isn’t a big problem for me, and I don’t really need to pursue it. I’ll just input O. spectabilis manually for the time being. Thanks for looking into it.
Is anyone else noting the suggestions of Anseriformes being really bad now? They were reliable before but right now the CV is having trouble recognizing Mallard and Canada Goose.
Yeah, I’m concerned about vision accuracy on avian hybrids & related species. It’s turning out to be a hard problem for the vision system. We don’t train on subspecies, and it may be that we shouldn’t train on any infraspecific taxa. I’m planning to do experiments next month to decide whether we should exclude avian hybrids in future models, or otherwise treat them differently.
The background is that previous versions of the model had far fewer avian hybrid photos to train on. We had a large growth in the number of avian hybrid identifications in the past year and for the first time, our (capped) training data had as many Mallard x American Black Duck photos as Mallard photos and American Black Duck photos. We didn’t exclude them from the training dataset because it wasn’t a known problem, but obviously I’m re-evaluating that now.
Is there any way certain taxa could be trained above the genus level? Many (most?) ciliates (to say nothing of other microbes) are not possible to identify to species with home equipment. Even genus can be quite challenging at times, sometimes impossible without complex staining techniques using chemicals not available to the amateur.
I don’t think that’s possible, since taxa above genus level have many genera under them, which might have completely different habitus. The AI still will suggest families and the like if it finds multiple genera that look similar to the observation.
Genus level is exactly the level I would like to see (this is the typical goal of microbial ID). But since I see no computer vision note on the genus pages, I assume they are not being trained at that level.
“We’ve also released a new feature for taxon pages on the website which allows you to see which taxa are included in the model. This badge only appears on species pages, not pages of genera, families, etc.”
yes it can. To quote Alex who runs it, “If enough photos are present at the genus level, but not enough photos for any of the descendent species, then the genus will be placed in the training set”
I posted the quote in my previous reply to explain why the tag symbol doesn’t show up on the genus pages