Thanks everyone for the responses. I think at this point, due to the 30-minute delay, the most effective way for me to respond is to create a big comment and just tag people at relevant sections. If you do not like me tagging you, please message me and I will remove you from this post and not tag you again.
@JeremyHussell thanks for the comment about Wikipedia, it is a good point that Wikipedia won’t allow unsourced ID tips. @tiagolubiana, I did not mean to come across as rude calling myself and most of the other complainers lazy for not editing Wikipedia. Definitely appreciate the work you’ve done on Wikipedia.
@tiagolubiana I apologize, I did not mean to disregard people’s feelings. However, I do think it is factually correct that the dissenters are making incorrect assumptions. I think that is demonstrated by your willingness to concede points 1 and 2 (though I must add, your stressing of “eventually” shows you are not familiar with the current techniques that can be used to great effect on targeted applications such as iNaturalist’s project). Whether they are cold and heartless is a subjective matter and I do not believe it is fair to use those terms pejoratively at the AI, especially considering how cold and heartless the primary literature often is.
I actually anticipated your counterpoint on point 3 in my final paragraph (“I understand that many of you have ethical concerns…”), so I do not believe I missed the point.
On point 4, the assumptions are not questionable because while humans are not optional, “humans performing this task” is just as optional as using an LLM. I agree that humans get motivated by working on iNat, and I would be a lot more motivated to provide ID comments if this feature is implemented.
@Masebrock, your example of a genetic fallacy is actually an ad hominem fallacy, which I think may speak for itself (and @tiagolubiana may want to reevaluate referencing it to debunk my 6th point).
In contrast, I have absolutely applied the genetic fallacy correctly. We must judge things on their own merits. We cannot dismiss a good thing because it comes from a bad company without committing a genetic fallacy (error in reasoning caused by evaluating the source of the thing and concluding that the thing cannot be good). Plus, you are overlooking the extremely salient point I already made–iNaturalist has collaborated with Google for a very long time without any problems, so it is not reasonable to say Google has “proven itself untrustworthy” in the context of iNaturalist. If anything, Google has proven itself to be a valuable partner.
Google’s own AI hallucinates. And the iNat partnership is with Google. Whatever “mitigation” you’re suggesting exists is apparently far from sufficient. Instead of saying “this time it will be different”, first develop an AI that has solved this problem, then get back to me.
Methods of dramatically decreasing hallucination have been public for years. See A Comprehensive Survey of Hallucination Mitigation Techniques in Large Language Models for an analysis of more than 30 methods to reduce hallucinations. It’s harder to implement these in a general generative AI model like Google’s public-facing model, but when you tailor Google’s LLM to a specific domain and then use one or more easy-to-implement mitigation techniques, this is really not a problem at all.
The only problem that would remain is the one @JeremyHussell points out:
It absolutely will [add incorrect ID information], because people on iNaturalist do, and make comments laying out their (flawed) reasoning on iNaturalist.
This could indeed be a problem, but it can be mitigated if users can provide feedback to the AI outputs (as the blog post indicates). However,
LLMs are probabilistic: they will sometimes put out one line of argument, and sometimes another, based on the frequency with which they see those lines of argument in the context they’re in.
I think this would work to our advantage. I have yet to see an example taxon where anywhere near the majority of ID comments were incorrect. I suspect that the correct or useful ID comments outweigh the incorrect ones by a substantial margin. Once again, any errors could be corrected by user input.
@calvertm, the environmental concerns have been addressed. In the US, LLMs are between 40 and 4,400 x more environmentally friendly than human labor. Reconciling the contrasting narratives on the environmental impact of large language models | Scientific Reports. There are some caveats there, but I have not seen any convincing reason to believe iNaturalist’s usecase would be less environmentally friendly. I hope this can help keep you on iNaturalist–looks like we will need more help in the coming days.
@andreass1 thanks for the response. I agree that AI is not making valuable art. I think you are drawing an equivalence between the general-use generative AI models and what iNaturalist is suggesting. Your point that these models are creating additional information is moot in this scenario, where the model creates information that otherwise would be created by a human. And it takes far fewer resources to correct a sentence or two of LLM output than to write everything from scratch. As I have noted many times now, there are ways to mitigate hallucinations, especially in a domain-specific application like what iNaturalist is proposing. Tailored LLMs with guiderails do not produce AI slop.
For instance, one LLM used in my field is PQAI, a free and open-source prior art searcher. (Edit suggested by @spidercat: prior art just means previous inventions, either in patents or other literature, not traditional art). It uses an LLM to take a natural language query, perform a prior art search, and pull up relevant papers and patents along with AI summaries. It has proven to be far better at finding relevant prior art than traditional searching resources that cost thousands of dollars, democratizing access to essential information for innovators. I could go on with many other examples of high-power, high-accuracy, low-hallucination applications that do not produce AI slop. iNaturalist should have no problem creating a solution that works.
With all due respect, I still do not see evidence that the people threatening to leave understand the underlying technology. I empathize with those who plan to stay but still have reservations. While I appreciate all of you, have learned from each of you, and do not mean to insult you by this, I also feel tired of responding to the handful of arguments that miss what I’m saying and get basic facts wrong. It also is demoralizing for my well-researched, factually accurate comments to get ratioed by factually inaccurate responses that fail to address my points. At this point, I recognize that this is not worth my effort, and I will let iNaturalist handle their own PR and use my free time to add identifications to make up for the upcoming exodus…