What is this - iNaturalist and generative AI?

What statements have suggested to you that this is what is planned? (And what precisely do you think is being proposed? How specifically will it hinder your ability to do research or the reliability of iNat data? I ask not in the interest of defending generative AI, but because there has been a lot of speculation and people are making very different assumptions about what this will look like. To understand your concerns, we have to know what you think is going to happen.)

I am not a defender of generative AI. I hate the way it is currently being forced upon us by countless websites and programs and I have major concerns about ethics, hidden bias, lack of reliability, eroding of human connections and agency, etc. I would be loudly protesting if I thought that the proposal was, say, to implement a chatbot that would mediate our interactions with iNat. I am not thrilled that this is where development energy is being directed rather than on myriad other improvements that we have been asking for for years.

But there are lots of ways that generative AI can be implemented, and I am trying very hard to trust staff and remain open to the idea that an optional custom-designed model intended for limited, specific purposes might be able to offer tools for searching or analyzing iNat’s data and texts in ways not possible with other technologies.

iNat already uses AI in the form of its computer vision. While I have something of a love-hate relationship with the CV, which performs quite poorly for my main taxon of interest, and while I am concerned by the widespread uncritical use of its suggestions and the way the new app seems to encourage this even more than was the case previously – on the whole I think iNat has done a good job striking a balance between making the AI assistance available without allowing it to dominate over human interactions or remove individual agency.

Please also note that iNat has only committed to producing a demo application of generative AI, perhaps something similar to the Vision Language demo they presented last year. This is not in any way the same thing as making it part of iNat’s interface. The fact that the previous experimental development did not become a new fixed feature of iNat suggests that staff is capable of making good judgments about what is acceptable and desirable for users and is not set on forcing innovations upon us that are neither useful nor wanted.

15 Likes

In my opinion bringing up the CV model doesn’t help the discussion. People (or most people in this thread) aren’t worried about “AI” in general, they are worried about gen AI. Part of the problem is still a lot of uncertainty. Yes, the staff have given some additional details in the thread, but they are buried in a 500 comment post. While I want to assume good intentions and trust the iNat team, we all have different ways to get there and not everyone will trust things “blindly”. The lack of another official announcement sharing what they do know, certainly gives me pause, and while I’m waiting for the Q&A session and the brainstorming to be done, the implementation of the genAI is just one part of it:

While I understand there are a lot of things up in the air, I also think they could be more transparent about the grant (which I — and maybe many of us — don’t understand how the process works):

I honestly also don’t understand why the wiki many have asked for hasn’t been addressed at all.

3 Likes

You have 2 tools just below the latest comment.
Most engaged / engaging comments is an acceptable option (whups 600 comments down to 100) ? The ‘hamburger’ icon :grinning_face:

1 Like

While slightly helpful, that is not an “official” announcement. It doesn’t get delivered the same way and doesn’t “notify” people the same way a blog post does. And like you said, it’s still buried in between other comments. More importantly it doesn’t answer many of the questions I made above.

3 Likes

That’s perfectly reasonable, and I don’t think anyone is demanding blind trust. Note that I was responding to a post by someone who said that generative AI will make iNat unusable for them but when asked if they would care to explain, essentially told us to shut up because it is none of our business.

My point was that it is surely a bit easier for staff to respond to concerns if they know more details about what those concerns are. Given that there has been a lot of discussion about all sorts of things it is likely that people are coming from different places with different understandings of what the situation is and different needs. Part of the conversation is going to have to be sorting out those various positions. Stating that one has a blanket mistrust of generative AI without providing more specifics isn’t a great basis for figuring out where to go next.

The CV is not irrelevant. AI (both generative and non-generative) encompasses a wide range of technologies with their own different biases, problems, and risks. It is not as simple as all generative AI being inherently harmful while non-generative image recognition like the CV is fundamentally innocous. The CV is a black box. It is subject to biases resulting from choices about what to include in its training set. It is susceptible to misuse – people uncritically using it to make IDs for themselves and others – and to circular reasoning whereby bad data can, under certain circumstances, end up being fed back into the training and perpetuating misconceptions and wrong IDs. Implementing the CV required making decisions about how to manage these risks: how it should be integrated into the interface, how much freedom and power the algorithm should have, how much uncertainty should be allowed in its predictions, how user interactions with it should be framed. It is quite conceivable that a team with a different vision for iNat might have made other choices, assigned the CV a much more prominent role, weighted results in problematic ways, etc. These are some of the same things that would need to be considered before any implementation of a generative AI tool.

So if one wants to get an idea about whether staff are likely to make responsible decisions as part of the current project, looking to past decision-making and implementation of other AI tools may provide some insights into their philosophy and approach. Some may find reason for reassurance in past behavior, others may not. But since I doubt that anyone on staff knows at this point exactly what the “demo” will look like, I think it would be difficult at present for them to provide concrete details beyond the fact that they are exploring possibilities and taking feedback from the community about their concerns and what would be acceptable and what not.

I read most of the blog post as standard grant verbiage of the sort that is fairly normal in science and research – that is, framing plans so as to highlight elements that are key to the purpose outlined in the call for submissions; this does not necessarily correspond with how much of a role they will later play in the activities made possible by the grant. It also doesn’t mean that the ideas sketched out in the grant proposal will be implemented in that exact form – the purpose of the proposal is to convince sponsors that one has ideas that are worth pursuing, but in practice, it may take a different direction as one figures out what is feasible. Given that the blog was likely written in some haste and repeats many of the ideas from the post on the Vision Language demo, I suspect that it reflects what was written in the grant proposal rather than being a well-considered presentation for the iNat community about what the plans are now that the grant has been received.

I am not a researcher but I work in academia, where the strategies for procuring funding are presumably not unlike the strategies used by non-profits. Since a lot of funding is in the form of project-based grants that cover expenses for, at most, a year or two, you are more-or-less constantly looking for the next source(s) of funding. Many grants are quite competitive, so you cast your net wide, apply for lots of different things, and hope that at least one of them is successful. Since iNat is already firmly anchored in digital technologies (app-based mobile interface) and AI development (the CV), continuing to explore the possibilities of AI is one way to help it position itself as relevant and interesting for further investments. And at the moment, generative AI is what is felt to be cutting-edge and therefore where a good portion of the money is. Nobody is going to give a grant to support the creation of a wiki with identification information. Or to help iNat develop a strategy for better user onboarding and recruiting IDers. Generally grants are based around the idea of some tangible outcome (“deliverable”) – this means that most grants are not simply going to be grants to cover running expenses so that iNat can continue as usual, without developing something new or producing results that wouldn’t have been possible otherwise.

7 Likes

I work in Computer science in academia. I disagree this is standard verbiage. I also don’t think I would get funding without the additional details I’m asking about in my post. Me saying “I don’t understand the Google grant proposal process” is giving them the benefit of the doubt. I would really like an explanation about how that works, and what was submitted as part of the proposal. If it works some other way, then I’d like to know that too so I that I can stop trying to fill in the blanks with my assumptions and my experience.

This is what we are asking. Simply be transparent about what was submitted.

I am not saying it is irrelevant, I am saying it is not helpful. People tolerate some level of “AI”, but this thread should make it clear there is a line being crossed and gen AI is on the side many of us aren’t comfortable with. The differences between these two have been discussed in this thread already so I won’t repeat that. Telling people they shouldn’t worry about gen AI because iNat implemented the CV thoughtfully, doesn’t address the worries and issues that are specific about gen AI. People’s concerns about genAI in iNat are more likely going to be assuaged by the staff than by well-intentioned users trying to interpret things between the lines.

My point is that we can either continue to guess what they may or may not do (based or not on the current features and the current CV and some of the staff posts) or the team could be transparent about the proposal and remove some of the uncertainty. And I certainly can sympathise with the concerned users and see where they are coming from, even if I try to give the benefit of the doubt to the iNat team.

This is not what I’ve said. To clarify, this is a feature people have requested and could be implemented regardless of and independently from the entire genAI issue.

11 Likes

I would be equally upset by either a person or LLM that paraphrased my descriptions/explanations, made some mistakes in phrasing, muddled what I stated, placed statements out of context, and then cited me as the source (or cited someone else as the source who didn’t contribute at all). The LLM’s we’ve seen demonstrated here on this thread are great examples of why people don’t want LLMs implemented here–because they did all or some of those things.

9 Likes

It’s very good to know that this grant from Google doesn’t seem to have any strings attached, not does it lead necessarily to the forced use of large language models in species suggestions.

Given the loss of trust that the purported use of the Google grant has brought about, plus the possibility of future reoccurrences of this or a similar issue, what I see as necessary is the establishment of a board of trustees to oversee major decisions. After all, iNat is a charity, and being accountable to the US government is not enough; it should be accountable first and foremost to all of those who commit their time and effort to fine-comb the world with their observations and to those who curate them. A board of trustees, with individual trustees being biologists, educators, and the like, could monitor operations and serve as a body for consultation to sound out any major proposals, acting as a guarantor for the statutes of the charity.

6 Likes

iNaturalist is already governed by a Board of Directors. if you don’t like the makeup of the Board, then you can always look at the organization’s bylaws to see how it is organized and work within those rules to influence / make the changes that you would like to see.

6 Likes

It is not very clear how many of these 7 work in the field of natural sciences, and rely on iNat data to this avail.

I don’t think the bylaws are posted on the iNaturalist site, but can be found here: https://rct.doj.ca.gov/Verification/Web/Download.aspx?saveas=1769427+FD.pdf&document_id=09027b8f804efde8

7 Likes

They were great examples of the results (good and bad) that could be obtained by a generic LLM in a very basic and simple initial experiment. I’m not sure though that they have very much to do with what could be obtained in a carefully planned and implemented trial involving a highly specialised LLM dedicated specifically to iNat :wink: .

6 Likes

They did a good job of illustrating (using actual LLMs) the concerns that many people have been voicing–that was my point. These are the sorts of things that real LLMs do, and these are the sorts of things that would need to be watched out for with the demo that is developed. In that context, I think they did a good job of illustrating several concerns.

3 Likes

Regarding the plagiarism comments, it does make me feel uncomfortable, but there are four safeguards that could be put into place that would completely alleviate my discomfort with an AI using my words.

  1. The AI says that the information was aggregated (paraphrased) from comments by [users]. I would rather this than a misquote, though adding quotation marks around direct quotes would be ideal.
  2. A statement provided with the AI summary that distinctly says that information may be misquoted or paraphrased incorrect.
  3. Links to the original posts that had the original wording and the context of that wording. This may need to be accompanied by duplication detection so that only unique comments are linked to. Even better is if each word or phrase was either hyperlinked to the original comment or hovering over the text showed a popup of the full original comment.
  4. A flag option to indicate when the AI is misquoting a user. This would allow further improvements to the AI to help guide it to quoting the source correctly. Alternatively, this could also be used to stop the AI from quoting the particular comment entirely if all instances of the use of the comment are misquoted in a severe enough way.
10 Likes

Just as a general comment, I wonder if it would be constructive to refocus on listing what would tip you over the edge allow you to tolerate or make you comfortable with the use of AI in the proposed way. I’ve created a new topic to hopefully help with that: https://forum.inaturalist.org/t/what-would-make-ai-tolerable-for-you/66987

I hope it’s helpful. 600 comments is a lot to wade through. Even the 100 comments provided by the “hamburger” is a lot to wade through at this point. The useful ideas about what we think the future AI should look like are buried in lengthy discussions. Pulling these out might provide more fruitful discussion.

At the end of the day, it would take a pretty substantial abuse to kick me off iNaturalist. A statement saying that the AI is the absolute authority on the matter might cross that line and get me upset. But, my personal opinion on the AI using my words is that it’s not worth getting offended over. I understand if you feel differently, but I simply don’t think the impact will be that bad. If the AI tools turns out to be bad, I just won’t use it. If people start spouting nonsense they learned from the AI, I’ll correct them. That’s my perspective on it. If it’s overly simplistic, than it won’t be the first or last time I’ve been wrong.

6 Likes

This is how I structured how I imagined the output might look above which includes most of those, but including everything requires a decent amount of text, which would either have to be small fineprint or take up a fair amount of space on the CV suggestion page on a small phone screen.

Should also add that it’s only in eastern North America as that eliminates other similar species like Red-crowned Woodpecker. But that goes to my point that more clarity requires increasingly more text.

1 Like

As it stands, from the perspective of an amateur science communicator who teaches (mostly) beginners how to get started with IDing the plants and fungi around them, I think a generative AI tool that explains how the app reached an ID is dangerous because it removes an opportunity for someone to use their brain and figure it out themselves.
Honestly, I prefer books and keys whenever reasonable. I know that’s old-fashioned, but there’s something to be said for what is and is not condusive to learning, memorization, and critical thinking. The process of figuring it out yourself IS the process of learning. I’m not comfortable teaching folks to use a tool that takes away what I see as an important opportunity to hone observational and reasoning skills. It’s the opposite direction I want to be pointing people in, especially if they’re beginners.

I don’t want a machine telling someone to pay attention to a stripe of color on a frog’s side. I want folks to use their eyes and their brains to look and read and make connections, so they can figure things out manually, the hard way, with an abundance of caution. I want folks looking at their own photo of a frog and comparing it to suggested species which they then also have to learn about. I want folks comparing and contrasting the physical characteristics of species so that they develop an eye for it, and get better over time.

I absolutely have concerns about Google being involved because of their history of messing up smaller companies, and I have concerns about the very real toll that generative AI takes on the environment, but the effect that some generative AI tools have on our ability to think for ourselves is far more grim in my mind, and I won’t support it any more than I support students in school using ChatGPT to outline their essays for them. I just think it’s repulsive.
And then there’s issues around plagiarism and consent, which are also very serious and we’ve seen in this very thread how upsetting it can be to have your voice stolen.
I really just hate all of this, and I wanted to share my two cents even if it makes me feel like an old man yelling at clouds.

11 Likes

i understand the desire to have a separate clean thread to collect a specific subset of ideas, but i think folks have demonstrated over there that they aren’t willing to allow that sort of discussion to occur. rather than flagging posts and being accused of suppressing discussion, and to help prevent folks from just duplicating thought here and there, i would suggest that the other thread be merged into this one.

maybe when there’s a clearer plan from iNat staff, then it might be time to have a clean thread to discuss new things.

4 Likes

I think that’s true, but I think useful information can still be extracted despite the noise. I don’t think this would be possible (or at least not easy) if the threads were merged into one. My preference would be to allow the noise but maintain the distinct threads, at least for now. If you feel any of my attempts to redirect towards the original post went too far, feel free to let me know.

The lists created so far have been very interesting and insightful. I think the aggregated list also provides a useful picture of what people want in the gray area between the simple “for” and “against”.

2 Likes

1000%!!! So many employees at compainies that have “deployed” AI are saying that it literally just made their jobs harder, because now they have to fix all the mistakes the AI made!

Why can’t we just be given the option to have a “users guide to identification” page on the taxon? It would be completely free, done by volunteers like the rest of the identifications and observations on the site, and wouldn’t have ANY of the moral and efficiency problems AI does.

Hell, I’ve already written two identification guides since joining this site, one for Small-flower and common pawpaws, the other for American persimmons. Why not just encourage more users to write identification guides?

Why go money-grubbing to google for AI crap that’s going to harm the environment, pollute minority neighborhoods, and provide false information, when the users of the site are right here, willing and able to provide the information ourselves?

7 Likes