Catherine Gracey
1. Searching / Information Retrieval
2. Publishing
Searching for information is the primary way in which people are using GenAI [a]
In many cases, this is sufficient, but there are also some issues with relying on GenAI for information.
Before I explain, let's rewind and explore some of the different tools for finding scholarly information.
"artificial intelligence" AND "diagnosis"
*the word diagnosis doesn't actually appear, but ML is used to determine that this is about diagnosis
| Databases | Search Engines | LLMs | RAG | |
|---|---|---|---|---|
| Ease of search | Skill required | Easy | Easy | Easy |
| Ease of interpretation | Skill required | Moderate | Easy | Easy |
| Transparency | High | Moderate | Low | Low or Moderate |
| Reliability | High | Low | Low | Moderate |
Why do transparency & reliability matter in information retrevial?
By handing off autonomy to these systems, we introduce the possibility for censorship & bias
If a system determines something isn't relevant, or search engine optomized, or is controversial, it simply doesn't present it to you, meaning you could only be getting half the story
[a]
Use tools that have RAG capabilities, and search Scholarly sources
Otherwise you might be trusting a Reddit user
Make sure the original source contains what the GenAI output is saying it does (have hallucinations or misrepresentations occurred?)
Please don't cite OpenAI or Microsoft - cite the human authors who wrote the linked work (think of the h-index!!)
AI-generated content has been showing up in the published literature for a while now
However, different publishers/journals have different policies, some indicating GenAI should be used very minimally or not at all
[a]
[a]
[b]
[b]
[a]
[a]
[a]
[a]
[b]
If publishers hold exclusive rights to your work, they have the authority to license it for various uses, including AI training, and financially benefit from these deals.
– Dede Dawson, 2024
Publishers can do this because authors have signed away the exclusive rights to their work in many cases
Authors did not explicitly consent to the sale of their work to AI companies, but had signed their rights away
There is very limited (or no) ability to opt out as authors
You don't want to finish an article only to realize you've accidentally broken a policy
If you are accused of using AI in a way you weren't supposed to, it can be helpful to have proof of work documents at the ready
Get a second set of eyes on documents you're asked to sign
[a]
[b]