Cheuk Ting Ho
Developer advocate / Data Scientist - support open-source and building the community.
Grab the slides: slides.com/cheukting_ho/ai-shadows/
A lot of example
prompt
new content
model
New research (not peer-reviewed) from UCSD found that on GPT-4, 62% of the generated code contains API misuses
We incorporated more human feedback, including feedback submitted by ChatGPT users, to improve GPT-4’s behavior.
A lot of example
prompt
new content
model
Prompt injection attacks, as the name suggests, involve maliciously inserting prompts or requests in interactive systems to manipulate or deceive users, potentially leading to unintended actions or disclosure of sensitive information.
It’s similar to something like an SQL injection attack in that a command is embedded in something that seems like a normal input at the start.
SQL injection is still included in the Open Worldwide Application Security Project (OWASP) Top 10 list of security vulnerabilities.
Credit: Anne Guérin
A lot of example
prompt
new content
model
Noelle Martin was 17 when she discovered that her face had been edited onto naked photos of someone else...
her screen had been flooded by deepfake pornographic imagery –
featuring her face – created by an unknown group of “nameless, faceless” sexual predators.
Australia’s eSafety Commissioner has already received a number of complaints about non-consensual distribution of deepfake intimate images, and expects this type of abuse to grow in volume as artificial intelligence (AI) technology becomes more accessible.
Identity thieves?
Fake news?
Grab the slides: slides.com/cheukting_ho/ai-shadows/
By Cheuk Ting Ho
Developer advocate / Data Scientist - support open-source and building the community.