Archive

Discover and discuss technology tools

Explore the Tiscuss archive by category or keyword, then jump into conversations around what matters most.

Search and filters
Reset
Active: any category / query: Chatbots / page 1 of 1 / 2 total
AI Tools

How Do Developers Correct AI LLMs When They Spread Misinformation?

I watched Last Week Tonight's piece on AI chatbots today, and it got me thinking about that old screenshot of a Google search in which Gemini recommends adding "1/8 cup of non-toxic glue" to pizza in order to make the cheese better stick to the slice. When something like this goes viral, I have to assume (though I could be wrong) that an employee at Google specifically goes out of their way to address that topic in particular. The image is a meme, of course, but I imagine Google wouldn't be keen to leave themselves open to liability if their LLM recommends that users consume glue. Does the developer "talk" to the LLM to correct it about that specific case? Do they compile specific information about (e.g.) pizza construction techniques and feed it that data to bring it to the forefront? Do their actions correct only the case in question, or do they make changes to the LLM that affects its accuracy more broadly (e.g. "teaching" the LLM to recognize that some Reddit comments are jokes)? On a more heavy note, the LWT piece includes several stories of chatbots encouraging users to self-harm. How does the process differ when developers are trying to prevent an LLM from giving that sort of response?

Global · General · Apr 29, 2026
AI Marketing

Snapchat Introduces AI Chat Ads for Conversational Marketing

Snapchat unveils AI Powered Chat Advertisements for Enhanced Engagement Snapchat has recently launched a groundbreaking feature called AI Chat Ads, designed to …

Global · Marketers · Apr 29, 2026
PreviousPage 1 / 1Next