Over the past year, AI researchers have found that when AI chatbots such as ChatGPT find themselves unable to answer questions that satisfy users’ requests, they tend to offer false answers. In a new study, as part of a program aimed at stopping chatbots from lying or making up answers, a research team added Chain of Thought (CoT) windows. These force the chatbot to explain its reasoning as it carries out each step on its path to finding a final answer to a query.
Experiments show adding CoT windows to chatbots teaches them to lie less obviously
Tech News
-
Highlights
Free Dark Web Monitoring Stamps the $17 Million Credentials Markets
-
Highlights
Smart buildings: What happens to our free will when tech makes choices for us?
-
Apps
Screenshots have generated new forms of storytelling, from Twitter fan fiction to desktop film
-
Highlights
Darknet markets generate millions in revenue selling stolen personal data, supply chain study finds
-
Security
Privacy violations undermine the trustworthiness of the Tim Hortons brand
-
Featured Headlines
Why Tesla’s Autopilot crashes spurred the feds to investigate driver-assist technologies – and what that means for the future of self-driving cars