Typos and slang in patient messages can trip up AI models, leading to inconsistent medical recommendations

A large language model (LLM) deployed to make treatment recommendations can be tripped up by nonclinical information in patient messages, like typos, extra white space, missing gender markers, or the use of uncertain, dramatic, and informal language, according to a study by MIT researchers.

This post was originally published on this site

Skip The Dishes Referral Code

KeyLegal.ca - Consult a Lawyer Online in a variety of legal subjects