Every time you swipe a credit card, most major financial companies will tap machine learning to generate a risk score—a likelihood that the purchase is fraudulent.
But over time, scammers may switch up their tricks or buying habits might change. Suddenly, the machine learning model is operating in a different environment than that on which it was trained. Maybe it’s flagging legitimate transactions or missing actual fraud as a result.
This is a phenomenon called “model drift”—a mismatch that emerges over time between what an AI system was trained to do and how it operates in the real world. And it’s not just a problem for credit card fraud detection, it can affect models of all kinds, including LLMs, according to Helen Gu, founder and CEO at InsightFinder, which works with Visa on avoiding these kinds of outcomes.
“Model drift is hard to detect because it’s not something you can actually clearly describe using one metric, and it’s basically a running metric…you have to compute this metric over a period of time,” Gu said.
It’s one of a few reasons that AI models might seem to degrade over time without constant monitoring and tweaking. As more businesses move from AI prototypes to full-fledged production, fending off these sorts of real-world quality issues may be more top of mind.
“Back in 2023, the focus was more around communicating that there is something called hallucination and why it could happen,” Amit Paka, founder and COO of the AI observability platform Fiddler. “Now [the conversation with clients is] more around, ‘What’s the quality of the hallucination metric that you have?’”
Keep reading here.—PK
|