Originally published on Towards AI.
One of the major challenges in using LLMs in business is that LLMs hallucinate. How can you entrust your clients to a chatbot that can go mad and tell them something inappropriate at any moment? Or how can you trust your corporate AI assistant if it makes things up randomly?
That’s a problem, especially given that an LLM can’t be fired or held accountable.
Leave a reply