The tendency of LLMs to generate plausible-sounding but factually incorrect or fabricated content. Hallucination is a primary enterprise risk for generative AI and is mitigated through RAG, grounding techniques, output validation, and human-in-the-loop review processes.
Réservez une consultation pour discuter de l'application des concepts IA à vos défis.