The research and engineering discipline of ensuring that AI systems pursue goals and exhibit behaviours that match human intentions and values. Misalignment risks range from models following instructions too literally to more speculative long-term risks discussed in AI safety literature.
Réservez une consultation pour discuter de l'application des concepts IA à vos défis.