The research and engineering discipline of ensuring that AI systems pursue goals and exhibit behaviours that match human intentions and values. Misalignment risks range from models following instructions too literally to more speculative long-term risks discussed in AI safety literature.
Boek een consultatie om te bespreken hoe AI-concepten op uw uitdagingen van toepassing zijn.