A technique that normalises activations across the batch dimension during training, accelerating convergence and reducing sensitivity to hyperparameters. Though largely superseded by layer norm in transformers, batch norm remains prevalent in convolutional vision architectures.
Réservez une consultation pour discuter de l'application des concepts IA à vos défis.