Batch Size — The number of training examples processed before updating model weights. Larger batches use more GPU memory but train faster. Smaller batches introduce noise that can help generalization. Typical sizes: 16, 32, 64, 128, 256. The right batch size depends on your GPU memory and dataset.
Part of the XLUXX AI Encyclopedia — the most comprehensive AI reference on the web.

Leave a Reply