From ebfcab79dd98cbfadaf2dd3fa1a7d9b7f52ffd2d Mon Sep 17 00:00:00 2001 From: ninggu Date: Tue, 24 Feb 2026 16:42:04 -0800 Subject: [PATCH] Update llama3.1-405b 32 node README.md --- .../nemo-pretraining-gke/32node-FP8CS-GBS2048/recipe/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/training/a4x/llama3-1-405b/nemo-pretraining-gke/32node-FP8CS-GBS2048/recipe/README.md b/training/a4x/llama3-1-405b/nemo-pretraining-gke/32node-FP8CS-GBS2048/recipe/README.md index 485aa912..0ffded1f 100644 --- a/training/a4x/llama3-1-405b/nemo-pretraining-gke/32node-FP8CS-GBS2048/recipe/README.md +++ b/training/a4x/llama3-1-405b/nemo-pretraining-gke/32node-FP8CS-GBS2048/recipe/README.md @@ -83,7 +83,7 @@ gcloud container clusters get-credentials $CLUSTER_NAME --region $CLUSTER_REGION ### Configure and submit a pretraining job -#### Using 16 node (64 gpus) fp8 precision +#### Using 32 node (128 gpus) fp8 precision To execute the job with the default settings, run the following command from your client: