Add MaxText Llama 3.1 70B training with GCS recipe#56
Add MaxText Llama 3.1 70B training with GCS recipe#56raymondzouu wants to merge 1 commit intomainfrom
Conversation
698777c to
57ff296
Compare
f8dcddc to
51d9cee
Compare
| @@ -0,0 +1,11 @@ | |||
| python3 benchmarks/benchmark_runner.py xpk \ | |||
| project=$PROJECT \ | |||
| zone=$ZONE \ | |||
There was a problem hiding this comment.
Not sure why we need the .sh script here. This is not blocking, if you think its needed, go ahead with it.
LGTM.
2246636 to
c568eb0
Compare
c568eb0 to
fe2b53f
Compare
mkmg
left a comment
There was a problem hiding this comment.
Thanks for making all those changes; it's looking really good! Just a few more comments, but I think we are close!
f9d5280 to
9f57b66
Compare
mkmg
left a comment
There was a problem hiding this comment.
Please see the last few comments, but overall LGTM! Thanks for your work on this!
9f57b66 to
e52b1b8
Compare
e52b1b8 to
4aea327
Compare
| device_type=v6e-256 \ | ||
| num_slices=1 \ | ||
| cluster_name=$CLUSTER \ | ||
| base_output_directory=$OUTPUT_DIR \ |
There was a problem hiding this comment.
I just noticed that there are some env variables here that users are not instructed to populate above.
| ``` | ||
| export RECIPE_REPO="path-to-this-recipe-repo" # Update | ||
|
|
||
| cd ~/xpk |
There was a problem hiding this comment.
Also we never instruct the user to clone XPK.
There was a problem hiding this comment.
In step 1 we link to the instructions for cloning xpk and installing the dependencies.
4aea327 to
2e13786
Compare
No description provided.