Skip to content

Commit

Permalink
Use linux.4xlarge.memory instead of linux.12xlarge (pytorch#6896)
Browse files Browse the repository at this point in the history
* Try linux.4xlarge.memory

* Testing

* Ready to land

* More tests

* Forget one line change

* Use linux.8xlarge.memory for llama3_2_text_decoder

* Increase timeout value

* Skip llama3_2_text_decoder because it takes too long to export

* Ready to land
  • Loading branch information
huydhn authored Nov 26, 2024
1 parent 0c83c59 commit aadf2ee
Show file tree
Hide file tree
Showing 2 changed files with 12 additions and 9 deletions.
20 changes: 11 additions & 9 deletions .ci/scripts/gather_test_models.py
Original file line number Diff line number Diff line change
Expand Up @@ -20,16 +20,16 @@
CUSTOM_RUNNERS = {
"linux": {
# This one runs OOM on smaller runner, the root cause is unclear (T163016365)
"w2l": "linux.12xlarge",
"ic4": "linux.12xlarge",
"resnet50": "linux.12xlarge",
"llava": "linux.12xlarge",
"llama3_2_vision_encoder": "linux.12xlarge",
# "llama3_2_text_decoder": "linux.12xlarge", # TODO: re-enable test when Huy's change is in / model gets smaller.
"w2l": "linux.4xlarge.memory",
"ic4": "linux.4xlarge.memory",
"resnet50": "linux.4xlarge.memory",
"llava": "linux.4xlarge.memory",
"llama3_2_vision_encoder": "linux.4xlarge.memory",
"llama3_2_text_decoder": "linux.4xlarge.memory",
# This one causes timeout on smaller runner, the root cause is unclear (T161064121)
"dl3": "linux.12xlarge",
"emformer_join": "linux.12xlarge",
"emformer_predict": "linux.12xlarge",
"dl3": "linux.4xlarge.memory",
"emformer_join": "linux.4xlarge.memory",
"emformer_predict": "linux.4xlarge.memory",
}
}

Expand All @@ -39,10 +39,12 @@
"linux": {
"mobilebert": 90,
"emformer_predict": 360,
"llama3_2_text_decoder": 360,
},
"macos": {
"mobilebert": 90,
"emformer_predict": 360,
"llama3_2_text_decoder": 360,
},
}

Expand Down
1 change: 1 addition & 0 deletions examples/models/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,6 +19,7 @@
"llama2": ("llama", "Llama2Model"),
"llama": ("llama", "Llama2Model"),
"llama3_2_vision_encoder": ("llama3_2_vision", "FlamingoVisionEncoderModel"),
# TODO: This take too long to export on both Linux and MacOS (> 6 hours)
# "llama3_2_text_decoder": ("llama3_2_vision", "Llama3_2Decoder"),
"lstm": ("lstm", "LSTMModel"),
"mobilebert": ("mobilebert", "MobileBertModelExample"),
Expand Down

0 comments on commit aadf2ee

Please sign in to comment.