From 7941b5bab709a5b3286a501048bdc085787e46f0 Mon Sep 17 00:00:00 2001 From: Katrina Cruz Date: Thu, 30 May 2024 10:21:28 -0700 Subject: [PATCH] Change instance count for ACA workloadProfiles to be minimum 0 and maximum 1 --- .../infra/provision/finetuning.parameters.json | 2 +- .../infra/provision/inference.parameters.json | 2 +- configs/llama-v2-7b/infra/provision/finetuning.parameters.json | 2 +- configs/llama-v2-7b/infra/provision/inference.parameters.json | 2 +- configs/llama-v3-8b/infra/provision/finetuning.parameters.json | 2 +- configs/llama-v3-8b/infra/provision/inference.parameters.json | 2 +- .../mistral-7b-v02/infra/provision/finetuning.parameters.json | 2 +- .../mistral-7b-v02/infra/provision/inference.parameters.json | 2 +- configs/mistral-7b/infra/provision/finetuning.parameters.json | 2 +- configs/mistral-7b/infra/provision/inference.parameters.json | 2 +- configs/phi-1_5/infra/provision/finetuning.parameters.json | 2 +- configs/phi-1_5/infra/provision/inference.parameters.json | 2 +- configs/phi-2/infra/provision/finetuning.parameters.json | 2 +- configs/phi-2/infra/provision/inference.parameters.json | 2 +- .../zephyr-7b-beta/infra/provision/finetuning.parameters.json | 2 +- .../zephyr-7b-beta/infra/provision/inference.parameters.json | 2 +- 16 files changed, 16 insertions(+), 16 deletions(-) diff --git a/configs/Phi-3-mini-4k-instruct/infra/provision/finetuning.parameters.json b/configs/Phi-3-mini-4k-instruct/infra/provision/finetuning.parameters.json index 851f4db..fd4019d 100644 --- a/configs/Phi-3-mini-4k-instruct/infra/provision/finetuning.parameters.json +++ b/configs/Phi-3-mini-4k-instruct/infra/provision/finetuning.parameters.json @@ -12,7 +12,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "timeout": { "value": 10800 diff --git a/configs/Phi-3-mini-4k-instruct/infra/provision/inference.parameters.json b/configs/Phi-3-mini-4k-instruct/infra/provision/inference.parameters.json index a0c419d..73b2076 100644 --- a/configs/Phi-3-mini-4k-instruct/infra/provision/inference.parameters.json +++ b/configs/Phi-3-mini-4k-instruct/infra/provision/inference.parameters.json @@ -11,7 +11,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "location": { "value": null diff --git a/configs/llama-v2-7b/infra/provision/finetuning.parameters.json b/configs/llama-v2-7b/infra/provision/finetuning.parameters.json index a5681d6..d3f569c 100644 --- a/configs/llama-v2-7b/infra/provision/finetuning.parameters.json +++ b/configs/llama-v2-7b/infra/provision/finetuning.parameters.json @@ -12,7 +12,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "timeout": { "value": 10800 diff --git a/configs/llama-v2-7b/infra/provision/inference.parameters.json b/configs/llama-v2-7b/infra/provision/inference.parameters.json index a0c419d..73b2076 100644 --- a/configs/llama-v2-7b/infra/provision/inference.parameters.json +++ b/configs/llama-v2-7b/infra/provision/inference.parameters.json @@ -11,7 +11,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "location": { "value": null diff --git a/configs/llama-v3-8b/infra/provision/finetuning.parameters.json b/configs/llama-v3-8b/infra/provision/finetuning.parameters.json index 2310573..2dddb94 100644 --- a/configs/llama-v3-8b/infra/provision/finetuning.parameters.json +++ b/configs/llama-v3-8b/infra/provision/finetuning.parameters.json @@ -12,7 +12,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "timeout": { "value": 10800 diff --git a/configs/llama-v3-8b/infra/provision/inference.parameters.json b/configs/llama-v3-8b/infra/provision/inference.parameters.json index a0c419d..73b2076 100644 --- a/configs/llama-v3-8b/infra/provision/inference.parameters.json +++ b/configs/llama-v3-8b/infra/provision/inference.parameters.json @@ -11,7 +11,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "location": { "value": null diff --git a/configs/mistral-7b-v02/infra/provision/finetuning.parameters.json b/configs/mistral-7b-v02/infra/provision/finetuning.parameters.json index 300e0a1..6476f35 100644 --- a/configs/mistral-7b-v02/infra/provision/finetuning.parameters.json +++ b/configs/mistral-7b-v02/infra/provision/finetuning.parameters.json @@ -12,7 +12,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "location": { "value": null diff --git a/configs/mistral-7b-v02/infra/provision/inference.parameters.json b/configs/mistral-7b-v02/infra/provision/inference.parameters.json index a0c419d..73b2076 100644 --- a/configs/mistral-7b-v02/infra/provision/inference.parameters.json +++ b/configs/mistral-7b-v02/infra/provision/inference.parameters.json @@ -11,7 +11,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "location": { "value": null diff --git a/configs/mistral-7b/infra/provision/finetuning.parameters.json b/configs/mistral-7b/infra/provision/finetuning.parameters.json index 4c1f50f..b446d9a 100644 --- a/configs/mistral-7b/infra/provision/finetuning.parameters.json +++ b/configs/mistral-7b/infra/provision/finetuning.parameters.json @@ -12,7 +12,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "location": { "value": null diff --git a/configs/mistral-7b/infra/provision/inference.parameters.json b/configs/mistral-7b/infra/provision/inference.parameters.json index a0c419d..73b2076 100644 --- a/configs/mistral-7b/infra/provision/inference.parameters.json +++ b/configs/mistral-7b/infra/provision/inference.parameters.json @@ -11,7 +11,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "location": { "value": null diff --git a/configs/phi-1_5/infra/provision/finetuning.parameters.json b/configs/phi-1_5/infra/provision/finetuning.parameters.json index 8e21ea4..0afc3a7 100644 --- a/configs/phi-1_5/infra/provision/finetuning.parameters.json +++ b/configs/phi-1_5/infra/provision/finetuning.parameters.json @@ -12,7 +12,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "timeout": { "value": 10800 diff --git a/configs/phi-1_5/infra/provision/inference.parameters.json b/configs/phi-1_5/infra/provision/inference.parameters.json index 54ab90a..b83dfd1 100644 --- a/configs/phi-1_5/infra/provision/inference.parameters.json +++ b/configs/phi-1_5/infra/provision/inference.parameters.json @@ -11,7 +11,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "location": { "value": null diff --git a/configs/phi-2/infra/provision/finetuning.parameters.json b/configs/phi-2/infra/provision/finetuning.parameters.json index e57bca7..53e2f03 100644 --- a/configs/phi-2/infra/provision/finetuning.parameters.json +++ b/configs/phi-2/infra/provision/finetuning.parameters.json @@ -12,7 +12,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "timeout": { "value": 10800 diff --git a/configs/phi-2/infra/provision/inference.parameters.json b/configs/phi-2/infra/provision/inference.parameters.json index 54ab90a..b83dfd1 100644 --- a/configs/phi-2/infra/provision/inference.parameters.json +++ b/configs/phi-2/infra/provision/inference.parameters.json @@ -11,7 +11,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "location": { "value": null diff --git a/configs/zephyr-7b-beta/infra/provision/finetuning.parameters.json b/configs/zephyr-7b-beta/infra/provision/finetuning.parameters.json index df8068a..2bad80f 100644 --- a/configs/zephyr-7b-beta/infra/provision/finetuning.parameters.json +++ b/configs/zephyr-7b-beta/infra/provision/finetuning.parameters.json @@ -12,7 +12,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "timeout": { "value": 10800 diff --git a/configs/zephyr-7b-beta/infra/provision/inference.parameters.json b/configs/zephyr-7b-beta/infra/provision/inference.parameters.json index a0c419d..73b2076 100644 --- a/configs/zephyr-7b-beta/infra/provision/inference.parameters.json +++ b/configs/zephyr-7b-beta/infra/provision/inference.parameters.json @@ -11,7 +11,7 @@ ] }, "maximumInstanceCount": { - "value": 2 + "value": 1 }, "location": { "value": null