diff --git a/src/documentation/elo_qa_eval.ipynb b/src/documentation/elo_qa_eval.ipynb index ef2b2c55..4b71af6f 100644 --- a/src/documentation/elo_qa_eval.ipynb +++ b/src/documentation/elo_qa_eval.ipynb @@ -448,7 +448,7 @@ "outputs": [], "source": [ "newly_added_models = [\n", - " Llama3InstructModel(name=\"llama-3.1-70b-instruct\", client=aa_client),\n", + " Llama3InstructModel(name=\"llama-3.3-70b-instruct\", client=aa_client),\n", "]\n", "\n", "for model in newly_added_models:\n", diff --git a/src/intelligence_layer/core/model.py b/src/intelligence_layer/core/model.py index 3c691808..c1612174 100644 --- a/src/intelligence_layer/core/model.py +++ b/src/intelligence_layer/core/model.py @@ -261,7 +261,7 @@ def __init__( ) if name not in [model["name"] for model in self._client.models()]: warnings.warn( - "The provided model is not a recommended model for this model class." + "The provided model is not a recommended model for this model class. " "Make sure that the model you have selected is suited to be use for the prompt template used in this model class." ) self._complete: Task[CompleteInput, CompleteOutput] = _Complete( @@ -414,7 +414,7 @@ def __init__( ) -> None: if name not in self.RECOMMENDED_MODELS or name == "": warnings.warn( - "The provided model is not a recommended model for this model class." + "The provided model is not a recommended model for this model class. " "Make sure that the model you have selected is suited to be use for the prompt template used in this model class." ) super().__init__(name, client)