From 93a056fe4249281107baafc2a936e34f7f59e111 Mon Sep 17 00:00:00 2001 From: Albert Villanova del Moral <8515462+albertvillanova@users.noreply.github.com> Date: Thu, 12 Dec 2024 16:30:45 +0100 Subject: [PATCH] Fix imports from model_config (#443) --- docs/source/using-the-python-api.mdx | 2 +- src/lighteval/main_endpoint.py | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/docs/source/using-the-python-api.mdx b/docs/source/using-the-python-api.mdx index 2e160a679..8c44050f4 100644 --- a/docs/source/using-the-python-api.mdx +++ b/docs/source/using-the-python-api.mdx @@ -11,7 +11,7 @@ After that, simply run the pipeline and save the results. ```python import lighteval from lighteval.logging.evaluation_tracker import EvaluationTracker -from lighteval.models.model_config import VLLMModelConfig +from lighteval.models.vllm.vllm_model import VLLMModelConfig from lighteval.pipeline import ParallelismManager, Pipeline, PipelineParameters from lighteval.utils.utils import EnvConfig from lighteval.utils.imports import is_accelerate_available diff --git a/src/lighteval/main_endpoint.py b/src/lighteval/main_endpoint.py index 952aae074..be75b711a 100644 --- a/src/lighteval/main_endpoint.py +++ b/src/lighteval/main_endpoint.py @@ -93,7 +93,7 @@ def openai( Evaluate OPENAI models. """ from lighteval.logging.evaluation_tracker import EvaluationTracker - from lighteval.models.model_config import OpenAIModelConfig + from lighteval.models.endpoints.openai_model import OpenAIModelConfig from lighteval.pipeline import EnvConfig, ParallelismManager, Pipeline, PipelineParameters env_config = EnvConfig(token=TOKEN, cache_dir=cache_dir) @@ -317,7 +317,7 @@ def tgi( import yaml from lighteval.logging.evaluation_tracker import EvaluationTracker - from lighteval.models.model_config import TGIModelConfig + from lighteval.models.endpoints.tgi_model import TGIModelConfig from lighteval.pipeline import EnvConfig, ParallelismManager, Pipeline, PipelineParameters env_config = EnvConfig(token=TOKEN, cache_dir=cache_dir)