Skip to content

Commit

Permalink
Fix imports from model_config (#443)
Browse files Browse the repository at this point in the history
  • Loading branch information
albertvillanova authored Dec 12, 2024
1 parent 0135c2e commit 93a056f
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 3 deletions.
2 changes: 1 addition & 1 deletion docs/source/using-the-python-api.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ After that, simply run the pipeline and save the results.
```python
import lighteval
from lighteval.logging.evaluation_tracker import EvaluationTracker
from lighteval.models.model_config import VLLMModelConfig
from lighteval.models.vllm.vllm_model import VLLMModelConfig
from lighteval.pipeline import ParallelismManager, Pipeline, PipelineParameters
from lighteval.utils.utils import EnvConfig
from lighteval.utils.imports import is_accelerate_available
Expand Down
4 changes: 2 additions & 2 deletions src/lighteval/main_endpoint.py
Original file line number Diff line number Diff line change
Expand Up @@ -93,7 +93,7 @@ def openai(
Evaluate OPENAI models.
"""
from lighteval.logging.evaluation_tracker import EvaluationTracker
from lighteval.models.model_config import OpenAIModelConfig
from lighteval.models.endpoints.openai_model import OpenAIModelConfig
from lighteval.pipeline import EnvConfig, ParallelismManager, Pipeline, PipelineParameters

env_config = EnvConfig(token=TOKEN, cache_dir=cache_dir)
Expand Down Expand Up @@ -317,7 +317,7 @@ def tgi(
import yaml

from lighteval.logging.evaluation_tracker import EvaluationTracker
from lighteval.models.model_config import TGIModelConfig
from lighteval.models.endpoints.tgi_model import TGIModelConfig
from lighteval.pipeline import EnvConfig, ParallelismManager, Pipeline, PipelineParameters

env_config = EnvConfig(token=TOKEN, cache_dir=cache_dir)
Expand Down

0 comments on commit 93a056f

Please sign in to comment.