Skip to content

Commit

Permalink
clean after rebase
Browse files Browse the repository at this point in the history
  • Loading branch information
Valery Chernov committed Jan 9, 2024
1 parent 9129ef0 commit 86d63fa
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 5 deletions.
5 changes: 2 additions & 3 deletions serve/mlc_serve/engine/base.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,12 +4,11 @@
from enum import Enum
from abc import ABC, abstractmethod

from typing import List, Callable, Any, Optional, Dict, Tuple
from typing import List, Callable, Any, Optional, Dict
import inspect
import numpy as np

from .sampling_params import SamplingParams, SamplingType
from ..api.protocol import LogprobsContent

RequestId = str

Expand Down Expand Up @@ -174,7 +173,7 @@ class SequenceOutput:
finish_reason: Optional[FinishReason] = None
# Number of generated tokens so far
num_generated_tokens: int = 0
logprob_info: Optional[LogprobsContent] = None
logprob_info: Optional[RawLogprobsInfo] = None

@property
def is_finished(self) -> bool:
Expand Down
2 changes: 0 additions & 2 deletions serve/mlc_serve/model/paged_cache_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,6 @@
from .paged_cache_manager import KVCache, CacheManager
from .tokenizer import HfTokenizerModule, ConversationTemplate, Tokenizer
from ..engine import (
RequestId,
SamplingType,
MLCServeEngineConfig,
SamplingParams,
Expand All @@ -29,7 +28,6 @@
DecodeRequest,
PrefillRequest,
TextGenerationResult,
LOGPROBS_TYPE
)
from ..engine.model_module import ModelModule

Expand Down

0 comments on commit 86d63fa

Please sign in to comment.