diff --git a/backends/exllamav2/model.py b/backends/exllamav2/model.py index 5d7bd15..97e54ad 100644 --- a/backends/exllamav2/model.py +++ b/backends/exllamav2/model.py @@ -1242,9 +1242,6 @@ class ExllamaV2Container: grammar_handler = ExLlamaV2Grammar() banned_strings = [] - # TODO: Not used for some reason? - generate_window = max(params.generate_window, self.config.max_seq_len // 8) - self.assign_gen_params( params, gen_settings, diff --git a/common/sampling.py b/common/sampling.py index 5eb2dc8..fc9f9bc 100644 --- a/common/sampling.py +++ b/common/sampling.py @@ -41,12 +41,6 @@ class BaseSamplerRequest(BaseModel): ge=0, ) - generate_window: Optional[int] = Field( - default_factory=lambda: get_default_sampler_value("generate_window", 512), - examples=[512], - ge=0, - ) - stop: Optional[Union[str, List[Union[str, int]]]] = Field( default_factory=lambda: get_default_sampler_value("stop", []), validation_alias=AliasChoices("stop", "stop_sequence"), diff --git a/sampler_overrides/sample_preset.yml b/sampler_overrides/sample_preset.yml index d85229c..9225976 100644 --- a/sampler_overrides/sample_preset.yml +++ b/sampler_overrides/sample_preset.yml @@ -14,9 +14,6 @@ max_tokens: min_tokens: override: 0 force: false -generate_window: - override: 512 - force: false stop: override: [] force: false