Skip to content

Commit f1dfbc5

Browse files
py4Pooya Moradi
andauthored
[Test] Fix broken tests due to upstream change. (#1228)
Signed-off-by: Pooya Moradi <pooyam@google.com> Co-authored-by: Pooya Moradi <pooyam@google.com>
1 parent 6570a1e commit f1dfbc5

File tree

6 files changed

+31
-16
lines changed

6 files changed

+31
-16
lines changed

tests/runner/test_kv_cache_manager.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,9 @@ def setup_method(self):
4949
swap_space=4,
5050
cache_dtype="auto",
5151
)
52-
scheduler_config = SchedulerConfig(max_num_seqs=16, )
52+
scheduler_config = SchedulerConfig(max_num_seqs=16,
53+
max_model_len=1024,
54+
is_encoder_decoder=False)
5355
parallel_config = ParallelConfig(
5456
pipeline_parallel_size=1,
5557
tensor_parallel_size=1,

tests/runner/test_multimodal_manager.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -42,7 +42,9 @@ def setup_method(self):
4242
swap_space=4,
4343
cache_dtype="auto",
4444
)
45-
scheduler_config = SchedulerConfig(max_num_seqs=16, )
45+
scheduler_config = SchedulerConfig(max_num_seqs=16,
46+
max_model_len=1024,
47+
is_encoder_decoder=False)
4648
parallel_config = ParallelConfig(
4749
pipeline_parallel_size=1,
4850
tensor_parallel_size=1,

tests/runner/test_speculative_decoding_manager.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,9 @@ def setup_method(self):
4141
swap_space=4,
4242
cache_dtype="auto",
4343
)
44-
scheduler_config = SchedulerConfig(max_num_seqs=16, )
44+
scheduler_config = SchedulerConfig(max_num_seqs=16,
45+
max_model_len=1024,
46+
is_encoder_decoder=False)
4547
parallel_config = ParallelConfig(
4648
pipeline_parallel_size=1,
4749
tensor_parallel_size=1,

tests/runner/test_structured_decoding_manager.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,9 @@ def setup_method(self):
3939
swap_space=4,
4040
cache_dtype="auto",
4141
)
42-
scheduler_config = SchedulerConfig(max_num_seqs=16, )
42+
scheduler_config = SchedulerConfig(max_num_seqs=16,
43+
max_model_len=1024,
44+
is_encoder_decoder=False)
4345
parallel_config = ParallelConfig(
4446
pipeline_parallel_size=1,
4547
tensor_parallel_size=1,

tests/runner/test_tpu_runner.py

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,9 @@ def setup_method(self):
3434
swap_space=4,
3535
cache_dtype="auto",
3636
)
37-
scheduler_config = SchedulerConfig(max_num_seqs=16, )
37+
scheduler_config = SchedulerConfig(max_num_seqs=16,
38+
max_model_len=1024,
39+
is_encoder_decoder=False)
3840
parallel_config = ParallelConfig(
3941
pipeline_parallel_size=1,
4042
tensor_parallel_size=1,
@@ -131,7 +133,9 @@ def setup_method(self):
131133
swap_space=4,
132134
cache_dtype="auto",
133135
)
134-
scheduler_config = SchedulerConfig(max_num_seqs=16, )
136+
scheduler_config = SchedulerConfig(max_num_seqs=16,
137+
max_model_len=1024,
138+
is_encoder_decoder=False)
135139
parallel_config = ParallelConfig(
136140
pipeline_parallel_size=1,
137141
tensor_parallel_size=1,

tests/spec_decode/test_eagle3.py

Lines changed: 13 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -38,16 +38,19 @@ def _create_proposer(
3838
num_speculative_tokens=num_speculative_tokens,
3939
)
4040

41-
vllm_config = VllmConfig(
42-
model_config=model_config,
43-
cache_config=CacheConfig(block_size=16),
44-
speculative_config=speculative_config,
45-
device_config=DeviceConfig(device="tpu"),
46-
parallel_config=ParallelConfig(pipeline_parallel_size=1,
47-
tensor_parallel_size=1),
48-
load_config=LoadConfig(),
49-
scheduler_config=SchedulerConfig(max_num_batched_tokens=8192,
50-
max_num_seqs=128))
41+
vllm_config = VllmConfig(model_config=model_config,
42+
cache_config=CacheConfig(block_size=16),
43+
speculative_config=speculative_config,
44+
device_config=DeviceConfig(device="tpu"),
45+
parallel_config=ParallelConfig(
46+
pipeline_parallel_size=1,
47+
tensor_parallel_size=1),
48+
load_config=LoadConfig(),
49+
scheduler_config=SchedulerConfig(
50+
max_num_batched_tokens=8192,
51+
max_num_seqs=128,
52+
max_model_len=model_config.max_model_len,
53+
is_encoder_decoder=False))
5154

5255
# Mock the runner, as the proposer needs it for initialization
5356
mock_runner = mock.MagicMock()

0 commit comments

Comments
 (0)