@@ -44,11 +44,11 @@ def run_worker(backend: str, cfg: dict, results_path: str) -> None:
4444 import aiohttp
4545 import torch
4646
47- from benchmarks .sglang_vs_vllm .metrics_collector import (
47+ from benchmarks .sglang_benchmarks .metrics_collector import (
4848 BenchmarkRun , RequestMetrics , StepMetrics ,
4949 get_gpu_memory_usage_nvidia_smi ,
5050 )
51- from benchmarks .sglang_vs_vllm .config import generate_benchmark_prompts
51+ from benchmarks .sglang_benchmarks .config import generate_benchmark_prompts
5252
5353 logger .info (f"[{ backend } ] Worker PID={ os .getpid ()} GPUs={ torch .cuda .device_count ()} " )
5454
@@ -217,7 +217,7 @@ async def _run_unsloth() -> BenchmarkRun:
217217 packed_tensors_from_tokenized_results ,
218218 packed_tensors_to_dir ,
219219 )
220- from benchmarks .sglang_vs_vllm .unsloth_sglang_service import UnslothSGLangService
220+ from benchmarks .sglang_benchmarks .unsloth_sglang_service import UnslothSGLangService
221221
222222 unsloth_port = cfg .get ("unsloth_port" , 8300 )
223223 unsloth_lora_rank = cfg .get ("unsloth_lora_rank" , 1 )
@@ -552,7 +552,7 @@ def main():
552552 return
553553
554554 # ---- Orchestrator mode ----------------------------------------
555- from benchmarks .sglang_vs_vllm .metrics_collector import (
555+ from benchmarks .sglang_benchmarks .metrics_collector import (
556556 BenchmarkRun , StepMetrics , RequestMetrics ,
557557 generate_comparison_report_multi ,
558558 )
@@ -641,7 +641,7 @@ def main():
641641
642642
643643def _dict_to_run (d : dict ):
644- from benchmarks .sglang_vs_vllm .metrics_collector import BenchmarkRun , StepMetrics , RequestMetrics
644+ from benchmarks .sglang_benchmarks .metrics_collector import BenchmarkRun , StepMetrics , RequestMetrics
645645 run = BenchmarkRun (backend = d ["backend" ], model = d ["model" ],
646646 dataset = d .get ("dataset" , "" ), server_startup_time = d .get ("server_startup_s" , 0 ))
647647 run .start_time = 0.0
0 commit comments