@@ -61,8 +61,8 @@ def git_url(self) -> str:
6161 return "https://github.com/intel/compute-benchmarks.git"
6262
6363 def git_hash (self ) -> str :
64- # Nov 17, 2025
65- return "932ae79f7cca7e156285fc10a59610927c769e89 "
64+ # Dec 17, 2025
65+ return "420549188cd8900c27cf9b04fd859ebe81876a99 "
6666
6767 def setup (self ) -> None :
6868 if options .sycl is None :
@@ -182,7 +182,7 @@ def benchmarks(self) -> list[Benchmark]:
182182 GraphApiSinKernelGraph (self , runtime , with_graphs , num_kernels )
183183 )
184184
185- # Add ULLS benchmarks
185+ # Add ULLS benchmarks
186186 for runtime in list (RUNTIMES ):
187187 if runtime == RUNTIMES .SYCL :
188188 benches .append (
@@ -355,6 +355,36 @@ def createTorchMultiQueueBench(variant_name: str, **kwargs):
355355 ),
356356 ]
357357
358+ # Add TorchSlmSize benchmarks
359+ for runtime in filter (lambda x : x != RUNTIMES .UR , RUNTIMES ):
360+
361+ def createTorchSlmSizeBench (variant_name : str , ** kwargs ):
362+ return TorchSlmSize (
363+ self ,
364+ runtime ,
365+ variant_name ,
366+ PROFILERS .TIMER ,
367+ ** {** kwargs , "warmupIterations" : 1 },
368+ )
369+
370+ benches += [
371+ createTorchSlmSizeBench (
372+ "small" ,
373+ batchSize = 512 ,
374+ slmNum = 1 ,
375+ ),
376+ createTorchSlmSizeBench (
377+ "medium" ,
378+ batchSize = 512 ,
379+ slmNum = 1024 ,
380+ ),
381+ createTorchSlmSizeBench (
382+ "max" ,
383+ batchSize = 512 ,
384+ slmNum = - 1 ,
385+ ),
386+ ]
387+
358388 # Add UR-specific benchmarks
359389 benches += [
360390 # TODO: multithread_benchmark_ur fails with segfault
@@ -810,25 +840,31 @@ def _bin_args(self, run_trace: TracingType = TracingType.NONE) -> list[str]:
810840 return [f"--{ k } ={ v } " for k , v in self ._rr_params .items ()]
811841
812842
813- class TorchMultiQueue (ComputeBenchmark ):
843+ class TorchBenchmark (ComputeBenchmark ):
814844 def __init__ (
815- self , suite , runtime : RUNTIMES , variant_name : str , profiler_type , ** kwargs
845+ self ,
846+ suite ,
847+ runtime : RUNTIMES ,
848+ bench_name : str ,
849+ variant_name : str ,
850+ profiler_type ,
851+ ** kwargs ,
816852 ):
817853 self ._variant_name = variant_name
818- self ._smq_params = kwargs
854+ self ._torch_params = kwargs
819855 self ._iterations_regular = 1000
820856 self ._iterations_trace = 10
821857 super ().__init__ (
822858 suite ,
823859 f"torch_benchmark_{ runtime .value } " ,
824- "KernelSubmitMultiQueue" ,
860+ bench_name ,
825861 runtime ,
826862 profiler_type ,
827863 )
828864
829865 def name (self ):
830866 ret = []
831- for k , v in self ._smq_params .items ():
867+ for k , v in self ._torch_params .items ():
832868 ret .append (f"{ k } { v } " )
833869 ret .sort ()
834870 return self ._bench_name + " " + ", " .join (ret )
@@ -848,10 +884,38 @@ def _supported_runtimes(self) -> list[RUNTIMES]:
848884 def _bin_args (self , run_trace : TracingType = TracingType .NONE ) -> list [str ]:
849885 iters = self ._get_iters (run_trace )
850886 return [f"--iterations={ iters } " ] + [
851- f"--{ k } ={ v } " for k , v in self ._smq_params .items ()
887+ f"--{ k } ={ v } " for k , v in self ._torch_params .items ()
852888 ]
853889
854890
891+ class TorchMultiQueue (TorchBenchmark ):
892+ def __init__ (
893+ self , suite , runtime : RUNTIMES , variant_name : str , profiler_type , ** kwargs
894+ ):
895+ super ().__init__ (
896+ suite ,
897+ runtime ,
898+ "KernelSubmitMultiQueue" ,
899+ variant_name ,
900+ profiler_type ,
901+ ** kwargs ,
902+ )
903+
904+
905+ class TorchSlmSize (TorchBenchmark ):
906+ def __init__ (
907+ self , suite , runtime : RUNTIMES , variant_name : str , profiler_type , ** kwargs
908+ ):
909+ super ().__init__ (
910+ suite ,
911+ runtime ,
912+ "KernelSubmitSlmSize" ,
913+ variant_name ,
914+ profiler_type ,
915+ ** kwargs ,
916+ )
917+
918+
855919class QueueInOrderMemcpy (ComputeBenchmark ):
856920 def __init__ (self , bench , isCopyOnly , source , destination , size , profiler_type ):
857921 self ._is_copy_only = isCopyOnly
0 commit comments