@@ -95,7 +95,7 @@ def pre_hook(ctx: Dict[str, Any], *args: Any, **kwargs: Any) -> None:
9595 )
9696 ctx ["model_name" ] = kwargs .get ("model" , "" )
9797 set_span_attribute (
98- ctx ["span" ], AttributeKeys .GEN_AI_REQUEST_MODEL , ctx ["model_name" ]
98+ ctx ["span" ], AttributeKeys .JUDGMENT_LLM_MODEL_NAME , ctx ["model_name" ]
9999 )
100100
101101 def post_hook (ctx : Dict [str , Any ], result : Message ) -> None :
@@ -112,17 +112,19 @@ def post_hook(ctx: Dict[str, Any], result: Message) -> None:
112112 _extract_anthropic_tokens (result .usage )
113113 )
114114 set_span_attribute (
115- span , AttributeKeys .GEN_AI_USAGE_INPUT_TOKENS , prompt_tokens
115+ span ,
116+ AttributeKeys .JUDGMENT_USAGE_NON_CACHED_INPUT_TOKENS ,
117+ prompt_tokens ,
116118 )
117119 set_span_attribute (
118- span , AttributeKeys .GEN_AI_USAGE_OUTPUT_TOKENS , completion_tokens
120+ span , AttributeKeys .JUDGMENT_USAGE_OUTPUT_TOKENS , completion_tokens
119121 )
120122 set_span_attribute (
121- span , AttributeKeys .GEN_AI_USAGE_CACHE_READ_INPUT_TOKENS , cache_read
123+ span , AttributeKeys .JUDGMENT_USAGE_CACHE_READ_INPUT_TOKENS , cache_read
122124 )
123125 set_span_attribute (
124126 span ,
125- AttributeKeys .GEN_AI_USAGE_CACHE_CREATION_INPUT_TOKENS ,
127+ AttributeKeys .JUDGMENT_USAGE_CACHE_CREATION_INPUT_TOKENS ,
126128 cache_creation ,
127129 )
128130 set_span_attribute (
@@ -133,7 +135,7 @@ def post_hook(ctx: Dict[str, Any], result: Message) -> None:
133135
134136 set_span_attribute (
135137 span ,
136- AttributeKeys .GEN_AI_RESPONSE_MODEL ,
138+ AttributeKeys .JUDGMENT_LLM_MODEL_NAME ,
137139 result .model ,
138140 )
139141
@@ -169,7 +171,7 @@ def pre_hook(ctx: Dict[str, Any], *args: Any, **kwargs: Any) -> None:
169171 )
170172 ctx ["model_name" ] = kwargs .get ("model" , "" )
171173 set_span_attribute (
172- ctx ["span" ], AttributeKeys .GEN_AI_REQUEST_MODEL , ctx ["model_name" ]
174+ ctx ["span" ], AttributeKeys .JUDGMENT_LLM_MODEL_NAME , ctx ["model_name" ]
173175 )
174176 ctx ["accumulated_content" ] = ""
175177
@@ -197,17 +199,21 @@ def yield_hook(inner_ctx: Dict[str, Any], chunk: RawMessageStreamEvent) -> None:
197199 _extract_anthropic_tokens (usage_data )
198200 )
199201 set_span_attribute (
200- span , AttributeKeys .GEN_AI_USAGE_INPUT_TOKENS , prompt_tokens
202+ span ,
203+ AttributeKeys .JUDGMENT_USAGE_NON_CACHED_INPUT_TOKENS ,
204+ prompt_tokens ,
201205 )
202206 set_span_attribute (
203- span , AttributeKeys .GEN_AI_USAGE_OUTPUT_TOKENS , completion_tokens
207+ span , AttributeKeys .JUDGMENT_USAGE_OUTPUT_TOKENS , completion_tokens
204208 )
205209 set_span_attribute (
206- span , AttributeKeys .GEN_AI_USAGE_CACHE_READ_INPUT_TOKENS , cache_read
210+ span ,
211+ AttributeKeys .JUDGMENT_USAGE_CACHE_READ_INPUT_TOKENS ,
212+ cache_read ,
207213 )
208214 set_span_attribute (
209215 span ,
210- AttributeKeys .GEN_AI_USAGE_CACHE_CREATION_INPUT_TOKENS ,
216+ AttributeKeys .JUDGMENT_USAGE_CACHE_CREATION_INPUT_TOKENS ,
211217 cache_creation ,
212218 )
213219 set_span_attribute (
@@ -279,7 +285,7 @@ def pre_hook(ctx: Dict[str, Any], *args: Any, **kwargs: Any) -> None:
279285 )
280286 ctx ["model_name" ] = kwargs .get ("model" , "" )
281287 set_span_attribute (
282- ctx ["span" ], AttributeKeys .GEN_AI_REQUEST_MODEL , ctx ["model_name" ]
288+ ctx ["span" ], AttributeKeys .JUDGMENT_LLM_MODEL_NAME , ctx ["model_name" ]
283289 )
284290
285291 def post_hook (ctx : Dict [str , Any ], result : Message ) -> None :
@@ -296,17 +302,19 @@ def post_hook(ctx: Dict[str, Any], result: Message) -> None:
296302 _extract_anthropic_tokens (result .usage )
297303 )
298304 set_span_attribute (
299- span , AttributeKeys .GEN_AI_USAGE_INPUT_TOKENS , prompt_tokens
305+ span ,
306+ AttributeKeys .JUDGMENT_USAGE_NON_CACHED_INPUT_TOKENS ,
307+ prompt_tokens ,
300308 )
301309 set_span_attribute (
302- span , AttributeKeys .GEN_AI_USAGE_OUTPUT_TOKENS , completion_tokens
310+ span , AttributeKeys .JUDGMENT_USAGE_OUTPUT_TOKENS , completion_tokens
303311 )
304312 set_span_attribute (
305- span , AttributeKeys .GEN_AI_USAGE_CACHE_READ_INPUT_TOKENS , cache_read
313+ span , AttributeKeys .JUDGMENT_USAGE_CACHE_READ_INPUT_TOKENS , cache_read
306314 )
307315 set_span_attribute (
308316 span ,
309- AttributeKeys .GEN_AI_USAGE_CACHE_CREATION_INPUT_TOKENS ,
317+ AttributeKeys .JUDGMENT_USAGE_CACHE_CREATION_INPUT_TOKENS ,
310318 cache_creation ,
311319 )
312320 set_span_attribute (
@@ -317,7 +325,7 @@ def post_hook(ctx: Dict[str, Any], result: Message) -> None:
317325
318326 set_span_attribute (
319327 span ,
320- AttributeKeys .GEN_AI_RESPONSE_MODEL ,
328+ AttributeKeys .JUDGMENT_LLM_MODEL_NAME ,
321329 result .model ,
322330 )
323331
@@ -354,7 +362,7 @@ def pre_hook(ctx: Dict[str, Any], *args: Any, **kwargs: Any) -> None:
354362 )
355363 ctx ["model_name" ] = kwargs .get ("model" , "" )
356364 set_span_attribute (
357- ctx ["span" ], AttributeKeys .GEN_AI_REQUEST_MODEL , ctx ["model_name" ]
365+ ctx ["span" ], AttributeKeys .JUDGMENT_LLM_MODEL_NAME , ctx ["model_name" ]
358366 )
359367 ctx ["accumulated_content" ] = ""
360368
@@ -382,17 +390,21 @@ def yield_hook(inner_ctx: Dict[str, Any], chunk: RawMessageStreamEvent) -> None:
382390 _extract_anthropic_tokens (usage_data )
383391 )
384392 set_span_attribute (
385- span , AttributeKeys .GEN_AI_USAGE_INPUT_TOKENS , prompt_tokens
393+ span ,
394+ AttributeKeys .JUDGMENT_USAGE_NON_CACHED_INPUT_TOKENS ,
395+ prompt_tokens ,
386396 )
387397 set_span_attribute (
388- span , AttributeKeys .GEN_AI_USAGE_OUTPUT_TOKENS , completion_tokens
398+ span , AttributeKeys .JUDGMENT_USAGE_OUTPUT_TOKENS , completion_tokens
389399 )
390400 set_span_attribute (
391- span , AttributeKeys .GEN_AI_USAGE_CACHE_READ_INPUT_TOKENS , cache_read
401+ span ,
402+ AttributeKeys .JUDGMENT_USAGE_CACHE_READ_INPUT_TOKENS ,
403+ cache_read ,
392404 )
393405 set_span_attribute (
394406 span ,
395- AttributeKeys .GEN_AI_USAGE_CACHE_CREATION_INPUT_TOKENS ,
407+ AttributeKeys .JUDGMENT_USAGE_CACHE_CREATION_INPUT_TOKENS ,
396408 cache_creation ,
397409 )
398410 set_span_attribute (
0 commit comments