@@ -235,7 +235,9 @@ impl Infer {
235235 let start_time = Instant :: now ( ) ;
236236
237237 if self . is_splade ( ) && normalize {
238- metrics:: counter!( "te_request_failure" , "err" => "model_type" ) . increment ( 1 ) ;
238+ let counter = metrics:: counter!( "te_request_failure" , "err" => "model_type" ) ;
239+ counter. increment ( 1 ) ;
240+
239241 let message = "`normalize` is not available for SPLADE models" . to_string ( ) ;
240242 tracing:: error!( "{message}" ) ;
241243 return Err ( TextEmbeddingsError :: Backend ( BackendError :: Inference (
@@ -294,14 +296,16 @@ impl Infer {
294296 let total_time = start_time. elapsed ( ) ;
295297
296298 // Metrics
297- metrics:: counter!( "te_embed_success" ) . increment ( 1 ) ;
298- metrics:: histogram!( "te_embed_duration" ) . record ( total_time. as_secs_f64 ( ) ) ;
299- metrics:: histogram!( "te_embed_tokenization_duration" )
300- . record ( response. metadata . tokenization . as_secs_f64 ( ) ) ;
301- metrics:: histogram!( "te_embed_queue_duration" )
302- . record ( response. metadata . queue . as_secs_f64 ( ) ) ;
303- metrics:: histogram!( "te_embed_inference_duration" )
304- . record ( response. metadata . inference . as_secs_f64 ( ) ) ;
299+ let counter = metrics:: counter!( "te_embed_success" ) ;
300+ counter. increment ( 1 ) ;
301+ let histogram = metrics:: histogram!( "te_embed_duration" ) ;
302+ histogram. record ( total_time. as_secs_f64 ( ) ) ;
303+ let histogram = metrics:: histogram!( "te_embed_tokenization_duration" ) ;
304+ histogram. record ( response. metadata . tokenization . as_secs_f64 ( ) ) ;
305+ let histogram = metrics:: histogram!( "te_embed_queue_duration" ) ;
306+ histogram. record ( response. metadata . queue . as_secs_f64 ( ) ) ;
307+ let histogram = metrics:: histogram!( "te_embed_inference_duration" ) ;
308+ histogram. record ( response. metadata . inference . as_secs_f64 ( ) ) ;
305309
306310 Ok ( response)
307311 }
0 commit comments