@@ -574,8 +574,6 @@ def _get_stats(self,
574574 num_waiting = len (self .scheduler .waiting )
575575
576576 # Iteration stats if we have scheduler output.
577- num_prompt_tokens = 0
578- num_generation_tokens = 0
579577 num_prompt_tokens_lst = []
580578 num_generation_tokens_lst = []
581579 request_n = []
@@ -594,13 +592,7 @@ def _get_stats(self,
594592 for scheduled_seq_group in
595593 scheduler_outputs .scheduled_seq_groups
596594 ]
597- num_prompt_tokens = sum (num_prompt_tokens_lst )
598- num_generation_tokens = sum (
599- scheduled_seq_group .seq_group .num_seqs ()
600- for scheduled_seq_group in
601- scheduler_outputs .scheduled_seq_groups )
602595 else :
603- num_generation_tokens = scheduler_outputs .num_batched_tokens
604596 num_generation_tokens_lst = [
605597 seq .get_output_len () for scheduled_seq_group in
606598 scheduler_outputs .scheduled_seq_groups for seq in
@@ -651,8 +643,6 @@ def _get_stats(self,
651643 gpu_cache_usage = gpu_cache_usage ,
652644 cpu_cache_usage = cpu_cache_usage ,
653645 finished_reason_lst = finished_reason_lst ,
654- num_prompt_tokens = num_prompt_tokens ,
655- num_generation_tokens = num_generation_tokens ,
656646 num_prompt_tokens_lst = num_prompt_tokens_lst ,
657647 num_generation_tokens_lst = num_generation_tokens_lst ,
658648 request_n = request_n ,
0 commit comments