Skip to content

Commit

Permalink
use python int lists for free block instead of torch tensor
Browse files Browse the repository at this point in the history
  • Loading branch information
tohtana committed Jan 16, 2024
1 parent 22895da commit bfdb5db
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions mii/batching/ragged_batching.py
Original file line number Diff line number Diff line change
Expand Up @@ -226,7 +226,7 @@ def _generate_output(self, r: Request) -> bool:
self.result_queues[r.tid].put_nowait(output)

def _schedule_token_gen(self, requests: List[Request]) -> None:
free_blocks = self.inference_engine.free_blocks.min().item()
free_blocks = min(self.inference_engine.free_blocks)
conf_manager = self.inference_engine._config.state_manager

num_schedulable = min(len(requests), conf_manager.max_ragged_sequence_count)
Expand All @@ -247,7 +247,7 @@ def _schedule_token_gen(self, requests: List[Request]) -> None:
self.scheduled_requests.append(r)

def _schedule_prompts(self, requests: List[Request]) -> None:
free_blocks = self.inference_engine.free_blocks.min().item()
free_blocks = min(self.inference_engine.free_blocks)
conf_manager = self.inference_engine._config.state_manager

for r in requests:
Expand Down

0 comments on commit bfdb5db

Please sign in to comment.