You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thus, each time the len(tokenizer) is called, row __len__ is called rather than the CachedTokenizer's.
To fixed this problem of this issue, the get_cached_tokenizer adapter should be applied by the tokenizer setter of LLM.
The text was updated successfully, but these errors were encountered:
With the change of pr #5207, tested on qwen1.5-0.5B with the AdvertiseGen dataset, the generation decreased to 47 seconds, which cost less time than the solution of #5206 .
Your current environment
same to the issue #5206
🐛 Describe the bug
As the basic reason of the issue reported by #5206, the tokenizer setter of the LLM will override the cached tokenizer inited by llm_engine.
Thus, each time the
len(tokenizer)
is called, row__len__
is called rather than the CachedTokenizer's.To fixed this problem of this issue, the
get_cached_tokenizer
adapter should be applied by the tokenizer setter of LLM.The text was updated successfully, but these errors were encountered: