File tree Expand file tree Collapse file tree 1 file changed +13
-5
lines changed
Filter options
Expand file tree Collapse file tree 1 file changed +13
-5
lines changed
Original file line number Diff line number Diff line change @@ -1551,7 +1551,9 @@ def logit_bias_processor(
1551
1551
"utf-8" , errors = "ignore"
1552
1552
)
1553
1553
text_offset = len (prompt ) + len (
1554
- self .detokenize (completion_tokens [:returned_tokens ])
1554
+ self .detokenize (completion_tokens [:returned_tokens ]).decode (
1555
+ "utf-8" , errors = "ignore"
1556
+ )
1555
1557
)
1556
1558
token_offset = len (prompt_tokens ) + returned_tokens
1557
1559
logits = self ._scores [token_offset - 1 , :]
@@ -1789,13 +1791,19 @@ def logit_bias_processor(
1789
1791
]
1790
1792
all_logprobs = Llama .logits_to_logprobs (self ._scores )[token_offset :]
1791
1793
# TODO: may be able to change this loop to use np.take_along_dim
1792
- for token , token_str , logprobs_token in zip (
1793
- all_tokens , all_token_strs , all_logprobs
1794
+ for idx , ( token , token_str , logprobs_token ) in enumerate (
1795
+ zip ( all_tokens , all_token_strs , all_logprobs )
1794
1796
):
1795
1797
if token == self .token_bos ():
1796
1798
continue
1797
- text_offsets .append (text_offset )
1798
- text_offset += len (token_str )
1799
+ text_offsets .append (
1800
+ text_offset
1801
+ + len (
1802
+ self .detokenize (all_tokens [:idx ]).decode (
1803
+ "utf-8" , errors = "ignore"
1804
+ )
1805
+ )
1806
+ )
1799
1807
tokens .append (token_str )
1800
1808
sorted_logprobs = list (
1801
1809
sorted (
You can’t perform that action at this time.
0 commit comments