Setting the attention_mask
and past_key_values
simultaneously will cause the error.
#34835
Closed
4 tasks done
Labels
System Info
transformers version: 4.46.3
python version: 3.8.0
System: Ubuntu 20.04
Who can help?
@ArthurZucker @stevhliu @zucchini-nlp
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
The above code can reproduce my issue. I set the
attention_mask
andpast_key_values
simultaneously. Although removing one of them can skip the bug, I need to set theattention_mask
andpast_key_values
due to the research goal. So what should I do? I have read themodeling_llama.py
carefully, maybe the code in this line is crucial.Expected behavior
I want a feasible solution :)
The text was updated successfully, but these errors were encountered: