You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We did not use monkey patch for qwen. we directly load Qwen2ForCausalLM_RingAttn in easy_context/modeling_qwen2.py
No, this monkey patch does not have effect for Qwen2ForCausalLM_RingAttn
Then how is zigzag_ring_attn performed during benchmarking for llava_qwen based models?
The only place we need zigzag_ring_attn is during text training and evaluation on V-NIAH. We do not use zigzag_ring_attn for other benchmarks in LMMs-Eval, or image-text training.
Hello, thanks for your great work. I have some little questions.
When testing a Qwen2 based model, like
llava_qwen
orlmms-lab/LongVA-7B
, on V-NIAH benchmark,there is a function apply_seq_parallel_monkey_patch("zigzag_ring_attn", "llama").
The text was updated successfully, but these errors were encountered: