Update src/axolotl/monkeypatch/llama_attn_hijack_xformers.py
Browse files
src/axolotl/monkeypatch/llama_attn_hijack_xformers.py
CHANGED
|
@@ -25,7 +25,6 @@ def hijack_llama_sdp_attention():
|
|
| 25 |
transformers.models.llama.modeling_llama.LlamaAttention.forward = (
|
| 26 |
sdp_attention_forward
|
| 27 |
)
|
| 28 |
-
logging.info("Replaced attention with sdp_attention")
|
| 29 |
|
| 30 |
|
| 31 |
def xformers_forward(
|
|
|
|
| 25 |
transformers.models.llama.modeling_llama.LlamaAttention.forward = (
|
| 26 |
sdp_attention_forward
|
| 27 |
)
|
|
|
|
| 28 |
|
| 29 |
|
| 30 |
def xformers_forward(
|