From 3f2f11a7029735f4de9b23cb74381a73df67870d Mon Sep 17 00:00:00 2001 From: Daniel King Date: Mon, 9 Oct 2023 18:32:49 -0700 Subject: [PATCH] precommit --- llmfoundry/models/layers/attention.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/llmfoundry/models/layers/attention.py b/llmfoundry/models/layers/attention.py index 00cb36d215..39fa7162ac 100644 --- a/llmfoundry/models/layers/attention.py +++ b/llmfoundry/models/layers/attention.py @@ -213,7 +213,8 @@ def flash_attn_fn( try: from flash_attn import bert_padding, flash_attn_interface # type: ignore # yapf: disable # isort: skip except: - raise RuntimeError('Please install flash-attn==1.0.9 or flash-attn==2.3.2') + raise RuntimeError( + 'Please install flash-attn==1.0.9 or flash-attn==2.3.2') check_valid_inputs(query, key, value)