We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 20bda42 commit 2d2325cCopy full SHA for 2d2325c
backends/python/server/text_embeddings_server/models/__init__.py
@@ -15,11 +15,11 @@
15
torch.set_grad_enabled(False)
16
17
FLASH_ATTENTION = True
18
-# try:
19
-from text_embeddings_server.models.flash_bert import FlashBert
20
-# except ImportError as e:
21
-# logger.warning(f"Could not import Flash Attention enabled models: {e}")
22
-# FLASH_ATTENTION = False
+try:
+ from text_embeddings_server.models.flash_bert import FlashBert
+except ImportError as e:
+ logger.warning(f"Could not import Flash Attention enabled models: {e}")
+ FLASH_ATTENTION = False
23
24
if FLASH_ATTENTION:
25
__all__.append(FlashBert)
0 commit comments