rogerxfeng8
commited on
Commit
•
bef256d
1
Parent(s):
e203eff
Update modeling_phi3_small.py
Browse files- modeling_phi3_small.py +0 -3
modeling_phi3_small.py
CHANGED
@@ -215,9 +215,6 @@ class Phi3SmallSelfAttention(nn.Module):
|
|
215 |
f"Layer {layer_idx + 1} is using dense attention since it is divisible by "
|
216 |
f"{self.config.dense_attention_every_n_layers}"
|
217 |
)
|
218 |
-
# use warnings to allow the modeling use different flash attention implementation later
|
219 |
-
if not is_flash_attention_available:
|
220 |
-
logger.warning_once("Flash Attention is not available, but is needed for dense attention")
|
221 |
else:
|
222 |
# BlockSparse related Parameters
|
223 |
self.blocksparse_params = BlockSparseParams.from_config(config)
|
|
|
215 |
f"Layer {layer_idx + 1} is using dense attention since it is divisible by "
|
216 |
f"{self.config.dense_attention_every_n_layers}"
|
217 |
)
|
|
|
|
|
|
|
218 |
else:
|
219 |
# BlockSparse related Parameters
|
220 |
self.blocksparse_params = BlockSparseParams.from_config(config)
|