make style
This commit is contained in:
parent
8e4733b3c3
commit
e29dc97215
|
@ -297,8 +297,8 @@ class AttentionBlock(nn.Module):
|
||||||
)
|
)
|
||||||
elif not torch.cuda.is_available():
|
elif not torch.cuda.is_available():
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
"torch.cuda.is_available() should be True but is False. xformers' memory efficient attention is only"
|
"torch.cuda.is_available() should be True but is False. xformers' memory efficient attention is"
|
||||||
" available for GPU "
|
" only available for GPU "
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
try:
|
try:
|
||||||
|
@ -461,8 +461,8 @@ class BasicTransformerBlock(nn.Module):
|
||||||
)
|
)
|
||||||
elif not torch.cuda.is_available():
|
elif not torch.cuda.is_available():
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
"torch.cuda.is_available() should be True but is False. xformers' memory efficient attention is only"
|
"torch.cuda.is_available() should be True but is False. xformers' memory efficient attention is"
|
||||||
" available for GPU "
|
" only available for GPU "
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
try:
|
try:
|
||||||
|
|
Loading…
Reference in New Issue