mirror of
https://github.com/deepseek-ai/DeepSeek-VL2.git
synced 2025-02-22 13:49:00 -05:00
optimize lib dependencies
This commit is contained in:
parent
32d92585ed
commit
cf21982aa5
@ -1,4 +1,3 @@
|
|||||||
from attrdict import AttrDict
|
|
||||||
from dataclasses import dataclass
|
from dataclasses import dataclass
|
||||||
import logging
|
import logging
|
||||||
import gc
|
import gc
|
||||||
|
@ -13,7 +13,6 @@ from timm.layers import (
|
|||||||
)
|
)
|
||||||
from timm.models._manipulate import named_apply, checkpoint_seq, adapt_input_conv
|
from timm.models._manipulate import named_apply, checkpoint_seq, adapt_input_conv
|
||||||
from transformers.modeling_utils import is_flash_attn_2_available
|
from transformers.modeling_utils import is_flash_attn_2_available
|
||||||
from xformers.ops import memory_efficient_attention
|
|
||||||
from functools import partial
|
from functools import partial
|
||||||
|
|
||||||
|
|
||||||
@ -134,6 +133,8 @@ class Attention(nn.Module):
|
|||||||
self.proj_drop = nn.Dropout(proj_drop) if proj_drop > 0. else nn.Identity()
|
self.proj_drop = nn.Dropout(proj_drop) if proj_drop > 0. else nn.Identity()
|
||||||
|
|
||||||
def forward(self, x: torch.Tensor) -> torch.Tensor:
|
def forward(self, x: torch.Tensor) -> torch.Tensor:
|
||||||
|
from xformers.ops import memory_efficient_attention
|
||||||
|
|
||||||
B, N, C = x.shape
|
B, N, C = x.shape
|
||||||
qkv = self.qkv(x).reshape(B, N, 3, self.num_heads, self.head_dim)
|
qkv = self.qkv(x).reshape(B, N, 3, self.num_heads, self.head_dim)
|
||||||
|
|
||||||
|
@ -4,7 +4,6 @@ xformers>=0.0.21
|
|||||||
timm>=0.9.16
|
timm>=0.9.16
|
||||||
accelerate
|
accelerate
|
||||||
sentencepiece
|
sentencepiece
|
||||||
attrdict
|
|
||||||
einops
|
einops
|
||||||
|
|
||||||
# for gradio demo
|
# for gradio demo
|
||||||
|
Loading…
Reference in New Issue
Block a user