Skip to content

KeyError: 'sdpa' #105

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wangjin0811 opened this issue Mar 11, 2025 · 4 comments
Open

KeyError: 'sdpa' #105

wangjin0811 opened this issue Mar 11, 2025 · 4 comments

Comments

@wangjin0811
Copy link

Traceback (most recent call last):
File "/data/omniparse/server.py", line 88, in
main()
File "/data/omniparse/server.py", line 53, in main
load_omnimodel(args.documents, args.media, args.web)
File "/data/omniparse/omniparse/init.py", line 48, in load_omnimodel
shared_state.model_list = load_all_models()
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/marker/models.py", line 74, in load_all_models
order = setup_order_model(device, dtype)
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/marker/models.py", line 61, in setup_order_model
model = load_order_model()
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/surya/model/ordering/model.py", line 27, in load_model
model = OrderVisionEncoderDecoderModel.from_pretrained(checkpoint, config=config, torch_dtype=dtype)
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/transformers/models/vision_encoder_decoder/modeling_vision_encoder_decoder.py", line 379, in from_pretrained
return super().from_pretrained(pretrained_model_name_or_path, *model_args, **kwargs)
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/transformers/modeling_utils.py", line 262, in _wrapper
return func(*args, **kwargs)
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/transformers/modeling_utils.py", line 4185, in from_pretrained
model = cls(config, *model_args, **model_kwargs)
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/transformers/models/vision_encoder_decoder/modeling_vision_encoder_decoder.py", line 199, in init
decoder = AutoModelForCausalLM.from_config(config.decoder)
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/transformers/models/auto/auto_factory.py", line 440, in from_config
return model_class._from_config(config, **kwargs)
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/transformers/modeling_utils.py", line 262, in _wrapper
return func(*args, **kwargs)
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/transformers/modeling_utils.py", line 1501, in _from_config
model = cls(config, **kwargs)
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/surya/model/ordering/decoder.py", line 495, in init
self.model = MBartOrderDecoderWrapper(config)
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/surya/model/ordering/decoder.py", line 480, in init
self.decoder = MBartOrderDecoder(config)
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/surya/model/ordering/decoder.py", line 294, in init
self.layers = nn.ModuleList([MBartOrderDecoderLayer(config) for _ in range(config.decoder_layers)])
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/surya/model/ordering/decoder.py", line 294, in
self.layers = nn.ModuleList([MBartOrderDecoderLayer(config) for _ in range(config.decoder_layers)])
File "/data/anaconda3/envs/omniparse-venv/lib/python3.10/site-packages/surya/model/ordering/decoder.py", line 209, in init
self.self_attn = MBART_ATTENTION_CLASSES[config._attn_implementation](
KeyError: 'sdpa'

@760485464
Copy link

有解决吗

@CrushLSG
Copy link

我也是同样的问题,谁能解决一下呐,

@MockingBill
Copy link

我也是同样的问题

@Rogerskelamen
Copy link

add one line in model.py which is a surya package source file(some path like $HOME/miniconda3/envs/omniparse-venv/lib/python3.10/site-packages/surya/model/ordering/model.py):

+ config._attn_implementation = "eager" # or "flash_attention_2", it depends on the situation
model = OrderVisionEncoderDecoderModel.from_pretrained(checkpoint, config=config, torch_dtype=dtype)

It works for me.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants