74 Star 572 Fork 1.1K

Ascend/pytorch

 / 详情

【BUG】module 'torch_npu' has no attribute '_npu_rotary_embedding'

DONE
需求
创建于  
2025-03-24 13:40

一、需求场景&价值

版本信息:
torch: 2.5.1
torch-npu: 2.5.1rc1
cann: 8.0.RC2.2
vllm-ascend: 0.7.3 rc1
vlllm: 0.7.3

报错如下:

[rank1]: Traceback (most recent call last):
[rank1]:   File "/home/hwtest/hzy/swift-grpo-npu/swift/cli/rlhf.py", line 5, in <module>
[rank1]:     rlhf_main()
[rank1]:   File "/home/hwtest/hzy/swift-grpo-npu/swift/llm/train/rlhf.py", line 96, in rlhf_main
[rank1]:     return SwiftRLHF(args).main()
[rank1]:   File "/home/hwtest/hzy/swift-grpo-npu/swift/llm/base.py", line 47, in main
[rank1]:     result = self.run()
[rank1]:   File "/home/hwtest/hzy/swift-grpo-npu/swift/llm/train/sft.py", line 137, in run
[rank1]:     trainer = trainer_cls(
[rank1]:   File "/home/hwtest/hzy/swift-grpo-npu/swift/trainers/rlhf_trainer/grpo_trainer.py", line 226, in __init__
[rank1]:     self.prepare_vllm(model, fast_infer_device)
[rank1]:   File "/home/hwtest/hzy/swift-grpo-npu/swift/trainers/rlhf_trainer/grpo_trainer.py", line 375, in prepare_vllm
[rank1]:     self.engine = cls(
[rank1]:   File "/home/hwtest/hzy/swift-grpo-npu/swift/llm/infer/infer_engine/grpo_vllm_engine.py", line 88, in __init__
[rank1]:     self._prepare_engine()
[rank1]:   File "/home/hwtest/hzy/swift-grpo-npu/swift/llm/infer/infer_engine/grpo_vllm_engine.py", line 93, in _prepare_engine
[rank1]:     engine = LLM(**self.engine_args.__dict__)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/utils.py", line 1022, in inner
[rank1]:     return fn(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/entrypoints/llm.py", line 242, in __init__
[rank1]:     self.llm_engine = self.engine_class.from_engine_args(
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/engine/llm_engine.py", line 489, in from_engine_args
[rank1]:     engine = cls(
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/engine/llm_engine.py", line 276, in __init__
[rank1]:     self._initialize_kv_caches()
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/engine/llm_engine.py", line 421, in _initialize_kv_caches
[rank1]:     self.model_executor.determine_num_available_blocks())
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/executor/uniproc_executor.py", line 132, in determine_num_available_blocks
[rank1]:     a, b = super().determine_num_available_blocks()
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/executor/executor_base.py", line 102, in determine_num_available_blocks
[rank1]:     results = self.collective_rpc("determine_num_available_blocks")
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/executor/uniproc_executor.py", line 56, in collective_rpc
[rank1]:     answer = run_method(self.driver_worker, method, args, kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/utils.py", line 2196, in run_method
[rank1]:     return func(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context
[rank1]:     return func(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm_ascend/worker/worker.py", line 227, in determine_num_available_blocks
[rank1]:     self.model_runner.profile_run()
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context
[rank1]:     return func(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm_ascend/worker/model_runner.py", line 1360, in profile_run
[rank1]:     self.execute_model(model_input, kv_caches, intermediate_tensors)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context
[rank1]:     return func(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm_ascend/worker/model_runner.py", line 1140, in execute_model
[rank1]:     hidden_or_intermediate_states = model_executable(
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1736, in _wrapped_call_impl
[rank1]:     return self._call_impl(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1747, in _call_impl
[rank1]:     return forward_call(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/model_executor/models/qwen2.py", line 486, in forward
[rank1]:     hidden_states = self.model(input_ids, positions, kv_caches,
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/compilation/decorators.py", line 172, in __call__
[rank1]:     return self.forward(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/model_executor/models/qwen2.py", line 348, in forward
[rank1]:     hidden_states, residual = layer(
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1736, in _wrapped_call_impl
[rank1]:     return self._call_impl(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1747, in _call_impl
[rank1]:     return forward_call(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/model_executor/models/qwen2.py", line 247, in forward
[rank1]:     hidden_states = self.self_attn(
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1736, in _wrapped_call_impl
[rank1]:     return self._call_impl(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1747, in _call_impl
[rank1]:     return forward_call(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/model_executor/models/qwen2.py", line 178, in forward
[rank1]:     q, k = self.rotary_emb(positions, q, k)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1736, in _wrapped_call_impl
[rank1]:     return self._call_impl(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1747, in _call_impl
[rank1]:     return forward_call(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm/model_executor/custom_op.py", line 25, in forward
[rank1]:     return self._forward_method(*args, **kwargs)
[rank1]:   File "/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/site-packages/vllm_ascend/ops/rotary_embedding.py", line 45, in rope_forward_oot
[rank1]:     torch_npu._npu_rotary_embedding(
[rank1]: AttributeError: module 'torch_npu' has no attribute '_npu_rotary_embedding'
[ERROR] 2025-03-24-13:34:28 (PID:553687, Device:2, RankID:-1) ERR99999 UNKNOWN applicaiton exception
/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/tempfile.py:869: ResourceWarning: Implicitly cleaning up <TemporaryDirectory '/root/.cache/modelscope/hub/tmp/hf_datasets-_l2pfbk8'>
  _warnings.warn(warn_message, ResourceWarning)
/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/tempfile.py:869: ResourceWarning: Implicitly cleaning up <TemporaryDirectory '/root/.cache/modelscope/hub/offload_cache/tmp1r8fi0zv'>
  _warnings.warn(warn_message, ResourceWarning)
/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/tempfile.py:869: ResourceWarning: Implicitly cleaning up <TemporaryDirectory '/tmp/tmp9cop1ilc'>
  _warnings.warn(warn_message, ResourceWarning)
[ERROR] 2025-03-24-13:34:29 (PID:553686, Device:1, RankID:-1) ERR99999 UNKNOWN applicaiton exception
/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/tempfile.py:869: ResourceWarning: Implicitly cleaning up <TemporaryDirectory '/root/.cache/modelscope/hub/tmp/hf_datasets-b8nnvzw8'>
  _warnings.warn(warn_message, ResourceWarning)
/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/tempfile.py:869: ResourceWarning: Implicitly cleaning up <TemporaryDirectory '/root/.cache/modelscope/hub/offload_cache/tmpqyned81w'>
  _warnings.warn(warn_message, ResourceWarning)
/root/miniconda3/envs/hzy_dev_vllm/lib/python3.10/tempfile.py:869: ResourceWarning: Implicitly cleaning up <TemporaryDirectory '/tmp/tmpm5z4x890'>
  _warnings.warn(warn_message, ResourceWarning)

如何解决该异常报错?是否torch_npu缺少算子?

二、需求建议实现的规格

三、竞品比较(选填)

评论 (6)

阿郎小哥 创建了需求 1个月前
阿郎小哥 修改了标题 1个月前
阿郎小哥 修改了描述 1个月前
展开全部操作日志

请使用vllm_ascend支持的版本


是按这个装的,难道我的cann版本比8.0.0低嘛?

这里torch_npu >= 2.5.1.dev20250308

我尝试了这个版本,一样的报错。
而且我安装的torch-npu 2.5.1 rc1要比这个版本高

如果使用他们推荐的版本还有问题,请咨询下他们版本适配情况
https://gitee.com/ascend/pytorch/releases/tag/v6.0.0.alpha001-pytorch2.5.1
torch-npu 2.5.1 rc1 在2024年就有了,不会高于 2.5.1.dev20250308

哦哦好的,明白了,感谢

huangyunlong 任务状态TODO 修改为WIP 1个月前
huangyunlong 任务状态WIP 修改为DONE 1个月前

登录 后才可以发表评论

状态
负责人
项目
里程碑
Pull Requests
关联的 Pull Requests 被合并后可能会关闭此 issue
分支
优先级
预计工期 (小时)
开始日期   -   截止日期
-
置顶选项
参与者(2)
huangyunlong-huangyunlong2022 855604 zhiyong git 1743153795
Python
1
https://gitee.com/ascend/pytorch.git
git@gitee.com:ascend/pytorch.git
ascend
pytorch
pytorch

搜索帮助