diff --git a/eagle/model/modeling_qwen2_kv.py b/eagle/model/modeling_qwen2_kv.py index 9b8f69de..4f1f91d0 100644 --- a/eagle/model/modeling_qwen2_kv.py +++ b/eagle/model/modeling_qwen2_kv.py @@ -29,7 +29,7 @@ from transformers.activations import ACT2FN from transformers.cache_utils import Cache, DynamicCache, StaticCache -from transformers.generation import GenerationMixin +from transformers.generation.utils import GenerationMixin from transformers.modeling_attn_mask_utils import AttentionMaskConverter from transformers.modeling_outputs import ( BaseModelOutputWithPast, diff --git a/eagle/model/modeling_qwen3_kv.py b/eagle/model/modeling_qwen3_kv.py index 4a3aeca5..5a22c94c 100644 --- a/eagle/model/modeling_qwen3_kv.py +++ b/eagle/model/modeling_qwen3_kv.py @@ -26,7 +26,7 @@ from transformers.activations import ACT2FN from transformers.cache_utils import Cache, DynamicCache -from transformers.generation import GenerationMixin +from transformers.generation.utils import GenerationMixin from transformers.integrations import use_kernel_forward_from_hub from transformers.masking_utils import create_causal_mask, create_sliding_window_causal_mask from transformers.modeling_flash_attention_utils import FlashAttentionKwargs diff --git a/setup.py b/setup.py index f83163de..2d466c2a 100644 --- a/setup.py +++ b/setup.py @@ -9,18 +9,18 @@ author_email='yuhui.li@stu.pku.edu.cn', url='https://github.com/SafeAILab/EAGLE', packages=find_packages(), - install_requires=[ - 'torch==2.0.1', - 'transformers==4.46.2', - 'accelerate==0.21.0', - 'fschat==0.2.31', - 'gradio==3.50.2', - 'openai==0.28.0', - 'anthropic==0.5.0', - 'sentencepiece==0.1.99', - 'protobuf==3.19.0', - 'wandb' - ], + install_requires=[ + 'torch>=2.6.0', + 'transformers>=4.53.3,<5', + 'accelerate>=0.26.0', + 'fschat==0.2.31', + 'gradio==3.50.2', + 'openai==0.28.0', + 'anthropic==0.5.0', + 'sentencepiece==0.1.99', + 'protobuf==3.19.0', + 'wandb' + ], classifiers=[ 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers',