Skip to content

fix: mllama e2e pytorch flow fix #3397

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Apr 11, 2025

Conversation

yechank-nvidia
Copy link
Collaborator

@yechank-nvidia yechank-nvidia commented Apr 9, 2025

This PR is the bugfix for the e2e pytorch flow of mllama.

python3 quickstart_advanced.py --model_dir meta-llama/Llama-3.2-11B-Vision --enable_chunked_prefill --enable_overlap_scheduler
=>
AttributeError: 'MllamaForConditionalGeneration' object has no attribute 'infer_max_seq_len'

MllamaForCausalLM is not based on DecoderModelForCausalLM which causes AttributeError of infer_max_seq_len. Copied infer_max_seq_len from DecoderModelForCausalLM.

@yechank-nvidia
Copy link
Collaborator Author

/bot run

@tensorrt-cicd
Copy link
Collaborator

PR_Github #1553 [ run ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #1553 [ run ] completed with state SUCCESS
/LLM/main/L0_MergeRequest_PR pipeline #1160 completed with status: 'FAILURE'

@yechank-nvidia
Copy link
Collaborator Author

/bot run

@tensorrt-cicd
Copy link
Collaborator

PR_Github #1560 [ run ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #1560 [ run ] completed with state SUCCESS
/LLM/main/L0_MergeRequest_PR pipeline #1166 completed with status: 'SUCCESS'

@kaiyux kaiyux force-pushed the mllama_torch_seqlen branch from 9a3c14a to f8e810c Compare April 11, 2025 09:20
@kaiyux kaiyux enabled auto-merge (squash) April 11, 2025 09:20
@kaiyux
Copy link
Member

kaiyux commented Apr 11, 2025

/bot reuse-pipeline

@tensorrt-cicd
Copy link
Collaborator

PR_Github #1911 [ reuse-pipeline ] triggered by Bot

@tensorrt-cicd
Copy link
Collaborator

PR_Github #1911 [ reuse-pipeline ] completed with state SUCCESS
Reusing PR_Github #1560 for commit f8e810c

@kaiyux kaiyux merged commit 5bc6f09 into NVIDIA:main Apr 11, 2025
3 checks passed
@yechank-nvidia yechank-nvidia deleted the mllama_torch_seqlen branch April 14, 2025 07:43
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants