-
Notifications
You must be signed in to change notification settings - Fork 810
Fix missing import in Gemma2-2B model #16684
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
It didn't add the necessary config and conversion script when landing the PR
As a result, our CI is timing out due to
```
# test_qnn_delegate.py:6430-6433
p = subprocess.Popen(cmds, stdout=subprocess.DEVNULL)
with Listener((self.ip, self.port)) as listener:
conn = listener.accept() # ← BLOCKS forever waiting for connection
p.communicate()
```
https://hud.pytorch.org/hud/pytorch/executorch/main/1?per_page=50&name_filter=test-static-llama-qnn-linux&mergeEphemeralLF=true
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/16684
Note: Links to docs will display an error until the docs builds have been completed. ⏳ No Failures, 19 PendingAs of commit a5fc1c8 with merge base af69d08 ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
|
Follow-up to #16624 cc @jethroqti |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Pull request overview
This PR adds missing files for the Gemma2-2B model that were not included in the original PR. These files are essential for model weight conversion and configuration, and their absence was causing CI timeouts during QNN delegate testing.
Changes:
- Added weight conversion script with Gemma2-specific weight mappings for transformer parameter conversion
- Added 2B model configuration with Gemma2-specific settings (softcapping, sliding window, layer types)
- Added module initialization with Gemma2Model class definition and exported utilities
Reviewed changes
Copilot reviewed 3 out of 3 changed files in this pull request and generated no comments.
| File | Description |
|---|---|
| examples/models/gemma2/convert_weights.py | Implements weight conversion from HuggingFace checkpoints to ExecuTorch format with Gemma2-specific layer normalizations |
| examples/models/gemma2/config/2b_config.json | Defines Gemma2-2B model architecture parameters including attention/logit softcapping and sliding window configuration |
| examples/models/gemma2/init.py | Exports Gemma2Model class and convert_weights utility for external use |
💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Pull request overview
Copilot reviewed 3 out of 3 changed files in this pull request and generated no new comments.
💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.
|
@pytorchbot label "release notes: qualcomm" |
|
Added Gemma2 data. |
Follow-up to #16624 where it didn't
git addthe necessary folderexamples/models/gemma2. As a result, it didn't have the necessary config and conversion script when landing the PROur CI is timing out due to
https://hud.pytorch.org/hud/pytorch/executorch/main/1?per_page=50&name_filter=test-static-llama-qnn-linux&mergeEphemeralLF=true