Merged
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/9634
Note: Links to docs will display an error until the docs builds have been completed. ❌ 1 New FailureAs of commit ccf664e with merge base 4b8ac94 ( NEW FAILURE - The following job has failed:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
metascroy
approved these changes
Mar 26, 2025
10408cd to
ccf664e
Compare
GregoryComer
approved these changes
Mar 26, 2025
Contributor
Author
|
phi_4_mini ci tests are failing with: there is no error message with the run, so I don't think anything is failing but just that this tests is getting killed. Running it locally on my laptop seems to pass |
jackzhxng
reviewed
Mar 26, 2025
Contributor
jackzhxng
left a comment
There was a problem hiding this comment.
Seems fine, doesn't seem like it should have affected the phi test. If it's consistently getting killed it might be OOMing
pytorchmergebot
pushed a commit
to pytorch/pytorch
that referenced
this pull request
Apr 1, 2025
From latest viable/strict: https://hud.pytorch.org/hud/pytorch/executorch/viable%2Fstrict/1?per_page=50 Fixes #144480 This commit has important CI stability fixes, such as pytorch/executorch#9561 and pytorch/executorch#9634 Pull Request resolved: #150308 Approved by: https://github.com/jathu, https://github.com/malfet
kirklandsign
pushed a commit
that referenced
this pull request
Apr 11, 2025
…ntization for Example Models (#9634) ### Summary Changes: 1. When initializing Llama2 for aot_compiler, since checkpoints can only e downloaded from hugging face, we initialize llama2 with uninitialized weights. The problem with this is that when running quantization, we can run into errors with the histogram if the unitialized values are nan. We fix this by initializing the weights with zeros if no check point is provided. This enforces that quantization step can still work. 2. Quant Type in AoT compiler. When looking at the model options available to XNNPACK, everything is quantized with per-tensor static quantization. This isn't the best option for all the models available. For example transformer based models like Llama and MobileBert would likely prefer dynamically quantized per channel weights, where has CNN like MobileNet would prefer statically quantized per channel weights. We add this type of Quant Type to the existing models options. This also helps with Test Timeouts. per-tensor static quantization on a model like llama can take a long time due to the introduction of MANY q/dq nodes, and the complex partitions it creates. As a result, proposing partitions can take a long time due to the constant BFS to find the largest possible partition. By specifying the more apt quantization scheme like dynamic per-channel quantization, we can avoid this complexity. Overall this should help with flakey [nan, nan] errors in the quantization histogram, and it should also help with CI timing out. ### Test plan OSS XNNPACK CI for all model delegation cc @digantdesai @cbilgin
amathewc
pushed a commit
to amathewc/pytorch
that referenced
this pull request
Apr 17, 2025
From latest viable/strict: https://hud.pytorch.org/hud/pytorch/executorch/viable%2Fstrict/1?per_page=50 Fixes pytorch#144480 This commit has important CI stability fixes, such as pytorch/executorch#9561 and pytorch/executorch#9634 Pull Request resolved: pytorch#150308 Approved by: https://github.com/jathu, https://github.com/malfet
huydhn
pushed a commit
to pytorch/pytorch
that referenced
this pull request
May 16, 2025
From latest viable/strict: https://hud.pytorch.org/hud/pytorch/executorch/viable%2Fstrict/1?per_page=50 Fixes #144480 This commit has important CI stability fixes, such as pytorch/executorch#9561 and pytorch/executorch#9634 Pull Request resolved: #150308 Approved by: https://github.com/jathu, https://github.com/malfet
atalman
pushed a commit
to pytorch/pytorch
that referenced
this pull request
May 21, 2025
…53750) * Update ExecuTorch pin to latest viable/strict 3/28/2025 (#150308) From latest viable/strict: https://hud.pytorch.org/hud/pytorch/executorch/viable%2Fstrict/1?per_page=50 Fixes #144480 This commit has important CI stability fixes, such as pytorch/executorch#9561 and pytorch/executorch#9634 Pull Request resolved: #150308 Approved by: https://github.com/jathu, https://github.com/malfet * Use new hash from #150722 * Update executorch.txt --------- Co-authored-by: Mergen Nachin <mnachin@meta.com>
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Summary
Changes:
Overall this should help with flakey [nan, nan] errors in the quantization histogram, and it should also help with CI timing out.
Test plan
OSS XNNPACK CI for all model delegation
cc @digantdesai @cbilgin