Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Less flaky for TimmBackboneModelTest::test_batching_equivalence #35971

Merged
merged 3 commits into from
Jan 30, 2025

Conversation

ydshieh
Copy link
Collaborator

@ydshieh ydshieh commented Jan 30, 2025

What does this PR do?

As in the (previous) comment, Timm model in transformers has no _init_weights method, and it relies the (some how different) timm's way of weight init.

This seems giving somehow larger values and differences. Decorated is_flaky and use 1e-5 still fails with 1% ratio.

Let's just use a 1e-4

@ydshieh ydshieh changed the title Less flaky for ``` Less flaky for TimmBackboneModelTest::test_batching_equivalence Jan 30, 2025
@ydshieh ydshieh requested a review from zucchini-nlp January 30, 2025 14:44
@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

Copy link
Member

@zucchini-nlp zucchini-nlp left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks! I assume the pass rate is 100% now with a larger tolerance?

@ydshieh
Copy link
Collaborator Author

ydshieh commented Jan 30, 2025

0% In the sense of running 2000 times.

(but when it comes to 1 millions times, I won't promise it's 0% 😆 )

@ydshieh ydshieh merged commit 5757681 into main Jan 30, 2025
26 checks passed
@ydshieh ydshieh deleted the circleci_debug_base_timm_2 branch January 30, 2025 15:56
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants