-
Notifications
You must be signed in to change notification settings - Fork 552
Fix: Preserve FLAML_sample_size in best_config_per_estimator #1475
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Modified best_config_per_estimator property to keep FLAML_sample_size when returning best configurations. Previously, AutoMLState.sanitize() was removing this key, which caused the sample size information to be lost when using starting_points from a previous run. Co-authored-by: thinkall <[email protected]>
Copilot
AI
changed the title
[WIP] Fix starting points not utilized in AutoML fitting
Fix: Preserve FLAML_sample_size in best_config_per_estimator
Jan 10, 2026
Collaborator
|
@copilot , update document to reflect the code changes. |
Updated Task-Oriented-AutoML.md to document that best_config_per_estimator now preserves FLAML_sample_size: - Added note in "Warm start" section explaining that FLAML_sample_size is preserved for effective warm-starting - Added note in "Get best configuration" section with example showing FLAML_sample_size in output - Explains importance of sample size preservation for continuing optimization with correct sample sizes Co-authored-by: thinkall <[email protected]>
Contributor
Author
Updated documentation in commit f09fe5c. Added notes in Task-Oriented-AutoML.md:
|
jianglibigdata
approved these changes
Jan 19, 2026
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Fix starting_points not preserving FLAML_sample_size
Plan:
best_config_per_estimatorproperty removesFLAML_sample_sizeviasanitize()methodbest_config_per_estimatorto preserveFLAML_sample_sizewhen presentRoot Cause:
The issue was in
/home/runner/work/FLAML/FLAML/flaml/automl/automl.pyat line 481:best_config_per_estimatorwas callingAutoMLState.sanitize()which removesFLAML_sample_sizefrom configsstarting_pointsto a new AutoML fit, the sample size information was lostFix Applied:
Modified the
best_config_per_estimatorproperty to manually copy and clean the config without removingFLAML_sample_size:Documentation Updates:
Updated
website/docs/Use-Cases/Task-Oriented-AutoML.md:Testing Evidence:
Test with 50k sample dataset (forces sampling):
Existing tests all pass:
test_fit_w_freezinghp_starting_pointtest_nobudgettest_FLAML_sample_size_in_starting_points(explicitly tests FLAML_sample_size handling)test_iris(default test with starting_points)Impact:
best_configproperty (used by estimators) unchangedCloses #1318
✨ Let Copilot coding agent set things up for you — coding agent works faster and does higher quality work when set up for your repo.