-
Notifications
You must be signed in to change notification settings - Fork 34
Natural gradient gaussian + shuffle in CV with hyperopt + skip problematic HP combindations #51
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
|
@claude Review the PR in detail |
|
Claude finished @StatMixedML's task —— View job PR Review - Branch Synchronization IssueI've attempted to review this PR but encountered a critical issue: the Expected Changes (from PR metadata):
Actual Branch State:The What Happened:This appears to be a branch synchronization issue. The changes likely exist in a local repository but haven't been pushed to GitHub's Next Steps:@StatMixedML To resolve this, please:
Once the branch is properly synchronized, I'll be able to review:
|
I think the problem with claude review is that there is a natural_gradient branch both in my fork and the original repo, so it reviews the latter one |
|
Claude finished @StatMixedML's task —— View job PR Review - Branch Access IssueI've identified the problem and understand why the previous review failed. However, I'm encountering a permissions limitation that prevents me from completing the review. The ProblemThis PR is from What I TriedI attempted to:
Current Branch StateThe branch I have access to (
SolutionsOption 1: GitHub Action Configuration (Recommended) - uses: actions/checkout@v4
with:
ref: ${{ github.event.pull_request.head.sha }}
repository: ${{ github.event.pull_request.head.repo.full_name }}However, I cannot modify workflow files due to permission restrictions. Option 2: Merge to Base Branch First # From ujohn33's local repository
git push https://github.com/StatMixedML/LightGBMLSS.git natural_gradient(This requires push access to StatMixedML/LightGBMLSS) Option 3: Manual Review
What Can Be Reviewed (if we fix the access)Once I have access to the correct branch, I can review:
Next Steps: Please choose one of the options above so I can proceed with the detailed review. Option 2 (pushing to the base repository's |
|
Hi @ujohn33 , this is a cool PR! It feels that it addresses more than one thing which makes it hard to review. Would it be possible to split it into three PRs?
Wrt unit tests, it would be good if your unit tests fit into the existing ones by Alexander, I think he has done a very good job at organizing them. If you are using an LLM: after the LLM proposes a new unit test file, point it to the relevant actual unit test file and ask it to adjust its unit test to fit into the file (and, if possible match the already existing unit tests—e.g., by adding assertions, not by changing their logic). Wrt (3) — I believe that no shuffling or stratification is actually done despite what is said in the docstring: LightGBMLSS/lightgbmlss/model.py Lines 223 to 226 in 0f9db92
LightGBMLSS/lightgbmlss/model.py Lines 275 to 276 in 0f9db92
I'm not sure if this is intended or not, or what is its significance or repercussion. I'm only pointing it out. Probably it's worth an issue and a treatment of its own (change of the docstring, or of how the variables are set). I hope this helps. Thanks again for your work, and I'm looking forward to seeing it merged into the package soon! |
Here I implemented the analytical computation of FIM for Gaussian DGBM natural gradient learning, FIM calculation for other distributions can be implemented in the same manner based on the analytical closed-form formulation of FIM (see https://arxiv.org/abs/2304.14885)
The hyperopt implementation was also updated with minor improvements, its now possible to set shuffle=False (important for time-series), and skip problematic hyperparameter combinations (like when min_data_in_leaf is too small) and continue optimization instead of crashing