Improved logging to support lm-eval>=0.4.8 #154
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Current code base is based on
lm-eval<=0.4.7.lm-evaladopts the best practice of logging in0.4.8(EleutherAI/lm-evaluation-harness#2203), making it incompatible with the current version of evalchemy. Minimal changes needs to be made to support newer versions (and to adopt better logging practices). I have no problem running this withlm-eval==0.4.9.2andvllm==0.13.0with this patch; also, these changes are backward-compatible.lm-eval 0.4.8can be particularly helpful. It starts to supportvllm 0.7+(EleutherAI/lm-evaluation-harness#2706), with easier local data parallel setup. I've also included a section on how to run data-parallel with vllm for faster evaluation.This is similar to #124, but I think it's better to decouple the logging so that people can know where the message is coming from.
I also encourage the authors set up version control and publish this on pypi; it would make this wonderful tool be more easier to use by everyone!