-
Notifications
You must be signed in to change notification settings - Fork 93
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
support external models: vllm huggingface openai #97
Conversation
This prompt is used for SFT in Internlm-math-plus models. |
fixed |
Other model APIs e.g. Claude, Gemini, and Mistral will soon be tested and implemeted. |
For now, Gemini and Claude APIs only support one sequence output, so the |
Hi @hsz0403 , thanks again for the efforts! I am currently reviewing your code in order to merge and publish in the upcoming release. I was wondering if your code changes in any way the previous instructions to run external models, that is, e.g. run Do your additional code require different steps to run, or would the original steps still work? |
Hi, My previous changes include the file |
Yes, the CI run itself is a bit buggy right now. For this PR we don't need the second CI check to pass. |
Thanks for confirming. Since there are no additional steps, I suppose running the code in this PR is as simple as going to E.g. in the first step, when trying your code by running
which I suppose can be fixed by changing the import statements in your |
Sorry for this mistake, each model should work well now in codespace |
Thanks! Can you update the requirement in |
The code runs well, thank you! I will now go a pass reviewing it for code cleanness. Meanwhile I notice that there are unused package imports, could you remove them and only keep the necessary ones, as you go through the requirements? |
Hi, the unused package imports is because models not declared in |
Sounds great, thanks! |
By the way, is the requirement list in |
I will have this PR merged in the next official release. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I made some minor comments in the code too. Thank you for the nice work!
A meta comment is that let us try to not leave commented-out code in there. If they are necessary we should make them options for users to pick. If they are not we should just delete the dead code. Also let's make sure the requirements are comprehensive -- more in the specific comments.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Let's double check that all comments above are addressed. You can "resolve" a comment after you address it.
@hsz0403 Let me know when it's ready for final review. I will merge it and publish with the upcoming new release. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Changes look great, merging soon. Thanks!
#97) * support external models: vllm huggingface openai * fix prompt for internlm * support gemini claude * fix_bugs * fix_model_bugs * update_requirement * add requirements in README * fixed by suggestions * removed comments
Hi, I tested current sota llms for lean code and made a simple prompt to output tactics via vllm, huggingface and openai.