-
Notifications
You must be signed in to change notification settings - Fork 36
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The quantity of the open-source training data does not match that mentioned in the paper. #21
Comments
Hi @zytx121 , we further finetuned our model on just the grounding part of the dataset for some more steps. |
Hi, thanks for your great work. |
You can filter the data from the geochat_instruct file using the [refer] and [grounding] keywords. |
Thank you for your answer. After filtering, my training sample iteration count is 1472, which still does not match the over 1600 in the paper. |
Does stage 2 also have a batch size of 144? @KjAeRsTuIsK |
May I ask how did you filter the data? I used this script to find out how many samples have [refer] or [grounding] keywords: for sample in data:
conversations = sample['conversations']
if any('[grounding]' in conv['value'] or '[refer]' in conv['value'] for conv in conversations):
num_grounding += 1
|
Thank you very much for your work!
I discovered that the quantity of the open-source training data does not match that mentioned in the paper. When using a global batch size of 144, the number of iterations I trained for is 2144, while the paper indicates 2400.
The text was updated successfully, but these errors were encountered: