-
Notifications
You must be signed in to change notification settings - Fork 99
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
What is the Expected Format for Training Dataset? #81
Comments
I continue my attempts to get to the truth and still need help. If you've faced the same questions and managed to find answers, please share your experience in this issue.finetune.sh → main.py → train/train.py → VLAConsumerDataset → HDF5VLADataset.parse_hdf5_file I think I've figured out how the fine-tuning process works at the file interaction level: inetune.sh: This script launches the finetuning process for the model. It sets up the necessary environment variables and calls
main.py: The main entry script, which processes the provided arguments and initiates the training process by calling the
train/train.py: Contains the
train/dataset.py: This file defines the
data/hdf5_vla_dataset.py: Contains the
Based on the last point, I currently assume that the dataset should preferably be in the exact format expected for further processing (as shown in the last code segment). At the moment, I am working on a script to transform |
I want to train the model to perform tasks using a dual-arm robot. I am working with my Kuavo robot from Leju. According to the guide, I need to collect my own dataset. Currently, I can successfully gather data in the
.bag
format, but I am unsure about the required data format for proper training (what format does the script expect?).Could you please advise if there are any datasets used by the developers for pre-training that I can download to understand the required structure? If they are publicly available, please share the link.
Alternatively, point me to this information in the
README.md
if I have missed it.Thanks ❤️
The text was updated successfully, but these errors were encountered: