-
Notifications
You must be signed in to change notification settings - Fork 61
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How should we change the siam-mot code and configuration to train multiple class tracking? #40
Comments
I suggest you have a look at these two issues in which the very same question is being discussed: |
Hi, @mondrasovic |
Well, training this architecture with limited resources is pretty problematic. I can say this from my experience, and trust me, my experience is not negligible. I did a substantial portion of my experiments as part of my Ph.D. in deep learning using this architecture. So, not only did I receive additional validation from other people, but I also had to be more confident in my results since it wasn't just my hobby. Regarding the reproducibility aspect, it seems like it is impossible to achieve the very same results. Although you can get close, still no one I have discussed this issue with would consider it to be "within acceptable range". On top of all this, another researcher with his team from the Netherlands also discussed this topic with me and they couldn't reach the same performance level. He eventually became an opponent of my dissertation thesis, just a fun fact. As for the annotations. It is quite straightforward, I believe. Add annotations you want your model to use. If you want to train solely on This is what the configuration looks like in my case: DATASETS:
ROOT_DIR: "../../datasets"
TRAIN: ("MOT17",) Have a look here. I provide the important part of the code for clarity below. dataset_maps['MOT17'] = [
'MOT17',
'anno.json',
'splits.json',
'video'
] This tells you that the dataset key, specifically Later in the code, as demonstrated here, you can find that these values are used as follows: dataset_folder, anno_file, split_file, modality = dataset_maps[dataset_key] It speaks for itself. The value consists of a tuple with And as far as SOLVER:
VIDEO_CLIPS_PER_BATCH: 3 then the effective batch size is equal to Batch size is actually the culprit of having a hard time reliably reproducing the results. I experimented with gradient accumulation but to no avail, although it did help a little. So, if you are a mortal, you will probably have to stick to a single-digit number, and that is as good as you can get unless you try some powerful hardware. |
Hi @mondrasovic
|
My use case was aimed at vehicle tracking, not people, so it actually did not bother me that much that my results were slightly inferior. Furthermore, my experiments had to produce notable improvements in relative terms, not absolute. I actually worked with a completely different dataset as my main objective, but still, I played with MOT17 a lot, as you can imagine. I had to quantify the effect of my modifications on the underlying SiamMOT model in as many ways as possible.
|
Hi @mondrasovic |
hi how you are trying to improve this network? |
I have no clues. |
No description provided.
The text was updated successfully, but these errors were encountered: