-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Questions on FELT Dataset and Baseline Trackers in Your Research #6
Comments
@MR-Vico Hi, thanks for your attention to our work. |
Thank you very much for your prompt and helpful response. I am especially grateful for providing the updated link to the evaluation tool on Dropbox, which is extremely valuable for advancing my work. I also appreciate the additional information regarding the training setup and default configurations used for the baseline trackers. As for the sequences with an initial bounding box of (0, 0, 0, 0), here is a list of the specific videos where this was observed: In /test – 10 sequences:
In /train1 – 1 sequence:
In /train2 – 10 sequences:
When analyzing the code of the evaluation tool, I noticed that frames without targets are filtered out, which actually solves the problem of invalid initial bounding boxes. This insight was incredibly helpful. Thank you once again for your time and assistance. Your support has significantly advanced my research efforts. |
@MR-Vico It's cool. The training samples are randomly sampled from the given training videos, thus, the bounding box of (0, 0, 0, 0) doesn't influence the training. For tracking, only a non-empty rectangle will be used as the starting position (more details can be found in our source code). The evaluation will filter out all zero annotations. |
When analyzing the evaluation tool further, I noticed that the actual tracking does not seem to take place within the evaluation tool itself. It is purely an evaluation of the tracking results. I would like to ask about the specific tracking methods. In particular, I would like to know whether a specific tracking tool was used or whether the trackers were applied directly to the FELT dataset. Since the evaluation tool stores the results in the form of bounding boxes in .txt files, I assume that the trackers were applied directly to the dataset without the use of a tracking tool. Please correct me if I am wrong here. How were initial bounding boxes with the coordinates (0, 0, 0, 0) handled during tracking? Do I understand correctly that the tracker starts tracking with the first image in which the object is present? The following groundtruth.txt for the sequence dvSave-2022_10_27_18_58_42 illustrates my question: Groundtruth:
Results for the DiMP tracker on this sequence:
The timestamps for the results are listed as follows:
In order to reproduce the tracking results, I would need this additional information about how the trackers were applied. Especially for the DiMP tracker. |
@MR-Vico You can take a closer look at the code under this path: AMTTrack/lib/test/evaluation/tracker. py. We first find the first target box that is not (0,0,0,0) based on the gt and use it as a template,and directly replace the initial frames where the target disappears with (0,0,0,0). This has nothing to do with the evaluation toolkit. |
I would like to thank you for your earlier answers. I really appreciate your time and effort in answering my questions. I’m reaching out once more as I find myself needing a bit more clarification to fully understand the process, particularly with respect to re-training the baseline trackers. I apologize for having to ask again, but this is vital for replicating the results in your study. I hope you won’t mind clarifying a few more points. From your initial response, I understand that specific details about the re-training process are not provided, and you mentioned following the default settings to train the tracker. Based on this, I attempted to reproduce results using the DiMP tracker as an example. However, I encountered a few issues that I hope you can help clarify: Which repository was used for the tracker?
Was the FELT dataset actually used to train the baseline trackers?
Thank you again for your time and for sharing your excellent work. I greatly appreciate any further details you can provide. Please let me know if additional context or clarification is needed on my end. |
@MR-Vico The pytracking is a collection of trackers, not a GitHub for a single tracker. We merge the two modalities into one representation and re-training on the FELT SOT dataset. For the details of how to adapt the pytracking for new datasets, please refer to their tutorial. If you want to understand the tracking processing clearly, you may need to read their source code first and then run the basic rgb-based datasets (e.g., LaSOT or GOT-10k). |
@MR-Vico Our tracker is improved based on OSTrack and CEUTrack. For a fair comparison with other baselines, we retrain them based on their own configs using APS, DVS, and APS+DVS, respectively. For example, the config of DiMP is set as dimp50.py. Simply put, we only replaced the input with our own dataset while keeping other parameters unchanged. |
Dear FELT research team,
First, I would like to express my sincere gratitude for your outstanding work on the FELT dataset and your paper, which has become an invaluable reference for my current research. I am currently writing a paper that builds upon your methods, and I am very impressed by the comprehensive approach your team has taken to address the challenges of long-term frame-event tracking.
However, I have a few questions regarding the re-training of the baseline trackers on the FELT dataset. The paper mentions that 15 baseline trackers were re-trained, but I couldn’t find specific details about the re-training process. Could you clarify the methodology you used? Any additional information about the training configuration or settings would be very helpful for my research.
Additionally, I noticed that the evaluation tool for FELT is hosted on Baidu, which unfortunately I cannot access. Would it be possible to provide access to the tool through another platform or method?
Finally, I encountered sequences in the FELT dataset with bounding boxes initialized at (0, 0, 0, 0). Could you explain how the trackers handle such initializations and how these cases are evaluated within your framework, given my limited access to the evaluation tool?
Thank you very much for your time and assistance. I greatly appreciate any insights you could provide, and I look forward to your response.
The text was updated successfully, but these errors were encountered: