-
Notifications
You must be signed in to change notification settings - Fork 42
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
problems of optical flow results when finetuning on real scene? #19
Comments
Do you use ground truth flow label of your real scene to train FastFlowNet in a supervised manner or in an unsupervised way? For optical flow visualization, the scale factor will be normalized in current code, so there is no difference whether you multiply the optical flow result by div_flow(20), you can modify the code to meet you need. |
hi, |
I think taking FlowNet2's prediction as ground truth will lead to error accumulation, I suggest you try to adopt RAFT's prediction as ground truth label. For training FastFlowNet, you should normalize input images to [0, 1] and reduce the one channel mean value, also the flow ground truth should be divided by div_flow(20). Data augmentation like geometry and color augmentation should be adopted for better generalization. It is good to pretrain FastFlowNet in a self-supervised manner which does not have domain gap, I will release the training code if my under reviewed paper is published. |
Ok, thanks for your timely reply.
3)haven't tried self-supervised manner in real-scene dataset, I will try this. |
Could you please exchange the details of training FastFlowNet?
The problem we are facing now is that we used CHAIRS for training and MPI-SIntel as the verification set, and found that EPE has always been 13.5px, and the loss is very small, only 0.3. I checked the gradient, did the return, and also tried to adjust the learning rate, found no effect. @poincarelee |
Hi,
![flownet2_1857](https://user-images.githubusercontent.com/38481402/160764942-1075090b-9f9f-4df6-befa-f61f2f38e34c.png)
![fastflow_train_1857_347](https://user-images.githubusercontent.com/38481402/160764988-262f2ac2-f71c-4ee9-aea5-652bf2359e33.png)
have you tried training on real scene such as market or subway? I have finetuned the model according to IRR-PWC by using your './checkpoints/fastflownet_ft_mix.pth' in subway real scene, but the results are much worse than flownet2's.
And I met another weird problem: during predicting, whether I multiply the optical flow result by div_flow(20), there seems no difference on the flow-png(flow result transferred to png).
The text was updated successfully, but these errors were encountered: