-
Notifications
You must be signed in to change notification settings - Fork 34
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The AUC of FF++(c23) #12
Comments
My results on c23 are very similar to yours. Reporting the results of c23 in the paper are the consensus in the field of Deepfake Detection, and the author did not do so. |
@Vodka70 would you tell me how to test FF++ dataset ? should we revise the code in inference._dataset.py !? |
Yes, you should revise the code of data path in inference.dataset |
Hi
Did you train from scratch by using resnet34/efficient b4 then test on
FF++ or CDF ??my result aren't as good as paper
many thanks
Vodka70 ***@***.***> 於 2022年10月19日 週三 凌晨1:40寫道:
… Yes, you should revise the code of data path in inference.dataset
—
Reply to this email directly, view it on GitHub
<#12 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AYVDSKCZ2HO7LQTSCV6FKXLWD3OINANCNFSM55KCAO4A>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
@YU-SHAO-XU Sorry for the late reply. I re-trained from scratch by ImageNet pretrained efficientnet-b4. And test on FF++ and CDF. My results aren't also as good as those in the paper. FF-AUC: 0.9936, FF-DF-AUC:0.9999, FF-FS-AUC:0.9986, FF-F2F-AUC:0.9979, FF-NT-AUC:0.9780(which is obviously weaker than the performance reported in the paper). Besides, my CDF-AUC is only 0.8995, which is not as good as the reported performance (0.9382). @mapooon Would you like to provide some suggestions about this? Is tihs a normal experiment results due to some random factors? Many thanks~ |
Try to install packages with |
I install packages in requirements.txt one by one by efficientnet-b4 and the result is 99.36 (FF++) and 90.06 (CDF) respectively. |
I used the weights provided by the author to test on the DFDC image data set, and ACC was extremely poor.@YU-SHAO-XU |
@githuboflk |
Thanks to an enthusiastic collaborator, we have found that there is a bug in crop_dlib_ff.py. And we have just fixed the bug, so please try again from the point of executing preprocessing. |
After modifying the num_frames error of crop_dlib_ff.py and retraining the model, the AUC of CDF still did not reach the 0.9318 reported in the paper, but 0.9123. Is there anything else that needs to be modified? In addition, it is found that many of the faces recognized by retinaface are faces in the background. Will this problem lead to low test results? @mapooon |
Would anyone happen to be conducting cross-manipulation evaluation on FF++(c40)? I obtained results that were not as favorable as I had hoped. I would greatly appreciate any assistance or insights. |
hi, have you solved this problem? Can you tell me the reason for this problem? |
The authors' newly published model does not address this issue either |
I used the weights provided by the author, extracted the last fully connected, and trained a classifier as simple as it was in the rest of the original network, but the results for the CDF does not go above 91.90, Would you help me with this problem ? |
@Elahe-sd Is your batchsize 32? A low batchsize may result in a decrease in the auc. |
Actually, it was 16, then, after your suggestion, I changed it and checked, it is still the same |
|
I run the inference on Celeb-DF-v2, and FF++ (c23) with your pre-trained model. I get the similar results on the former (AUC:0.9381, AP:0.9669). However, the video-level AUC on the latter is only 0.9051, and AP is 0.9797. I also separately test the performance on Deepfakes, FaceSwap, Face2Face, NeuralTextures and FaceShifter. The results are as follows: DF-AUC:0.9856/AP:0.9900; FS-AUC:0.9698/AP:0.9747; F2F-AUC:0.9094/AP:0.9192; NT-AUC:0.8254/AP:0.8427; FSh-AUC:0.8351/AP:0.8427. Except for Deepfakes, other performance has some difference with that your reported in the supplementary. Do you run the inference on FF++ c23 dataset? I do not think the c23 videos can result the significant performance drop. Would you like to provide some suggestions to solve this problem?
The text was updated successfully, but these errors were encountered: