Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The AUC of FF++(c23) #12

Open
WeinanGuan opened this issue Aug 2, 2022 · 17 comments
Open

The AUC of FF++(c23) #12

WeinanGuan opened this issue Aug 2, 2022 · 17 comments

Comments

@WeinanGuan
Copy link

I run the inference on Celeb-DF-v2, and FF++ (c23) with your pre-trained model. I get the similar results on the former (AUC:0.9381, AP:0.9669). However, the video-level AUC on the latter is only 0.9051, and AP is 0.9797. I also separately test the performance on Deepfakes, FaceSwap, Face2Face, NeuralTextures and FaceShifter. The results are as follows: DF-AUC:0.9856/AP:0.9900; FS-AUC:0.9698/AP:0.9747; F2F-AUC:0.9094/AP:0.9192; NT-AUC:0.8254/AP:0.8427; FSh-AUC:0.8351/AP:0.8427. Except for Deepfakes, other performance has some difference with that your reported in the supplementary. Do you run the inference on FF++ c23 dataset? I do not think the c23 videos can result the significant performance drop. Would you like to provide some suggestions to solve this problem?

@LOOKCC
Copy link

LOOKCC commented Aug 29, 2022

My results on c23 are very similar to yours. Reporting the results of c23 in the paper are the consensus in the field of Deepfake Detection, and the author did not do so.

@YU-SHAO-XU
Copy link

@Vodka70 would you tell me how to test FF++ dataset ? should we revise the code in inference._dataset.py !?

@WeinanGuan
Copy link
Author

Yes, you should revise the code of data path in inference.dataset

@YU-SHAO-XU
Copy link

YU-SHAO-XU commented Oct 19, 2022 via email

@WeinanGuan
Copy link
Author

WeinanGuan commented Dec 18, 2022

@YU-SHAO-XU Sorry for the late reply. I re-trained from scratch by ImageNet pretrained efficientnet-b4. And test on FF++ and CDF. My results aren't also as good as those in the paper. FF-AUC: 0.9936, FF-DF-AUC:0.9999, FF-FS-AUC:0.9986, FF-F2F-AUC:0.9979, FF-NT-AUC:0.9780(which is obviously weaker than the performance reported in the paper). Besides, my CDF-AUC is only 0.8995, which is not as good as the reported performance (0.9382). @mapooon Would you like to provide some suggestions about this? Is tihs a normal experiment results due to some random factors? Many thanks~

@mapooon
Copy link
Owner

mapooon commented Dec 20, 2022

Try to install packages with requirements.txt we released at:
https://drive.google.com/file/d/14ZjiBFG825qP8ZKE_XGzQ6yob4XHx_Ml/view?usp=share_link
After executing the docker, you can use it as follows:
pip install -r requirements.txt.
Please let me know if you success in reproducing the results. We will update the repository.

@YU-SHAO-XU
Copy link

YU-SHAO-XU commented Dec 24, 2022

I install packages in requirements.txt one by one by efficientnet-b4 and the result is 99.36 (FF++) and 90.06 (CDF) respectively.
How about you @Vodka70 ?!
batch size=16, CDF AUC 92.1...

@githuboflk
Copy link

I used the weights provided by the author to test on the DFDC image data set, and ACC was extremely poor.@YU-SHAO-XU

@YU-SHAO-XU
Copy link

@githuboflk
I used the weights provided by the author is as good as paper say. If i train from scratch, the result is 99.36 (FF++) and 90.06 (CDF) respectively.

@mapooon
Copy link
Owner

mapooon commented Feb 4, 2023

Thanks to an enthusiastic collaborator, we have found that there is a bug in crop_dlib_ff.py. And we have just fixed the bug, so please try again from the point of executing preprocessing.

@angelalife
Copy link

After modifying the num_frames error of crop_dlib_ff.py and retraining the model, the AUC of CDF still did not reach the 0.9318 reported in the paper, but 0.9123. Is there anything else that needs to be modified? In addition, it is found that many of the faces recognized by retinaface are faces in the background. Will this problem lead to low test results? @mapooon

@Blosslzy
Copy link

Would anyone happen to be conducting cross-manipulation evaluation on FF++(c40)? I obtained results that were not as favorable as I had hoped. I would greatly appreciate any assistance or insights.

@lihanzhe
Copy link

I run the inference on Celeb-DF-v2, and FF++ (c23) with your pre-trained model. I get the similar results on the former (AUC:0.9381, AP:0.9669). However, the video-level AUC on the latter is only 0.9051, and AP is 0.9797. I also separately test the performance on Deepfakes, FaceSwap, Face2Face, NeuralTextures and FaceShifter. The results are as follows: DF-AUC:0.9856/AP:0.9900; FS-AUC:0.9698/AP:0.9747; F2F-AUC:0.9094/AP:0.9192; NT-AUC:0.8254/AP:0.8427; FSh-AUC:0.8351/AP:0.8427. Except for Deepfakes, other performance has some difference with that your reported in the supplementary. Do you run the inference on FF++ c23 dataset? I do not think the c23 videos can result the significant performance drop. Would you like to provide some suggestions to solve this problem?

hi, have you solved this problem? Can you tell me the reason for this problem?

@lihanzhe
Copy link

The authors' newly published model does not address this issue either

@Elahe-sd
Copy link

Elahe-sd commented May 7, 2024

I used the weights provided by the author, extracted the last fully connected, and trained a classifier as simple as it was in the rest of the original network, but the results for the CDF does not go above 91.90, Would you help me with this problem ?

@lihanzhe
Copy link

lihanzhe commented May 7, 2024

@Elahe-sd Is your batchsize 32? A low batchsize may result in a decrease in the auc.

@Elahe-sd
Copy link

Elahe-sd commented May 8, 2024

@Elahe-sd Is your batchsize 32? A low batchsize may result in a decrease in the auc.

Actually, it was 16, then, after your suggestion, I changed it and checked, it is still the same

@Qetx-Jul-27
Copy link

@Elahe-sd您的批次大小是 32 吗?批次大小过低可能会导致 AUC 下降。
可以告诉我ff++下的val.json train.json 和 test.json怎么来的吗?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

10 participants