Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Image Generation Performance Evaluation #58

Open
mmderakhshani opened this issue Feb 5, 2025 · 12 comments
Open

Image Generation Performance Evaluation #58

mmderakhshani opened this issue Feb 5, 2025 · 12 comments

Comments

@mmderakhshani
Copy link

mmderakhshani commented Feb 5, 2025

Hi @Sierkinhane,

Thank you for providing this amazing GitHub repository. Please let me know which checkpoint and configuration you used to compute the results for tables 2 and 3 related to FID performance evaluation and GenEval evaluation. I also would like to know which split of MSCOCO 30k you use. Any link for this link would be appreciated.

I am trying to replicate your number, and I would appreciate having access to your evaluation script.

Best,
Mohammad

@mmderakhshani
Copy link
Author

I am getting the following FID score: FID: 26.207129500696738.

I have computed it on the following dataset: https://huggingface.co/datasets/stasstaf/MS-COCO-validation using the standard Github code https://github.com/mseitzer/pytorch-fid. Could you please tell me how I can reproduce your numbers?

@Sierkinhane
Copy link
Collaborator

Hi, we used this split https://github.com/boomb0om/text2image-benchmark and we followed PixArt to fine-tune our model on a coco-like dataset (e.g., openimages) for evaluating MSCOCO FID. The other evaluations were assessed using the released checkpoints in GitHub.

@mmderakhshani
Copy link
Author

Is this finetuning dataset/script available? I want to provide a solid comparison with your work and that is why I am asking for it.

@mmderakhshani
Copy link
Author

Or would it be possible if I ask you to share with me the check point after finetuning? I do really need to have this one as I need to do inference on this model. Thanks.

@Sierkinhane
Copy link
Collaborator

Hi, I can share it with you via email. (sierkinhane@gmail.com)

@dhevarghese
Copy link

Hi,

Following up on the evaluations, I’m working on reproducing the results from Table 3 (GenEval). Here are the results I obtained:

Results:

Summary
=======
Total images: 553
Total prompts: 553
% correct images: 56.42%
% correct prompts: 56.42%
Task breakdown
==============
two_object       = 66.67% (66 / 99)
color_attr       = 34.00% (34 / 100)
colors           = 87.23% (82 / 94)
counting         = 47.50% (38 / 80)
position         = 13.00% (13 / 100)
single_object    = 98.75% (79 / 80)
Overall score (avg. over tasks): 0.57858

For image generation, I used the following parameters:

  • Guidance scale: 5
  • Generation timesteps: 50
  • Seed: 42
  • Checkpoint: Pretrained showo from the config (showlab/show-o)

Could you confirm if these settings align with those used in the original evaluation? If there are any additional details or adjustments I should consider, I’d appreciate your guidance.

Thanks!

@Sierkinhane
Copy link
Collaborator

Hi, we got a GenEval score around 0.53 when setting fewer inference steps (<=25) and more inference steps would obtain better performance.

@dhevarghese
Copy link

For the sake of reproducibility of the results in the paper, I'd like to achieve the score mentioned in the paper (0.68). Would it be possible to provide the hyperparmeters that were used to get that score? As it is much higher than what I get in my test runs with higher inference steps (0.57 maximum)

@Sierkinhane
Copy link
Collaborator

Hi, you should use this checkpoint https://huggingface.co/showlab/show-o-512x512.

@dhevarghese
Copy link

Thank you! I tried using this checkpoint, but something seems to be off. Below are the GenEval results I obtained:

Summary
=======
Total images: 553
Total prompts: 553
% correct images: 5.42%
% correct prompts: 5.42%

Task breakdown
==============
two_object       = 1.01% (1 / 99)
color_attr       = 0.00% (0 / 100)
colors           = 9.57% (9 / 94)
counting         = 0.00% (0 / 80)
position         = 0.00% (0 / 100)
single_object    = 25.00% (20 / 80)

Overall score (avg. over tasks): 0.05931

Would you be able to check if this checkpoint reproduces the reported score of 0.68? Are there any specific settings I might be missing? Any advice would be greatly appreciated.

@Sierkinhane
Copy link
Collaborator

The score is very weird. Can you check if the images were correctly generated? Besides, you must use this config https://github.com/showlab/Show-o/blob/main/configs/showo_demo_512x512.yaml

@dhevarghese
Copy link

Thank you so much for your guidance! It turns out the issue was with the config. I was able to reproduce the reported score using the provided configuration file. I appreciate your help!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants