Image-to-Image
Transformers
English
multimodal
frankzeng commited on
Commit
99a965e
Β·
verified Β·
1 Parent(s): 2588e10

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -12,7 +12,7 @@ library_name: transformers
12
  * Apr 25, 2025: πŸ‘‹ We release the inference code and model weights of Step1X-Edit. [inference code](https://github.com/stepfun-ai/Step1X-Edit)
13
  * Apr 25, 2025: πŸŽ‰ We have made our technical report available as open source. [Read](https://arxiv.org/abs/2504.17761)
14
 
15
- ## Image Edit Demos
16
 
17
  <div align="center">
18
  <img width="720" alt="demo" src="assets/image_edit_demo.gif">
@@ -31,8 +31,7 @@ images using a DiT-based network.More details please refer to our [technical rep
31
 
32
 
33
  ## Benchmark
34
- We release [GEdit-Bench](https://huggingface.co/datasets/stepfun-ai/GEdit-Bench) as a new benchmark, grounded in real-world usages is developed to support more authentic and comprehensive evaluation. This benchmark, which is carefully curated to reflect actual user editing needs and a wide range of editing scenarios, enables more authentic and comprehensive evaluations of image editing models.
35
- The evaluation process and related code can be found in [GEdit-Bench/EVAL.md](GEdit-Bench/EVAL.md). Part results of the benchmark are shown below:
36
  <div align="center">
37
  <img width="1080" alt="results" src="assets/eval_res_en.png">
38
  </div>
 
12
  * Apr 25, 2025: πŸ‘‹ We release the inference code and model weights of Step1X-Edit. [inference code](https://github.com/stepfun-ai/Step1X-Edit)
13
  * Apr 25, 2025: πŸŽ‰ We have made our technical report available as open source. [Read](https://arxiv.org/abs/2504.17761)
14
 
15
+ <!-- ## Image Edit Demos -->
16
 
17
  <div align="center">
18
  <img width="720" alt="demo" src="assets/image_edit_demo.gif">
 
31
 
32
 
33
  ## Benchmark
34
+ We release [GEdit-Bench](https://huggingface.co/datasets/stepfun-ai/GEdit-Bench) as a new benchmark, grounded in real-world usages is developed to support more authentic and comprehensive evaluation. This benchmark, which is carefully curated to reflect actual user editing needs and a wide range of editing scenarios, enables more authentic and comprehensive evaluations of image editing models. Part results of the benchmark are shown below:
 
35
  <div align="center">
36
  <img width="1080" alt="results" src="assets/eval_res_en.png">
37
  </div>