Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Object-driven Text-to-Image Synthesis via Adversarial Training (Obj-GAN; CVPR 2019) #19

Open
hcchiu1202 opened this issue Jun 2, 2019 · 0 comments

Comments

@hcchiu1202
Copy link

0. 論文

タイトル: Object-driven Text-to-Image Synthesis via Adversarial Training
リンク: https://arxiv.org/abs/1902.10740
著者: Wenbo Li, Pengchuan Zhang, Lei Zhang, Qiuyuan Huang, Xiaodong He, Siwei Lyu, Jianfeng Gao
所属: University at Albany, SUNY, Microsoft Research AI, Microsoft, JD AI Research
発表年: 2019
掲載: CVPR

1. どんなもの?

First, generate a semantic layout (with class labels, bounding boxes, shapes of objects) from the text.
Then, using the both the layout and text as input, generate each object separately with attentions towards different levels, especially useful is the Object-driven attention.

image

2. 先行研究と比べてどこがすごい?

Taking the usual bi-LSTM encoded word/sentence as input, giving +27% Inception score and -11% FID score compared to AttnGAN.

3. 技術や手法のキモはどこ?

Generation of the lay-out without extra input/label.
The novel object-driven attentive generative network and the object-wise discriminator.
(Details to be updated shortly.)

4. どうやって有効だと検証した?

Will be updated shortly.

5. 議論はある?

Will be updated shortly.

6. 次に読むべき論文は?

Will be updated shortly.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant