è«–æ–‡ã¸ã®ãƒªãƒ³ã‚¯
[2303.00262] Collage Diffusion
[2303.00262] Collage Diffusion
[2301.09515] StyleGAN-T: Unlocking the Power of GANs for Fast Large-Scale Text-to-Image Synthesis
Axel Sauer12 Tero Karras2 Samuli Laine2 Andreas Geiger1 Timo Aila2
1 University of Tubingen, T ¨ ubingen AI Center ¨
2 NVIDIA. Correspondence to: Axel Sauer a.sauer@uni-tuebingen.de
2023/01/23
GANã§é«˜å“質ãªtext-to-imagehttps://t.co/ZtAXenb3oF
— mi141 (@mi141) January 25, 2023
拡散モデルよりもéžå¸¸ã«é«˜é€Ÿã ãžã¨ã„ã†ä¸»å¼µã¨ã€åœ°å‘³ã«ç”Ÿæˆç”»åƒé–“ã®å†…挿ãŒã‚¹ãƒ ーズã«ã§ãã‚‹ã“ã¨ã‚’アピールã—ã¦ã¾ã™ã。GANã®è«–æ–‡ã¯ãŠä¹…ã—ã¶ã‚Šãªã‚“ã§ã™ãŒã€ãƒ†ã‚ã‚¹ãƒˆæƒ…å ±ã®ä½¿ã„æ–¹ã«èˆˆå‘³ãŒã‚ã£ãŸã®ã§ãƒã‚§ãƒƒã‚¯ã—ã¾ã—ãŸï¼ˆç¶šï¼‰ pic.twitter.com/ozVqHS9SWS
プãƒã‚¸ã‚§ã‚¯ãƒˆãƒšãƒ¼ã‚¸ sites.google.com
[2211.09800] InstructPix2Pix: Learning to Follow Image Editing Instructions
Tim Brooks Aleksander Holynski Alexei A. Efros
University of California, Berkeley
2022/11/17
プãƒã‚¸ã‚§ã‚¯ãƒˆãƒšãƒ¼ã‚¸
äººé–“ã®æŒ‡ç¤ºã‹ã‚‰æ•°ç§’ã§ç”»åƒç·¨é›†ã§ãã‚‹ 「InstructPix2Pixã€ã®ã‚³ãƒ¼ãƒ‰ãŒå…¬é–‹ï¼
— ã‚„ã¾ã‹ãš (@Yamkaz) January 20, 2023
「ã²ã¾ã‚りã¨ãƒãƒ©ã‚’交æ›ã€ã€Œç©ºã«èбç«ã‚’è¿½åŠ ã€ã€Œé›ªãŒé™ã£ã¦ã„ãŸã‚‰ï¼Ÿã€ãªã©ã‚’入力ã™ã‚‹ã¨ç”»åƒãŒç·¨é›†ã§ãã‚‹
Project: https://t.co/H3iY5qQpXm
code: https://t.co/sZueY6UG4R
demo: https://t.co/8bPVHiWkB3 pic.twitter.com/ox5pUUtOh8
高性能ãªPix2Pixを実ç¾ã™ã‚‹ãŸã‚ã«ã€Diffusion Modelをファインãƒãƒ¥ãƒ¼ãƒ‹ãƒ³ã‚°ã€‚
教師データã¯ã€GPT-3ã€Stable Diffusionを組ã¿åˆã‚ã›ã¦ç”Ÿæˆã€‚
Instruct Pix2Pixã®Google Colab版ãŒã‚ã£ãŸã®ã§è©¦ã—ã¦ã¿ãŸã€‚凄ã„ã§ã™ãŒã€ç°¡å˜ã«ç´°ã‹ã„ã¨ã“ã‚ã¾ã§æ€ã„通りã¨ã¯ãªã‹ãªã‹ã„ã‹ãªã„ã§ã™ãhttps://t.co/tQ9Izr5vAw pic.twitter.com/3XEX6wHtZi
— ã‹ã‚‰ã‚ã’ (@karaage0703) January 21, 2023
å¦ç¿’ã®è€ƒãˆæ–¹ã¯ã€ä»¥ä¸‹ã«è¿‘ã„部分ãŒã‚ã‚‹ã‹ã‚‚
[2301.07093] GLIGEN: Open-Set Grounded Text-to-Image Generation
Hao Su, Jianwei Niu, Xuefeng Liu, Qingfeng Li, Jiahe Cui, Ji Wan
全員 Beihang University(北京大å¦ï¼‰
2020/04/22
プãƒã‚¸ã‚§ã‚¯ãƒˆãƒšãƒ¼ã‚¸
物体ã®ä½ç½®ã‚„説明をæ¡ä»¶ã¨ã—ãŸç”»åƒç”Ÿæˆã‚’ã€å¦ç¿’æ¸ˆã¿æ‹¡æ•£ãƒ¢ãƒ‡ãƒ«ã‚’活用ã—ã¦å®Ÿç¾ã€‚æ¡ä»¶æƒ…å ±ã¯å…¨ã¦ãƒˆãƒ¼ã‚¯ãƒ³åŒ–ã—ã€æ‹¡æ•£ãƒ¢ãƒ‡ãƒ«ã«è¿½åŠ ã—ãŸattention層を介ã—ã¦å…¥åŠ›ã€‚ã“ã®å±¤ã ã‘fine-tune。https://t.co/11JFTKPA5j
— mi141 (@mi141) January 19, 2023
(引用ã•れã¦ãªã„ãŒï¼‰PITIã«ä¼¼ã¦ã‚‹ãŒã€æ¡ä»¶ã®å…¥ã‚Œæ–¹ãŒç•°ãªã‚‹ã€‚https://t.co/e8rVszOxzJ pic.twitter.com/OmREZEOCOt
[2205.12952] Pretraining is All You Need for Image-to-Image Translation
Tengfei Wang1 , Ting Zhang2, Bo Zhang2, Hao Ouyang1, Dong Chen2, Qifeng Chen1, Fang Wen2
1 The Hong Kong University of Science and Technology
2 Microsoft Research Asia
2022/5/25
 プãƒã‚¸ã‚§ã‚¯ãƒˆãƒšãƒ¼ã‚¸
 基盤モデルを使ã£ã¦ã€I2Iã®æ€§èƒ½é«˜ã‚ã‚‹ã¿ãŸã„ãªï¼Ÿå¾Œã§èªã‚€