Show simple item record

dc.contributor.authorHan, Evans Xu
dc.contributor.authorZhang, Alice
dc.contributor.authorZhu, Haiyi
dc.contributor.authorShen, Hong
dc.contributor.authorLiang, Paul Pu
dc.contributor.authorHsieh, Jane
dc.date.accessioned2025-10-07T21:23:50Z
dc.date.available2025-10-07T21:23:50Z
dc.date.issued2025-09-27
dc.identifier.isbn979-8-4007-2037-6
dc.identifier.urihttps://hdl.handle.net/1721.1/163075
dc.descriptionUIST ’25, Busan, Republic of Koreaen_US
dc.description.abstractState-of-the-art visual generative AI tools hold immense potential to assist users in the early ideation stages of creative tasks — offering the ability to generate (rather than search for) novel and unprecedented (instead of existing) images of considerable quality that also adhere to boundless combinations of user specifications. However, many large-scale text-to-image systems are designed for broad applicability, yielding conventional output that may limit creative exploration. They also employ interaction methods that may be difficult for beginners. Given that creative end-users often operate in diverse, context-specific ways that are often unpredictable, more variation and personalization are necessary. We introduce POET, a real-time interactive tool that (1) automatically discovers dimensions of homogeneity in text-to-image generative models, (2) expands these dimensions to diversify the output space of generated images, and (3) learns from user feedback to personalize expansions. An evaluation with 28 users spanning four creative task domains demonstrated POET’s ability to generate results with higher perceived diversity and help users reach satisfaction in fewer prompts during creative tasks, thereby prompting them to deliberate and reflect more on a wider range of possible produced results during the co-creative process. Focusing on visual creativity, POET offers a first glimpse of how interaction techniques of future text-to-image generation tools may support and align with more pluralistic values and the needs of end-users during the ideation stages of their work.en_US
dc.publisherACM|The 38th Annual ACM Symposium on User Interface Software and Technologyen_US
dc.relation.isversionofhttps://doi.org/10.1145/3746059.3747710en_US
dc.rightsCreative Commons Attributionen_US
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/en_US
dc.sourceAssociation for Computing Machineryen_US
dc.titlePOET: Supporting Prompting Creativity and Personalization with Automated Expansion of Text-to-Image Generationen_US
dc.typeArticleen_US
dc.identifier.citationEvans Xu Han, Alice Qian Zhang, Haiyi Zhu, Hong Shen, Paul Pu Liang, and Jane Hsieh. 2025. POET: Supporting Prompting Creativity and Personalization with Automated Expansion of Text-to-Image Generation. In Proceedings of the 38th Annual ACM Symposium on User Interface Software and Technology (UIST '25). Association for Computing Machinery, New York, NY, USA, Article 162, 1–18.en_US
dc.contributor.departmentProgram in Media Arts and Sciences (Massachusetts Institute of Technology)en_US
dc.identifier.mitlicensePUBLISHER_POLICY
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/ConferencePaperen_US
eprint.statushttp://purl.org/eprint/status/NonPeerRevieweden_US
dc.date.updated2025-10-01T07:54:07Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2025-10-01T07:54:08Z
mit.licensePUBLISHER_CC
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record