Abstract
Creating visual content such as product advertisements for marketing purposes has attracted research attention recently. Traditionally, such visuals showcase the product against a specific backdrop while adhering to a consistent corporate style to maintain brand visual identity. We present BrandDiffusion, a marketing image generation framework that leverages the power of pre-trained Stable Diffusion models while allowing designers to generate ideas by adapting prompt-controllable models to capture the intended marketing style. In addition, it is also capable of automatic product placement which incorporates harmonization of the product image with the generated background, guided by saliency detection, BLIP captioning, and SDEdit denoising process. We showcase the capabilities of the proposed framework through quantitative and qualitative evaluations, highlighting the quality of the generated content, its relevancy to user prompts, and a strong brand coherency.
Original language | English |
---|---|
Title of host publication | McGE '24: Proceedings of the 2nd International Workshop on Multimedia Content Generation and Evaluation: New Methods and Practice |
Publisher | Association for Computing Machinery |
Pages | 72-77 |
Number of pages | 6 |
ISBN (Print) | 9798400711947 |
DOIs | |
Publication status | Published - 28 Oct 2024 |
Event | 32nd ACM International Conference on Multimedia 2024 - Melbourne, Australia Duration: 28 Oct 2024 → 1 Nov 2024 Conference number: 32 https://icmsaust.com.au/event/acm-international-conference-for-multimedia-2024/ |
Conference
Conference | 32nd ACM International Conference on Multimedia 2024 |
---|---|
Abbreviated title | MM '24 |
Country/Territory | Australia |
City | Melbourne |
Period | 28/10/24 → 1/11/24 |
Internet address |
Keywords
- image generation
- visual style
- image harmonization
- diffusion models
- stable diffusion