« Back to List Implementing RAG in Multimodal Architectures for Image Generation
Time / Place:
⏱️ 09/27 (Fri.) 13:30-14:00 at R3 - 1st Conference Room
(學術活動中心 2 樓)
Abstract:
RAG (Retrieval-Augmented Generation) has become a mainstream method for enhancing responses from large language models. In recent years, multimodal LLM has been developing increasingly well, and similar approaches in multimodal LLM application scenarios and implementation methods are worth exploring. This presentation will discuss the implementation of multimodal LLM in the RAG process, and share practical applications using image generation as a case study.
Biography:
- 吳振和 Wu, Cheng-Ho
Website: https://www.facebook.com/chengho.wu?locale=zh_TW - cacaFly / Cloud AI+ Center, VP
- CH is the VP of Technology at cacaFly Cloud Intelligence Center. He has previously developed apps that have garnered millions of downloads and reached top ranks on Google and Apple charts, as well as managing high-traffic network services with billions of data flow. He is also a speaker at TEDx and AWS SUMMIT, as well as a speaker at the Taiwan Artificial Intelligence Conference and an adjunct lecturer at National Taiwan University of Science and Technology.