Leveraging AI and Diffusion Models for Anime Art Creation: A Study on Style Transfer and Image Quality Evaluation
- School of Art Design and Media, Sanda University, China
ghinishen@163.com - School of Information Science and Technology, Sanda University, China
snluo@sandau.edu.cn - Shanghai Technical Institute of Electronics & Information College, China
FL0514@126.com - School of Information Science and Technology, Sanda University, China
ad88105506@163.com
Abstract
The remarkable advancements in artificial intelligence (AI)-driven image generation technologies have brought about a profound transformation across various industries, particularly in new media, video production, and gaming. AI-generated content (AIGC) has emerged as a game-changing, cost-efficient solution for companies seeking high-quality visual assets while operating within constrained budgets and having limited access to traditional human resources. Through the use of sophisticated algorithms, AIGC enables the creation of stunning visuals without relying on conventional, labor-intensive workflows. Among the most prominent techniques, diffusion models have played a pivotal role in the development of AI image generation tools, giving rise to both proprietary platforms like Midjourney and open-source alternatives such as Stable Diffusion. These technologies continue to evolve, benefiting from the collaborative contributions of global programming communities. This study focuses on advancing the capabilities of Stable Diffusion, an open-source AI image generation model, to address prevalent challenges in style consistency and image quality. By integrating Python and harnessing cutting-edge AI techniques, such as DreamBooth and embedding methods, the research aims to enhance the model's ability to replicate and embed distinct artistic styles. Specifically, the study targets the unique art style of the popular mobile game "Arknights" as a training objective, applying advanced techniques to refine the system's output. The proposed approach demonstrates significant improvements over the baseline model, showcasing enhanced performance in generating style-consistent anime imagery. This research contributes to the evolving landscape of AI-driven art generation, offering novel insights into the application of diffusion-based technologies within creative industries. By utilizing DreamBooth and embedding for style transfer and injection, the study achieves notable efficiency, drastically reducing the time required to train a new model. Ultimately, this work paves the way for more specialized and customizable AI systems in art creation, pushing the boundaries of what AI can achieve in the realm of creative expression.
Key words
AI-Generated Content (AIGC); Diffusion Model; Academic Affairs Management
How to cite
Shen, C., Luo, S., Fan, L., Dai, C.: Leveraging AI and Diffusion Models for Anime Art Creation: A Study on Style Transfer and Image Quality Evaluation. Computer Science and Information Systems