Limitations of AI-Generated Content and Strategies for Enhancement
Main Article Content
Keywords
AI-generated content (AIGC), limitations, enhancement strategies
Abstract
With the rapid advancement of artificial intelligence (AI) technologies, tools such as Doubao and DeepSeek have become indispensable elements in everyday life. Many users turn to these AI platforms for secondary editing of images and videos or to generate visuals that align with specific textual descriptions. However, the outcomes frequently fall short of expectations: the generated content often bears little resemblance to the user’s prompts and is riddled with unacceptable flaws. To mitigate these issues and enable AI tools to produce satisfying results that meet given specifications, this paper explores the underlying principles of AI content generation. It proposes methods to enhance the accuracy of outputs and reduce inherent defects, concluding with a discussion of the study’s limitations.
References
- [1] Goodfellow, I. J., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A. and Bengio, Y. Generative adversarial nets. In Advances in neural information processing systems, Montreal, QC, Canada, 2014; p. 2672−2680.
- [2] Miao, F. C. Examination of the Technique Principle of Generative AI and Its Educational Applicability. Modern Educational Technology. 2023, 33(11), pp. 5-18.
- [3] Zhang, L. Y., Yang, S., Wang, W. J., Gao, X. and Liu, J. Y. AIGC-Based Image and Video Generation Method: A Review. Journal of Computer-Aided Design & Computer Graphics. 2025, 37(3), pp. 361-384.
- [4] Zeng, C., Ge, Y. J., Zhao, L. C. and Wang, Q. Survey of Multimodal Vision-Language Representation Learning Models and Their Adversarial Examples Attack and Defense Techniques. Journal of Computer Research and Development. 2025, 62(09), pp. 2208-2232.
- [5] Shao, S. Y., Du, Y. and Fan, X. L. Non-Autoregressive Sign Language Translation Technology Based on Transformer and Multimodal Alignment. Journal of Electronics & Information Technology. 2024, 46(7), pp. 2932-2941.
- [6] Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P. and Clark, J. Learning transferable visual models from natural language supervision. In International conference on machine learning, Virtual Conference, 2021; pp. 8748-8763.
- [7] Jiang, L. Interpreting and Addressing the “Uncanny Valley Effect” in AI-Generated Art. Journal of Nanjing University of the Arts(Fine Arts & Design). 2025(4), pp. 187-191.
