Taking a huge step forward in artificial intelligence, AMD has launched its Nitro Diffusion model, designed to revolutionize image creation and visual content creation. According to AMD, this model demonstrates advancements in AI technology that promise high-quality and versatile image output.
A revolution in image creation
The area of generative AI has experienced revolutionary changes, with diffusion models emerging as a superior technology, especially in image generation. These models can perform complex tasks such as text-to-image compositing, image-to-image conversion, and image inpainting. AMD’s introduction of the Nitro Diffusion model aims to further push the boundaries of these capabilities, opening up new opportunities across a variety of fields from entertainment to scientific visualization.
AMD Nitro Diffusion Model
AMD’s innovative model is built on two popular open source models: Stable Diffusion 2.1 and PixArt-Sigma. Leveraging the UNet architecture and integrating advanced text encoders such as CLIP and Diffusion Transformer, this model promises improved efficiency and high image quality. The implementation leverages PyTorch, the HuggingFace Accelerate library, and precomputed latent representations to optimize training throughput on the AMD Instinct MI250 accelerator.
Open source and developer involvement
AMD has made these models and their code available to the open source community to further advance advances in generative AI. This move invites developers and researchers to explore and expand the potential of AI-based image generation. Completed model files and code instructions can be accessed via AMD’s Hugging Face model card and GitHub repository. Developers are also encouraged to utilize the AMD Developer Cloud to remotely access AMD GPUs for testing and development purposes.
For additional information on the features and performance of these models, AMD provides a detailed technical blog that can be accessed here.
Image source: Shutterstock