Crypto Flexs
  • DIRECTORY
  • CRYPTO
    • ETHEREUM
    • BITCOIN
    • ALTCOIN
  • BLOCKCHAIN
  • EXCHANGE
  • ADOPTION
  • TRADING
  • HACKING
  • SLOT
  • CASINO
Crypto Flexs
  • DIRECTORY
  • CRYPTO
    • ETHEREUM
    • BITCOIN
    • ALTCOIN
  • BLOCKCHAIN
  • EXCHANGE
  • ADOPTION
  • TRADING
  • HACKING
  • SLOT
  • CASINO
Crypto Flexs
Home»ADOPTION NEWS»AMD improves AI algorithm efficiency with innovative depth pruning method
ADOPTION NEWS

AMD improves AI algorithm efficiency with innovative depth pruning method

By Crypto FlexsJune 8, 20243 Mins Read
Facebook Twitter Pinterest LinkedIn Tumblr Email
AMD improves AI algorithm efficiency with innovative depth pruning method
Share
Facebook Twitter LinkedIn Pinterest Email





AMD, a leading semiconductor supplier, has made significant progress in optimizing the hardware efficiency of artificial intelligence (AI) algorithms. According to AMD.com, the company’s latest research paper is ‘Unified progressive depth pruner for CNN and Vision Transformer‘ was accepted at the prestigious AAAI 2024 conference. In this paper, we introduce a new depth pruning method designed to improve performance across a variety of AI models.

Motivation for model optimization

Deep neural networks (DNNs) have become essential for a variety of industrial applications and require continuous model optimization. In this context, techniques such as model pruning, quantization, and efficient model design are very important. Existing channel-wise pruning methods face the problem of depth-wise convolutional layers due to sparse computations and few parameters. Additionally, these methods often suffer from high parallel computing demands, resulting in suboptimal hardware utilization.

To address these issues, the AMD research team proposed DepthShrinker and Layer-Folding techniques to optimize MobileNetV2 by reducing model depth through reparameterization. Despite their potential, these methods have limitations, such as potential loss of accuracy and constraints in certain regularization layers such as LayerNorm, making them unsuitable for vision transducer models.

Innovative depth pruning approach

AMD’s new depth pruning method introduces a progressive training strategy and a new block pruning technique that can optimize both CNN and vision transformer models. This approach ensures high utilization of baseline model weights, thereby increasing accuracy. Additionally, this method can effectively prune the vision transformer model by processing existing regularization layers, including LayerNorm.

The AMD deep pruning strategy converts complex, slow blocks into simpler, faster blocks through block merging. This involves replacing activation layers with ID layers and LayerNorm layers with BatchNorm layers to facilitate reparameterization. The reparameterization technique then merges the BatchNorm layers, adjacent convolutional layers, or fully connected layers and skips the connections.

Core technology

The depth pruning process includes four main steps: supernet training, subnet discovery, subnet training, and subnet merging. Initially, a supernet is constructed based on the base model by incorporating block modifications. After learning the supernet, a search algorithm identifies the optimal subnet. We then apply an incremental training strategy to optimize the subnets while minimizing accuracy loss. Finally, the subnets are merged into a shallower model using a reparameterization technique.

Benefits and Performance

AMD’s depth pruning method offers several key contributions:

  • A unified and efficient depth pruning method for CNN and vision transformer models.
  • An incremental training strategy for subnet optimization combined with a new block pruning strategy using reparameterization.
  • A comprehensive experiment showing good pruning performance across a variety of AI models.

Experimental results show that AMD’s method achieves up to 1.26x speedup on the AMD Instinct™ MI100 GPU accelerator, with only 1.9% degradation in top-1 accuracy. This approach has been tested on several models, including ResNet34, MobileNetV2, ConvNeXtV1, and DeiT-Tiny, showing its versatility and efficiency.

In conclusion, AMD’s integrated depth pruning method represents a significant advance in optimizing AI model performance. Its applicability to both CNNs and vision transformer models highlights its potential impact on future AI developments. AMD plans to further apply this method to more transformer models and tasks.

Image source: Shutterstock

. . .

tag


Share. Facebook Twitter Pinterest LinkedIn Tumblr Email

Related Posts

As you challenge the mixed technology signal, OnDo Price Hovers challenges the August Bullish predictions.

August 7, 2025

XRP Open Interests decrease by $ 2.4B after recent sale

July 30, 2025

KAITO unveils Capital Launchpad, a Web3 crowdfunding platform that will be released later this week.

July 22, 2025
Add A Comment

Comments are closed.

Recent Posts

Did you miss the TRON ‘S (TRX) 100X? Ruvi AI (Ruvi)

August 9, 2025

Re -creation attack in ERC -721 -Ackee Blockchain

August 8, 2025

The New Bybit Web3 Is Here–Fueling On-Chain Thrills With $200,000 Up For Grabs

August 8, 2025

Stella (XLM) Eye 35% Rally and Ripple and SEC END 5 years legal battle

August 8, 2025

Builders Are Proving What’s Possible With CARV’s AI Stack

August 8, 2025

Caldera Announces Partnership With EigenCloud To Integrate EigenDA V2

August 7, 2025

Are Monero in danger? Five orphan blocks were found during the Cubic Mining War.

August 7, 2025

One Card To Seamlessly Bridge Web3 Assets And Real-World Spending

August 7, 2025

Coinbase’s USDC fee, encryption or other banks?

August 7, 2025

Protocol Update 001 -scale L1

August 7, 2025

As you challenge the mixed technology signal, OnDo Price Hovers challenges the August Bullish predictions.

August 7, 2025

Crypto Flexs is a Professional Cryptocurrency News Platform. Here we will provide you only interesting content, which you will like very much. We’re dedicated to providing you the best of Cryptocurrency. We hope you enjoy our Cryptocurrency News as much as we enjoy offering them to you.

Contact Us : Partner(@)Cryptoflexs.com

Top Insights

Did you miss the TRON ‘S (TRX) 100X? Ruvi AI (Ruvi)

August 9, 2025

Re -creation attack in ERC -721 -Ackee Blockchain

August 8, 2025

The New Bybit Web3 Is Here–Fueling On-Chain Thrills With $200,000 Up For Grabs

August 8, 2025
Most Popular

Saylor highlights Bitcoin’s $1 million potential, pointing to its current price.

February 23, 2024

Crypto Stock Recap: Wall Street Doesn’t Treat All Bitcoin Miners Equally

December 10, 2023

Sui lost a $2 peg in the 20% slide Post-Insider Dump Saga.

October 23, 2024
  • Home
  • About Us
  • Contact Us
  • Disclaimer
  • Privacy Policy
  • Terms and Conditions
© 2025 Crypto Flexs

Type above and press Enter to search. Press Esc to cancel.