A Case for Dynamic Activation Quantization in CNNs
Karl Taht, Surya Narayanan, Rajeev Balasubramonian University of Utah
A Case for Dynamic Activation Quantization in CNNs Karl Taht, Surya - - PowerPoint PPT Presentation
A Case for Dynamic Activation Quantization in CNNs Karl Taht, Surya Narayanan, Rajeev Balasubramonian University of Utah Overview Background Proposal Search Space Architecture Results Future Work Improving CNN Efficiency
Karl Taht, Surya Narayanan, Rajeev Balasubramonian University of Utah
Concept: Add lightweight predictor here Save computations here
within the image (>55% in our tests)
the outside!
N = 25 N = 10 N = 8 N = 5 N = 2 Concept: Scale Feature Maps Proportionally
[ 0 1 0 1 ] [ 0 1 0 0 ] [ 0 0 1 0 ] [ 0 0 0 1 ]
permutations of top, bottom, left, and right crops encoded as a vector: [ TOP , BOTTOM , LEFT , RIGHT ]
exploit image-based information to enhance pruning options.
[ 1 0 0 0 ]
[ 1 0 1 1 ]
Weight Set Number of Edges Cropped
1 bit 2 bit 2 bit 2 bit 2 bit 1 bit 1 bit 1 bit 8 bit
Inputs W e i g h t s Outputs
1 bit 2 bit 2 bit 2 bit 2 bit 1 bit 1 bit 1 bit
8 columns
16 bit 10 bit 5 columns 8 bit ADC (Multiplexed)
8 x ADC Operations
5 x ADC Operations
1 bit 2 bit 2 bit 2 bit 2 bit 1 bit 1 bit 1 bit
8 columns
16 bit 10 bit 5 columns 8 bit ADC (Multiplexed) Can vary shift amount to compute fixed point computations with different exponents
1 bit 2 bit 2 bit 2 bit 2 bit 1 bit 1 bit 1 bit 8 bit
k-bit inputs Outputs 1...1010101 0...1000110 1...0011111 1 .. 1 1 1 1 1 .. 1 1 1 1 .. 1 1 1 1 .. 1 1 1 1 1 k .. 6 7 5 3 1 4 2 Time Step
Buffered Input
1 bit 2 bit 2 bit 2 bit 2 bit 1 bit 1 bit 1 bit 8 bit
k-bit inputs Outputs 1...1010101 0...1000110 1...0011111 1 .. 1 1 1 1 1 .. 1 1 1 1 .. 1 1 1 1 .. 1 1 1 1 1 k .. 6 7 5 3 1 4 2 Time Step
Buffered Input
few gradient-based kernels
Original Sobel Gradient