• Learning low-precision neural networks without Straight-Through Estimator(STE)

    Charlotte Christopherson
    Charlotte Christopherson
    Zhi-Gang Liu, Matthew Mattina
    The Straight-Through Estimator (STE) is widely used for back-propagating gradients through the quantization function, but the STE technique lacks a complete theoretical understanding. We propose an alternative methodology…
    • Learning low-precision neural networks without Straight-Through Estimator(STE).pdf
    • over 1 year ago
    • Arm Research
    • Resources
  • Efficient Winograd or Cook-Toom Convolution Kernel Implementation on Widely Used Mobile CPUs

    Charlotte Christopherson
    Charlotte Christopherson
    Partha Maji, Andrew Mundy, Ganesh Dasika, Jesse Beu, Matthew Mattina, Robert Mullins
    The Winograd or Cook-Toom class of algorithms help to reduce the overall compute complexity of many modern deep convolutional neural networks (CNNs). Although there has been…
    • Efficient Winograd or Cook-Toom Convolution Kernel Implementation on Widely Used Mobile CPUs.pdf
    • over 1 year ago
    • Arm Research
    • Resources
  • Measuring scheduling efficiency of RNNs for NLP applications

    Charlotte Christopherson
    Charlotte Christopherson
    Urmish Thakker, Ganesh Dasika, Jesse Beu, Matthew Mattina
    Recurrent neural networks (RNNs) have shown state of the art results for speech recognition, natural language processing, image captioning and video summarizing applications. Many of these applications…
    • Measuring scheduling efficiency of RNNs for NLP applications.pdf
    • over 1 year ago
    • Arm Research
    • Resources
  • FixyNN: Efficient Hardware for Mobile Computer Vision via Transfer Learning

    Charlotte Christopherson
    Charlotte Christopherson
    Paul N. Whatmough, Chuteng Zhou, Patrick Hansen, Shreyas Kolala Venkataramanaiah, Jae-sun Seo, Matthew Mattina
    The computational demands of computer vision tasks based on state-of-the-art Convolutional Neural Network (CNN) image classification far exceed…
    • FixyNN- Efficient Hardware for Mobile Computer Vision via Transfer Learning.pdf
    • over 1 year ago
    • Arm Research
    • Resources
  • Ternary Hybrid Neural-Tree Networks for Highly Constrained IoT Applications

    Charlotte Christopherson
    Charlotte Christopherson
    Dibakar Gope, Ganesh Dasika, Matthew Mattina
    Machine learning-based applications are increasingly prevalent in IoT devices. The power and storage constraints of these devices make it particularly challenging to run modern neural networks, limiting the…
    • Ternary Hybrid Neural-Tree Networks for Highly Constrained IoT Applications.pdf
    • over 1 year ago
    • Arm Research
    • Resources
  • Euphrates: Algorithm-SoC Co-Design for Low-Power Mobile Continuous Vision

    Charlotte Christopherson
    Charlotte Christopherson
    Yuhao Zhu, Anand Samajdar, Matthew Mattina, Paul Whatmough
    Continuous computer vision (CV) tasks increasingly rely on convolutional neural networks (CNN). However, CNNs have massive compute demands that far exceed the performance and energy constraints…
    • Euphrates- Algorithm-SoC Co-Design for Low-Power Mobile Continuous Vision.pdf
    • over 1 year ago
    • Arm Research
    • Resources
  • Mobile Machine Learning Hardware at Arm: A Systems-on-Chip (SoC) Perspective

    Charlotte Christopherson
    Charlotte Christopherson
    Yuhao Zhu, Matthew Mattina, Paul Whatmough
    Machine learning is playing an increasingly significant role in emerging mobile application domains such as AR/VR, ADAS, etc. Accordingly, hardware architects have designed customized hardware for machine learning…
    • Mobile Machine Learning Hardware at ARM- A Systems-on-Chip (SoC) Perspective.pdf
    • over 1 year ago
    • Arm Research
    • Resources
  • Energy Efficient Hardware for On-Device CNN Inference via Transfer Learning

    Charlotte Christopherson
    Charlotte Christopherson
    Paul Whatmough, Chuteng Zhou, Patrick Hansen, Matthew Mattina
    On-device CNN inference for real-time computer vision applications can result in computational demands that far exceed the energy budgets of mobile devices. This paper proposes FixyNN, a co-designed…
    • Energy Efficient Hardware for On-Device CNN Inference via Transfer Learning.pdf
    • over 1 year ago
    • Arm Research
    • Resources
  • SCALE-Sim: Systolic CNN Accelerator Simulator

    Charlotte Christopherson
    Charlotte Christopherson
    Ananda Samajdar, Yuhao Zhu, Paul Whatmough, Matthew Mattina, Tushar Krishna
    Systolic Arrays are one of the most popular compute substrates within Deep Learning accelerators today, as they provide extremely high efficiency for running dense matrix multiplications…
    • SCALE-Sim- Systolic CNN Accelerator Simulator.pdf
    • over 1 year ago
    • Arm Research
    • Resources
  • Efficient and Robust Machine Learning for Real-World Systems

    Charlotte Christopherson
    Charlotte Christopherson
    Franz Pernkopf, Wolfgang Roth, Matthias Zoehrer, Lukas Pfeifenberger, Guenther Schindler, Holger Froening, Sebastian Tschiatschek, Robert Peharz, Matthew Mattina, Zoubin Ghahramani
    While machine learning is traditionally a resource intensive task, embedded systems…
    • Efficient and Robust Machine Learning.pdf
    • over 1 year ago
    • Arm Research
    • Resources
  • View related content throughout Arm Research
  • More
  • Cancel