Essays about: "Quantization Aware Training"

Found 3 essays containing the words Quantization Aware Training.

  1. 1. Utilizing energy-saving techniques to reduce energy and memory consumption when training machine learning models : Sustainable Machine Learning

    University essay from KTH/Skolan för elektroteknik och datavetenskap (EECS)

    Author : Khalid El Yaacoub; [2024]
    Keywords : Sustainable AI; Machine learning; Quantization-Aware Training; Model Distillation; Quantized Distillation; Siamese Neural Networks; Continual Learning; Experience Replay; Data Efficient AI; Energy Consumption; Energy-Savings; Sustainable ML; Computation resources; Hållbar maskin inlärning; Hållbar AI; Maskininlärning; Quantization-Aware Training; Model Distillation; Quantized Distillation; siamesiska neurala nätverk; Continual Learning; Experience Replay; Dataeffektiv AI; Energiförbrukning; Energibesparingar; Beräkningsresurser;

    Abstract : Emerging machine learning (ML) techniques are showing great potential in prediction performance. However, research and development is often conducted in an environment with extensive computational resources and blinded by prediction performance. READ MORE

  2. 2. Mixed Precision Quantization for Computer Vision Tasks in Autonomous Driving

    University essay from KTH/Skolan för elektroteknik och datavetenskap (EECS)

    Author : Sri Janani Rengarajan; [2022]
    Keywords : Quantization; Neural Networks; Quantization Aware Training; Mixed precision; Semantic segmentation; Hessian; Kvantisering; Neurala nätverk; Kvantiseringsmedveten träning; Blandad precision; Semantisk segmentering; Hessian;

    Abstract : Quantization of Neural Networks is popular technique for adopting computation intensive Deep Learning applications to edge devices. In this work, low bit mixed precision quantization of FPN-Resnet18 model trained for the task of semantic segmentation is explored using Cityscapes and Arriver datasets. READ MORE

  3. 3. QPLaBSE: Quantized and Pruned Language-Agnostic BERT Sentence Embedding Model : Production-ready compression for multilingual transformers

    University essay from KTH/Skolan för elektroteknik och datavetenskap (EECS)

    Author : Sarthak Langde; [2021]
    Keywords : Transformers; LaBSE; Quantization; Pruning; PyTorch; TensorFlow; ONNX; Transformatorer; LaBSE; Kvantisering; Beskärning; PyTorch; TensorFlow; ONNX;

    Abstract : Transformer models perform well on Natural Language Processing and Natural Language Understanding tasks. Training and fine-tuning of these models consume a large amount of data and computing resources. Fast inference also requires high-end hardware for user-facing products. READ MORE