Search alternatives:
quantization » quantitative (Expand Search)
Showing 61 - 80 results of 314 for search 'quantization (efficient OR efficiency)', query time: 0.11s Refine Results
  1. 61

    Energy efficiency joint optimization algorithm for low-precision ADC massive MIMO systems by Haiyan CAO, Wenjuan HU, Yingjuan ZHI, Fangmin XU, Xin FANG

    Published 2020-06-01
    “…An uplink energy efficiency optimization algorithm based on alternating optimization algorithm was proposed for large-scale multiple-input multiple-output (MIMO) systems with low-precision analog to digital converters (ADC).That is,based on the maximum ratio combining (MRC) reception,an optimization model aiming at maximizing energy efficiency was established,and then the original fractional form optimization problem was converted into an equivalent subtractive form according to the nature of the fractional programming.In the case where the number of quantization bits was constant,the iterative optimization of the transmission power and the pilot length were performed.Simulation results show that the proposed algorithm has better spectral efficiency and energy efficiency.…”
    Get full text
    Article
  2. 62
  3. 63

    Evaluation of fluxon synapse device based on superconducting loops for energy efficient neuromorphic computing by Ashwani Kumar, Uday S. Goteti, Ertugrul Cubukcu, Robert C. Dynes, Duygu Kuzum

    Published 2025-02-01
    “…Here, we evaluate performance prospects of a new approach based on disordered superconducting loops with Josephson-junctions for energy efficient neuromorphic computing. Synaptic weights can be stored as internal trapped fluxon states of three superconducting loops connected with multiple Josephson-junctions (JJ) and modulated by input signals applied in the form of discrete fluxons (quantized flux) in a controlled manner. …”
    Get full text
    Article
  4. 64

    Efficient Limb Range of Motion Analysis from a Monocular Camera for Edge Devices by Xuke Yan, Linxi Zhang, Bo Liu, Guangzhi Qu

    Published 2025-01-01
    “…Our model uses a compact neural network architecture with 8-bit quantized parameters for enhanced memory efficiency and reduced latency. …”
    Get full text
    Article
  5. 65
  6. 66

    AFQSeg: An Adaptive Feature Quantization Network for Instance-Level Surface Crack Segmentation by Shaoliang Fang, Lu Lu, Zhu Lin, Zhanyu Yang, Shaosheng Wang

    Published 2025-05-01
    “…To address these issues, this paper proposes a crack detection model based on adaptive feature quantization, which primarily consists of a maximum soft pooling module, an adaptive crack feature quantization module, and a trainable crack post-processing module. …”
    Get full text
    Article
  7. 67

    NeuBridge: bridging quantized activations and spiking neurons for ANN-SNN conversion by Yuchen Yang, Jingcheng Liu, Chengting Yu, Chengyi Yang, Gaoang Wang, Aili Wang

    Published 2025-01-01
    “…Spiking neural networks (SNNs) offer a promising avenue for energy-efficient computations on neuromorphic hardware, leveraging the unique advantages of spike-based signaling. …”
    Get full text
    Article
  8. 68

    Decentralized non-convex online optimization with adaptive momentum estimation and quantized communication by Yunshan Lv, Hailing Xiong, Fuqing Zhang, Shengying Dong, Xiangguang Dai

    Published 2025-03-01
    “…To solve the problem over a communication-efficient manner, we propose a novel quantized decentralized adaptive momentum gradient descent algorithm based on the adaptive momentum estimation methods, where quantified information is exchanged between agents. …”
    Get full text
    Article
  9. 69

    Randomized Quantization for Privacy in Resource Constrained Machine Learning at-the-Edge and Federated Learning by Ce Feng, Parv Venkitasubramaniam

    Published 2025-01-01
    “…Through rigorous theoretical analysis and extensive experiments on benchmark datasets, we demonstrate that these methods significantly enhance the utility-privacy trade-off and computational efficiency in both ML-at-the-edge and FL systems. RQP-SGD is evaluated on MNIST and the Breast Cancer Diagnostic dataset, showing an average 10.62% utility improvement over the deterministic quantization-based projected DP-SGD while maintaining (1.0, 0)-DP. …”
    Get full text
    Article
  10. 70

    Convolution Smooth: A Post-Training Quantization Method for Convolutional Neural Networks by Yongyuan Chen, Zhendao Wang

    Published 2025-01-01
    “…Convolutional neural network (CNN) quantization is an efficient model compression technique primarily used for accelerating inference and optimizing resources. …”
    Get full text
    Article
  11. 71

    Lost-minimum post-training parameter quantization method for convolutional neural network by Fan ZHANG, Yun HUANG, Zizhuo FANG, Wei GUO

    Published 2022-04-01
    “…To solve the problem that that no dataset is available for model quantization in data-sensitive scenarios, a model quantization method without using data sets was proposed.Firstly, according to the parameters of batch normalized layer and the distribution characteristics of image data, the simulated input data was obtained by error minimization method.Then, by studying the characteristics of data rounding, a factor dynamic rounding method based on loss minimization was proposed.Through quantitative experiments on classification models such as GhostNet and target detection models such as M2Det, the effectiveness of the proposed quantification method for image classification and target detection models was verified.The experimental results show that the proposed quantization method can reduce the model size by about 75%, effectively reduce the power loss and improve the computing efficiency while basically maintaining the accuracy of the original model.…”
    Get full text
    Article
  12. 72

    Hierarchical Mixed-Precision Post-Training Quantization for SAR Ship Detection Networks by Hang Wei, Zulin Wang, Yuanhan Ni

    Published 2024-10-01
    “…However, limited satellite platform resources present a significant challenge. Post-training quantization (PTQ) provides an efficient method for pre-training neural networks to effectively reduce memory and computational resources without retraining. …”
    Get full text
    Article
  13. 73

    A Novel Method to Improve the Efficiency and Performance of Cloud-Based Visual Simultaneous Localization and Mapping by Omar M. Salih, Hussam Rostum, József Vásárhelyi

    Published 2024-11-01
    “…This paper proposes a novel solution to improve the efficiency and performance of exchanging data between the unmanned aerial vehicle (UAV) and the cloud server. …”
    Get full text
    Article
  14. 74

    CSCP-YOLO: A Lightweight and Efficient Algorithm for Real-Time Steel Surface Defect Detection by Chenglong Wang, Heng Wang, Yimin Jiang, Lei Yu, Xueting Wang

    Published 2025-01-01
    “…To accelerate defect detection, the SPPF-DW module efficiently fuses multi-scale defect features through optimized computational pathways. …”
    Get full text
    Article
  15. 75

    Fully Quantized Matrix Arithmetic-Only BERT Model and Its FPGA-Based Accelerator by Hiroshi Fuketa, Toshihiro Katashita, Yohei Hori, Masakazu Hioki

    Published 2025-01-01
    “…In this paper, we propose a fully quantized matrix arithmetic-only BERT (FQ MA-BERT) model to enable efficient natural language processing. …”
    Get full text
    Article
  16. 76

    Research of channel quantization and feedback strategies based on multiuser diversity MIMO-OFDM systems by LIANG Xue-jun, ZHU Guang-xi, SU Gang, WANG De-sheng

    Published 2009-01-01
    “…Firstly, a quantization method was proposed by quantized value indicating the modulation level instead of the full values of channel quality information(CQI) and the achievable average spectrum efficiency showed no loss compared with perfect case.Secondly, employment of the integrated design that combined with opportunistic, best, and hybrid feedback scheme was considered and the close-form expression of average spectrum efficiency was deduced in various case.Finally, the calculation of optimal feedback parameters was confirmed from two aspects of feedback channel capacity and capacity relative loss.Extensive simulations were presented to evaluate these proposed strategies.The results match with the numeral analysis very well.The proposed partial feedback schemes can reduce the feedback load greatly with the same system capability, only if the feedback parameters be chosen properly.Wherein, the hybrid feedback combined with quantization performs best and provides the instruction to design the channel feedback of practical systems.…”
    Get full text
    Article
  17. 77

    Optimizing Deep Learning Models for Resource‐Constrained Environments With Cluster‐Quantized Knowledge Distillation by Niaz Ashraf Khan, A. M. Saadman Rafat

    Published 2025-05-01
    “…These results highlight the feasibility of CQKD for efficient deep learning model deployment in low‐resource environments.…”
    Get full text
    Article
  18. 78

    Study of algorithmic approaches to digital signal filtering and the influence of input quantization on output accuracy by Shermuradova Malika, Gadoeva Mavlyuda, Rahmatov Shukhrat, Abdullaev Uktamjon, Aralov Dilshod

    Published 2025-01-01
    “…The research supports the broader integration of AI-driven technologies in modern automation systems, paving the way for more adaptive, efficient, and fault-tolerant control mechanisms in complex environments.…”
    Get full text
    Article
  19. 79

    Smoothed per-tensor weight quantization: a robust solution for neural network deployment by Xin Chang

    Published 2025-07-01
    “…This paper introduces a novel method to improve quantization outcomes for per-tensor weight quantization, focusing on enhancing computational efficiency and compatibility with resource-constrained hardware. …”
    Get full text
    Article
  20. 80

    Two-dimensional materials based two-transistor-two-resistor synaptic kernel for efficient neuromorphic computing by Qian He, Hailiang Wang, Yishu Zhang, Anzhe Chen, Yu Fu, Guodong Xue, Kaihui Liu, Shiman Huang, Yang Xu, Bin Yu

    Published 2025-05-01
    “…Here, we develop a 16 × 16 computing kernel based on two-transistor-two-resistor unit with three-dimensional heterogeneous integration compatibility to boost energy efficiency and computing performance. We demonstrate the 4-bit weight characteristics of artificial synapses with low stochasticity. …”
    Get full text
    Article