-
ArrhythmiaVision: Resource-Conscious Deep Learning Models with Visual Explanations for ECG Arrhythmia Classification
Authors:
Zuraiz Baig,
Sidra Nasir,
Rizwan Ahmed Khan,
Muhammad Zeeshan Ul Haque
Abstract:
Cardiac arrhythmias are a leading cause of life-threatening cardiac events, highlighting the urgent need for accurate and timely detection. Electrocardiography (ECG) remains the clinical gold standard for arrhythmia diagnosis; however, manual interpretation is time-consuming, dependent on clinical expertise, and prone to human error. Although deep learning has advanced automated ECG analysis, many…
▽ More
Cardiac arrhythmias are a leading cause of life-threatening cardiac events, highlighting the urgent need for accurate and timely detection. Electrocardiography (ECG) remains the clinical gold standard for arrhythmia diagnosis; however, manual interpretation is time-consuming, dependent on clinical expertise, and prone to human error. Although deep learning has advanced automated ECG analysis, many existing models abstract away the signal's intrinsic temporal and morphological features, lack interpretability, and are computationally intensive-hindering their deployment on resource-constrained platforms. In this work, we propose two novel lightweight 1D convolutional neural networks, ArrhythmiNet V1 and V2, optimized for efficient, real-time arrhythmia classification on edge devices. Inspired by MobileNet's depthwise separable convolutional design, these models maintain memory footprints of just 302.18 KB and 157.76 KB, respectively, while achieving classification accuracies of 0.99 (V1) and 0.98 (V2) on the MIT-BIH Arrhythmia Dataset across five classes: Normal Sinus Rhythm, Left Bundle Branch Block, Right Bundle Branch Block, Atrial Premature Contraction, and Premature Ventricular Contraction. In order to ensure clinical transparency and relevance, we integrate Shapley Additive Explanations and Gradient-weighted Class Activation Mapping, enabling both local and global interpretability. These techniques highlight physiologically meaningful patterns such as the QRS complex and T-wave that contribute to the model's predictions. We also discuss performance-efficiency trade-offs and address current limitations related to dataset diversity and generalizability. Overall, our findings demonstrate the feasibility of combining interpretability, predictive accuracy, and computational efficiency in practical, wearable, and embedded ECG monitoring systems.
△ Less
Submitted 30 April, 2025;
originally announced May 2025.
-
A Systematic Literature Review of Parameter-Efficient Fine-Tuning for Large Code Models
Authors:
Md Zahidul Haque,
Saima Afrin,
Antonio Mastropaolo
Abstract:
The rise of Artificial Intelligence (AI)-and particularly Large Language Models (LLMs) for code-has reshaped Software Engineering (SE) by enabling the automation of tasks such as code generation, bug detection, and repair. However, these models require significant computational resources for training and fine-tuning, posing challenges for real-world adoption in resource-constrained environments. T…
▽ More
The rise of Artificial Intelligence (AI)-and particularly Large Language Models (LLMs) for code-has reshaped Software Engineering (SE) by enabling the automation of tasks such as code generation, bug detection, and repair. However, these models require significant computational resources for training and fine-tuning, posing challenges for real-world adoption in resource-constrained environments. To address this, the research community has increasingly turned to Parameter-Efficient Fine-Tuning (PEFT)-a class of techniques that enables the adaptation of large models by updating only a small subset of parameters, rather than the entire model. In this Systematic Literature Review (SLR), we examine the growing application of PEFT techniques-across a wide range of software engineering tasks. We analyze how these methods are used to optimize various deep learning (DL) architectures, focusing on their impact on both performance and efficiency. Our study synthesizes findings from 27 peer-reviewed papers, identifying patterns in configuration strategies and adaptation trade-offs. The outcome of this review is a comprehensive taxonomy that categorizes PEFT usage by task type, distinguishing between generative (e.g., Code Summarization) and non-generative (e.g., Code Clone Detection) scenarios. Our findings aim to inform future research and guide the practical deployment of PEFT in sustainable, AI-powered software development. Our artifacts are publicly available at https://github.com/alvi75/SLR-PEFT
△ Less
Submitted 9 May, 2025; v1 submitted 29 April, 2025;
originally announced April 2025.
-
BD Currency Detection: A CNN Based Approach with Mobile App Integration
Authors:
Syed Jubayer Jaman,
Md. Zahurul Haque,
Md Robiul Islam,
Usama Abdun Noor
Abstract:
Currency recognition plays a vital role in banking, commerce, and assistive technology for visually impaired individuals. Traditional methods, such as manual verification and optical scanning, often suffer from limitations in accuracy and efficiency. This study introduces an advanced currency recognition system utilizing Convolutional Neural Networks (CNNs) to accurately classify Bangladeshi bankn…
▽ More
Currency recognition plays a vital role in banking, commerce, and assistive technology for visually impaired individuals. Traditional methods, such as manual verification and optical scanning, often suffer from limitations in accuracy and efficiency. This study introduces an advanced currency recognition system utilizing Convolutional Neural Networks (CNNs) to accurately classify Bangladeshi banknotes. A dataset comprising 50,334 images was collected, preprocessed, and used to train a CNN model optimized for high performance classification. The trained model achieved an accuracy of 98.5%, surpassing conventional image based currency recognition approaches. To enable real time and offline functionality, the model was converted into TensorFlow Lite format and integrated into an Android mobile application. The results highlight the effectiveness of deep learning in currency recognition, providing a fast, secure, and accessible solution that enhances financial transactions and assistive technologies.
△ Less
Submitted 25 February, 2025;
originally announced February 2025.
-
Multi-Class Plant Leaf Disease Detection: A CNN-based Approach with Mobile App Integration
Authors:
Md Aziz Hosen Foysal,
Foyez Ahmed,
Md Zahurul Haque
Abstract:
Plant diseases significantly impact agricultural productivity, resulting in economic losses and food insecurity. Prompt and accurate detection is crucial for the efficient management and mitigation of plant diseases. This study investigates advanced techniques in plant disease detection, emphasizing the integration of image processing, machine learning, deep learning methods, and mobile technologi…
▽ More
Plant diseases significantly impact agricultural productivity, resulting in economic losses and food insecurity. Prompt and accurate detection is crucial for the efficient management and mitigation of plant diseases. This study investigates advanced techniques in plant disease detection, emphasizing the integration of image processing, machine learning, deep learning methods, and mobile technologies. High-resolution images of plant leaves were captured and analyzed using convolutional neural networks (CNNs) to detect symptoms of various diseases, such as blight, mildew, and rust. This study explores 14 classes of plants and diagnoses 26 unique plant diseases. We focus on common diseases affecting various crops. The model was trained on a diverse dataset encompassing multiple crops and disease types, achieving 98.14% accuracy in disease diagnosis. Finally integrated this model into mobile apps for real-time disease diagnosis.
△ Less
Submitted 26 August, 2024;
originally announced August 2024.
-
E-Commerce Product Recommendation System based on ML Algorithms
Authors:
Md. Zahurul Haque
Abstract:
Algorithms are used in eCommerce product recommendation systems. These systems just recently began utilizing machine learning algorithms due to the development and growth of the artificial intelligence research community. This project aspires to transform how eCommerce platforms communicate with their users. We have created a model that can customize product recommendations and offers for each uni…
▽ More
Algorithms are used in eCommerce product recommendation systems. These systems just recently began utilizing machine learning algorithms due to the development and growth of the artificial intelligence research community. This project aspires to transform how eCommerce platforms communicate with their users. We have created a model that can customize product recommendations and offers for each unique customer using cutting-edge machine learning techniques, we used PCA to reduce features and four machine learning algorithms like Gaussian Naive Bayes (GNB), Random Forest (RF), Logistic Regression (LR), Decision Tree (DT), the Random Forest algorithms achieve the highest accuracy of 99.6% with a 96.99 r square score, 1.92% MSE score, and 0.087 MAE score. The outcome is advantageous for both the client and the business. In this research, we will examine the model's development and training in detail and show how well it performs using actual data. Learning from machines can change of eCommerce world.
△ Less
Submitted 14 July, 2024;
originally announced July 2024.
-
Bengali License Plate Recognition: Unveiling Clarity with CNN and GFP-GAN
Authors:
Noushin Afrin,
Md Mahamudul Hasan,
Mohammed Fazlay Elahi Safin,
Khondakar Rifat Amin,
Md Zahidul Haque,
Farzad Ahmed,
Md. Tanvir Rouf Shawon
Abstract:
Automated License Plate Recognition(ALPR) is a system that automatically reads and extracts data from vehicle license plates using image processing and computer vision techniques. The Goal of LPR is to identify and read the license plate number accurately and quickly, even under challenging, conditions such as poor lighting, angled or obscured plates, and different plate fonts and layouts. The pro…
▽ More
Automated License Plate Recognition(ALPR) is a system that automatically reads and extracts data from vehicle license plates using image processing and computer vision techniques. The Goal of LPR is to identify and read the license plate number accurately and quickly, even under challenging, conditions such as poor lighting, angled or obscured plates, and different plate fonts and layouts. The proposed method consists of processing the Bengali low-resolution blurred license plates and identifying the plate's characters. The processes include image restoration using GFPGAN, Maximizing contrast, Morphological image processing like dilation, feature extraction and Using Convolutional Neural Networks (CNN), character segmentation and recognition are accomplished. A dataset of 1292 images of Bengali digits and characters was prepared for this project.
△ Less
Submitted 17 December, 2023;
originally announced December 2023.
-
JutePestDetect: An Intelligent Approach for Jute Pest Identification Using Fine-Tuned Transfer Learning
Authors:
Md. Simul Hasan Talukder,
Mohammad Raziuddin Chowdhury,
Md Sakib Ullah Sourav,
Abdullah Al Rakin,
Shabbir Ahmed Shuvo,
Rejwan Bin Sulaiman,
Musarrat Saberin Nipun,
Muntarin Islam,
Mst Rumpa Islam,
Md Aminul Islam,
Zubaer Haque
Abstract:
In certain Asian countries, Jute is one of the primary sources of income and Gross Domestic Product (GDP) for the agricultural sector. Like many other crops, Jute is prone to pest infestations, and its identification is typically made visually in countries like Bangladesh, India, Myanmar, and China. In addition, this method is time-consuming, challenging, and somewhat imprecise, which poses a subs…
▽ More
In certain Asian countries, Jute is one of the primary sources of income and Gross Domestic Product (GDP) for the agricultural sector. Like many other crops, Jute is prone to pest infestations, and its identification is typically made visually in countries like Bangladesh, India, Myanmar, and China. In addition, this method is time-consuming, challenging, and somewhat imprecise, which poses a substantial financial risk. To address this issue, the study proposes a high-performing and resilient transfer learning (TL) based JutePestDetect model to identify jute pests at the early stage. Firstly, we prepared jute pest dataset containing 17 classes and around 380 photos per pest class, which were evaluated after manual and automatic pre-processing and cleaning, such as background removal and resizing. Subsequently, five prominent pre-trained models -DenseNet201, InceptionV3, MobileNetV2, VGG19, and ResNet50 were selected from a previous study to design the JutePestDetect model. Each model was revised by replacing the classification layer with a global average pooling layer and incorporating a dropout layer for regularization. To evaluate the models performance, various metrics such as precision, recall, F1 score, ROC curve, and confusion matrix were employed. These analyses provided additional insights for determining the efficacy of the models. Among them, the customized regularized DenseNet201-based proposed JutePestDetect model outperformed the others, achieving an impressive accuracy of 99%. As a result, our proposed method and strategy offer an enhanced approach to pest identification in the case of Jute, which can significantly benefit farmers worldwide.
△ Less
Submitted 28 May, 2023;
originally announced August 2023.
-
Efficient approach of using CNN based pretrained model in Bangla handwritten digit recognition
Authors:
Muntarin Islam,
Shabbir Ahmed Shuvo,
Musarrat Saberin Nipun,
Rejwan Bin Sulaiman,
Jannatul Nayeem,
Zubaer Haque,
Md Mostak Shaikh,
Md Sakib Ullah Sourav
Abstract:
Due to digitalization in everyday life, the need for automatically recognizing handwritten digits is increasing. Handwritten digit recognition is essential for numerous applications in various industries. Bengali ranks the fifth largest language in the world with 265 million speakers (Native and non-native combined) and 4 percent of the world population speaks Bengali. Due to the complexity of Ben…
▽ More
Due to digitalization in everyday life, the need for automatically recognizing handwritten digits is increasing. Handwritten digit recognition is essential for numerous applications in various industries. Bengali ranks the fifth largest language in the world with 265 million speakers (Native and non-native combined) and 4 percent of the world population speaks Bengali. Due to the complexity of Bengali writing in terms of variety in shape, size, and writing style, researchers did not get better accuracy using Supervised machine learning algorithms to date. Moreover, fewer studies have been done on Bangla handwritten digit recognition (BHwDR). In this paper, we proposed a novel CNN-based pre-trained handwritten digit recognition model which includes Resnet-50, Inception-v3, and EfficientNetB0 on NumtaDB dataset of 17 thousand instances with 10 classes.. The Result outperformed the performance of other models to date with 97% accuracy in the 10-digit classes. Furthermore, we have evaluated the result or our model with other research studies while suggesting future study
△ Less
Submitted 19 September, 2022;
originally announced September 2022.
-
Traffic model of LTE using maximum flow algorithm with binary search technique
Authors:
Md. Zahurul Haque,
Md. Rafiqul Isla
Abstract:
Inrecent time a rapid increase in the number of smart devices and user applications have generated an intensity volume of data traffic from/to a cellular network. So the Long Term Evaluation(LTE)network is facing some issuesdifficulties ofthebase station and infrastructure in terms of upgrade and configuration becausethere is no concept of BSC (Base Station Controller) of 2G and RNC (Radio Network…
▽ More
Inrecent time a rapid increase in the number of smart devices and user applications have generated an intensity volume of data traffic from/to a cellular network. So the Long Term Evaluation(LTE)network is facing some issuesdifficulties ofthebase station and infrastructure in terms of upgrade and configuration becausethere is no concept of BSC (Base Station Controller) of 2G and RNC (Radio Network Controller) of 3G to control several BTS/NB. Only 4G (LTE) all the eNBs areinterconnected for traffic flow from UE (user equipment) to core switch. Determination of capacity of a linkof such a network is a challenging job since each node offers its own traffic andat the same time conveys traffic of other nodes.In this paper, we apply maximum flow algorithm including the binary search techniqueto solve the traffic flow of radio networkandinterconnected eNBs of the LTE network. The throughput of the LTE network shown graphically under the QPSK and 16-QAM
△ Less
Submitted 28 September, 2020;
originally announced September 2020.
-
TensorFlow Estimators: Managing Simplicity vs. Flexibility in High-Level Machine Learning Frameworks
Authors:
Heng-Tze Cheng,
Zakaria Haque,
Lichan Hong,
Mustafa Ispir,
Clemens Mewald,
Illia Polosukhin,
Georgios Roumpos,
D Sculley,
Jamie Smith,
David Soergel,
Yuan Tang,
Philipp Tucker,
Martin Wicke,
Cassandra Xia,
Jianwei Xie
Abstract:
We present a framework for specifying, training, evaluating, and deploying machine learning models. Our focus is on simplifying cutting edge machine learning for practitioners in order to bring such technologies into production. Recognizing the fast evolution of the field of deep learning, we make no attempt to capture the design space of all possible model architectures in a domain- specific lang…
▽ More
We present a framework for specifying, training, evaluating, and deploying machine learning models. Our focus is on simplifying cutting edge machine learning for practitioners in order to bring such technologies into production. Recognizing the fast evolution of the field of deep learning, we make no attempt to capture the design space of all possible model architectures in a domain- specific language (DSL) or similar configuration language. We allow users to write code to define their models, but provide abstractions that guide develop- ers to write models in ways conducive to productionization. We also provide a unifying Estimator interface, making it possible to write downstream infrastructure (e.g. distributed training, hyperparameter tuning) independent of the model implementation. We balance the competing demands for flexibility and simplicity by offering APIs at different levels of abstraction, making common model architectures available out of the box, while providing a library of utilities designed to speed up experimentation with model architectures. To make out of the box models flexible and usable across a wide range of problems, these canned Estimators are parameterized not only over traditional hyperparameters, but also using feature columns, a declarative specification describing how to interpret input data. We discuss our experience in using this framework in re- search and production environments, and show the impact on code health, maintainability, and development speed.
△ Less
Submitted 8 August, 2017;
originally announced August 2017.
-
Wide & Deep Learning for Recommender Systems
Authors:
Heng-Tze Cheng,
Levent Koc,
Jeremiah Harmsen,
Tal Shaked,
Tushar Chandra,
Hrishi Aradhye,
Glen Anderson,
Greg Corrado,
Wei Chai,
Mustafa Ispir,
Rohan Anil,
Zakaria Haque,
Lichan Hong,
Vihan Jain,
Xiaobing Liu,
Hemal Shah
Abstract:
Generalized linear models with nonlinear feature transformations are widely used for large-scale regression and classification problems with sparse inputs. Memorization of feature interactions through a wide set of cross-product feature transformations are effective and interpretable, while generalization requires more feature engineering effort. With less feature engineering, deep neural networks…
▽ More
Generalized linear models with nonlinear feature transformations are widely used for large-scale regression and classification problems with sparse inputs. Memorization of feature interactions through a wide set of cross-product feature transformations are effective and interpretable, while generalization requires more feature engineering effort. With less feature engineering, deep neural networks can generalize better to unseen feature combinations through low-dimensional dense embeddings learned for the sparse features. However, deep neural networks with embeddings can over-generalize and recommend less relevant items when the user-item interactions are sparse and high-rank. In this paper, we present Wide & Deep learning---jointly trained wide linear models and deep neural networks---to combine the benefits of memorization and generalization for recommender systems. We productionized and evaluated the system on Google Play, a commercial mobile app store with over one billion active users and over one million apps. Online experiment results show that Wide & Deep significantly increased app acquisitions compared with wide-only and deep-only models. We have also open-sourced our implementation in TensorFlow.
△ Less
Submitted 24 June, 2016;
originally announced June 2016.