PhD Student · Computer Science
University of Texas at Dallas · Dallas, Texas
I am a PhD student in Computer Science at UT Dallas with expertise in deep learning and NLP. My research focuses on uncertainty quantification and mechanistic interpretability for large language models — understanding how LLMs represent knowledge and quantifying their confidence to make them more trustworthy and interpretable. I am also working on improving model performance in resource-constrained environments.
Currently a Research Assistant at UT Dallas, contributing to projects funded by NIST and NSF. Previously, I worked as an ML engineer developing Bangla OCR systems, knowledge mining pipelines with LLMs, and deploying multi-framework models at scale with NVIDIA Triton. I also served as a Teaching Assistant at UT Dallas covering Big Data Management, Programming Language Paradigms, and Database Design.
Started as Research Assistant at UT Dallas, working on projects funded by NIST, NSF, and other government organizations.
Paper published at AISTATS 2025: Evidential Uncertainty Probes for Graph Neural Networks (28th International Conference on Artificial Intelligence and Statistics).
Completed M.Sc. in Computer Science (Intelligent Systems Track) at the University of Texas at Dallas. GPA: 3.85 / 4.00.
Paper published: A Multifaceted Evaluation of Representation of Graphemes for Practically Effective Bangla OCR in IJDAR (Springer, 2024).
Started PhD in Computer Science at the University of Texas at Dallas. Research focus: Uncertainty Quantification and Evidential Deep Learning.
Paper accepted at ICPR 2022: Rethinking Task-Incremental Learning Baselines (26th International Conference on Pattern Recognition, IEEE).
Research on uncertainty quantification and mechanistic interpretability for large language models. Investigating how LLMs internally represent and process knowledge, and developing methods to better calibrate their confidence, making them safer and more transparent in real-world use.
Exploring techniques to improve model performance in resource-constrained environments including model compression and distillation techniques to enable capable AI in settings with limited compute, memory, or energy budgets.
Developed novel grapheme representation methods for Bangla OCR addressing the challenge of conjunct characters. Created large-scale synthetic datasets and benchmark protocols for Bangla script recognition.
The 28th International Conference on Artificial Intelligence and Statistics (AISTATS) · 2025
Accurate quantification of both aleatoric and epistemic uncertainties is essential when deploying Graph Neural Networks (GNNs) in high-stakes applications such as drug discovery and financial fraud detection, where reliable predictions are critical. Although Evidential Deep Learning (EDL) efficiently quantifies uncertainty using a Dirichlet distribution over predictive probabilities, existing EDL-based GNN (EGNN) models require modifications to the network architecture and retraining, failing to take advantage of pre-trained models. We propose a plug-and-play framework for uncertainty quantification in GNNs that works with pre-trained models without the need for retraining. Our Evidential Probing Network (EPN) uses a lightweight Multi-Layer-Perceptron (MLP) head to extract evidence from learned representations, allowing efficient integration with various GNN architectures. We further introduce evidence-based regularization techniques, referred to as EPN-reg, to enhance the estimation of epistemic uncertainty with theoretical justifications. Extensive experiments demonstrate that the proposed EPN-reg achieves state-of-the-art performance in accurate and efficient uncertainty quantification, making it suitable for real-world deployment.
International Journal on Document Analysis and Recognition (IJDAR), Springer · 2024
Bangla Optical Character Recognition (OCR) poses a unique challenge due to the presence of hundreds of diverse conjunct characters formed by the combination of two or more letters. In this paper, we propose two novel grapheme representation methods that improve the recognition of these conjunct characters and the overall performance of OCR in Bangla. We have utilized the popular Convolutional Recurrent Neural Network architecture and implemented our grapheme representation strategies to design the final labels of the model. Due to the absence of a large-scale Bangla word-level printed dataset, we created a synthetically generated Bangla corpus containing 2 million samples that are representative and sufficiently varied in terms of fonts, domain, and vocabulary size to train our Bangla OCR model. To test the various aspects of our model, we have also created 6 test protocols. Finally, to establish the generalizability of our grapheme representation methods, we have performed training and testing on external handwriting datasets. Experimental results proved the effectiveness of our novel approach. Furthermore, our synthetically generated training dataset and the test protocols are made available to serve as benchmarks for future Bangla OCR research.
26th International Conference on Pattern Recognition (ICPR), IEEE · 2022
It is common to have continuous streams of new data that need to be introduced in the system in real-world applications. The model needs to learn newly added capabilities (future tasks) while retaining old knowledge (past tasks). In this study, we present a simple yet effective adjustment network (SAN) for task-incremental learning that achieves near state-of-the-art performance while using minimal architectural size without using memory instances. We investigate this approach on both 3D point cloud object (ModelNet40) and 2D image (CIFAR10, CIFAR100, MiniImageNet, MNIST, PermutedMNIST, notMNIST, SVHN, and FashionMNIST) recognition tasks and establish a strong baseline result for a fair comparison with existing methods. On both 2D and 3D domains, we observe that SAN is primarily unaffected by different task orders in a task-incremental setting.
Intelligent Systems and Sustainable Computing, Springer Nature · 2022
Keyboards are the primary devices for interaction with computer platforms and have been a central topic in HCI research for decades. The study of their layout design is important in deciding their efficiency, practicality and adoption. Multiple keyboard layouts have been developed for Bangla without any rigorous study for their comparison. In this paper, we take a quantitative data-driven approach to compare their efficiency. Our evaluation strategy is based on the key-pair stroke timing with data collected from standard QWERTY English keyboards. Our experiments conclude that the Bijoy keyboard layout is the most efficient design for Bangla among the four layouts studied. This quantitative approach can lay the groundwork for further study of these layouts based on other criteria.
University of Texas at Dallas
University of Texas at Dallas
REEA Digital Ltd.
Apurba Technologies Ltd.
Department of ECE, North South University
Department of ECE, North South University
When not in the lab, I enjoy reading science fiction and popular science books, watching films and anime series, and exploring new places. I'm a fan of speculative fiction and love tracking my watch history.
Feel free to reach out for research collaborations, academic inquiries, or a conversation about AI and deep learning.
pritom.saha@utdallas.edu© 2025 Pritom Saha