Pritom Saha

PhD Student · Computer Science

Pritom Saha

University of Texas at Dallas  ·  Dallas, Texas

I am a PhD student in Computer Science at UT Dallas with expertise in deep learning and NLP. My research focuses on uncertainty quantification and mechanistic interpretability for large language models — understanding how LLMs represent knowledge and quantifying their confidence to make them more trustworthy and interpretable. I am also working on improving model performance in resource-constrained environments.

Currently a Research Assistant at UT Dallas, contributing to projects funded by NIST and NSF. Previously, I worked as an ML engineer developing Bangla OCR systems, knowledge mining pipelines with LLMs, and deploying multi-framework models at scale with NVIDIA Triton. I also served as a Teaching Assistant at UT Dallas covering Big Data Management, Programming Language Paradigms, and Database Design.

News & Updates

Jan 2026

Started as Research Assistant at UT Dallas, working on projects funded by NIST, NSF, and other government organizations.

May 2025

Paper published at AISTATS 2025: Evidential Uncertainty Probes for Graph Neural Networks (28th International Conference on Artificial Intelligence and Statistics).

May 2025

Completed M.Sc. in Computer Science (Intelligent Systems Track) at the University of Texas at Dallas. GPA: 3.85 / 4.00.

Aug 2023

Started PhD in Computer Science at the University of Texas at Dallas. Research focus: Uncertainty Quantification and Evidential Deep Learning.

Dec 2022

Paper accepted at ICPR 2022: Rethinking Task-Incremental Learning Baselines (26th International Conference on Pattern Recognition, IEEE).

Research

Uncertainty Quantification & Interpretability

Research on uncertainty quantification and mechanistic interpretability for large language models. Investigating how LLMs internally represent and process knowledge, and developing methods to better calibrate their confidence, making them safer and more transparent in real-world use.

Uncertainty Quantification Mechanistic Interpretability LLMs Evidential Deep Learning

Resource-Efficient AI

Exploring techniques to improve model performance in resource-constrained environments including model compression and distillation techniques to enable capable AI in settings with limited compute, memory, or energy budgets.

Model Compression Quantization Efficient Inference LLMs

OCR & Document Analysis

Developed novel grapheme representation methods for Bangla OCR addressing the challenge of conjunct characters. Created large-scale synthetic datasets and benchmark protocols for Bangla script recognition.

Bangla OCR CRNN Handwriting Recognition Synthetic Data

Publications

AISTATS 2025 Conference

Evidential Uncertainty Probes for Graph Neural Networks

The 28th International Conference on Artificial Intelligence and Statistics (AISTATS) · 2025

Yu, L., Li, K., Saha, P. K., Lou, Y., & Chen, F.

Accurate quantification of both aleatoric and epistemic uncertainties is essential when deploying Graph Neural Networks (GNNs) in high-stakes applications such as drug discovery and financial fraud detection, where reliable predictions are critical. Although Evidential Deep Learning (EDL) efficiently quantifies uncertainty using a Dirichlet distribution over predictive probabilities, existing EDL-based GNN (EGNN) models require modifications to the network architecture and retraining, failing to take advantage of pre-trained models. We propose a plug-and-play framework for uncertainty quantification in GNNs that works with pre-trained models without the need for retraining. Our Evidential Probing Network (EPN) uses a lightweight Multi-Layer-Perceptron (MLP) head to extract evidence from learned representations, allowing efficient integration with various GNN architectures. We further introduce evidence-based regularization techniques, referred to as EPN-reg, to enhance the estimation of epistemic uncertainty with theoretical justifications. Extensive experiments demonstrate that the proposed EPN-reg achieves state-of-the-art performance in accurate and efficient uncertainty quantification, making it suitable for real-world deployment.

IJDAR 2024 Journal

A Multifaceted Evaluation of Representation of Graphemes for Practically Effective Bangla OCR

International Journal on Document Analysis and Recognition (IJDAR), Springer · 2024

Roy, K., Hossain, M. S., Saha, P. K., Rohan, S., Ashrafi, I., et al.

Bangla Optical Character Recognition (OCR) poses a unique challenge due to the presence of hundreds of diverse conjunct characters formed by the combination of two or more letters. In this paper, we propose two novel grapheme representation methods that improve the recognition of these conjunct characters and the overall performance of OCR in Bangla. We have utilized the popular Convolutional Recurrent Neural Network architecture and implemented our grapheme representation strategies to design the final labels of the model. Due to the absence of a large-scale Bangla word-level printed dataset, we created a synthetically generated Bangla corpus containing 2 million samples that are representative and sufficiently varied in terms of fonts, domain, and vocabulary size to train our Bangla OCR model. To test the various aspects of our model, we have also created 6 test protocols. Finally, to establish the generalizability of our grapheme representation methods, we have performed training and testing on external handwriting datasets. Experimental results proved the effectiveness of our novel approach. Furthermore, our synthetically generated training dataset and the test protocols are made available to serve as benchmarks for future Bangla OCR research.

ICPR 2022 Conference

Rethinking Task-Incremental Learning Baselines

26th International Conference on Pattern Recognition (ICPR), IEEE · 2022

Sazzad Hossain, M., Saha, P., Faisal Chowdhury, T., Rahman, S., Rahman, F., & Mohammed, N.

It is common to have continuous streams of new data that need to be introduced in the system in real-world applications. The model needs to learn newly added capabilities (future tasks) while retaining old knowledge (past tasks). In this study, we present a simple yet effective adjustment network (SAN) for task-incremental learning that achieves near state-of-the-art performance while using minimal architectural size without using memory instances. We investigate this approach on both 3D point cloud object (ModelNet40) and 2D image (CIFAR10, CIFAR100, MiniImageNet, MNIST, PermutedMNIST, notMNIST, SVHN, and FashionMNIST) recognition tasks and establish a strong baseline result for a fair comparison with existing methods. On both 2D and 3D domains, we observe that SAN is primarily unaffected by different task orders in a task-incremental setting.

Springer 2022 Book Chapter

Analysis of Bangla Keyboard Layouts Based on Keystroke Dynamics

Intelligent Systems and Sustainable Computing, Springer Nature · 2022

Rohan, S., Roy, K., Saha, P. K., Hossain, S., Rahman, F., & Mohammed, N.

Keyboards are the primary devices for interaction with computer platforms and have been a central topic in HCI research for decades. The study of their layout design is important in deciding their efficiency, practicality and adoption. Multiple keyboard layouts have been developed for Bangla without any rigorous study for their comparison. In this paper, we take a quantitative data-driven approach to compare their efficiency. Our evaluation strategy is based on the key-pair stroke timing with data collected from standard QWERTY English keyboards. Our experiments conclude that the Bijoy keyboard layout is the most efficient design for Bangla among the four layouts studied. This quantitative approach can lay the groundwork for further study of these layouts based on other criteria.

Experience

Research Assistant

University of Texas at Dallas

Jan. 2026 – Present
  • Working on multiple research projects funded by the National Institute of Standards and Technology (NIST), National Science Foundation (NSF), and other prestigious government organizations.

Teaching Assistant

University of Texas at Dallas

Aug. 2023 – Dec. 2025
  • Served as TA for the Big Data Management and Analytics, Programming Language Paradigms, and Database Design courses.
  • Held weekly office hours and graded assignments.

Software Engineer (AI/ML)

REEA Digital Ltd.

Apr. 2023 – Jul. 2023
  • Developed a knowledge mining solution utilizing Azure Cognitive Services and the OpenAI GPT model API.

Machine Learning Engineer

Apurba Technologies Ltd.

Sep. 2021 – Mar. 2023
  • Trained and developed novel Bangla word-level OCR and Handwriting Recognition (OHR) deep learning models as part of the EBLICT project under the Ministry of ICT, Bangladesh.
  • Deployed multi-framework models for inference serving on CPU and GPU servers using NVIDIA Triton Inference Server, FastAPI, and Flask.
  • Conducted document denoising work to improve word segmentation and recognition accuracy.
  • Contributed to a Bangla Text-to-Speech deep learning model and an NLP project for medical note similarity using state-of-the-art language models.

Research Assistant

Department of ECE, North South University

Dec. 2020 – Aug. 2021
  • Explored deep learning techniques using synthetic data to address real-world data scarcity.
  • Developed and studied quantized models for reduced memory footprint and faster CPU inference.

Undergraduate Teaching Assistant

Department of ECE, North South University

Feb. 2020 – Sep. 2021
  • TA for Object-Oriented Programming and Microprocessor Interfacing & Embedded Systems courses.
  • Held weekly office hours and graded assignments.

Education

Ph.D. in Computer Science
University of Texas at Dallas
Fall 2023 – Present
GPA: 3.85 / 4.00
  • Uncertainty Quantification: Research on UQ for convolutional and graph neural networks using evidential deep learning.
  • Mechanistic Interpretability: Investigating how neural networks internally represent and process information to improve model transparency.
  • Cybersecurity (NIST-funded): Developed an active learning framework for cybersecurity threat detection.
  • HAYSTAC (IARPA-funded): Applied constraint optimization techniques for optimal path routing.
M.Sc. in Computer Science (Intelligent Systems Track)
University of Texas at Dallas
Fall 2023 – Spring 2025
GPA: 3.85 / 4.00
  • Relevant Coursework: Machine Learning, Artificial Intelligence, Natural Language Processing, Statistics for AI and ML.
B.Sc. in Computer Science & Engineering
North South University
Spring 2017 – Spring 2021
CGPA: 3.74 / 4.00 Magna Cum Laude
  • Thesis: Speaker verification system using deep learning from raw audio waveform.
  • Projects: Sketch Colorization (GAN), Ashbab e-commerce with AR, Line Follower Robot.

Skills

Certifications

Mathematics for Machine Learning & Data Science
DeepLearning.AI — Coursera · Jul 2023
Machine Learning Specialization
DeepLearning.AI & Stanford Online — Coursera · Feb 2023
Deep Learning Specialization
DeepLearning.AI — Coursera · Sep 2020

Extracurricular

Vice Chair, Robotics & Automation Society
IEEE NSU Student Branch Chapter
Aug 2020 – Dec 2021
Content Writing Coordinator
IEEE North South University Student Branch
  • Publication Coordinator, IEEE Student Professional Awareness Conference (SPAC) 2019.
  • Content Writing Coordinator, IEEE Student Transition and Elevation Partnership (STEP) 2019.
Jan 2019 – Aug 2020

Awards

Finalist — Robi Datathon 2.0 2022
Champion — Whoosh!! Without Wheels, IEEE NSU Student Branch 2019
Runner Up — IoT For Tomorrow, IEEE NSU Student Branch 2019
1st Runner Up — Electrathon, IEEE NSU Student Branch 2018
3rd Place — Quiz Competition, Transparency International 2012

Interests

When not in the lab, I enjoy reading science fiction and popular science books, watching films and anime series, and exploring new places. I'm a fan of speculative fiction and love tracking my watch history.

Get in Touch

Feel free to reach out for research collaborations, academic inquiries, or a conversation about AI and deep learning.

pritom.saha@utdallas.edu

© 2025 Pritom Saha