Theses and Dissertations

Advisor

Ramkumar, Mahalingam

Committee Member

Pillai, Nisha

Committee Member

Chen, Zhiqian

Date of Degree

8-7-2025

Original embargo terms

Immediate Worldwide Access

Document Type

Graduate Thesis - Open Access

Major

Computer Science

Degree Name

Master of Science (M.S.)

College

James Worth Bagley College of Engineering

Department

Department of Computer Science and Engineering

Abstract

This thesis introduces a novel interpretability framework for transformer encoder models by hypothesizing that their internal embedding updates define a state transition system. We propose that transformers implicitly learn token-to-token influence dynamics, which can be analyzed using graph-theoretic methods. To validate this, we construct transition graphs from embedding changes and rank token importance using the PageRank algorithm. We develop two transformer models from scratch: a self-supervised model that incorporates serotype tokens to learn contextualized embeddings, and a supervised model initialized with these embeddings. Cosine similarity analysis of their token influence patterns reveals strong structural alignment, with values exceeding 93%. This architecture-aware approach reframes attention as a dynamic transition process, providing deeper insight into how information propagates within transformer models. Our findings high- light the potential of graph-based analysis for uncovering interpretable structures in modern deep learning architectures.

Sponsorship (Optional)

USDA

Share

COinS