Essa Jan

Essa Jan

1st Year Masters @ Brown University

I am a first year MS CS student at Brown University, advised by Stephen Bach. In my previous, truly amazing life, I completed my undergrad in CS at the LUMS, where I worked on the alignment of language models and was fortunate to be advised by Dr. Fareed Zaffar, Dr. Yasir Zaki, and Faizan Ahmed.

I am interested in building safer and generalizable AI models. My research interests span several areas within AI safety and model understanding, including reasoning, control mechanisms and the ability of large language models to adapt and improve over time. As these models become increasingly capable, ensuring that they generalize reliably across diverse tasks and contexts is critical for robust and safe deployment. Therefore, I am particularly interested in developing methods to uncover model limitations, support adaptive learning, and steer them toward safer and more aligned behaviour.

News

August 2025
Our recent work on Evaluating the Transferability of Injected Knowledge in LLMs made it into the EMNLP Findings track.
May 2025
Working on understanding the steering of the model behaviour.
Dec 2024
Our paper on MultitaskBench was accepted at COLING 25, focusing on safety gaps in LLM fine-tuning.

Research

MultitaskBench

MultitaskBench: Unveiling and Mitigating Safety Gaps in LLMs Fine-tuning

*Essa Jan, *Nouar Aldahoul, Moiz Ali, Faizan Ahmad, Fareed Zaffar, Yasir Zaki

2025 • COLING 25

Investigated how fine-tuning on downstream tasks affects the safety guardrails of large language models. We develop a comprehensive benchmark to evaluate safety degradation and the robustness of different safety solution across multiple task domains.

Knowledge Injection

Data Doping or True Intelligence? Evaluating the Transferability of Injected Knowledge in LLMs

*Essa Jan, *Moiz Ali, Fareed Zaffar, Yasir Zaki

2025 • Under Review

We evaluate how LLMs internalize and transfer knowledge injected during fine-tuning through surface-level and deep comprehension tasks. Our study explores whether model size and task demands impact the depth and generalizability of learned knowledge.

YovaCLIP

YovaCLIP: Multimodal Framework for Forest Fire Detection

Essa Jan, Zubair Anwar, Murtaza Taj

2024 • Collaboration with WWF

A multimodal pipeline combining vision language models for real-time early detection of smoke and fire in forest environments. Currently deployed across 600+ square kilometer of forest, processed 10,000+ hours of live footage.

Projects

TradeSnap.ai

2024 • LLM-Based Trading Platform

A LLM based stock trading platform that reduces the entire trading process to a single multilingual conversational interface, significantly lowering barriers to entry for retail traders especially in the developing world.

Urban Power Consumption and Poverty Mapping

2023 • Machine Learning for Grid Management

Scrapped data and developed ML models to analyze electricity consumption of 3 million consumers, predicting feeder overloads enhanced grid management and identifying poverty hot spots.

SafeNet: AI-Powered Content Moderation

2023 • Social Media Monitoring Tool

An LLM powered content moderation system to identify and flag harmful/toxic social media content. Built using Kaggle Jigsaw challenge dataset.

Teaching

CS-334: Principles and Techniques of Data Science

Spring 2025 Teaching Assistant

AI 500: Foundations of AI

Fall 2024 Graduate Course Teaching Assistant

CS100: Computational Problem Solving

Spring 2024 Head Teaching Assistant