AI Engineer & Researcher

Building intelligent
systems that matter

I'm Amal Joe R S — an AI engineer specializing in LLMs, multimodal models, and reinforcement learning. Currently finishing my M.Tech at IIT Bombay with research at IBM and collaborations with Meta and BharatGen.

Current CPI 9.77 / 10
Focus LLMs & VLMs
Based in Mumbai, India
01

Education

M.Tech, CSE
Indian Institute of Technology Bombay
2024 — 2026
9.77
CPI out of 10
B.Tech, CSE
Mar Baselios College of Engineering (APJ Abdul Kalam TU)
2019 — 2023
9.31
CGPA out of 10
02

Experience

Starting Jul 2026
AI Research Engineer
IBM Research
Nov 2025 — Present
Freelance AI Engineer
Financial Reporting & Agentic AI
Working with a company that helps MNCs file annual financial reports globally. Integrating agentic AI into existing workflows to automate iXBRL tagging and fine-tuning custom models for in-house use cases.
May 2025 — Aug 2025
AI Research Intern
IBM Research
Designed a dynamic data sampling strategy for multi-lingual multi-task LLM training. Built the "Online Data Loader" — a novel PyTorch framework for dynamic data mixing, achieving 5–10% improvement over state-of-the-art sampling methods on IBM's real-world use cases.
Jul 2023 — Jul 2024
Software Engineer
Qburst Technologies
Reduced frontend load times via caching and compression. Split monolithic React apps into micro frontends using module federation, cutting development time by 20–30%. R&D on microanimations with CSS/JS.
Jun 2021 — Jul 2022
Flutter Developer Intern
Accubits Technologies
Developed cross-platform mobile applications for iOS and Android using Flutter.
Sep 2020 — Nov 2020
Flutter Developer Intern
Leimo Technologies
Developed the company's primary mobile application with detailed user analytics, crash analytics, and end-to-end Play Store management.
03

Projects

IBM Research
Online Data Loader Framework
5–10% downstream improvement
A novel framework built on PyTorch for dynamic data mixing during LLM training. Supports static and dynamic sampling strategies with on-the-fly mixture adjustment — integrated with IBM's internal training stack.
PyTorch Data Loading LLM Training Framework
Research
Few-Shot Distillation: ICL into Weights
+8.4pp over standard SFT; beats 8-shot teacher at zero-shot
A drop-in extension for standard SFT that brings the benefits of few-shot ICL into fine-tuning. Uses the base model as a "free teacher": an 8-shot context produces rich token-level soft labels that supervise a LoRA student alongside cross-entropy loss — capturing dark knowledge and task-specific registers that one-hot labels discard. Evaluated on GSM8K across multiple model families, FSD consistently outperformed standard SFT with gains up to +8.4 pp, with the Qwen3-8B variant surpassing its own 8-shot teacher — all at zero-shot inference speeds.
Knowledge Distillation In-Context Learning Fine-tuning LoRA Qwen3 PyTorch GSM8K
Research
Image-Conditioned Code Generation via RLVR
120% improvement over base model
Trained a VLM to generate code from UI images and PDFs using reinforcement learning with verifiable rewards — without SFT. One of the first works exploring RLVR for image-to-code tasks.
RLVR VLM PyTorch Code Generation
BharatGen Collab
Online Data Mixing for LLM Pre-training
Implemented online data mixing for pre-training BharatGen models, eliminating the ablation costs of iterating over different static mixtures — saving hundreds of hours of compute while delivering best-in-class performance.
LLM Pre-training Data Mixing PyTorch Multi-lingual
Research
Layout-Aware Document Image Translation
38.5% relative BLEU improvement
Built a document image translation pipeline with a novel layout detection module that models layout as a token classification problem, grouping related tokens into segments for better translation.
NLP Layout Detection Translation Transformers
IBM Research Collab
Data-Efficient Instruction Fine-Tuning
20–30% gain over random subset
Developed an algorithm using submodular functions to select the optimal data subset yielding highest returns. The resulting models consistently outperformed random subset baselines on downstream tasks.
Fine-tuning Submodular Optimization LLMs
Meta Collab
Explainability Framework for Legal LLMs
35–40% reduction in hallucinations
Built a RAG-based framework that retrieves relevant legal documents for each query, making chatbot responses more accurate and explainable. A joint initiative of IIT Bombay, NLU Bangalore, and Meta.
RAG LangChain Legal AI LLMs
Research
Contextual Knowledge for OCR Accuracy
10–15% accuracy boost
Combined predictions from TrOCR (vision) and RoBERTa (language) through an innovative multi-model logits fusion approach, significantly improving OCR accuracy on handwritten text.
TrOCR RoBERTa Logits Fusion OCR
Project
Micrograd — Autograd Engine
A compact autograd engine implementing backpropagation (reverse-mode autodiff) over a dynamically constructed DAG. Includes a small-scale neural networks library with a PyTorch-like API.
Python Autograd Neural Networks From Scratch
Project
Real-Time People Counter
Real-time people counting system using OpenCV with SSD + MobileNet for detection and a centroid tracker for tracking. Features alert systems for occupancy compliance and threaded video processing.
OpenCV MobileNet SSD Python
Project
Automatic Attendance Tracking
Automated attendance system for online classes using a Selenium web scraping bot, React admin panel, and Flutter mobile app. Received a special appreciation award from the college.
Selenium React Flutter Automation
App
Crossroads '23 — Ticketing System
Built the cultural fest website and an offline automation system to streamline ticketing for 2000+ participants with ID card scanning to prevent fraudulent re-entry.
Web Dev Automation ID Scanning
04

Skills

ML & AI
LLMs VLMs / MLLMs Reinforcement Learning RAG Fine-tuning GenAI Data Sampling Model Training Evaluation & Benchmarking
🔧
Libraries & Tools
PyTorch HuggingFace LangChain Pandas NumPy TensorFlow Keras Git Linux Firebase MongoDB
💻
Programming
Python C Bash Java JavaScript TypeScript Dart React Flutter
05

Achievements & Position of Responsibilities

Achievements

🏆
Selected as Applied Scientist Intern at Amazon (2025)
🥇
Proficiency Award — Department Topper, BTech (2022)
Best Performing Student Award, BTech (2023)
🏅
Baselian Award Nomination — Overall Best Student, BTech
🚀
Intel Student Ambassador & Microsoft Learn Student Ambassador

Position of Responsibilities

🎓
Department Representative, CSE (BTech)
🎉
Technical Head of Crossroads — Cultural Fest (BTech)
💡
General Convenor of Hash — Tech Fest (BTech)
📋
Class Representative — 3rd Year (BTech)

Let's connect

Open to research collaborations, AI engineering roles, and interesting problems in LLMs and multimodal AI.