I'm Amal Joe R S — an AI engineer specializing in LLMs, multimodal models, and reinforcement learning. Currently
finishing my M.Tech at IIT Bombay with research at IBM and collaborations with Meta and BharatGen.
Current CPI9.77 / 10
FocusLLMs & VLMs
Based inMumbai, India
01
Education
M.Tech, CSE
Indian Institute of Technology Bombay
2024 — 2026
9.77
CPI out of 10
B.Tech, CSE
Mar Baselios College of Engineering (APJ Abdul Kalam TU)
2019 — 2023
9.31
CGPA out of 10
02
Experience
Starting Jul 2026
AI Research Engineer
IBM Research
Nov 2025 — Present
Freelance AI Engineer
Financial Reporting & Agentic AI
Working with a company that helps MNCs file annual financial reports globally.
Integrating agentic AI into existing workflows to automate iXBRL tagging and
fine-tuning custom models for in-house use cases.
May 2025 — Aug 2025
AI Research Intern
IBM Research
Designed a dynamic data sampling strategy for multi-lingual
multi-task LLM training. Built the "Online Data Loader" — a novel PyTorch framework for dynamic data mixing,
achieving 5–10% improvement over state-of-the-art sampling methods on IBM's real-world use
cases.
Jul 2023 — Jul 2024
Software Engineer
Qburst Technologies
Reduced frontend load times via caching and compression. Split monolithic React apps
into micro frontends using module federation, cutting development time by 20–30%. R&D on
microanimations with CSS/JS.
Jun 2021 — Jul 2022
Flutter Developer Intern
Accubits Technologies
Developed cross-platform mobile applications for iOS and Android using Flutter.
Sep 2020 — Nov 2020
Flutter Developer Intern
Leimo Technologies
Developed the company's primary mobile application with detailed user analytics,
crash analytics, and end-to-end Play Store management.
03
Projects
IBM Research
Online Data Loader Framework
5–10% downstream improvement
A novel framework built on PyTorch for dynamic data mixing during LLM training.
Supports static and dynamic sampling strategies with on-the-fly mixture adjustment — integrated with IBM's
internal training stack.
+8.4pp over standard SFT; beats 8-shot teacher at zero-shot
A drop-in extension for standard SFT that brings the benefits of few-shot ICL into
fine-tuning. Uses the base model as a "free teacher": an 8-shot context produces rich token-level soft labels
that supervise a LoRA student alongside cross-entropy loss — capturing dark knowledge and task-specific
registers that one-hot labels discard. Evaluated on GSM8K across multiple model families, FSD consistently
outperformed standard SFT with gains up to +8.4 pp, with the Qwen3-8B variant surpassing its own 8-shot
teacher — all at zero-shot inference speeds.
Trained a VLM to generate code from UI images and PDFs using reinforcement learning
with verifiable rewards — without SFT. One of the first works exploring RLVR for image-to-code tasks.
RLVRVLMPyTorchCode Generation
BharatGen Collab
Online Data Mixing for LLM Pre-training
Implemented online data mixing for pre-training BharatGen models, eliminating the
ablation costs of iterating over different static mixtures — saving hundreds of hours of compute while
delivering best-in-class performance.
Built a document image translation pipeline with a novel layout detection module that
models layout as a token classification problem, grouping related tokens into segments for better translation.
NLPLayout DetectionTranslationTransformers
IBM Research Collab
Data-Efficient Instruction Fine-Tuning
20–30% gain over random subset
Developed an algorithm using submodular functions to select the optimal data subset
yielding highest returns. The resulting models consistently outperformed random subset baselines on downstream
tasks.
Fine-tuningSubmodular OptimizationLLMs
Meta Collab
Explainability Framework for Legal LLMs
35–40% reduction in hallucinations
Built a RAG-based framework that retrieves relevant legal documents for each query,
making chatbot responses more accurate and explainable. A joint initiative of IIT Bombay, NLU Bangalore, and
Meta.
Combined predictions from TrOCR (vision) and RoBERTa (language) through an innovative
multi-model logits fusion approach, significantly improving OCR accuracy on handwritten text.
TrOCRRoBERTaLogits FusionOCR
Project
Micrograd — Autograd Engine
A compact autograd engine implementing backpropagation (reverse-mode autodiff) over a
dynamically constructed DAG. Includes a small-scale neural networks library with a PyTorch-like API.
Real-time people counting system using OpenCV with SSD + MobileNet for detection and a
centroid tracker for tracking. Features alert systems for occupancy compliance and threaded video processing.
OpenCVMobileNetSSDPython
Project
Automatic Attendance Tracking
Automated attendance system for online classes using a Selenium web scraping bot,
React admin panel, and Flutter mobile app. Received a special appreciation award from the college.
Built the cultural fest website and an offline automation system to streamline
ticketing for 2000+ participants with ID card scanning to prevent fraudulent re-entry.