AI Software Engineer building production AI systems - especially LLM-powered workflows, RAG, and applied machine learning - where reliability and real-world usability matter.
Currently at Long Health, I contribute to AI-driven document and workflow automation across backend services and intelligent pipelines that structure complex, unstructured data.
Open to conversations around AI Software Engineering and Applied AI/ML Engineering roles where building dependable AI systems at scale is a core requirement.
Combining research and production engineering to deliver reliable AI systems
I am an AI Software Engineer focused on building real-world AI applications, with an emphasis on LLM-powered systems, retrieval-augmented generation (RAG), and applied machine learning. My work sits at the intersection of software engineering and AI, where system reliability, model quality, and real-world usability are critical.
Currently at Long Health, I contribute to AI-driven document and workflow automation, working on backend services and intelligent pipelines that process and structure complex, unstructured data. My role involves integrating LLMs, data pipelines, and event-driven systems into production software, and I also contribute to frontend interfaces where needed to enable end-to-end delivery of AI features.
Alongside my industry work, I independently publish research in machine learning and AI systems, with peer-reviewed and archival publications (ICPR/Springer, IEEE Access, arXiv) spanning dynamic language modeling, time-series prediction, and applied deep learning. I'm particularly interested in translating research ideas into robust, production-ready systems.
I'm open to conversations around AI Software Engineering and Applied AI/ML Engineering roles where building dependable AI systems at scale is a core requirement.
Elected Full Member (peer-nominated and reviewed) based on demonstrated original research published in peer-reviewed journals.
Shipped document AI pipelines and LLM-powered physician assistants.
Built GPT-powered legal assistants with RAG & vector DBs.
Forecasted 911 call volume and optimized staffing using ML.
Boosted campaign ROI via LSTM & ETL optimization.
Built lane detection and GNN-based driver behavior forecasting.
Built LSTM models for IoT network health prediction at scale.
I’ve had the pleasure of working with Aravinda Jatavallabha as part of the SmartProtect Public Safety Solutions team, where he has served as a Data Science Intern over the past six months. Aravinda’s contributions to building out our predictive analytics framework for public safety have been both impactful and highly innovative. His work, especially in structuring and analyzing complex datasets, has significantly advanced our ability to deliver tailored, data-driven solutions to our clients.
Aravinda brings a deep technical expertise to his work, combined with a genuine commitment to understanding and solving client challenges in public safety. His attention to detail, analytical insight, and dedication have made him an invaluable asset to our team, and his contributions will continue to benefit both our company and the customers we serve.
I had the pleasure of working with Aravinda during his time at SmartProtect, where he made a lasting impact on our AI/ML initiatives. From day one, he brought thoughtfulness and a sharp technical mindset to every challenge. Whether it was building early prototypes or helping us think through how machine learning could meaningfully support public safety operations, Aravinda consistently delivered high-quality work and insightful ideas.
What stood out most was his willingness to dive deep, learn fast, and always look for ways to make the work better—not just from a technical standpoint, but in a way that aligned with the mission of serving first responders. He was a valued team member with great communication skills and someone I trusted to bring both integrity and innovation to the table.
Any team would be lucky to have Aravinda on board, and I’m excited to see where his journey takes him next.
Data Science Track
AI & Machine Learning
Deep Learning, Neural Networks, NLP
Data Science
Analytics, Mining, Visualization
Systems & Architecture
Databases, Algorithms, Software Engineering
Information Technology
Big Data Analytics
Minor Specialization
Software Engineering
Core Focus
AI & ML
Technical Electives
Long Health
Jun 2025 - Present · Full-time
San Jose, California, United States · Remote
Document Throughput
+35%
10K+ files weekly
Processing Latency
-30%
Async pipelines
App Performance
-40%
UI load time
LLM Copilot Impact
-50%
Documentation time
SmartProtect Public Safety Solutions
May 2024 - Jun 2025 · Internship (Part-time)
Wilmington, Delaware, United States · Remote
Forecast Accuracy
+20%
Model Performance
Operational Efficiency
18%
Overtime Reduction
Processing Speed
35%
Faster Retraining
North Carolina State University
Aug 2024 - May 2025 · Part-time
Raleigh, North Carolina, United States
Course Coverage
2
Advanced ML Courses
Teaching Duration
10
Months of Instruction
Jan 2025 - May 2025 · Under Prof. Thomas Price
Aug 2024 - Dec 2024 · Under Prof. Xipeng Shen
Defence Research and Development Organisation (DRDO)
Jan 2023 - Jun 2023 · Research Internship
Bengaluru, Karnataka, India · On-site
Model Performance
3.19
Language Model Perplexity (SOTA)
Processing Efficiency
40%
Reduction in retraining latency
Merkle
May 2022 - Jul 2022 · Internship
Bengaluru, Karnataka, India · Hybrid
Campaign Profitability
+10%
Revenue Optimization
Query Performance
40%
Latency Reduction
Data Scale
16M+
Records Processed
Manipal Institute of Technology
Mar 2021 - Jun 2022 · Part-time
Udupi, Karnataka, India · On-site
Model Accuracy
99.4%
IoT Network Prediction
Automation Impact
60%
Reduction in Manual Tasks
Building innovative solutions across AI, Machine Learning, and Full-Stack Development
A full-stack AI-powered app that simplifies health insurance documents using LLMs and Retrieval-Augmented Generation (RAG). Users can upload plans, ask natural-language questions, view smart summaries, compare multiple plans side-by-side, and export personalized PDF reports. Built with end-to-end semantic search, summarization, and secure document handling.
A campus-wide parking management system that tracks lot availability, zoning rules, permit assignments, and citations. It allows administrators to efficiently manage parking resources, issue fines, and generate reports to support data-driven decisions for better traffic control and user experience.
An automated cold outreach tool that combines LangChain's ChatGroq + LLaMA3 with ChromaDB to extract job descriptions, match user skills, and generate personalized emails using RAG. Includes an interactive Streamlit UI for seamless job-to-email generation.
A chatbot powered by LLMs (OpenAI GPT / LLaMA) and RAG, designed to retrieve, summarize, and answer complex legal queries from document repositories with high accuracy and fast vector-based search.
Built a full ML pipeline for predicting customer churn using Apache Airflow, AWS (S3, SageMaker, ECR), and Dockerized Flask APIs, enabling scalable deployment and real-time churn inference.
Implemented a hybrid SegNet + LSTM deep learning model to detect lane lines, compute lane curvature, and measure vehicle offset using OpenCV-based image processing.
Developed a CNN-based classifier to detect pneumonia (COVID-19) from chest X-ray images, achieving 95.28% training accuracy and 89.52% validation accuracy using preprocessed radiology data.
Built a robust hybrid forecasting model combining CNN and BiLSTM architectures to predict daily item-level sales across 10 stores. Leveraged the Kaggle Store-Item Demand Forecasting dataset (2013–2017) and benchmarked against models like XGBoost, ANN, and ARIMA. The hybrid model achieved the lowest MSE, improving forecasting precision and enabling optimized retail inventory decisions.
Implemented CycleGAN to translate images between domains without paired datasets—such as Monet paintings to real photographs and human faces to zombies. Trained on publicly available datasets and deployed the model for real-time translation, showcasing the power of unsupervised generative learning in computer vision tasks.
Used U-Net architecture to perform semantic segmentation on brain MRI images, detecting and outlining tumor regions. The project utilized the LGG MRI Segmentation dataset from Kaggle and focused on pixel-level mask prediction using FLAIR MRI sequences. Achieved high segmentation accuracy and visual interpretability for potential medical diagnostics.
Designed a hybrid recommender system that combines cosine similarity with sentiment analysis to suggest movies tailored to user preferences. Scraped metadata from TMDB and IMDB, offering dynamic updates, cast bios, trailers, and review sentiment. Upgraded from a static recommendation system to an interactive, emotionally aware movie discovery experience.
Research papers in AI and Machine Learning
A reproducible benchmark for deployment trust in LLM systems, measuring desired behavior rate, run stability, confidence dispersion, and cost-effectiveness across leading proprietary models.
Deployment trust in large language model (LLM) systems is shaped by behaviors that do not show up in single run accuracy numbers. In production, teams routinely need models to request missing information instead of guessing, behave consistently across repeated calls, refuse unsafe requests when required, and remain economical at scale. We present TrustBench, a reproducible benchmark designed around these practical requirements. For each model and prompt pair, TrustBench executes Nr = 5 trials under fixed settings and reports: desired behavior rate (DBR) from a per item majority vote, run stability as agreement with that majority label, confidence dispersion across reruns, and cost effectiveness from token based pricing. We evaluate five proprietary models: OpenAI GPT-4.1, GPT-4o, GPT-5; Claude Sonnet 4.5; and xAI Grok 4.1 (Fast Reasoning). On the evaluated ambiguity set, requesting clarification is rarely the dominant outcome, and some models never reach a majority clarify label. Safety behavior is highly stable across reruns, yet pricing differences are large enough to change value rankings even when behavioral rates are close. For the multi step set, the main tables report repeatability and confidence dispersion rather than verified correctness. Code, prompts, and per run logs are released to support reproducibility.
Comprehensive technical analysis of fairness auditing in healthcare ML systems, covering subgroup evaluation, counterfactual testing, lifecycle-based auditing, and cloud-native MLOps pipelines for continuous fairness monitoring.
The increasing deployment of machine learning (ML) systems in healthcare has raised significant concerns regarding algorithmic fairness, particularly in high-stakes clinical decision-making contexts. While prior research has extensively examined bias mitigation techniques and fairness metrics, comparatively less attention has been paid to systematic fairness auditing practices tailored to healthcare ML systems. Healthcare data often embed structural, historical, and measurement biases arising from unequal access to care, proxy outcome variables, and evolving clinical practices, which can propagate disparities across demographic and socioeconomic groups. This paper provides a comprehensive technical analysis of fairness auditing in healthcare machine learning systems, emphasizing the unique challenges posed by medical data, clinical workflows, and regulatory constraints. We review established and emerging auditing methodologies, including subgroup performance evaluation, counterfactual testing, and lifecycle-based auditing approaches spanning model development, deployment, and post-deployment monitoring. Additionally, we examine deployment considerations such as feedback loops, distributional shift, and compliance with healthcare regulations. We also explore cloud-native MLOps pipelines that support continuous fairness auditing in regulated healthcare environments. By synthesizing existing research and outlining open technical challenges, this work aims to inform the design of robust, deployment-aware fairness auditing frameworks for real-world healthcare ML systems.
Novel transformer framework with multi-head attention for e-commerce inventory optimization; achieves 23.7% reduction in stockouts and 18.4% decrease in excess inventory costs.
The rapid growth of e-commerce has introduced unprecedented complexity in inventory management, characterized by demand volatility, seasonal fluctuations, and dynamic market conditions. Traditional inventory replenishment methods often fail to capture intricate temporal patterns and long-range dependencies inherent in modern digital retail environments. This study investigates the application of transformer-based deep learning models for optimizing inventory replenishment decisions in dynamic e-commerce markets. We propose a novel framework that leverages multi-head attention mechanisms to process historical sales data, promotional activities, pricing dynamics, and external market signals. Our empirical analysis, conducted using data from multiple e-commerce platforms spanning 24 months, demonstrates that transformer models achieve a 23.7% reduction in stockout incidents and an 18.4% decrease in excess inventory costs compared to conventional forecasting approaches. The model exhibits superior performance in capturing demand seasonality, promotional lift effects, and product substitution patterns. Results indicate that attention mechanisms effectively identify critical demand drivers and their temporal relationships, enabling more accurate replenishment timing and quantity decisions. This research contributes to the growing body of knowledge on artificial intelligence applications in supply chain management and provides practical insights for e-commerce retailers seeking to enhance operational efficiency. The findings suggest that transformer architectures represent a significant advancement over recurrent neural networks and traditional time series methods for inventory optimization in complex, fast-paced digital marketplaces.
Study of input regurgitation risk and prompt-induced sanitization strategies for privacy-sensitive LLM deployments under HIPAA and GDPR.
Large Language Models (LLMs) are indispensable in privacy-sensitive fields like healthcare and hiring because of their remarkable natural language processing capabilities. However, the potential for input regurgitation and sensitive information in outputs raises serious privacy concerns, especially when it comes to regulations like GDPR and HIPAA. This study examines these privacy concerns by contrasting seven innovative models - GPT-3.5, GPT-4, GPT-4 Turbo, GPT-5.1, GPT-5.2, Gemini-2.5-pro, and Gemini-2.5-flash - and their propensity to retain and regurgitate PHI and personally identifiable information. We test these models on a combination of prompt-induced sanitization techniques and synthetic datasets to determine their ability to generate privacy-compliant outcomes. Our findings reveal that newer models, particularly GPT-5.2 and Gemini-2.5-pro, demonstrate enhanced privacy-preserving capabilities, with GPT-5.2 achieving near-zero leakage (0.0%) even with minimal prompt adjustments, while maintaining utility. Surprisingly, we discover that some of the advanced models have built-in privacy consciousness, in that sensitive information is sanitised even without instructions. This paper contains some useful recommendations and tips on increasing privacy compliance in LLM applications.
Practitioner-oriented survey of prompting strategies and security defenses for prompt injection and retrieval-augmented generation (RAG) systems.
Large Language Models (LLMs) are increasingly embedded in security-sensitive workflows such as incident triage, code review, threat hunting, and retrieval-augmented assistants. In these settings, prompting is not only a performance tool but also a security control surface: LLMs may consume untrusted content (tickets, logs, web pages, retrieved documents) that can adversarially manipulate outputs via prompt injection. This paper presents a structured, practitioner-oriented survey that unifies (i) core prompting methods (zero-shot, few-shot), (ii) reasoning-structured prompting (chains/trees/graphs of thought), (iii) robustness methods (self-consistency), (iv) efficiency workflows (skeleton-of-thought), and (v) automatic prompt optimization (APO) and learned prompting (directional stimulus prompting). We additionally synthesize cybersecurity-specific research on prompt injection, indirect prompt injection benchmarks, retrieval-augmented generation (RAG) poisoning, coordinated Prompt–RAG attacks, and interface-level defenses (structured queries). We provide a taxonomy, summary tables, and a decision guide for selecting prompting strategies under accuracy–cost–security constraints.
Investigation of dynamic representation learning in large language models using graph-based attention mechanisms for improved context understanding.
In the world of Large Language Modeling, incremental learning plays an important role in evolving data such as streaming text. We introduce an incremental learning approach for dynamic contextualized word embeddings in the setting of streaming data. We call the embeddings generated by our model as Incremental Dynamic Contextualized Word Embeddings (iDCWE). Our model introduces the incremental BERT (iBERT) (BERT stands for Bidirectional Encoder Representations from Transformers) to create a dynamic and incremental model to perform incremental training. Our model further captures the semantic drift of words using dynamic graphs. Our paper is the first in the line of research on (incremental) dynamic modeling of streaming text which we also refer to as Neural Dynamic Language Modeling. The performance of our model on the benchmark datasets is on par and even often outperforms the dynamic contextualized word embeddings which was the first paper to combine contextualization with dynamic word embeddings, to the best of our knowledge. Moreover, the compute time efficiency of our model is better than that of the aforementioned paper.
A Temporal Dynamic Graph Neural Network (TDGNN) framework designed to model real-world dynamic graphs by integrating time-aware message passing, graph topology, and point process theory to enhance prediction in social and interaction networks.
The utilization of Graph Neural Networks (GNN) in modeling real-world graph structures has shown promising results, making it a widely recognized method for extracting information from non- Euclidean data, such as social networks. For static networks, there are many different sophisticated GNN architectures and models; nevertheless, the development of comparable methods for dynamic graphs has been slow. Due to the dynamic nature of graphs, Dynamic Graph Neural Networks have recently attracted interest in various fields, especially social networks. Continuous Time Dynamic Graphs (CTDG) effectively incorporate temporal information, graph structure (topology), and node properties to record the continuous time progression of dynamic graphs. The computational and memory requirements for Dynamic Graph Neural Networks provide considerable difficulties. In order to address this, a novel deep learning methodology using a model called TDGNN (Temporal Dynamic Graph Neural Network) has been developed to effectively model dynamic graphs while including temporal data, graph structure, and node attributes. The model suggests an approach that has been shown to perform better than baseline methods when tested on benchmark datasets.
Time series analysis and prediction framework for optimizing network traffic offloading in 5G networks using software-defined networking.
The continuous growth of mobile traffic and limited spectrum resources limits the capacity and data rate. Heterogeneous Networks (HetNet) is a solution with multiple radio interfaces in smartphones to realize such demands. Simultaneous data transfer on Long Term Evolution (LTE) and WiFi has gained attention for data offloading in 5G HetNet. Maintaining the average throughput and minimum delay for LTE users is still a challenge in data offloading owing to the mobility and load in the network. This study explores the benefits of Software-Defined Networking (SDN) based multipath for data offloading schemes for LTE-WiFi integrated networks to maintain the user's average throughput based on channel quality classification. We classify future link qualities using deep learning models such as Long Short-Term Memory Networks (LSTM) and Bidirectional Long Short-Term Memory Networks (BLSTM). The received signal strength indicator (RSSI) and packet data rate (PDR) are parameters used in BLSTM. The results of the prediction were compared with those of state-of-the-art methods. We obtained a 2.1% better prediction than the state-of-the-art methods. The predicted results were used to offload the data using LTE and WiFi. The performance of HetNet was compared with the state-of-the-art method for average throughput, and with the proposed method, a 6.29% improvement was observed.
Novel approach to mitigating bias in computer vision models through adversarial debiasing and balanced representation learning.
This paper introduces a feature distillation framework that aims to learn fairer representations without significantly sacrificing task performance. We propose a Maximum Mean Discrepancy (MMD)-based loss to distill information from an unfair teacher network while encouraging feature invariance across protected attributes such as race or gender. Our method demonstrates marked reduction in disparity measures while maintaining competitive accuracy on standard computer vision benchmarks.
Integration of visual and textual cues for early detection of conversation derailment in multimodal AI systems.
This paper presents a hierarchical transformer-based framework that jointly models textual and visual modalities for detecting derailment in multimodal discussions. The proposed system integrates BERT-based text encoding with Faster R-CNN-derived visual features, achieving 71.0% accuracy and 78.3% AUC, outperforming text-only baselines by 6.1%. Our approach demonstrates the importance of multimodal cues in understanding conversation dynamics and early detection of potential derailments.
Development of a lightweight graph contrastive learning framework for efficient recommendation systems.
GraphNeural Networks (GNNs)have emerged as apotent framework for graph-structured recommendation tasks. Incorporating contrastive learning with GNNs has recently demon strated remarkable efficacy in addressing challenges posed by data sparsity, thanks to in novative data augmentation strategies. However, many existing methods employ stochastic perturbations (e.g., node or edge modifications) or heuristic approaches (e.g., clustering based augmentations) to generate contrastive views, which may distort semantic integrity or amplify noise. We introduce LightGCL, a novel and streamlined graph contrastive learn ing model to address these limitations. LightGCL utilizes Singular Value Decomposition (SVD) to achieve robust augmentation, facilitating structural refinement and global collab orative relation modeling without manual augmentation strategies. Extensive experiments on benchmark datasets showcase its substantial performance enhancements over state-of the-art methods. Further analysis highlights the model's resilience to challenges like data sparsity and bias related to item popularity.
Analysis of ethical considerations and policy implications in autonomous vehicle systems, focusing on Tesla's Autopilot implementation.
This case study delves into the ethical ramifications of incidents involving Tesla's Autopilot, emphasizing Tesla Motors' moral responsibility. Using a seven-step ethical decision-making process, we examine user behavior, system constraints, and regulatory implications. The analysis offers insights into ethical considerations in evolving technological landscapes and proposes a framework for evaluating autonomous system deployments in safety-critical applications.
Time series forecasting model for predicting flight delays using weather data and historical flight performance.
This research investigates flight delay trends by examining factors such as departure time, airline, and weather conditions. We employ time-series models including LSTM, Hybrid LSTM, and Bi-LSTM, comparing them with baseline regression models. Our approach focuses on identifying influential features in delay prediction, potentially informing flight planning strategies. The study demonstrates the effectiveness of deep learning approaches in capturing complex temporal patterns in aviation operations.
Machine learning approach to diabetes prognosis using patient data and clinical markers for early detection.
This study addresses the critical need for early diabetes detection using machine learning algorithms. We compare K-Nearest Neighbor, Random Forest, and Artificial Neural Network approaches, incorporating comprehensive preprocessing and feature engineering strategies. The Random Forest classifier achieved the highest accuracy of 87.89%, significantly outperforming traditional diagnostic methods and demonstrating the potential for ML-assisted medical decision-making.
Deep learning approach to automated bone age assessment using X-ray images for pediatric growth evaluation.
This research evaluates deep learning models for automated bone age assessment, comparing pre-trained architectures including VGG-16, InceptionV3, XceptionNet, and MobileNet. Our study focuses on pediatric X-ray images, developing a system that can accurately determine skeletal maturity without expert intervention. The results demonstrate the potential for AI-assisted growth evaluation in clinical settings, with particular emphasis on reducing assessment time while maintaining accuracy.
I'm open to conversations around AI Software Engineering and Applied AI/ML Engineering roles where building dependable AI systems at scale is a core requirement. Whether you're building innovative systems or solving real-world problems with data, I'd love to be a part of it. Let's chat about how I can help your team move faster with intelligent, scalable solutions.