Ayush Ranjan

Ayush Ranjan

Graduate Student At University of California

About Me

Hi, I’m Ayush Ranjan from Madhubani, Bihar, India. I build practical AI and backend systems- the digital “bridges” that connect products and people. I earned my B.Tech in Information Technology from Manipal University Jaipur and completed my Masters in Computer Science in 2025 at University of California Santa Cruz, where I specialized in AI systems and database architecture.

Currently, I'm a Backend Developer at Altimetrik working on Airbnb's Payments Incentives platform, where I engineer high-concurrency Java microservices handling 2.55 billion requests monthly and architected systems that surfaced $120M in dormant user credits.

My journey began at Capgemini, developing enterprise diagnostic systems for Mercedes-Benz. I led a micro-frontend architecture project that earned 3rd place at Innocircle 2022( Mercedes-Benz's internal innovation hackathon), reducing topology review time by 50%. That experience taught me how thoughtful engineering directly translates to business impact.

At UC Santa Cruz, I bridged academia and industry through AI research. As an AI Engineer Intern in the Information Retrieval & Knowledge Management Lab, I built a multimodal AI agent for smart wearables from the ground up-implementing wake word detection, real-time audio-visual processing, and intelligent query routing with 95% accuracy under Prof. Yi Zhang. I also served as a Graduate Researcher in the AI Explainability & Accountability (AIEA) Lab under Prof. Leilani H. Gilpin, focusing on Retrieval-Augmented Generation (RAG) systems and AI ethics.

I'm particularly passionate about making complex systems understandable and accessible. This shows in my teaching- I've been a Teaching Assistant for Database Systems four times and Software Engineering once, where I mentored students on everything from SQL optimization to Agile practices.

I specialize in building scalable backend systems and AI-augmented infrastructure. Whether it's processing billions of requests, designing intelligent agents, or extending databases with custom vector search capabilities, I focus on engineering solutions that are both technically robust and practically impactful.

Beyond code, I'm a passionate football (soccer) player and spectator- there's something pure about the game's flow. I'm an avid reader of technology blogs. And when I'm not on the field or reading around on Internet, you'll likely find me experimenting in the kitchen, where I approach cooking with the same curiosity I bring to engineering.

Work Experience

Altimetrik (Client: Airbnb)- San Francisco, CA, USA
Dropwizard, Kafka, Thrift RPC, MySQL, Redis, Bazel, K8s

Backend Developer

(Sep 2025 – Present)

  • Engineered high-concurrency Java microservices for Airbnb Payments Incentives, processing 2.55B requests/month (avg 5.06k QPS) using Kafka, Thrift RPC, Dropwizard, and Kubernetes.
  • Optimized developer workflows by contributing to internal Claude Code capabilities; engineered custom commands and skills to automate Payments-specific codebase navigation, analysis, and boilerplate generation.
  • Co-architected Thrift RPC endpoints that surfaced $120M in dormant user credits and integrated results into automated re-engagement email pipelines.
  • Developed a robust event-driven Producer-Consumer architecture using Kafka and Tempo to automate the lifecycle of Airbnb Virtual Credit Cards (VCC), ensuring 100% automated refund reconciliation upon card expiration.

AI Explainability and Accountability (AIEA) Lab - Santa Cruz, CA, USA
Python, Langgraph, Docker, Kubernetes, FastAPI, CI/CD

Graduate Researcher

(Oct 2024 – August 2025)
  • Conducted applied research to improve the reliability and explainability of LLM-based university chatbots across campus use cases (enrollment, deadlines, housing, course queries), leading to higher user satisfaction.
  • Designed and evaluated 10+ advanced RAG workflow architectures- Classic RAG, Chain-of-Thought, RARE RAG, Adaptive RAG, Corrective RAG, RAT RAG-using a comprehensive RAGAS evaluation framework.
  • Fine-tuned open-source LLMs to align with university-specific tone, structure, and factual accuracy, enabling domain adaptation for student and administrative queries.
  • Achieved consistent 35–50% performance improvement over baseline RAG systems, with approaches excelling by metric-faithfulness, answer relevancy, and context precision-based on query complexity and domain.
  • Developed a production deployment pipeline using Docker, Kubernetes, and FastAPI with automated CI/CD, load balancing, and monitoring for a scalable campus-wide chatbot implementation.

Information Retrieval and Knowledge Management Lab - Santa Cruz, CA, USA
Python, Flask,Dialogflow, LangGraph, Whisper, Computer Vision, Pinecone

AI Research Intern - Stealth Hardware Startup

(July 2024 - September 2024)
  • Built a 0-to-1 multimodal AI agent for smart wearables (camera-integrated earphones), implementing wake word detection (WWD), intent classification, and real-time audio-visual processing for calorie estimation, emergency response, and video summarization.
  • Designed an intelligent query routing system with 95% accuracy at classifying continuous vs. new queries, integrating Dialogflow for 8+ pre-built workflows (calorie estimation, contact calling, emergency location services) and custom LangGraph agents for open-domain conversations.
  • Engineered a real-time multimodal data fusion pipeline combining audio transcription (Whisper), computer vision (food segmentation, depth estimation), and vector similarity search with intelligent fallback to external tools (web search, OCR) when confidence dropped below the 0.8 threshold.
  • Developed a multi-threaded memory manager to asynchronously encode and cache historical observations (images, transcripts) into vector embeddings using Hugging Face Transformers, with persistent storage in Pinecone.
  • Integrated the prototype with a local edge pipeline (FFmpeg, Whisper, custom CV models), achieving sub-500ms inference latency for key commands and enabling real-time calorie detection via food segmentation and depth estimation.

Capgemini Technology Services India Limited - Mumbai, India
Java, Spring Boot, React, Micro-frontend, SQL, IBM Db2, MySQL, ETL, XML, Autosar, CI/CD

Associate Consultant

(Oct 2022 - Aug 2023)

  • Headed the Data Modeling team for Mercedes-Benz’s XDIS platform, driving backend schema evolution for vehicle network topology change requests (e.g., ECU reconfigurations, bus architecture edits).
  • Designed a lightweight ETL pipeline in Java to process large XML diagnostic files-extracting telemetry, transforming into updated entity models, and loading into IBM Db2-enabling seamless data migration.
  • Authored and tuned complex SQL queries and views in Db2 for schema validation, relational consistency checks, and historical topology comparisons supporting Change Request (CR) automation.
  • Achieved 3rd Place at Innocircle 2022 by implementing a micro frontend architecture that let users modify and review vehicle network topology changes, reducing process time by 50%+.
  • Built an AI-assisted validation system for 2,500+ historical CRs using Word2Vec and Sentence-BERT embeddings of symbolic topologies, flagging rare configurations and recommending optimal topologies to improve validation accuracy.

Senior Analyst

(July 2021 - Sep 2022)

  • Initially, worked as a Java Full Stack Developer on the 'Arek Oy' project, a Finnish company, on the development of banking system. This project primarily focused on frontend development and utilized a tech stack comprising React, Spring Boot, Redux, and GitLab for version control. It followed a Maven project structure, with a MySQL database as the data repository. Responsibilities included frontend development, tech stack integration, and codebase maintenance.
  • Later, transitioned to a Java Developer role at Mercedes-Benz Research & Development India's Project XDIS (Cross-platform Data Information System), a critical tool for vehicle diagnostics and automatic driving scenarios in Mercedes. Conducted comprehensive software analysis, programming, and proficiently handled testing and debugging. Contributed to creating well-designed, efficient, and testable code that contributed to project success.
  • XDIS is a core Java program with a swing-based user interface, integrated within the vehicle to download diagnostic data as well as to assist users to change the network topology of their cars. It followed a gradle project structure, with a IBM Db2 database as the data repository.
  • Dramatically optimized XML file migration time by an impressive 66.67%. Additionally, enhanced the tool's robustness by concurrently implementing indexing strategies for associated database tables.
  • Optimized export testing by developing a wrapper around the Autosar framework and implementing an efficient XML file import strategy, reducing overall testing time by 40% and improving export performance for individual modules by an average of 17%.

Senior Analyst Intern - Capgemini Technology Services India Limited - Pune

(Jan 2021 - May 2021)

  • Worked as a Java Full Stack Developer, collaboratively engaging in both frontend and backend development.
  • Utilized React for frontend development and Java Spring Boot for backend tasks.
  • Key project involved the creation of an online medical portal, catering to four distinct user roles: User, Doctor, Nurse, and Admin.
  • Ensured thorough documentation for this innovative digital solution.
  • Seamlessly integrated the frontend and backend via Axios, enhancing user experience and data security.
  • Rigorous testing procedures were executed, employing JUnit for the backend and Jasmine for the frontend.

Key Projects

Fathom Project

Fathom : Code-Aware Search Engine AI Software Engineering
Python, FastAPI, ChromaDB, Java, Protobuf

Dec 2025 -- Jan 2026

  • Designed and implemented Fathom, a code-aware search engine combining semantic (Tree-sitter + sentence-transformer), structural (SCIP), and literal search behind a unified FastAPI endpoint for LLM agents.
  • Built a project-aware indexing manager (SQLite Librarian) and a reproducible indexing pipeline producing embeddings in ChromaDB and SCIP .scip indices, enabling accurate go-to-definition and semantic retrieval.
  • Architected Fathom as an MCP-compatible tool for autonomous coding agents, improving repository grounding and reducing hallucination risk in code synthesis tasks.

View on Github


project name

Unveiling Glitches: A Deep Dive into Image Encoding Bugs within CLIPAI
Python, DAF, TCAC, DINOv2

Recieved A+ in CSE 290D Neural Computation at UCSC for this project.

  • The research project focuses on uncovering glitches in image encoding within CLIP, a model known for its integration of vision and language processing. By employing methodologies like the Discrepancy Analysis Framework (DAF) and the Transformative Caption Analysis for CLIP (TCAC), the study aims to evaluate CLIP's performance and identify areas for improvement.
  • The Discrepancy Analysis Framework (DAF) method is a systematic approach used to evaluate CLIP's performance by comparing its image similarity rankings with those of the DINOv2 model.
  • The Transformative Caption Analysis for CLIP (TCAC) method is an extension of the Discrepancy Analysis Framework (DAF) that focuses on evaluating CLIP's response to image transformations. This approach involves setting up various transformations to simulate real-world conditions, predicting caption probabilities before and after transformations, and manually inspecting images and captions for discrepancies.
  • Through systematic analysis, we reveal discrepancies in CLIP's interpretation of images compared to human perception, highlighting 14 systemic faults, including 4 novel faults.
  • By addressing these limitations, the study lays the groundwork for the development of more accurate image embedding models in artificial intelligence.
  • You can refer to this ppt or github repository to understand more.
You can read the research paper here

project name

Sentiment Analysis With CNN AI
Python, Pytorch, Spacy, GloVe, CNN, EMNLP

Recieved A+ ( 10/10 ) as my minor Project in my undergrad at Manipal

Supervised by: Shashank Sharma

  • The project is based on implementing Convolutional Neural Networks (CNN) for sentence classification, inspired by the paper "Convolutional Neural Networks for Sentence Classification" presented at EMNLP 2014. The aim is to extract features from text using different filter sizes and numbers to learn various n-gram features.
  • The project utilizes the Spacy library for various NLP tasks such as tokenization, lemmatization, part-of-speech tagging, entity recognition, and dependency parsing. The CNN architecture is designed with multiple convolutional layers of different filter sizes (2, 3, 4, and 5) to capture bi-gram, tri-gram, 4-gram, and 5-gram features respectively. Dropout regularization is applied to prevent overfitting.
  • The dataset is split into training, validation, and test sets, with the vocabulary size limited to 50,002 words, including special tokens for padding and unknown words. Each word is represented using one-hot encoding and passed through an embedding layer to convert them into word embeddings.
  • The forward method of the model involves passing the text data through the embedding layer, followed by separate convolutional layers for each filter size. After applying the convolutional layers, the output tensor is passed through ReLU activation to introduce non-linearity. The resulting feature maps are then concatenated and passed through a linear layer for sentiment prediction. The model is trained using the Adam optimizer and BCEWithLogitsLos function, which combines sigmoid activation and binary cross-entropy loss. The initial representations of word embeddings are obtained from pre-trained GloVe embeddings.
  • After training the model for just 5 epochs, it achieves a test accuracy of 87%, a validation accuracy of 89%, and a training accuracy of 88%. Overall, the project demonstrates the effectiveness of CNNs in text classification tasks using Pytorch.
More about this Project

project name

Enhancing Image Captioning with Attention MechanismsComputer Vision NLPDeep Learning
Python, Pytorch, LSTM, Resnet50, BLEU

  • Developed and implemented a baseline LSTM model with Resnet50 for feature extraction using an encoder-decoder architecture.
  • Integrated attention mechanisms to enhance the model's performance, achieving significant improvements with minimal training epochs.
  • Conducted extensive experiments, including custom data splits and benchmarks, evaluated using BLEU metrics.
  • Addressed and analyzed irregular validation loss, exploring learning rate schedules and comparing results with existing implementations.
  • Proposed future investigations into larger datasets and beam search strategies to improve inference.
  • Completed this project as part of my Advanced Computer Vision course, securing an A+ grade.

View on Github

project name

Video to Mp3 Converter Microservice Software Enginnering
Python, RabbitMQ, MongoDB, GridFS, Docker, Kubernetes, Minikube

  • Developed a microservices-based system with four services, including an authentication gateway, authorization service, video upload service, and converter service. The gateway authenticates users via an authorization service, generating JWT tokens for valid users, enabling secure video uploads. Video-to-MP3 conversion was facilitated using the Python library ”moviepy”.
  • Implemented asynchronous communication using RabbitMQ queues to facilitate seamless video processing and conversion to MP3, ensuring efficient task distribution among services.
  • Utilized MongoDB with GridFS for efficient storage and retrieval of large video files, overcoming MongoDB's 16MB size limit.
  • Managed video and audio file storage, handling, and conversion while ensuring data integrity and secure storage mechanisms.
  • Utilized Docker for containerization, Kubernetes for orchestration, and Minikube for local development, ensuring consistent and scalable deployment across environments.

View on Github

project name

Covid-19 Detection from CT-Scan Deep Learning
Python, Keras, CNN

I applied 4 layered CNN architecture using Keras on CT dataset of just 350 Ct-Images of 219 peoples to identify covid and non covid patients used 4 convolutional layer followed by Max-pooling. To calculate loss, I used binary cross-entropy. Used data augmentation and dropout to tackle overfitting. Got an training accuracy of 78.9% and test accuracy of 67.3% on such a small set of data.

View on Github

project name

Facial Attendance System AI Software Enginnering
Python, OpenCV, Haar-cascade, Tkinter, Google Text-to-Speech

Supervised by: Ginika Mahajan

Utilizing OpenCV, the system captures image frames and employs Haar features and Cascade Classifiers to detect faces within them. Identified faces are then compared against a database for identification. Attendance records of recognized individuals, including date and time stamps, are logged, with the capability to retrieve historical data using names.

The GUI, built with Tkinter, encompasses attendance features and admin query capabilities for login times. Integration of Google Text-to-Speech enhances user interaction with personalized welcome messages for recognized individuals.

View on Github

project name

Personal Website using React Software Engineering
React, Bootstrap, JavaScript, HTML, CSS, Github Pages

Designed and developed a personal website using React , Bootstrap, JavaScript, HTML and CSS. Deployed the website using Github Pages.

View on Github

Visit the Site

project name

Avoid-Obstacle Game Algorithm Design
Python, Pygame

Developed a Python(Pygame)-based interactive game featuring character movement control for left and right directions, allowing players to navigate and evade dynamically changing obstacle blocks. The game's challenge dynamically escalates as the speed of the blocks doubles with each successful evasion, providing an engaging and progressively challenging gaming experience.

View on Github

My GitHub Activity