Researcher Collab

About

I am Harinath Reddy Cingapuram, a research-focused AI/ML engineer and data scientist with experience across Fortune 500 companies and high-impact academic projects in generative AI and large language models. My work spans transfer learning, LLM alignment, and retrieval-augmented generation, with multiple peer-reviewed publications and ongoing conference submissions in applied machine learning and AI systems. I am particularly interested in building scalable, cloud-native ML pipelines and collaborating on research that bridges cutting-edge theory with real-world deployment on platforms such as AWS, Azure, and OpenShift.

Areas of Interest

My research focuses on advancing large language models through alignment techniques transfer learning and retrieval-augmented generation. I am particularly interested in bridging theoretical advances in generative AI with practical deployment challenges in cloud-native environments. My work spans supervised fine-tuning of LLMs diffusion models for image generation and building scalable ML pipelines on distributed systems. I am also exploring federated learning approaches and the intersection of AI systems with real-world applications in healthcare finance and enterprise solutions.

Stepwise Guidance for LLM Reasoning via Probe-and-Retrieve In-Context Learning

Springer

Large language models often plan solutions correctly yet stumble within individual reasoning steps, especially when guided by coarse, problem-level demonstrations. This paper introduces a probe- and-guide in-context learning framework that aligns guidance to the step granularity: the model first issues a brief probe for the next step, then retrieves and conditions on closely matched example steps from a cu- rated repository to execute the step with higher fidelity. The approach reduces irrelevant-example noise, improves single-step correctness with- out additional training, and slots into standard inference pipelines and search-based decoders, enhancing both candidate generation and verifi- cation. Evaluations across diverse mathematical reasoning settings show consistent gains over zero-shot and few-shot baselines, and the method composes naturally with tree-search strategies to further improve solu- tion quality while controlling token cost. The design is model-agnostic, training-free, and centers on LLM inference workflows, making it practi- cal for deployments that demand reliable, fine-grained reasoning.

Publish Year: 2026

Download PDF
Research Assistant Harinath Reddy Cingapuram Computer Science: Artificial Intelligence
University at Buffalo SUNY
Seeking for Research Collaborators
Open 3 weeks, 4 days ago

Looking for researchers interested in advancing large language model alignment techniques, reinforcement learning from human feedback (RLHF…

United States
Research Assistant Harinath Reddy Cingapuram Computer Science: Artificial Intelligence
University at Buffalo SUNY
Seeking for Research Collaborators
Open 3 weeks, 4 days ago

Looking for researchers interested in advancing large language model alignment techniques, reinforcement learning from human feedback (RLHF…

United States