Company Overview 10Pearls is an award-winning end-to-end digital innovation company that helps businesses imagine and build the future. We are proud to announce that 10Pearls was named as winner of the Best Tech Work Culture Timmy Award in Washington DC by Tech in Motion, recognized on the Inc. 5000 Fastest-Growing Companies List, and was ranked the #1 Most Diverse Midsize Company in Greater Washington. We partner with businesses to help them transform, scale, and accelerate by adopting digital and exponential technologies. Our work has ranged from creating highly usable, secure digital experiences, mobile and software products, to helping businesses modernize through cloud adoption and development and the digitalization of their business processes. Our clientele is highly diverse, including Global 1000 enterprises, mid-market businesses, and high-growth start-ups. But those are just the facts. What makes us unique is that we have true heart and soul. We have a strong focus on a double bottom line and actively support and engage with the communities where we live and work to make the world a better place. In a nutshell, we believe in doing well, while doing good, and know how to balance the two.
Role 10Pearls is seeking a Senior AI Engineer to design, build, and scale production‑grade Generative AI systems. This role is ideal for an engineer who has moved beyond experimentation and understands what it takes to operate LLM‑powered applications reliably in real‑world environments. You will work at the intersection of LLMOps, Retrieval‑Augmented Generation (RAG), orchestration frameworks, and cloud AI infrastructure, building intelligent workflows that power customer‑facing products across web, messaging, and enterprise platforms. This is a hands‑on engineering role with direct ownership of AI system design, performance, reliability, and evolution.
Responsibilities
Design and implement LLM‑powered applications using frameworks such as LangChain and LangGraph
Build, tune, and maintain RAG (Retrieval‑Augmented Generation) pipelines, including ingestion, chunking, embeddings, retrieval, and grounding
Integrate and manage Azure OpenAI models for scalable, production‑ready inference
Develop prompt engineering strategies, including prompt templates, versioning, and systematic evaluation
Implement LLMOps pipelines for deployment, monitoring, experimentation, and continuous improvement of AI systems
Optimize performance, latency, accuracy, and cost across LLM workflows
Work with Redis or similar caching/session stores to manage conversational state and improve response times
Collaborate closely with product managers, backend engineers, and frontend teams to ship AI‑driven features end‑to‑end
Monitor model and system behavior in production and proactively address quality, safety, and reliability concerns
Apply best practices around scalability, security, observability, and fault tolerance in AI systems
Requirements
Bachelor’s or Master’s degree in Computer Science, AI, Software Engineering, or a related field
4–8 years of experience in AI/ML engineering, backend engineering, or applied data systems
Strong hands‑on experience building LLM‑based or Generative AI applications
Excellent proficiency in Python, including building production APIs and services
Experience working with Azure OpenAI, Azure AI Search, or comparable cloud AI platforms
Solid understanding of RAG architectures, vector embeddings, and semantic search
Experience with LangChain, LangGraph, or similar AI orchestration frameworks
Exposure to LLMOps/MLOps practices, including deployment, monitoring, evaluation, and versioning
Strong problem‑solving skills and the ability to reason about trade‑offs in complex systems
Nice to Have
Experience designing multi‑agent or multi‑step AI workflows
Hands‑on work with vector databases, embedding optimization, and retrieval tuning
Experience optimizing LLM cost vs. performance trade‑offs at scale
Familiarity with cloud infrastructure, CI/CD pipelines, and observability tools
Experience working on customer‑facing AI products in regulated or production‑critical environments