Role overview
A role within a forward-thinking AI/ML-focused team committed to building scalable, production-grade AI workflows. If you’re passionate about Python, API-driven AI applications, LLMs, and responsible AI, we want you on the team.
What You’ll Do
- Design and develop AI workflows and production-grade APIs using Python (FastAPI/Flask).
- Lead prompt engineering initiatives and implement RAG (Retrieval-Augmented Generation) architectures and agentic AI patterns.
- Build, deploy, and maintain AI models across all environments (Dev/QA/Prod) with robust monitoring, testing, and documentation.
- Work with data processing pipelines and AI-enabled workflows; implement hybrid prompting techniques and advanced generative AI solutions. -Leverage GenAI tools and frameworks (LangChain, Hugging Face, LlamaIndex, etc.) and explore Gemini LLMs when applicable.
- Collaborate on model deployment and containerization (Docker), version control (Git), and CI/CD practices.
- Conduct code reviews, write unit/integration tests, and contribute to technical documentation.
- Stay at the cutting edge of Generative AI, NLP, ML frameworks (PyTorch, TensorFlow, Keras), and best practices.
What We’re Looking For Must Have:
- Strong hands-on Python experience.
- Hands-on API development in Python using FastAPI or Flask.
- Experience with data processing and AI-enabled workflows in Python.
- Knowledge of LLM concepts, prompt engineering, RAG, and agentic AI.
- Basic knowledge of Hybrid prompting techniques.
- Experience with Generative AI tools (LangChain, Hugging Face, LlamaIndex, etc.).
Hybrid in office 3x days a week in Irvine, TX