Senior AI Pipeline Engineer (Python, AWS, LLMs) – Document Intelligence / CLM
Actively Reviewing the ApplicationsInhubber
On-site
Posted 2 weeks ago
•
Apply by June 10, 2026
Job Description
About Inhubber
Inhubber is a security-first, AI-powered Contract Lifecycle Management (CLM) platform built for organizations with high compliance and data protection requirements. We combine end-to-end encryption and modern cloud architecture with advanced AI to extract, analyze, and generate contract intelligence.
Our platform processes sensitive legal documents for companies worldwide. We are now scaling our AI capabilities and looking for a senior engineer to take ownership of our production AI pipelines and help build the next generation of document intelligence.
The Role
We are looking for a hands-on Senior AI Pipeline Engineer (Python) who will:
What Youll Do
1. Own & Extend Production Pipelines
Weeks 1–2
Inhubber is a security-first, AI-powered Contract Lifecycle Management (CLM) platform built for organizations with high compliance and data protection requirements. We combine end-to-end encryption and modern cloud architecture with advanced AI to extract, analyze, and generate contract intelligence.
Our platform processes sensitive legal documents for companies worldwide. We are now scaling our AI capabilities and looking for a senior engineer to take ownership of our production AI pipelines and help build the next generation of document intelligence.
The Role
We are looking for a hands-on Senior AI Pipeline Engineer (Python) who will:
- Own and optimize our production extraction pipelines.
- Deliver new document-analysis pipelines end-to-end.
- Build the foundation for next-generation GenAI features (agentic contract drafting & interpretation).
What Youll Do
1. Own & Extend Production Pipelines
- Maintain and optimize our Python-based extraction pipelines (AWS Lambda + S3 + Docker components).
- Ensure stable document processing and downstream triggering.
- Improve observability: logging, metrics, alerting, traceability, cost monitoring.
- Debug and stabilize real-world failure modes in production.
- Design and implement end-to-end pipelines for new document families.
- Build evaluation datasets and regression tests.
- Prevent silent quality degradation through measurable metrics.
- Improve Q&A and structured extraction using LLMs.
- Implement structured outputs, retrieval (RAG where useful), and deterministic validation.
- Add robust failure handling (timeouts, retries, fallbacks, safe defaults).
- Build agentic building blocks in Python behind stable APIs.
- Contribute to a contract-generation/editing kernel (planner, drafter, risk checks).
- Collaborate with backend/frontend teams for clean integration.
- Ensure scalability, cost-efficiency, and security.
- Contribute to deployment/versioning/rollback strategies.
- Help define operational runbooks.
- Strong production-grade Python (clean architecture, testing, packaging, APIs).
- Experience owning code in production.
- AWS serverless (Lambda + S3 required; Step Functions/SQS/CloudWatch a plus).
- Docker and containerized services.
- Proven experience maintaining/debugging automated pipelines.
- Hands-on experience with LLMs (OpenAI/Azure OpenAI/Anthropic or similar):
- Structured outputs
- Prompt iteration
- Retrieval (RAG)
- Evaluation approaches
- Document AI experience (OCR, layout extraction, noisy PDFs).
- Evaluation-driven development (test sets, regression checks, quality metrics).
- Experience with cost/latency budgeting.
- Familiarity with TypeScript/Node.
- Experience integrating REST services.
Weeks 1–2
- Fully understand current pipeline architecture.
- Stabilize staging/local environments.
- Define quality, cost, and latency baselines.
- Improve logging and monitoring.
- Deliver a new production-ready pipeline for a new document family.
- Implement evaluation datasets and regression checks.
- Deploy with monitoring and rollback strategy.
- Improve Q&A/extraction accuracy measurably.
- Deliver a first version of a GenAI contract drafting kernel.
- Harden operations (cost controls, retries, fallbacks, documentation).
- Frontend: React (TypeScript)
- Backend: Java (JEE)
- AI Pipelines: Python (AWS Lambda), Dockerized OCR/NLP
- Storage: AWS S3
- LLMs: Azure-hosted LLM APIs
- Infrastructure: AWS + Azure (hybrid)
- Senior-level role
- Long-term collaboration preferred
- Strong ownership mindset required
- Experience in regulated/sensitive data environments is a strong plus
Required Skills
Documentation
Monitoring
Python
AWS
Editing
TypeScript
Docker
Interpretation
Azure
React
Testing
Lambda
NLP
Anthropic
Debugging
RAG
Validation
Regression
Risk
SQS
AWS Lambda
Packaging
OCR
Node
Clean architecture
Azure OpenAI
Traceability
PDFs
Rest services
Kernel
Pipeline architecture
Logging
OpenAI
JEE
Extraction
LLM APIs
Regression tests
AWS S3
GenAI
Retrieval
Java
LLMs
Budgeting
Storage
Observability
Rollback
Drafting
Planner
SAFe
LLM
Serverless
Step Functions
Quick Tip
Customize your resume and cover letter to highlight relevant skills for this position to increase your chances of getting hired.
Related Similar Jobs
View All
Door & Gates Engineer
H & J Martin Limited
Engineering
Repair
Safety
+31
Surface Treatment Specialist
ARVIND LTD CO
India
Full-Time
Safety
Safety Standards
Maintenance
+9
Java Software Engineer
Wenger & Watson
India
Full-Time
₹8–17 LPA
Java
Vice President Data Operations - Transformation & Automation
MSCI Inc.
India
Full-Time
Communication
Engineering
Team Management
+65
Program Business Analyst (Oklahoma REMOTE)
ICF
₹1–1 LPA
Reporting
Documentation
Compliance
+10
Share
Quick Apply
Upload your resume to apply for this position