Home Job Listings Categories Locations

Data / Generative AI Engineers

📍 India

Technology Sharc Hire

Job Description

Details:

- Contract Duration: Min 3–6 months - Work Timing: 8:00 AM – 4:00 PM EST - Start Timeline: Within 2 weeks

Position Overview

We are seeking experienced Data/GenAI Engineers to join our Professional Services

team on a contract basis. You will work directly on client engagements delivering

production-grade Generative AI solutions, including conversational AI assistants,

document processing automation, RAG (Retrieval-Augmented Generation) systems,

and AI-powered data analytics platforms. This role requires hands-on technical

execution, client interaction, and the ability to work independently within an agile

delivery framework.

Primary Responsibilities

GenAI Solution Development

● Design and implement production-ready Generative AI applications using

Amazon Bedrock, Anthropic Claude, and other foundation models

● Build and optimize RAG (Retrieval-Augmented Generation) pipelines with vector

databases (Weaviate, OpenSearch, Pinecone)

● Develop AI agents and multi-agent orchestration systems using frameworks like

LangChain, LlamaIndex, or custom implementations

● Create conversational AI interfaces with natural language understanding, intent

detection, and context management

● Implement prompt engineering strategies, few-shot learning, and fine-tuning

approaches for domain-specific applications

AWS Cloud Architecture & Development

● Build serverless architectures using AWS Lambda, API Gateway, Step Functions,

and EventBridge

● Design and implement data pipelines for AI model training, inference, and

feedback loops

● Develop RESTful APIs and WebSocket connections for real-time AI interactions

● Configure and optimize AWS services including S3, DynamoDB, RDS, SQS,

SNS, and CloudWatch

● Implement infrastructure-as-code using CloudFormation, CDK, or Terraform

Data Engineering & ML Operations

● Design and build data ingestion pipelines for structured and unstructured data

sources

● Implement ETL/ELT workflows for data preparation, cleaning, and transformation

● Create vector embeddings and semantic search capabilities for knowledge

retrieval

● Develop data validation, quality monitoring, and observability frameworks

● Optimize model inference performance, latency, and cost efficiency

Client Engagement & Delivery

● Participate in sprint planning, daily standups, and client review sessions

● Translate business requirements into technical specifications and implementation

plans

● Provide technical guidance and recommendations to clients on AI/ML best

practices

● Document architecture decisions, code, and deployment procedures

● Troubleshoot production issues and implement solutions quickly

Required Technical Skills (Priority Order)

Tier 1 - Critical Must-Haves

● Amazon Bedrock - Hands-on experience with foundation models (Claude, Nova,

Llama or others), model invocation, streaming responses, and guardrails

● Agent Frameworks & Orchestration - Production experience with LangChain,

LlamaIndex, Bedrock Agents, or custom multi-agent orchestration systems

● Python - Advanced proficiency with modern Python (3.9+), including async/await,

type hints, and testing frameworks (pytest, unittest)

● AWS Lambda & Serverless - Production experience building event-driven

architectures, function optimization, and cold start mitigation

● Vector Databases - Practical experience with at least one: Weaviate,

OpenSearch, Pinecone, Chroma, or FAISS for semantic search

● LLM Integration - Direct experience with LLM APIs (Anthropic, OpenAI, Cohere),

prompt engineering, and response parsing

● API Development - RESTful API design and implementation using FastAPI,

Flask, or similar frameworks

Tier 2 - Highly Valuable

● Amazon Bedrock AgentCore - Experience with AgentCore Runtime, Memory,

Gateway, and Observability for building production agent systems

● AWS API Gateway - Configuration, authorization, throttling, and integration with

Lambda/backend services

● DynamoDB - NoSQL data modeling, single-table design, GSI/LSI optimization,

and DynamoDB Streams

● AWS Step Functions - Workflow orchestration for complex AI pipelines and

multi-step processes

● Docker & Containers - Containerization, ECR, ECS/Fargate deployment for AI

workloads

● Data Processing - Experience with Pandas, PySpark, AWS Glue, or similar data

transformation tools

Tier 3 - Strong Differentiators

● RAG Architecture - End-to-end RAG system design including chunking

strategies, retrieval optimization, and context management

● Embedding Models - Working knowledge of text embeddings (Bedrock Titan,

OpenAI, Cohere) and embedding optimization

● AWS S3 & Data Lakes - S3 event notifications, lifecycle policies, and data lake

architecture patterns

● CloudWatch & Observability - Logging, metrics, alarms, and distributed tracing

for AI applications

● IAM & Security - AWS security best practices, least privilege access, secrets

management (Secrets Manager, Parameter Store)

● CI/CD Pipelines - Experience with CodePipeline, GitHub Actions, or GitLab CI for

automated deployments

Tier 4 - Nice to Have

● SageMaker - Model training, deployment, endpoints, and feature stores

● OpenSearch - Full-text search, vector search, and hybrid search implementations

● EventBridge - Event-driven architectures and cross-service integrations

● WebSockets - Real-time bidirectional communication for streaming AI responses

● AWS CDK - Infrastructure-as-code using Python or TypeScript CDK constructs

● Fine-tuning & Training - Experience with model fine-tuning, PEFT methods, or

custom model training

Required Experience & Qualifications

● 5+ years of software engineering experience with at least 2+ years focused on

AI/ML, data engineering, or cloud-native development

● 2+ years of hands-on AWS experience with production deployments

● 1+ years of direct Generative AI experience (LLMs, embeddings, RAG, agents)

● Proven track record delivering production AI applications from concept to

deployment

● Strong understanding of software engineering best practices (version control,

testing, code review, documentation)

● Experience working in agile/scrum environments with distributed teams

● Excellent problem-solving skills and ability to work independently with minimal

supervision

● Strong written and verbal communication skills for client-facing interactions

Preferred Qualifications

● AWS Certifications: Solutions Architect Associate/Professional, Machine

Learning Specialty, or Developer Associate

● Background in healthcare, financial services, or regulated industries with

understanding of compliance requirements (HIPAA, PCI-DSS, SOC 2)

● Contributions to open-source AI/ML projects or published technical content

● Experience with multi-tenant SaaS architectures and data isolation patterns

● Knowledge of cost optimization strategies for AI workloads (model selection,

caching, batching)

● Familiarity with frontend frameworks (React, Angular) for building AI-powered

UIs.

Project Examples You May Work On

● Building conversational AI assistants for customer service automation using

Bedrock and Anthropic Claude

● Implementing RAG systems for document processing, classification, and

intelligent search

● Developing AI-powered data extraction and validation pipelines for healthcare

claims processing

● Creating multi-agent systems for complex workflow automation and decision

support

● Building integration marketplaces connecting AI capabilities to third-party

platforms

● Designing voice AI solutions using Amazon Connect and Polly for customer

engagement

● Implementing AI-driven content recommendation and personalization engines.

Ready to Apply?

Don't miss this opportunity! Apply now and join our team.

Job Details

Posted Date: March 7, 2026
Job Type: Technology
Location: India
Company: Sharc Hire

Ready to Apply?

Don't miss this opportunity! Apply now and join our team.