Job Description

We are looking for a Lead Data Scientist - Vision & Multimodal AI to architect and build next-generation Vision-Language Model (VLM) systems at scale.

The Candidate Will Have Responsibilities Across The Following Functions

Architect and Build RLHF Frameworks:

  • Design end-to-end RLHF pipelines (SFT Reward Modelling PPO/DPO).
  • Develop scalable human feedback collection systems.
  • Implement preference modelling and ranking pipelines.
  • Optimise reward models for multimodal outputs (image + text).
  • Build automated evaluation frameworks.

Train And Fine-Tune OSS Vision-Language Models

  • Experience working with Qwen-VL, Llama, and GPT OSS.
  • Pretraining / instruction tuning multimodal models.
  • Parameter-efficient fine-tuning (LoRA, QLoRA).
  • Dataset curation and synthetic data generation.
  • Scaling training on multi-GPU / multi-node clusters.
  • Optimising for alignment, hallucination reduction, and safety.

Highly Scalable Deployment Of VLM Systems

  • Design distributed inference pipelines (GPU-optimised).
  • Model serving using vLLM and Triton Inference Server.
  • Optimise latency, throughput, and cost.
  • Implement batching, KV caching, quantisation, and tensor parallelism.
  • Deploy on a Kubernetes-based infrastructure.
  • Build monitoring for drift, performance, and hallucinations.

Multimodal AI System Design

  • Architect systems combining OCR, vision encoders, LLMs, and retrieval.
  • Implement retrieval-augmented multimodal pipelines.
  • Design evaluation benchmarks for VQA, grounding, and reasoning.
  • Ensure model safety and guardrails.

Technical Leadership

  • Lead a team of ML engineers and research scientists.
  • Define a technical roadmap for multimodal AI.
  • Review model architectures and code quality.
  • Collaborate with product and infrastructure teams.

Requirements

  • 6+ years in ML / AI.
  • 2+ years working with large-scale LLM or VLM systems.
  • Strong hands-on experience building RLHF pipelines (not just using libraries).
  • Deep PyTorch expertise.
  • Experience training models > 7B parameters.
  • Experience with distributed training (Deep Speed, FSDP).
  • Production-grade deployment experience handling 10k+ QPS workloads.
  • Strong understanding of transformer architectures.

This Role Requires Deep Expertise In

  • Architecting and implementing RLHF (Reinforcement Learning from Human Feedback) Frameworks.
  • Training and fine-tuning Open-Source Vision-Language Models (VLMs).
  • Deploying and scaling multimodal models to production, serving millions of requests.

This job was posted by Eucloid Careers from Eucloid Data Solutions.


Job Details

Role Level: Mid-Level Work Type: Full-Time
Country: India City: Chennai ,Tamil Nadu
Company Website: https://www.eucloid.com/ Job Function: Analyst
Company Industry/
Sector:
IT Services And IT Consulting Software Development And Technology Information And Internet

What We Offer


About the Company

Searching, interviewing and hiring are all part of the professional life. The TALENTMATE Portal idea is to fill and help professionals doing one of them by bringing together the requisites under One Roof. Whether you're hunting for your Next Job Opportunity or Looking for Potential Employers, we're here to lend you a Helping Hand.

Report

Disclaimer: talentmate.com is only a platform to bring jobseekers & employers together. Applicants are advised to research the bonafides of the prospective employer independently. We do NOT endorse any requests for money payments and strictly advice against sharing personal or bank related information. We also recommend you visit Security Advice for more information. If you suspect any fraud or malpractice, email us at abuse@talentmate.com.


Recent Jobs
View More Jobs
Talentmate Instagram Talentmate Facebook Talentmate YouTube Talentmate LinkedIn