Job Description

Job Description

Skills: Python, FastAPI, LLMs.

Responsibilities: Scrape public data, build pipelines (bronze/silver/gold), and create APIs.

What You Will Do

  • Bring in industry best-practices around creating and maintaining robust data pipelines for complex data projects with / without AI component:
  • programmatically retrieve (unstructured mostly) data from several static and real-time sources (incl. web scraping, API use)
  • Structure this data into a structured format o Harmonize the data, into a common format and store it in a dedicated database.
  • Schedule the different jobs into a dedicated pipeline
  • rendering results through dynamic interfaces incl. web / mobile / dashboard with ability to log usage and granular user feedbacks
  • performance tuning and optimal implementation of complex Python scripts, SQL,…
  • Industrialize ML / DL solutions and deploy and manage production services; proactively handle data issues arising on live apps
  • Perform ETL on large and complex datasets for AI applications - work closely with data scientists on performance optimization of large-scale ML/DL model finetuning
  • Build data tools to facilitate fast data cleaning and statistical analysis
  • Build and ensure data architecture is secure and compliant
  • Resolve issues escalated from Business and Functional areas on data quality, accuracy, and availability
  • Work closely with APAC IT Transformation and coordinate with a fully decentralized team across different locations in APAC and global HQ (Paris).

You should be

  • Expert in structured and unstructured data in traditional and Big data environments Oracle / SQLserver, MongoDB, Hive / Pig, BigQuery and Spark
  • Have excellent knowledge of Python programming both in traditional and distributed models (PySpark)
  • Expert in shell scripting and writing schedulers
  • Hands-on experience with Cloud - deploying complex data solutions in hybrid cloud / on-premise environment both for data extraction / storage and computation
  • Experience working on industry standard services like Message Queue, Redis, Elastic Search, Kafka, or Spark Streaming
  • Well versed with DevOps best practices like containerization, CICD pipeline (Jenkins and Maven)
  • Hands-on experience in deploying production apps using large volumes of data with state-of-the-art technologies like Dockers, Kubernetes and Kafka
  • Strong knowledge of data security best practices


Job Details

Role Level: Mid-Level Work Type: Full-Time
Country: India City: Bengaluru East ,Karnataka
Company Website: https://www.cloudesign.com Job Function: Information Technology (IT)
Company Industry/
Sector:
IT Services and IT Consulting

What We Offer


About the Company

Searching, interviewing and hiring are all part of the professional life. The TALENTMATE Portal idea is to fill and help professionals doing one of them by bringing together the requisites under One Roof. Whether you're hunting for your Next Job Opportunity or Looking for Potential Employers, we're here to lend you a Helping Hand.

Report

Disclaimer: talentmate.com is only a platform to bring jobseekers & employers together. Applicants are advised to research the bonafides of the prospective employer independently. We do NOT endorse any requests for money payments and strictly advice against sharing personal or bank related information. We also recommend you visit Security Advice for more information. If you suspect any fraud or malpractice, email us at abuse@talentmate.com.


Recent Jobs
View More Jobs
Talentmate Instagram Talentmate Facebook Talentmate YouTube Talentmate LinkedIn