We are looking for candidate who is a hands-on Data Engineer with 3–5 years of experience building scalable batch and real-time data pipelines and delivering production-grade data products on modern cloud platforms. The candidate will bring strong programming and data querying skills, along with a proven ability to improve performance through efficient storage and processing design. The candidate is comfortable defining data quality checks, implementing automated tests, and setting up reliable release processes to support continuous delivery. Strong problem-solver with clear communication, collaborative teamwork, analytical thinking, and a structured, quality-focused approach.
Your tasks and responsibilities:
You build and optimize scalable batch and streaming data pipelines using Spark 3.x with RDDs, DataFrames, Spark SQL, and Structured Streaming.
You develop and maintain robust Lakehouse solutions based on Medallion Architecture using Parquet and Delta formats.
You work with Databricks capabilities such as Workflows, SQL Warehouses/Endpoints, Delta Live Tables, Pipelines, Unity Catalog, and Auto Loader.
You contribute to data storage and partitioning strategies considering distribution, data skew, compaction, and overall big data storage efficiency.
You contribute to development using Python and functional programming principles and collaborate effectively with engineering teams using IntelliJ, PyCharm, Git, Azure DevOps, and GitHub Copilot.
You implement data quality and test strategies using pytest and Great Expectations.
You build and maintain CI/CD pipelines using Azure Pipelines YAML, ensuring continuous delivery and acceptance testing.
You collaborate with architects, analysts, and product teams to define scalable data models and reliable data processing solutions.
You support modern data platform evolution by contributing to best practices in data governance, storage design, and platform standardization.
To find out more about the specific business, have a look at Magnetic Resonance Imaging
Your qualifications and experience:
You have successfully completed a university degree (Bachelor/Masters/PhD) in Computer Science, Data Engineering, Data Science, Information Technology, or a similar field
You have 3 to 5 years of professional experience in big data engineering, distributed data processing, and cloud-based data platform development.
You have strong hands-on experience with Spark 3.x, including batch and streaming workloads using RDDs, DataFrames, and SQL.
You have solid experience with Databricks, including Workflows, SQL Warehouses, DLT, Pipelines, Unity Catalog, and Auto Loader.
You have strong knowledge of Lakehouse and Medallion Architecture, Delta/Parquet formats, partitioning, distribution, compaction, and performance tuning.
You are proficient in Python and have a good understanding of functional programming concepts.
You have strong SQL skills, including Spark SQL, HiveQL, and T-SQL.
You have practical experience in CI/CD, Azure Pipelines YAML, continuous delivery, and testing practices for data platforms.
You have experience with data validation and testing frameworks such as pytest and Great Expectations.
It is beneficial if you also have exposure to ADF, Synapse Pipelines, Airflow, Oozie, Scala, Java, NoSQL databases, Hadoop ecosystem components, data catalog tools, or cube technologies such as SSAS/AAS/Tabular.
Your attributes and skills:
Since the development teams are spread internationally across multiple locations, communication in English is not a problem for you.
You are proactive in solving complex engineering problems and delivering reliable, high-quality data solutions.
You present your ideas and technical results confidently and convincingly in cross-functional development teams.
Personally, you are characterized by strong teamwork and cooperation skills, analytical thinking, and a structured way of working.
You are passionate about scalable data systems, engineering excellence, and continuous improvement.
The highest quality standards in product development are a matter of course for you.
Medical Equipment Manufacturing Hospitals and Health Care and Biotechnology Research
What We Offer
About the Company
Searching, interviewing and hiring are all part of the professional life. The TALENTMATE Portal idea is to fill and help professionals doing one of them by bringing together the requisites under One Roof. Whether you're hunting for your Next Job Opportunity or Looking for Potential Employers, we're here to lend you a Helping Hand.
Disclaimer: talentmate.com is only a platform to bring jobseekers & employers together.
Applicants
are
advised to research the bonafides of the prospective employer independently. We do NOT
endorse any
requests for money payments and strictly advice against sharing personal or bank related
information. We
also recommend you visit Security Advice for more information. If you suspect any fraud
or
malpractice,
email us at abuse@talentmate.com.
You have successfully saved for this job. Please check
saved
jobs
list
Applied
You have successfully applied for this job. Please check
applied
jobs list
Do you want to share the
link?
Please click any of the below options to share the job
details.
Report this job
Success
Successfully updated
Success
Successfully updated
Thank you
Reported Successfully.
Copied
This job link has been copied to clipboard!
Apply Job
Upload your Profile Picture
Accepted Formats: jpg, png
Upto 2MB in size
Your application for Data Analysis And Simulation Professional
has been successfully submitted!
To increase your chances of getting shortlisted, we recommend completing your profile.
Employers prioritize candidates with full profiles, and a completed profile could set you apart in the
selection process.
Why complete your profile?
Higher Visibility: Complete profiles are more likely to be viewed by employers.
Better Match: Showcase your skills and experience to improve your fit.
Stand Out: Highlight your full potential to make a stronger impression.
Complete your profile now to give your application the best chance!