Job Description
Hiring for:
A subsidiary of a global software investor with a 100B+ market capitalization across portfolio companies.
Role:
Data Engineer - Pentaho Data Integration & Modernization - On-prem & Cloud
Experience:
3 to 6 years in Pentaho, overall could be more
Location:
Permanent Remote, India
Salary:
Based on fitment
Notice Period:
30 days preferred
Overview:
Our client is modernizing our data integration landscape and is seeking a skilled
Pentahoย Developer / Data Engineer
who is excited about transforming legacy ETL systems into scalable, cloud ready data pipelines using industry-leading technologies.
You will help maintain and enhance existing
Pentaho Data Integration (PDI/Kettle)
processes while playing a key role in their
migration to modern data engineering platforms.
You will collaborate with technical and
business teams across their portfolio to ensure data flows are robust, efficient, and future ready.
This role is ideal for a developer / data engineer who enjoys solving complex data problems and is
passionate about modernization, performance, and continuous improvement.
Key Responsibilities
Develop, maintain, and optimize ETL workflows using
Pentaho Data Integration
(PDI/Kettle) .
Document and analyze existing Pentaho jobs, data flows, dependencies, and
performance bottlenecks.
Contribute to and execute the migration strategy
from Pentaho to modern ETL/data
integration platforms , such as: Talend / Talend Cloud, Informatica Cloud (IICS), Azure Data Factory, AWS Glue, dbt (Data Build Tool), Snowflake pipelines (Tasks, Streams, Snowpipe).
Work with architects and engineering leads to shape target-state data integration
architecture.
Implement data validation, quality checks, and reconciliation processes.
Assist in building scalable, maintainable, and secure data pipelines across on-prem
and cloud environments.
Support production workloads, resolve incidents, and ensure reliability of critical
data processes.
Produce clear technical documentation and migration runbooks.
Required Qualifications
3 to 6+ years of hands-on experience with Pentaho PDI/Kettle or similar ETL platforms.
Strong SQL development skills and solid understanding of relational database concepts.
Experience with at least one modern ETL/cloud data integration tool, including: Talend, Informatica Cloud (IICS), Azure Data Factory, AWS Glue, dbt.
Understanding of cloud ecosystems (Azure, AWS, or GCP).
Familiarity with Git-based version control, CI/CD, and ETL/job scheduling tools.
Excellent problem-solving skills, attention to detail, and ability to work
collaboratively in a distributed team.
Preferred Skills
Experience with modern cloud data warehouses: Snowflake, BigQuery, Redshift,
etc.
Exposure to data modeling (dimensional, star schema, SCD patterns).
Python scripting for automation and data engineering tasks.
Experience supporting ETL modernization or cloud migration projects.
Ready to Apply?
Don't miss this opportunity! Apply now and join our team.
Job Details
Posted Date:
December 18, 2025
Job Type:
Technology
Location:
India
Company:
datavruti
Ready to Apply?
Don't miss this opportunity! Apply now and join our team.