Home Job Listings Categories Locations

Data Engineer - Pentaho Data Integration & Modernization - PERMANENT, REMOTE

๐Ÿ“ India

Technology datavruti

Job Description

Hiring for:

A subsidiary of a global software investor with a 100B+ market capitalization across portfolio companies. Role:

Data Engineer - Pentaho Data Integration & Modernization - On-prem & Cloud Experience:

3 to 6 years in Pentaho, overall could be more Location:

Permanent Remote, India Salary:

Based on fitment Notice Period:

30 days preferred

Overview: Our client is modernizing our data integration landscape and is seeking a skilled

Pentahoย Developer / Data Engineer

who is excited about transforming legacy ETL systems into scalable, cloud ready data pipelines using industry-leading technologies.

You will help maintain and enhance existing

Pentaho Data Integration (PDI/Kettle)

processes while playing a key role in their

migration to modern data engineering platforms.

You will collaborate with technical and

business teams across their portfolio to ensure data flows are robust, efficient, and future ready.

This role is ideal for a developer / data engineer who enjoys solving complex data problems and is

passionate about modernization, performance, and continuous improvement.

Key Responsibilities Develop, maintain, and optimize ETL workflows using

Pentaho Data Integration

(PDI/Kettle) . Document and analyze existing Pentaho jobs, data flows, dependencies, and

performance bottlenecks. Contribute to and execute the migration strategy

from Pentaho to modern ETL/data

integration platforms , such as: Talend / Talend Cloud, Informatica Cloud (IICS), Azure Data Factory, AWS Glue, dbt (Data Build Tool), Snowflake pipelines (Tasks, Streams, Snowpipe). Work with architects and engineering leads to shape target-state data integration

architecture. Implement data validation, quality checks, and reconciliation processes. Assist in building scalable, maintainable, and secure data pipelines across on-prem

and cloud environments. Support production workloads, resolve incidents, and ensure reliability of critical

data processes. Produce clear technical documentation and migration runbooks.

Required Qualifications 3 to 6+ years of hands-on experience with Pentaho PDI/Kettle or similar ETL platforms. Strong SQL development skills and solid understanding of relational database concepts. Experience with at least one modern ETL/cloud data integration tool, including: Talend, Informatica Cloud (IICS), Azure Data Factory, AWS Glue, dbt. Understanding of cloud ecosystems (Azure, AWS, or GCP). Familiarity with Git-based version control, CI/CD, and ETL/job scheduling tools. Excellent problem-solving skills, attention to detail, and ability to work

collaboratively in a distributed team.

Preferred Skills Experience with modern cloud data warehouses: Snowflake, BigQuery, Redshift,

etc. Exposure to data modeling (dimensional, star schema, SCD patterns). Python scripting for automation and data engineering tasks. Experience supporting ETL modernization or cloud migration projects.

Ready to Apply?

Don't miss this opportunity! Apply now and join our team.

Job Details

Posted Date: December 18, 2025
Job Type: Technology
Location: India
Company: datavruti

Ready to Apply?

Don't miss this opportunity! Apply now and join our team.