Home Job Listings Categories Locations

Azure Data Warehouse Developer

📍 India

Technology Ensar Solutions Inc

Job Description

Job Description: We are seeking an 5+ Years experienced Azure Data Warehouse Developer to join our growing data team. The ideal candidate will have a deep understanding of cloud-based data architecture, specifically within Microsoft Azure, and will play a key role in designing, implementing, and maintaining data warehouse solutions on Azure. You will collaborate with cross-functional teams to ensure scalable and high-performance data solutions.

Roles & Responsibilities: Designing and implementing dimensional models — star and snowflake schemas, slowly changing dimensions (SCD Types 1, 2, 3), and fact/dimension table strategies Writing complex, optimized T-SQL including window functions, CTEs, dynamic SQL, and query tuning Table design considerations within Fabric Warehouse: distribution strategies, indexing, partitioning, and statistics management Data quality enforcement through constraints, validation logic, and reconciliation frameworks Understanding of Fabric Warehouse's current limitations vs. Azure Synapse Dedicated SQL Pools, and how to work within them Hands-on Synapse Analytics experience — dedicated/serverless SQL pools, distribution strategies Deep ADLS Gen2 knowledge — hierarchical namespace, Parquet/Delta file formats, partitioning strategies Storage security — RBAC, ACLs, managed identities, private endpoints Managing large-scale file formats: Parquet, Delta, CSV, JSON, and Avro Experience connecting to source systems: databases, REST APIs, flat files, SaaS platforms Query and storage performance tuning — file size optimization, indexing, partitioning Hands-on experience with Fabric's T-SQL Warehouse and understanding how it differs from Synapse Knowledge of cross-database and cross-item queries spanning Lakehouse and Warehouse Understanding of Fabric's V-Order optimization and Delta Lake file management Identifying and resolving data skew, spill, and shuffle issues in Spark workloads Monitoring capacity usage, query execution plans, and resource contention Experience building semantic models directly on top of Fabric Warehouse or Lakehouse Understanding of when to choose Fabric DW over Lakehouse SQL endpoint for a given workload OneLake architecture — shortcuts to ADLS Gen2, Lakehouse vs. Warehouse trade-offs Medallion architecture implementation (Bronze/Silver/Gold) Lakehouse mirroring and real-time data ingestion patterns Fabric Data Pipelines, Notebooks (PySpark/SQL), and Dataflows Gen2 Delta Lake operations — OPTIMIZE, VACUUM, Z-ordering, time travel Incremental load strategies: watermark patterns, change data capture (CDC), and merge/upsert logic Implementing row-level security (RLS) and column-level security (CLS) in Fabric Warehouse Managing workspace roles, item permissions, and One Lake access controls Understanding of data lineage tracking and cataloging within Fabric Compliance awareness: GDPR, HIPAA, or industry-specific data handling requirements

Required Skills:

Direct Lake mode in Power BI and semantic model design End-to-end security model spanning ADLS and Fabric Row-level and column-level security in DW and semantic layers Sensitivity labels and Microsoft Information Protection Designing end-to-end solutions where ADLS underpins Fabric via One Lake shortcuts ELT/ETL pattern selection and pipeline orchestration Microsoft Purview for governance and lineage across ADLS and Fabric Multi-workspace and multi-domain architecture patterns for enterprise environments Disaster recovery, backup strategies, and SLA considerations within Fabric Fabric REST APIs and CI/CD using Azure DevOps or GitHub Comfortable with Fabric's rapid evolution and incomplete documentation Experience with integrating with Human in the middle workflows

Ready to Apply?

Don't miss this opportunity! Apply now and join our team.

Job Details

Posted Date: February 28, 2026
Job Type: Technology
Location: India
Company: Ensar Solutions Inc

Ready to Apply?

Don't miss this opportunity! Apply now and join our team.