⚡ New

Manager

PwC

BengaluruFull-timeMid LevelOn-site

Job Description

Job Role: Senior Data Engineer (Databricks | Insurance | Data Lakehouse) Location: Mumbai (Preferred) / Bangalore Experience: 7–12 years Start: Immediate Joiner Preferred Role Overview We are looking for a highly skilled Manager / Senior Data Engineer with deep expertise in Databricks data management , logical & physical data modelling , and insurance domain data workflows . The candidate will work on a strategic data migration initiative for a leading UK-based insurance company, moving data from Guidewire into Databricks Silver and Gold layers with strong governance, lineage, and scalability standards. Key Responsibilities Databricks Data Engineering & Management Design, build, and optimize Silver and Gold layer data pipelines in Databricks using PySpark, SQL, Delta Lake, and Workflow orchestration.

Implement data quality, lineage, schema evolution, and governance controls across curated layers. Optimize Databricks jobs for performance, scalability, and cost efficiency . Guidewire → Databricks Migration Lead the end-to-end migration of large-scale insurance data from Guidewire PolicyCenter/ClaimCenter/BillingCenter into Databricks.

Map and transform complex Guidewire entity structures into normalized and star-schema models. Data Modelling & Architecture Develop robust logical and physical data models aligned to insurance business processes. Build high-quality curated data marts (Gold) for analytics, reporting, pricing, underwriting, and claims.

Define standards for metadata, naming conventions, partitioning, and model documentation. Insurance Domain Expertise Understand core insurance data entities such as policy, claims, billing, customer, underwriting, rating, and product hierarchies . Apply domain knowledge to rationalize Guidewire data structures and create business-ready datasets.

Solutioning & Ideation Collaborate with client SMEs, architects, and business analysts to shape data solutions and propose design improvements. Ability to ** ideate, simplify complex data flows**, and contribute to overall solution architecture. Required Skills & Experience Technical 7–12 years of experience in data engineering, data modelling, and data management .

Strong hands-on experience in Databricks, Delta Lake, PySpark, Spark SQL, and ETL/ELT pipelines . Expertise in logical & physical data modelling (3NF, Star Schema, Data Vault preferred). Practical knowledge of Guidewire data model and prior migration experience (mandatory).

Experience working with large-scale insurance datasets Strong understanding of data quality frameworks , lineage, cataloging, and governance. Soft Skills Strong problem-solving and conceptualization / ideation capability. Excellent communication and stakeholder-management for UK client environment.

Ability to work in fast-paced delivery tracks with cross-functional global teams. Preferred Qualifications Certifications in Databricks , Azure/AWS , and Data Warehousing are added advantages. Experience delivering enterprise-grade data lakes or lakehouse architectures .

Why Join This Role? Work on a flagship insurance data modernisation project for a top UK carrier. Opportunity to shape enterprise-scale data models on the Databricks Lakehouse .

High-visibility role with strong career growth in insurance data engineering .

Posted 2 days ago

Related Jobs

Analyst

Hubell

Kochi Today 1 views
Full-time On-site Mid Level

Business Analyst

ATLANTA SYSTEMS PVT LTD

New Delhi Today 1 views
Full-time On-site Mid Level

Related Searches

Apply Now