Skip to Content

IND - Staff Engineer, Reliability

Apply
Share Job Back

Job Details

Location:
Hyderabad, Telangāna, IN
Category:
Information Technology
Employment Type:
Full time
Job Ref:
R2624792-333

IND - Staff Engineer, Reliability - GCC070

We’re determined to make a difference and are proud to be an insurance company that goes well beyond coverages and policies. Working here means having every opportunity to achieve your goals – and to help others accomplish theirs, too. Join our team as we help shape the future.

Key Responsibilities 

 

  • Data Reliability & Quality:Establish and enforce Data Service Level Objectives (SLOs) focused on data freshness, completeness, and accuracy across critical data products. 

  • Data Observability: Implement advanced data observability tools to monitor the entire data journey—from ingestion to consumption—detecting data quality anomalies, schema drifts, and pipeline delays in real-time. 

  • Pipeline Resiliency & Automation: Collaborate with Data Engineering to embed reliability patterns into data pipelines built using Informatica, Python/Pyspark, and running on platforms like Amazon EMR/Hadoop, Informatica and cloud native services. 

  • Toil Elimination in Data Operations: Automate data validation, data reprocessing, data backfilling, and other manual operational tasks within the data lifecycle to reduce toil and improve operational efficiency. 

  • Incident and Problem Management (Data Focus): Lead the response and resolution for data-related incidents (e.g., corrupt data, delayed reporting), ensuring fast recovery and effective post-incident reviews (blameless post-mortems). 

  • Runbook Creation & Automation (Data Focus): Develop and automate sophisticated, data-aware runbooks for common data pipeline failures, data quality issues, and data recovery scenarios. 

 

Required Skills & Experience 

 

  • 8+ year’s overall experience in an Infrastructure, Data or related technology organization with increasing responsibilities as a hands-on technologist. 

  • 2-3+ year experience in Data Engineering, Data Quality, or a specialized SRE role within an enterprise data environment. 

  • Hands-on experience with data warehousing and data lake technologies, including Snowflake, and cloud environments (AWS/GCP). 

  • Hands-on experience in pipeline development and support using technologies like Informatica, Python/Pyspark, and distributed compute (EMR/Hadoop). 

  • Experience in designing and implementing data quality checks, data validation frameworks, and data governance standards. 

  • Hands on experience in software or cloud engineering. Familiarity with cloud service providers and their core capabilities (compute, containers, databases, APIs etc.). 

  • In depth and hands on experience with data observability concepts and tools for monitoring data in motion and at rest (e.g., Monte Carlo, Bigeye, Astro Observe, Datafold, custom solutions). 

  • A strong understanding of the "data journey" and the impact of data issues on business outcomes. 

  • Expertise implementing AIOps to monitor, manage and self-heal data pipelines, using machine learning principles for anomaly detection.  

  • Experience with prompt engineering, implementing AWS or Google AI services, AI enabled automation for data quality, reliability and pipeline performance management. 

  • Expertise defining and implementing of DataOps practices

Apply
Share Job Back

A Glimpse Inside
The Hartford

About Us

We believe every day is a day to do right.

And that belief has guided us for over 200 years. Showing up for people isn’t just what we do, it’s who we are. We’re devoted to finding innovative ways to serve our customers, communities and employees – continually asking ourselves what more we can do.

And while how we contribute looks different for each of us, it’s these values that drive all of us to do more and to do better every day.