Vanguard logo
Tech Lead – Data Engineer - Assets and Cash Flow
full-timeMalvern

Summary

Location

Malvern

Type

full-time

Explore Jobs

About this role

Vanguard is seeking a Tech Lead that is passionate about Vanguard and our funds. In this role, you will lead the system which articulates assets and cash flow across each of Vanguard's lines of business. This role will function as a critical liaison with Finance and Technology teams and will engage with clients across the enterprise.

We’re seeking a hands-on Tech Lead to own the design, build, and reliability of our modern data ETL pipelines and lakehouse platform. This leader will partner with product, analytics, and platform engineering to deliver resilient, cost‑effective data solutions that power high‑quality insights. The ideal candidate is deeply experienced with AWS serverless data services—Lambda, Glue, and Step Functions—and has production expertise operating Apache Iceberg tables at scale (schema evolution, partitioning, compaction, snapshot management, and metadata performance). You will set engineering standards, mentor developers, and drive the roadmap for data ingestion, transformation, and consumption across batch and event-driven pipelines.

Core Responsibilities

  • Leads end‑to‑end architecture and delivery of data pipelines.

  • Defines patterns for ingestion, transformation, storage, and serving across batch/streaming workflows; selects the right tools and designs SLAs/SLOs for availability, latency, and cost.

  • Designs and implements serverless data solutions using AWS Lambda, Glue, and Step Functions.

  • Builds Glue Jobs (Spark/Python) and orchestrates state machines for complex dependency management, error handling, retries, and parallelization.

  • Owns Apache Iceberg table design and lifecycle.

  • Establishes standards for partitioning, clustering, snapshot retention, compaction, merge/delete, schema evolution, and Glue/AWS Lake Formation catalog management.

  • Ensures performance, reliability, and observability.

  • Implements robust monitoring, logging, and alerting (e.g., CloudWatch, metrics, tracing) and leads incident response, root‑cause analysis, and continuous improvement.

  • Elevates code into dev/test/prod safely and on schedule.

  • Champions CI/CD and IaC (CloudFormation/CDK/Terraform), automated testing (unit/integration/data quality), canary deployments, and rollback strategies.

  • Establishes data quality and governance controls.

  • Builds validation frameworks, contracts, and lineage; integrates with catalog, access controls (IAM/Lake Formation), encryption, and compliance requirements.

  • Mentors engineers and sets engineering standards.

  • Provides code reviews, technical coaching, and career growth guidance; fosters a culture of high‑quality design, documentation, and collaborative delivery.

  • Partners with stakeholders to translate business needs into scalable data products.

  • Works closely with analytics, product, and platform teams to prioritize backlogs, define SLAs, and deliver user‑centric datasets and APIs.

  • Optimizes cost and performance across the stack.

  • Tunes Glue (workers, DPUs, job bookmarks), Lambda (memory/timeout/concurrency), Iceberg (file sizing, compaction cadence), and storage/query engines (e.g., S3/Athena/EMR).

  • Participates in special projects and performs other duties as assigned.

Qualifications

  • 8+ years of related data engineering experience with a strong record of leading delivery for production ETL platforms.

  • 5+ years building on AWS with deep expertise in Lambda, Glue, and Step Functions (state machine patterns, error handling, retries, and parallel branches).

  • Strong proficiency in Python and SQL

  • Hands‑on experience with batch and event‑driven pipelines (e.g., Kinesis/Kafka), S3‑based data lakes, and Glue/Athena/EMR integration.

  • Solid understanding of lakehouse architecture, columnar formats (Parquet), and performance tuning (file sizing, predicate pushdown, metadata operations).

  • Familiarity with data quality frameworks, lineage/metadata management, and governance (IAM, Lake Formation, encryption at rest/in transit).

  • Excellent communication skills; able to explain complex technical concepts to non‑technical audiences and influence cross‑functional decisions.

  • Undergraduate degree or equivalent combination of training and experience; graduate degree preferred.

Preferred Experience

  • Operating Iceberg with query engines (Athena/Trino/Presto/Spark) and catalog integrations (Glue/HMS).

  • Experience in financial services or similarly regulated industries.

Special Factors

Sponsorship

Vanguard is not offering visa sponsorship for this position.

About Vanguard

At Vanguard, we don't just have a mission—we're on a mission.

To work for the long-term financial wellbeing of our clients. To lead through product and services that transform our clients' lives. To learn and develop our skills as individuals and as a team. From Malvern to Melbourne, our mission drives us forward and inspires us to be our best.

How We Work

Vanguard has implemented a hybrid working model for the majority of our crew members, designed to capture the benefits of enhanced flexibility while enabling in-person learning, collaboration, and connection. We believe our mission-driven and highly collaborative culture is a critical enabler to support long-term client outcomes and enrich the employee experience.

Other facts

Tech stack
AWS,Lambda,Glue,Step Functions,Python,SQL,Apache Iceberg,Data Engineering,ETL,Data Quality,Monitoring,CI/CD,IaC,Data Governance,Batch Processing,Event-Driven Pipelines

About Vanguard

We are a community of 50 million who think—and feel—differently about investing. Together, we’re changing the way the world invests.

For over 50 years, Vanguard has helped people pursue their financial goals with a spotlight on long-term value and low costs. We’ve made it a focus to put investors first, so whether you’re saving for your first home or investing for a comfortable retirement, we’re here to help you succeed.

Community guidelines: vgi.vg/sgl1

Team size: 10,001+ employees
LinkedIn: Visit
Industry: Financial Services
Founding Year: 1975

What you'll do

  • The Tech Lead will lead the architecture and delivery of data pipelines, defining patterns for ingestion, transformation, and storage. They will also ensure performance, reliability, and observability while mentoring engineers and collaborating with stakeholders.

Ready to join Vanguard?

Take the next step in your career journey

Frequently Asked Questions

What does a Tech Lead – Data Engineer - Assets and Cash Flow do at Vanguard?

As a Tech Lead – Data Engineer - Assets and Cash Flow at Vanguard, you will: the Tech Lead will lead the architecture and delivery of data pipelines, defining patterns for ingestion, transformation, and storage. They will also ensure performance, reliability, and observability while mentoring engineers and collaborating with stakeholders..

Why join Vanguard as a Tech Lead – Data Engineer - Assets and Cash Flow?

Vanguard is a leading Financial Services company.

Is the Tech Lead – Data Engineer - Assets and Cash Flow position at Vanguard remote?

The Tech Lead – Data Engineer - Assets and Cash Flow position at Vanguard is based in Malvern, Pennsylvania, United States. Contact the company through Clera for specific work arrangement details.

How do I apply for the Tech Lead – Data Engineer - Assets and Cash Flow position at Vanguard?

You can apply for the Tech Lead – Data Engineer - Assets and Cash Flow position at Vanguard directly through Clera. Click the "Apply Now" button above to start your application. Clera's AI-powered platform will help match your profile with this opportunity and guide you through the application process. You can also learn more about Vanguard on their website.