Principal Data Engineer (L5)
See yourself at Twilio
Join the Segment team as Twilio’s next Principal Data Engineer.
About the job
As a Principal Data Engineer, you will lead the technical direction of our core data processing infrastructure. You’ll be responsible for building and scaling large-scale distributed systems that support Segment’s customer data platform (CDP). This role demands deep expertise in data engineering, a strong product sense, and the ability to make high-leverage architectural decisions.
Our systems handle hundreds of thousands of events per second across real-time and batch processing environments. You'll collaborate closely with cross-functional teams to design and implement scalable data pipelines, high-throughput compute engines, and durable storage solutions that power advanced capabilities like audience segmentation, real-time personalization, and predictive insights
.
Responsibilities
In this role, you’ll:
Architect and implement large-scale data processing systems that power core Segment capabilities.
Design and optimize distributed compute frameworks leveraging Apache Spark, Scala, and cloud-native services.
Drive scalability and performance improvements across our data pipelines, enabling real-time and batch data processing.
Partner with product, infrastructure, and platform teams to build reusable systems that accelerate development across Segment.
Mentor and guide a team of data engineers, setting technical standards and fostering a culture of operational excellence.
Lead end-to-end delivery of major platform features and improvements with measurable business impact.
Ensure systems are designed for observability, maintainability, security, and compliance.
Qualifications
Twilio values diverse experiences from all kinds of industries, and we encourage everyone who meets the required qualifications to apply. If your career is just starting or hasn't followed a traditional path, don't let that stop you from considering Twilio. We are always looking for people who will bring something new to the table!
[CANDIDATES MUST MEET ALL REQUIRED QUALIFICATIONS IN THE JOB DESCRIPTION. IF ANY OF THESE QUALIFICATIONS CHANGE DURING THE HIRING PROCESS, THE JD MUST BE REVISED TO REFLECT THE NEW REQUIREMENTS]
Required:
8+ years of software engineering experience, with at least 5+ years focused on large-scale data engineering.
Proficiency in Scala or Java, with hands-on experience building systems using Apache Spark or similar distributed compute engines.
Deep understanding of distributed systems, stream processing, and data pipeline design patterns.
Experience processing large volumes of data using cloud-native and open-source big data tools (e.g., Spark, Kafka, Parquet, Delta Lake).
Track record of leading technical initiatives and collaborating across engineering and product teams.
Strong problem-solving skills and a bias toward ownership and execution.
Experience with AWS and infrastructure-as-code tools.
Excellent written and verbal communication skills.
Desired:
Experience with other data technologies such as Apache Flink, Trino, Snowflake, or Databricks.
Familiarity with data governance, quality frameworks, and GDPR/CCPA compliance.
Exposure to real-time personalization or ML-driven analytics use cases.
Experience working in globally distributed teams
Location
This role will be remote and based in the U.S.
Travel
We prioritize connection and opportunities to build relationships with our customers and each other. For this role, you may be required to travel occasionally to participate in project or team in-person meetings.
What We Offer
Working at Twilio offers many benefits, including competitive pay, generous time off, ample parental and wellness leave, healthcare, a retirement savings program, and much more. Offerings vary by location.
Compensation
*Please note this role is open to candidates outside of California, Colorado, Hawaii, Illinois, Maryland, Massachusetts, Minnesota, New Jersey, New York, Vermont, Washington D.C., and Washington State. The information below is provided for candidates hired in those locations only.
The estimated pay ranges for this role are as follows:
Based in Colorado, Hawaii, Illinois, Maryland, Massachusetts, Minnesota, Vermont or Washington D.C. : $167,800.00 - $209,700.00.
Based in New York, New Jersey, Washington State, or California (outside of the San Francisco Bay area): $177,700.00 - $222,100.00).
Based in the San Francisco Bay area, California: $197,400.00 - $246,700.00.
This role may be eligible to participate in Twilio’s equity plan and corporate bonus plan. All roles are generally eligible for the following benefits: health care insurance, 401(k) retirement account, paid sick time, paid personal time off, paid parental leave.
The successful candidate’s starting salary will be determined based on permissible, non-discriminatory factors such as skills, experience, and geographic location.
Applications for this role are intended to be accepted until June 1st, 2025, but may change based on business needs.
About the job
Apply for this position
Principal Data Engineer (L5)
See yourself at Twilio
Join the Segment team as Twilio’s next Principal Data Engineer.
About the job
As a Principal Data Engineer, you will lead the technical direction of our core data processing infrastructure. You’ll be responsible for building and scaling large-scale distributed systems that support Segment’s customer data platform (CDP). This role demands deep expertise in data engineering, a strong product sense, and the ability to make high-leverage architectural decisions.
Our systems handle hundreds of thousands of events per second across real-time and batch processing environments. You'll collaborate closely with cross-functional teams to design and implement scalable data pipelines, high-throughput compute engines, and durable storage solutions that power advanced capabilities like audience segmentation, real-time personalization, and predictive insights
.
Responsibilities
In this role, you’ll:
Architect and implement large-scale data processing systems that power core Segment capabilities.
Design and optimize distributed compute frameworks leveraging Apache Spark, Scala, and cloud-native services.
Drive scalability and performance improvements across our data pipelines, enabling real-time and batch data processing.
Partner with product, infrastructure, and platform teams to build reusable systems that accelerate development across Segment.
Mentor and guide a team of data engineers, setting technical standards and fostering a culture of operational excellence.
Lead end-to-end delivery of major platform features and improvements with measurable business impact.
Ensure systems are designed for observability, maintainability, security, and compliance.
Qualifications
Twilio values diverse experiences from all kinds of industries, and we encourage everyone who meets the required qualifications to apply. If your career is just starting or hasn't followed a traditional path, don't let that stop you from considering Twilio. We are always looking for people who will bring something new to the table!
[CANDIDATES MUST MEET ALL REQUIRED QUALIFICATIONS IN THE JOB DESCRIPTION. IF ANY OF THESE QUALIFICATIONS CHANGE DURING THE HIRING PROCESS, THE JD MUST BE REVISED TO REFLECT THE NEW REQUIREMENTS]
Required:
8+ years of software engineering experience, with at least 5+ years focused on large-scale data engineering.
Proficiency in Scala or Java, with hands-on experience building systems using Apache Spark or similar distributed compute engines.
Deep understanding of distributed systems, stream processing, and data pipeline design patterns.
Experience processing large volumes of data using cloud-native and open-source big data tools (e.g., Spark, Kafka, Parquet, Delta Lake).
Track record of leading technical initiatives and collaborating across engineering and product teams.
Strong problem-solving skills and a bias toward ownership and execution.
Experience with AWS and infrastructure-as-code tools.
Excellent written and verbal communication skills.
Desired:
Experience with other data technologies such as Apache Flink, Trino, Snowflake, or Databricks.
Familiarity with data governance, quality frameworks, and GDPR/CCPA compliance.
Exposure to real-time personalization or ML-driven analytics use cases.
Experience working in globally distributed teams
Location
This role will be remote and based in the U.S.
Travel
We prioritize connection and opportunities to build relationships with our customers and each other. For this role, you may be required to travel occasionally to participate in project or team in-person meetings.
What We Offer
Working at Twilio offers many benefits, including competitive pay, generous time off, ample parental and wellness leave, healthcare, a retirement savings program, and much more. Offerings vary by location.
Compensation
*Please note this role is open to candidates outside of California, Colorado, Hawaii, Illinois, Maryland, Massachusetts, Minnesota, New Jersey, New York, Vermont, Washington D.C., and Washington State. The information below is provided for candidates hired in those locations only.
The estimated pay ranges for this role are as follows:
Based in Colorado, Hawaii, Illinois, Maryland, Massachusetts, Minnesota, Vermont or Washington D.C. : $167,800.00 - $209,700.00.
Based in New York, New Jersey, Washington State, or California (outside of the San Francisco Bay area): $177,700.00 - $222,100.00).
Based in the San Francisco Bay area, California: $197,400.00 - $246,700.00.
This role may be eligible to participate in Twilio’s equity plan and corporate bonus plan. All roles are generally eligible for the following benefits: health care insurance, 401(k) retirement account, paid sick time, paid personal time off, paid parental leave.
The successful candidate’s starting salary will be determined based on permissible, non-discriminatory factors such as skills, experience, and geographic location.
Applications for this role are intended to be accepted until June 1st, 2025, but may change based on business needs.