Logo of Huzzle

Sr. Data Engineer I

image

GoodRx

Aug 30

  • Job
    Full-time
    Senior Level
  • San Francisco
  • Quick Apply

AI generated summary

  • You need a bachelor's in a related field, 6+ years in big data, expertise in pySpark, SQL, and AWS, experience with data models, Airflow, and tools like Kafka and Segment, plus relevant certifications.
  • You will collaborate with teams to define data needs, design and deploy data solutions, manage ETL processes, enhance data infrastructure, and ensure data integrity using cloud technologies.

Requirements

  • Bachelor’s degree in analytics, engineering, math, computer science, information technology or related discipline.
  • 6+ years professional experience in the big data space.
  • 6+ years' experience in engineering data pipelines using big data technologies (Spark, Flink etc...) on large scale data sets.
  • Expert knowledge in writing complex pySpark, SQL, dbt and ETL development with experience processing extremely large datasets.
  • Expert in applying SCD types on S3 data lake using Databricks/Delta Lake.
  • Experience with data model principles and data cataloging.
  • Experience with job scheduler Airflow or similar.
  • Demonstrated ability to analyze large data sets to identify gaps and inconsistencies, provide data insights, and advance effective product solutions.
  • Deep familiarity with AWS Services (S3, Event Bridge, Kinesis, Glue, EMR, Lambda).
  • Experience with data warehouse platforms such as Redshift, Databricks, Big Query, Snowflake.
  • Ability to quickly learn complex domains and new technologies.
  • Innately curious and organized with the drive to analyze data to identify deliverables, anomalies and gaps and propose solutions to address these findings.
  • Thrives in a fast-paced startup environment.
  • Experience with customer data platform tools such as Segment.
  • Experience with data streaming such as Kafka.
  • Experience using Jira, GitHub, Docker, CodeFresh, Terraform.
  • Experience contributing to full lifecycle deployments with a focus on testing and quality.
  • Experience with data quality processes, data quality checks, validations, data quality metrics definition and measurement.
  • AWS/Kafka/Databricks or similar certifications.

Responsibilities

  • Collaborate with product managers, data scientists, data analysts and engineers to define requirements and data specifications.
  • Plan, design, build, test and deploy data warehouse and data mart solutions.
  • Leading small to medium size projects, solve data problems through the documentation, design and creation of ETL jobs, data marts.
  • Works to increase the usage and value of the data warehouse and ensures the integrity of the data delivered.
  • Develops and implements standards, and promotes their use throughout the warehouse.
  • Develop, deploy and maintain data processing pipelines using cloud technology such as AWS, Kubernetes, Airflow, Redshift, Databricks, EMR.
  • Define and manage overall schedule and availability for a variety of data sets.
  • Work closely with other engineers to enhance infrastructure, improve reliability and efficiency.
  • Make smart engineering and product decisions based on data analysis and collaboration.
  • Act as an in house data expert and make recommendations regarding standards for code quality and timeliness.
  • Architect cloud-based data pipeline solutions to meet stakeholder needs.

FAQs

What is the primary responsibility of a Sr. Data Engineer I at GoodRx?

The primary responsibility is to collaborate with product managers, data scientists, data analysts, and engineers to build data pipelines and processes that integrate complex data stores to guide business decisions.

What qualifications are required for this role?

A Bachelor’s degree in analytics, engineering, math, computer science, information technology, or a related discipline, along with 6+ years of professional experience in the big data space and data pipeline engineering.

What technologies will I be working with as a Sr. Data Engineer I?

You will work with technologies such as AWS, Kubernetes, Airflow, Redshift, Databricks, EMR, and big data frameworks like Spark and Flink.

Is experience with cloud technology important for this position?

Yes, experience in cloud technologies, particularly AWS services, is crucial for developing, deploying, and maintaining data processing pipelines.

Are there opportunities for growth and development within the company?

Yes, GoodRx promotes growth and empowers employees to learn new technologies and enhance their skills within a nurturing environment.

What is the work environment like at GoodRx?

The work environment is fast-paced and friendly, allowing for collaboration across teams and projects.

What additional compensation does GoodRx offer?

GoodRx provides annual cash bonuses, equity grants, and generous benefits such as medical, dental, and vision insurance, 401(k) with company match, unlimited vacation, and various wellness programs.

How does GoodRx approach diversity and inclusion?

GoodRx is committed to creating a diverse and inclusive community by hiring and cultivating teams from various backgrounds and encourages individuals from underrepresented groups to apply.

What key skills are essential for this role?

Essential skills include expertise in pySpark, SQL, ETL development, data modeling principles, familiarity with job scheduling tools like Airflow, and strong analytical abilities to work with large datasets.

How does GoodRx accommodate candidates with disabilities during the recruitment process?

GoodRx is committed to providing reasonable accommodations for candidates with disabilities. Candidates can reach out for assistance at accommodations@goodrx.com.

Affordable healthcare for everyone.

Science & Healthcare
Industry
501-1000
Employees
2011
Founded Year

Mission & Purpose

GoodRx believes everyone deserves affordable and convenient healthcare. Its mission is to create better ways for people to find the right care at the best price, providing all Americans—regardless of income or insurance status—with the tools and knowledge to stay healthy. Since 2011, GoodRx has helped Americans save over $35 billion on prescriptions through its website, app, and savings cards. As a customer-first company, GoodRx focuses on solutions that address healthcare affordability and accessibility, striving to narrow the gap for the 46 million U.S. adults who struggle to afford needed care.