You are viewing a preview of this job. Log in or register to view more details about this job.

Data Quailty Intern

ABOUT THE ROLE

We’re on the lookout for an enthusiastic Data Intern ready to dive into the world of data quality assurance, ETL/ELT testing, and automated data validation! This role offers the chance to work with cutting-edge technologies like AWS, Snowflake, dbt, Apache Airflow, and Python. As part of our agile team, you’ll collaborate with cross-functional experts, gain hands-on experience in building and validating robust data pipelines, and play a key role in ensuring the integrity and accuracy of our data. If you’re eager to learn and ready to make an impact, this internship is for you!
This internship offers an hourly wage of $18.00.

 

HOW YOU’LL MAKE AN IMPACT

  • Actively participate in agile ceremonies (e.g., sprint planning, daily stand-ups, retrospectives) with product owners and engineers to understand data requirements and ensure proper test coverage.
  • Work closely with data engineers, analysts, and product managers to refine requirements, validate data processes, and ensure business needs are met.
  • Assist in validating data transformations and ensuring data accuracy across ETL pipelines by writing and executing SQL queries on large datasets.
  • Contribute to building automated test scripts using Python and PyTest for data validation, regression testing, and ETL process validation.
  • Support the testing of data pipelines developed using tools like dbt and Apache Airflow, ensuring smooth data ingestion and transformations.
  • Help monitor data pipeline performance, identify issues, and collaborate with the team to resolve them efficiently.
  • Create and maintain clear and detailed documentation for test cases, data pipelines, and test results, ensuring reproducibility and transparency.
  • Gain experience with cloud services like AWS (e.g., S3, EventBridge) and assist in testing data workflows in a cloud environment.

 

WHAT YOU’LL NEED

  • Must be available to work 24 hours per week Monday-Friday for the duration of the 16-week program.
  • Education: Currently pursuing a degree in Computer Science, Data Science, Engineering, or a related field.
  • SQL Proficiency: Ability to write basic SQL queries for data validation and analysis.
  • Programming Skills: Familiarity with Python for scripting and automation.
  • Analytical Skills: Strong problem-solving abilities and attention to detail for identifying and resolving data discrepancies.
  • Communication Skills: Excellent verbal and written communication skills for collaborating with team members and documenting workflows.
  • Eagerness to Learn: A passion for learning new tools, technologies, and methodologies in the data engineering and quality assurance space.

 

HOW YOU’LL LEVEL UP  

  • Cloud Platforms: Exposure to cloud environments like AWS or Google Cloud.
  • ETL/ELT Tools: Familiarity with tools like dbt, Apache Airflow, or Snowflake.
  • Automation Testing: Experience with test automation frameworks such as PyTest or similar.
  • Big Data Concepts: Understanding of data warehousing concepts like star schemas, snowflakes, and dimensional modeling.