You are viewing a preview of this job. Log in or register to view more details about this job.

Junior AWS Data Engineer

Junior AWS Data Engineer

Location: Reston, VA [ In person 5 days]

Job Summary:

We are seeking a motivated and technically strong Entry-Level AWS Data Engineer with strong hands-on experience in building and supporting cloud-based data pipelines. The ideal candidate should have solid experience with AWS data services—including S3, Lambda, Glue, Step Functions, EMR, RDS, Redshift, and Athena—along with strong skills in Python, PySpark, and SQL. In this role, you will design, develop, and maintain data ingestion, transformation, and orchestration workflows that support analytics, reporting, and operational use cases. You should be eager to learn, adaptable to evolving requirements, and capable of working both independently and collaboratively within cross-functional teams.

Tech Stack:

AWS – S3, SQS, SNS, Lambda, Glue, Step Functions, EMR, EC2, ECS Fargate, Oracle/PostgreSQL RDS, Redshift, Athena
Languages/Frameworks – Python, PySpark, SQL
Other – Data pipelines, ETL/ELT, Data orchestration, Distributed processing

Key Responsibilities:

•Develop and maintain data ingestion, transformation, and ETL/ELT pipelines using AWS Glue, Lambda, EMR (PySpark), and Step Functions.

•Build and optimize data processing jobs using Python, and SQL for various analytical and operational needs.

•Work with cross-functional teams to understand data requirements and translate them into scalable cloud data solutions.

•Implement data quality checks, logging, monitoring, and error handling to ensure reliability of data pipelines.

•Document pipeline designs, data flow diagrams, and operational procedures for team reference and knowledge sharing.

•Stay current with AWS data services and evolving data engineering tools.

Qualifications & Skills:

•Bachelor’s degree in Computer Science, Data Engineering, Information Systems, or a related field.

•Experience in data engineering with strong hands-on skills in Python and PySpark.

•Experience working with AWS data services such as Glue, Lambda, Step Functions, EMR, S3, RDS, Redshift, and Athena.

•Strong SQL proficiency and solid understanding of data modeling and ETL/ELT concepts.

•Ability to work independently and collaborate effectively in a fast-paced environment.

•Strong analytical, debugging, and problem-solving skills with high attention to detail.

•Good communication and teamwork skills.

•Willingness to learn new tools, cloud services, and data engineering patterns.