Data Engineer
Position Overview
We are seeking a skilled Data Engineer with 3 years of hands-on experience to join our dynamic data team. In this role, you will design, build, and maintain scalable data pipelines to support business analytics, machine learning initiatives, and real-time decision-making. The ideal candidate thrives in a fast-paced environment, collaborates effectively with data scientists and analysts, and is passionate about transforming raw data into actionable insights.
Location: Orlando, FL
Key Responsibilities
- Develop and optimize ETL/ELT pipelines using tools like Apache Airflow, Spark, or dbt to ingest, process, and transform large-scale data from diverse sources.
- Design and implement robust data architectures, including data lakes (e.g., AWS S3, Azure Data Lake) and warehouses (e.g., Snowflake, BigQuery).
- Ensure data quality, integrity, and security through automated testing, monitoring, and governance practices.
- Collaborate with cross-functional teams to define data requirements, resolve technical issues, and deliver reliable data solutions.
- Build and maintain infrastructure for real-time data streaming using Kafka or similar technologies.
- Perform data modeling, schema design, and optimization for performance and scalability.
- Document processes, contribute to code reviews, and mentor junior engineers on best practices.
- Stay updated on emerging data technologies and recommend improvements to our data ecosystem.
Required Qualifications
- Bachelor's degree in Computer Science, Information Technology, or a related field (or equivalent experience).
- 3+ years of professional experience as a Data Engineer or in a similar role.
- Proficiency in Python (or Java/Scala) for data processing and scripting.
- Strong SQL skills with experience in complex queries, joins, and optimization.
- Hands-on experience with cloud platforms (AWS, GCP, or Azure) and related services (e.g., Glue, Lambda, Databricks).
- Familiarity with version control systems like Git and CI/CD pipelines.
- Excellent problem-solving skills and ability to work independently or in agile teams.
Preferred Skills
- Experience with big data frameworks such as Apache Spark, Hadoop, or Flink.
- Knowledge of containerization and orchestration (Docker, Kubernetes).
- Exposure to data orchestration tools like Apache Airflow or Prefect.
- Understanding of data privacy regulations (GDPR, CCPA) and security best practices.
- Prior work in machine learning pipelines or feature engineering is a plus.
What We Offer
- Competitive salary and comprehensive benefits package, including health insurance, 401(k) matching, and unlimited PTO.
- Opportunities for professional growth through conferences, certifications, and internal training.
- A collaborative, inclusive culture with flexible work arrangements.
- Cutting-edge tools and projects that impact millions of users.