Match score not available

Data Engineer - AWS

Remote: 
Full Remote
Contract: 
Experience: 
Mid-level (2-5 years)
Work from: 
Connecticut (USA), United States

Offer summary

Qualifications:

8+ years of experience in data processing pipelines, Strong proficiency in AWS services, Experience with Databricks and Pyspark, Solid understanding of database design principles, Familiarity with version control systems and CI/CD.

Key responsabilities:

  • Design, develop, and deploy data pipelines on AWS
  • Implement data processing workflows using Databricks and SQL
  • Build orchestration workflows with Apache Airflow
  • Collaborate with teams to understand data needs
  • Optimize data pipelines for performance and cost-effectiveness
Tiger Analytics logo
Tiger Analytics XLarge https://www.tigeranalytics.com/
1001 - 5000 Employees
See more Tiger Analytics offers

Job description

Tiger Analytics is a fast-growing advanced analytics consulting firm. Our consultants bring deep expertise in Data Engineering, Data Science, Machine Learning and AI. We are the trusted analytics partner for multiple Fortune 500 companies, enabling them to generate business value from data. Our business value and leadership has been recognized by various market research firms, including Forrester and Gartner. We are looking for top-notch talent as we continue to build the best global analytics consulting team in the world.

As an AWS Data Engineer, you will be responsible for designing, building, and maintaining scalable data pipelines on AWS cloud infrastructure. You will work closely with cross-functional teams to support data analytics, machine learning, and business intelligence initiatives. The ideal candidate will have strong experience with AWS services, Databricks, and Snowflake.

Key Responsibilities:

  • Design, develop, and deploy end-to-end data pipelines on AWS cloud infrastructure using services such as Amazon S3, AWS Glue, AWS Lambda, Amazon Redshift, etc.
  • Implement data processing and transformation workflows using Databricks, Apache Spark, and SQL to support analytics and reporting requirements.
  • Build and maintain orchestration workflows using Apache Airflow to automate data pipeline execution, scheduling, and monitoring.
  • Collaborate with data scientists, analysts, and business stakeholders to understand data requirements and deliver scalable data solutions.
  • Optimize data pipelines for performance, reliability, and cost-effectiveness, leveraging AWS best practices and cloud-native technologies.

Requirements

  • 8+ years of experience building and deploying large-scale data processing pipelines in a production environment.
  • Hands-on experience in designing and building data pipelines
  • Strong proficiency in AWS services such as Amazon S3, AWS Glue, AWS Lambda, Amazon Redshift, etc.
  • Strong experience with Databricks, Pyspark for data processing and analytics.
  • Solid understanding of data modeling, database design principles, and SQL and Spark SQL.
  • Experience with version control systems (e.g., Git) and CI/CD pipelines.
  • Excellent communication skills and the ability to collaborate effectively with cross-functional teams.
  • Strong problem-solving skills and attention to detail.

Benefits

This position offers an excellent opportunity for significant career development in a fast-growing and challenging entrepreneurial environment with a high degree of individual responsibility.

Required profile

Experience

Level of experience: Mid-level (2-5 years)
Spoken language(s):
English
Check out the description to know which languages are mandatory.

Other Skills

  • Collaboration
  • Problem Solving
  • Verbal Communication Skills
  • Detail Oriented

Data Engineer Related jobs