Match score not available

Data Engineer

Remote: 
Full Remote
Contract: 
Experience: 
Mid-level (2-5 years)
Work from: 

Offer summary

Qualifications:

Bachelor’s or Master’s degree in Computer Science, Information Systems, or a related field., 3+ years of hands-on experience in data engineering and building ETL/ELT pipelines., Proficient in Python or Java for data processing and familiar with SQL and NoSQL databases., Experience with cloud-based data solutions (AWS, Azure, or GCP) and knowledge of distributed data processing..

Key responsabilities:

  • Build and maintain scalable ETL/ELT pipelines for machine learning datasets.
  • Connect to various data sources and ensure seamless data ingestion.
  • Implement validation rules and monitoring to maintain high data integrity.
  • Collaborate with data scientists and MLOps engineers to optimize data workflows and improve model performance.

Foxbox Digital logo
Foxbox Digital Scaleup https://www.foxbox.com/
51 - 200 Employees
See all jobs

Job description

Foxbox Digital is an award-winning digital product agency, headquartered in Chicago. We partner with clients ranging from start-ups to mid-sized businesses and everyone in between to design, develop, and deliver successful digital experiences.

We're a remote-first team of associates located in the United States and LATAM regions. Our mission is rooted in continuously engaging and assembling tech-enthusiasts together to build our global team.

Summary

We are seeking a passionate and skilled Data Engineer to join our dynamic team at Foxbox Digital. In this role, you will be responsible for designing, building, and maintaining the data pipelines and integrations that serve as the backbone for our AI/ML applications. You’ll work closely with data scientists, MLOps engineers, and software teams to ensure clean, reliable, and secure data is available for model training, inference, and reporting.

Responsibilities:

  • Data Pipeline Development: Build and maintain scalable ETL/ELT pipelines, transforming raw data into ready-to-use datasets for machine learning.
  • Integration & APIs: Connect to various data sources (e.g., Quickbase, REST APIs, databases) and ensure seamless data ingestion.
  • Data Quality & Governance: Implement validation rules, data profiling, and monitoring to maintain high data integrity.
  • Collaboration: Work with data scientists and MLOps engineers to optimize data workflows, improve model performance, and address bottlenecks.
  • Performance & Scalability: Identify and resolve issues around data throughput and latency, ensuring pipelines can handle production-level loads.
  • Security & Compliance: Enforce data privacy best practices and role-based access control in accordance with project or regulatory requirements.
  • Monitoring & Troubleshooting: Set up alerts, logs, and metrics to proactively detect and resolve pipeline or data-related issues.

Requirements

Who You are:

  • You have a Bachelor’s or Master’s degree in Computer Science, Information Systems, or a related field.
  • You have 3+ years of hands-on experience in data engineering, building ETL/ELT pipelines, and managing data workflows in a production environment.
  • You are proficient in Python or Java for data processing.
  • You are familiar with SQL and NoSQL databases.
  • You have experience with cloud-based data solutions (AWS, Azure, or GCP).
  • You have knowledge of distributed data processing (e.g., Spark, Hadoop, Databricks).
  • You understanding of containerization (e.g., Docker) and orchestration (e.g., Kubernetes) is a plus.
  • You excel at working cross-functionally and translating data requirements into actionable solutions.
  • You have the ability to debug complex data issues quickly and innovate new approaches to data challenges.
  • Experience with Airflow, Kafka, or Azure Data Factory is preferred.
  • Familiarity with CI/CD practices and Git-based workflows is preferred.
  • Exposure to MLOps tools like MLflow or Kubeflow is a plus.

Benefits

Technologies we use:
  • Data Processing & Pipelines: Apache Spark, Airflow, Kafka, Azure Data Factory
  • Cloud Platforms: AWS, Azure, GCP for compute, storage, and managed data services
  • Databases: SQL (PostgreSQL, MySQL), NoSQL (MongoDB, DynamoDB)
  • Containerization & Orchestration: Docker, Kubernetes
  • Version Control & CI/CD: Git, GitHub Actions, or Azure DevOps
Why Foxbox Digital
  • We offer continuous training and growth opportunities
  • Remote-first environment with a culture of collaboration and innovation.
  • Opportunity to work on a project that directly impacts business success.
  • You are part of a multicultural and collaborative team that is constantly growing.
  • Don’t be afraid to break things; we encourage risk-takers.
Diversity and Inclusion

Foxbox Digital is an LGBT company certified by the Illinois and National LGBT Chambers of Commerce. We are committed to working with diverse and inclusive teams to continue building the digital revolution.

Foxbox is committed to the principle of equal employment opportunity for all and team members with a work environment free of discrimination and harassment. All employment decisions at Foxbox are based on business needs, job requirements and individual qualifications, without regard to race, color, religion or belief, family or parental status, or any other status protected by the laws or regulations in the locations where we operate. Foxbox will not tolerate discrimination or harassment based on any of these characteristics. Foxbox encourages applicants of all ages.

Required profile

Experience

Level of experience: Mid-level (2-5 years)
Spoken language(s):
English
Check out the description to know which languages are mandatory.

Other Skills

  • Collaboration
  • Problem Solving

Data Engineer Related jobs