Bachelor's degree in a relevant technical discipline or equivalent experience required., Proficiency in SQL and experience in developing analytical solutions., Hands-on experience with Big Data technologies such as Hadoop and Spark., Strong understanding of ETL processes and cloud services like Azure and AWS..
Key responsabilities:
Collaborate with the Product Owner to align data solutions with business objectives.
Design and implement efficient data pipelines for data ingestion and transformation.
Develop and enforce data governance policies while monitoring data quality issues.
Utilize agile methodologies to drive improvements in data-driven applications.
Report This Job
Help us maintain the quality of our job listings. If you find any issues with this job post, please let us know.
Select the reason you're reporting this job:
At Cummins, we empower everyone to grow their careers through meaningful work, building inclusive and equitable teams, coaching, development and opportunities to make a difference. Across our entire organization, you'll find engineers, developers, and technicians who are innovating, designing, testing, and building. You'll also find accountants, marketers, as well as manufacturing, quality and supply chain specialists who are working with technology that's just as innovative and advanced.
From your first day at Cummins, we’re focused on understanding your talents, current skills and future goals – and creating a plan to get you there. Your journey begins with planning your development and connecting to diverse experiences designed to spur innovation. From our internships to our senior leadership roles, we attract, hire and reward the best and brightest from around the world and look to them for new ideas and fresh perspectives. Learn more about #LifeAtCummins at cummins.com/careers.
Product & Business Alignment – Collaborate with the Product Owner to align data solutions with business objectives and product vision.
Data Pipeline Development – Design, develop, and implement efficient data pipelines for ingesting, transforming, and transporting data into Cummins Digital Core (Azure DataLake, Snowflake) from various sources, including transactional systems (ERP, CRM).
Architecture & Standards Compliance – Ensure alignment with AAI Digital Core and AAI Solutions Architecture standards for data pipeline design, storage architectures, and governance processes.
Automation & Optimization – Implement and automate distributed data systems, ensuring reliability, scalability, and efficiency through monitoring, alerting, and performance tuning.
Data Quality & Governance – Develop and enforce data governance policies, including metadata management, access control, and retention policies, while actively monitoring and troubleshooting data quality issues.
Modeling & Storage – Design and implement conceptual, logical, and physical data models, optimizing storage architectures using distributed and cloud-based platforms (e.g., Hadoop, HBase, Cassandra, MongoDB, Accumulo, DynamoDB).
Documentation & Best Practices – Create and maintain data engineering documentation, including standard operating procedures (SOPs) and best practices, with guidance from senior engineers.
Tool Evaluation & Innovation – Support proof-of-concept (POC) initiatives and evaluate emerging data tools and technologies to enhance efficiency and effectiveness.
Testing & Troubleshooting – Participate in the testing, troubleshooting, and continuous improvement of data pipelines to ensure data integrity and usability.
Agile & DevOps Practices – Utilize agile development methodologies, including DevOps, Scrum, and Kanban, to drive iterative improvements in data-driven applications.
Responsibilities
Qualifications:
Bachelor's degree in a relevant technical discipline, or equivalent experience required.
This position may require licensing for compliance with export controls or sanctions regulations.
Competencies
System Requirements Engineering: Translate stakeholder needs into verifiable requirements, establish acceptance criteria, track requirement status, and assess impact changes.
Collaborates: Build partnerships and work collaboratively with others to meet shared objectives.
Communicates Effectively: Deliver multi-mode communications tailored to different audiences.
Customer Focus: Build strong customer relationships and provide customer-centric solutions.
Decision Quality: Make good and timely decisions that drive the organization forward.
Data Extraction: Perform ETL activities from various sources using appropriate tools and technologies.
Programming: Develop, test, and maintain code using industry standards, version control, and automation tools.
Quality Assurance Metrics: Measure and assess solution effectiveness using IT Operating Model (ITOM) standards.
Solution Documentation: Document knowledge gained and communicate solutions for improved productivity.
Solution Validation Testing: Validate configurations and solutions to meet customer requirements using SDLC best practices.
Data Quality: Identify, correct, and manage data flaws to support effective governance and decision-making.
Problem Solving: Use systematic analysis to determine root causes and implement robust solutions.
Values Differences: Recognize and leverage the value of diverse perspectives and cultures.
Qualifications
Preferred Experience:
Hands-on experience gained through internships, co-ops, student employment, or team-based extracurricular projects.
Proficiency in SQL query language and experience in developing analytical solutions.
Exposure to open-source Big Data technologies such as Spark, Scala/Java, MapReduce, Hive, HBase, and Kafka.
Familiarity with cloud-based, clustered computing environments and large-scale data movement applications.
Understanding of Agile software development methodologies.
Exposure to IoT technology and data-driven solutions.
Technical Skills
Programming Languages: Proficiency in Python, Java, and/or Scala.
Database Management: Expertise in SQL and NoSQL databases.
Big Data Technologies: Hands-on experience with Hadoop, Spark, Kafka, and similar frameworks.
Cloud Services: Experience with Azure, Databricks, and AWS platforms.
ETL Processes: Strong understanding of Extract, Transform, Load (ETL) processes.
Data Replication: Working knowledge of replication technologies like Qlik Replicate is a plus.
API Integration: Experience working with APIs to consume data from ERP and CRM systems.
Job Systems/Information Technology
Organization Cummins Inc.
Role Category Remote
Job Type Exempt - Experienced
ReqID 2411593
Relocation Package No
Required profile
Experience
Spoken language(s):
English
Check out the description to know which languages are mandatory.