Bachelor's degree in a relevant technical discipline or equivalent experience required., Intermediate experience in data engineering with hands-on knowledge of SPARK, Scala/Java, and SQL., Proficiency in programming languages such as Python, Java, and/or Scala, along with expertise in SQL and NoSQL databases., Strong understanding of Agile methodologies, including Scrum and Kanban, for iterative development..
Key responsabilities:
Lead projects for the design, development, and maintenance of a data and analytics platform.
Design, develop, test, and deploy scalable data pipelines for efficient data transport into cloud environments.
Implement data governance processes and continuously monitor data integrity issues.
Provide guidance and mentorship to junior team members, fostering skill development and knowledge sharing.
Report This Job
Help us maintain the quality of our job listings. If you find any issues with this job post, please let us know.
Select the reason you're reporting this job:
At Cummins, we empower everyone to grow their careers through meaningful work, building inclusive and equitable teams, coaching, development and opportunities to make a difference. Across our entire organization, you'll find engineers, developers, and technicians who are innovating, designing, testing, and building. You'll also find accountants, marketers, as well as manufacturing, quality and supply chain specialists who are working with technology that's just as innovative and advanced.
From your first day at Cummins, we’re focused on understanding your talents, current skills and future goals – and creating a plan to get you there. Your journey begins with planning your development and connecting to diverse experiences designed to spur innovation. From our internships to our senior leadership roles, we attract, hire and reward the best and brightest from around the world and look to them for new ideas and fresh perspectives. Learn more about #LifeAtCummins at cummins.com/careers.
GPP Database Link (https://cummins365.sharepoint.com/sites/CS38534/)
Job Summary
Leads projects for the design, development, and maintenance of a data and analytics platform. Effectively and efficiently processes, stores, and makes data available to analysts and other consumers. Works with key business stakeholders, IT experts, and subject-matter experts to plan, design, and deliver optimal analytics and data science solutions. Works on one or many product teams at a time. Though the role category is listed as Remote, this specific position is designated as Hybrid.
Key Responsibilities
Business Alignment & Collaboration – Partner with the Product Owner to align data solutions with strategic goals and business requirements.
Data Pipeline Development & Management – Design, develop, test, and deploy scalable data pipelines for efficient data transport into Cummins Digital Core (Azure DataLake, Snowflake) from various sources (ERP, CRM, relational, event-based, unstructured).
Architecture & Standardization – Ensure compliance with AAI Digital Core and AAI Solutions Architecture standards for data pipeline design and implementation.
Automation & Optimization – Design and automate distributed data ingestion and transformation systems, integrating ETL/ELT tools and scripting languages to ensure scalability, efficiency, and quality.
Data Quality & Governance – Implement data governance processes, including metadata management, access control, and retention policies, while continuously monitoring and troubleshooting data integrity issues.
Innovation & Tool Evaluation – Conduct proof-of-concept (POC) initiatives, evaluate new data tools, and provide recommendations for improvements in data management and integration.
Documentation & Best Practices – Maintain standard operating procedures (SOPs) and data engineering documentation to support consistency and efficiency.
Agile Development & Automation – Use Agile methodologies (DevOps, Scrum, Kanban) to drive automation in data integration, preparation, and infrastructure management, reducing manual effort and errors.
Coaching & Team Development – Provide guidance and mentorship to junior team members, fostering skill development and knowledge sharing.
Responsibilities
Competencies:
System Requirements Engineering: Translates stakeholder needs into verifiable requirements, tracks status, and assesses impact changes.
Collaborates: Builds partnerships and works collaboratively with others to meet shared objectives.
Communicates Effectively: Delivers multi-mode communications tailored to different audiences.
Customer Focus: Builds strong customer relationships and provides customer-centric solutions.
Decision Quality: Makes good and timely decisions that drive the organization forward.
Data Extraction: Performs ETL activities from various sources using appropriate tools and technologies.
Programming: Develops, tests, and maintains code using industry standards, version control, and automation tools.
Quality Assurance Metrics: Measures and assesses solution effectiveness using IT Operating Model (ITOM) standards.
Solution Documentation: Documents knowledge gained and communicates solutions for improved productivity.
Solution Validation Testing: Validates configurations and solutions to meet customer requirements using SDLC best practices.
Data Quality: Identifies, corrects, and manages data flaws to support effective governance and decision-making.
Problem Solving: Uses systematic analysis to determine root causes and implement robust solutions.
Values Differences: Recognizes and leverages the value of diverse perspectives and cultures.
Qualifications
Preferred Experience:
Technical Expertise – Intermediate experience in data engineering with hands-on knowledge of SPARK, Scala/Java, MapReduce, Hive, HBase, Kafka, and SQL.
Big Data & Cloud Solutions – Proven ability to design and develop Big Data platforms, manage large datasets, and implement clustered compute solutions in cloud environments.
Data Processing & Movement – Experience developing applications requiring large-scale file movement and utilizing various data extraction tools in cloud-based environments.
Business & Industry Knowledge – Familiarity with analyzing complex business systems, industry requirements, and data regulations to ensure compliance and efficiency.
Analytical & IoT Solutions – Experience building analytical solutions with exposure to IoT technology and its integration into data engineering processes.
Agile Development – Strong understanding of Agile methodologies, including Scrum and Kanban, for iterative development and deployment.
Technology Trends – Awareness of emerging technologies and trends in data engineering, with a proactive approach to innovation and continuous learning.
Technical Skills
Programming Languages: Proficiency in Python, Java, and/or Scala.
Database Management: Expertise in SQL and NoSQL databases.
Big Data Technologies: Hands-on experience with Hadoop, Spark, Kafka, and similar frameworks.
Cloud Services: Experience with Azure, Databricks, and AWS platforms.
ETL Processes: Strong understanding of Extract, Transform, Load (ETL) processes.
Data Replication: Working knowledge of replication technologies like Qlik Replicate is a plus.
API Integration: Experience working with APIs to consume data from ERP and CRM systems.
Education, Licenses, And Certifications
Bachelor's degree in a relevant technical discipline, or equivalent experience required.
This position may require licensing for compliance with export controls or sanctions regulations.
Job Systems/Information Technology
Organization Cummins Inc.
Role Category Remote
Job Type Exempt - Experienced
ReqID 2410680
Relocation Package Yes
Required profile
Experience
Spoken language(s):
English
Check out the description to know which languages are mandatory.
Remote: California (USA), Colorado (USA), Florida (USA), Georgia (USA), Idaho (USA), Illinois (USA), Massachusetts (USA), New Jersey (USA), New York (USA), Oregon (USA), Pennsylvania (USA), Texas (USA), Vermont (USA), Virginia (USA), Washington (USA)