COMPANY BACKGROUNDLOVE. It's what makes Subaru, Subaru®. As a leading auto brand in the US, we strive to be More Than a Car Company®. Subaru believes in being a positive force in the communities in which we live and work, not just with donations but with actions that set an example for others to follow. That's what we call our Subaru Love Promise®.
Subaru is a globally renowned automobile manufacturer known for its commitment to innovation, safety, and sustainability. With a rich history dating back to 1953, Subaru has consistently pushed the boundaries of automotive engineering to deliver vehicles that offer not only exceptional performance but also a unique blend of utility and adventure.
Subaru's company culture is built on collaboration, diversity, and a shared passion for our product. We foster an inclusive environment that encourages employees to bring their unique perspectives and talents to the table. Our team members are driven by a common goal: to create exceptional vehicles that inspire and delight our customers.
ROLE SUMMARY The Data Analytic Engineer will be responsible for developing and optimizing cloud-based Business Intelligence solutions. Advances data analytics capabilities and drives innovative solutions. Possesses deep technical expertise in data engineering and plays instrumental role in managing data integrations from on-premises Oracle systems, Cloud CRM (Dynamics), and telematics. Collaborates closely with Data Science and Enterprise Data Warehouse teams and business stakeholders.
Primary Responsibilities: Data Ingestion and Storage: - Designs, develops, and maintains scalable, efficient data pipelines using Data Factory, and Databricks, leveraging Py Spark for complex data transformations and large-scale processing.
- Builds and manages extract, transform, and load (ETL)/extract, load, transform (ELT) processes to seamlessly extract, transform, and load data from on-premises Oracle systems, customer relationship management (CRM) technology, and connected vehicles into data storage solutions, such as Data Lake Storage and SQL Database.
Data Engineering: - Creates high-code data engineering solutions using Databricks to clean, transform, and prepare data for in-depth analysis.
- Develops and manages data models, schemas, and data warehouses, utilizing Lakehouse Architecture to enhance advanced analytics and business intelligence.
- Leverages Unity Catalog to ensure unified data governance and management across the enterprise's data assets.
- Optimizes data storage, retrieval strategies, and query performance to drive scalability and efficiency in all data operations.
Data Integration: - Integrates and harmonizes data from diverse sources including on-premises databases, cloud services, application programming interfaces (APIs), and connected vehicle telematics.
- Ensures consistent data quality, accuracy, and reliability across all integrated data sources.
GitHub Development: - Utilizes GitHub for version control and collaborative development, implementing best practices for code management, testing, and deployment.
- Develops workflows for continuous integration (CI) and continuous deployment (CD), ensuring efficient delivery and maintenance of data solutions.
Additional Responsibilities: Collaboration and Leadership: - Works closely with Data Science, Enterprise Data Warehouse, and Data Visualization teams, as well as business stakeholders, to understand data requirements and deliver innovative solutions.
- Collaborates with cross-functional teams to troubleshoot and resolve data infrastructure issues, identifying and addressing performance bottlenecks.
- Provides technical leadership, mentorship, and guidance to junior data engineers, promoting a culture of continuous improvement and innovation.
Required Skills and Abilities - Technical Expertise: Extensive experience with Data Factory, Databricks, and Synapse, as well as proficiency in Python and PySpark.
- Data Integration: Experience integrating data from on-premises Oracle systems and connected vehicle data into cloud-based solutions.
- Lakehouse Architecture & Governance: Deep knowledge of Lakehouse Architecture and Unity Catalog for enterprise data governance.
- Version Control & Collaboration: Demonstrated proficiency in GitHub for development, collaboration, and deployment in large-scale environments.
- Infrastructure as Code (IaC): Experience with Infrastructure as Code tools such as Resource Manager (ARM) templates or terraform.
- Problem-Solving & Troubleshooting: Strong analytical skills with the ability to diagnose and resolve complex data infrastructure challenges.
- Collaboration: Proven ability to work effectively with Data Science teams, business stakeholders, and cross-functional teams to drive data-driven insights.
- Communication: Excellent verbal and written communication skills with the ability to translate technical concepts to non-technical stakeholders.
Education/Experience Requirements: BA/BS with 6 to 8 years of relevant experience.
Work Environment - Hybrid Role: Remote work 2 days per week (After 90 Days Onboarding)
- Travel Required: 25%
Compensation: The recruiting base salary range for this full-time position is $99, 700.00 - $140, 000.00/year. Within the range, individual pay is determined by factors, including job-related skills, experience, and relevant education or training. (Internal Job Grade: P3_T) In addition to competitive salary, Subaru offers an amazing benefits package that includes:
- Medical, Dental, Vision Plans
- Pension, Profit Sharing, and 401K Match Offerings
- 15 Vacation days, 9 Company Holidays, 5 Floating Holidays, and 5 Sick days.
- Tuition Reimbursement Program
- Vehicle Discount Programs
- See our Careers landing page for additional information about our compensation and benefit programs.
Please see the job description for required or recommended skills.
Please see the job description for benefits.