Backbone logoBackbone Jobs

Job Description

Job Description

Job Summary:

DENSO North America Enterprise Solutions Group is transforming its enterprise analytics and business intelligence capabilities through a comprehensive Data Lakehouse platform built on Databricks. This strategic initiative involves migrating legacy data warehouses, reporting systems, and analytical workloads from on-premise platforms (SQL Server data warehouses, Oracle analytics, iSeries reporting) and current BI tools (DOMO, Qlik) into a unified, governed analytical platform. The organization manages complex analytical data flows, dimensional modeling, reporting datasets, and advanced analytics capabilities across multiple business domains.

The E3 Data Engineer role serves as a technical leader in building and maintaining enterprise-scale analytical data solutions. This position focuses on designing data warehouse architectures, implementing ETL/ELT processes, creating dimensional models, and enabling self-service analytics within the modern Data Lakehouse framework. The role bridges traditional data warehousing and business intelligence expertise with cutting-edge lakehouse technologies, working within established governance frameworks to deliver scalable, performant analytical solutions that support strategic business decision-making across the enterprise.

Data Lakehouse Role Responsibilities:

Based on project needs and organizational priorities, the E3 Data Engineer may be assigned to fulfill one or more of the following platform or domain roles:

Platform Role Capabilities:

Platform Engineer

Build and maintain core platform infrastructure and data pipelines

Support catalog automation and orchestration across domains

Monitor job performance and compute resource usage

Collaborate with Unity Catalog admins for infrastructure provisioning

Platform Data Scientist

Conduct advanced analysis and modeling across domains using certified data

Build and maintain ML models within the Databricks ML platform

Share insights and collaborate with domain teams and governance leads

Release Manager

Coordinate promotion of notebooks, jobs, and data objects between environments

Oversee change control and deployment workflows

Manage versioning and release calendar coordination

Validate content before production deployment

Domain Role Capabilities:

Domain Editor

Develop data transformations and ingestion logic for specific domains

Write and maintain bronze/silver dataset pipelines

Coordinate with domain stewards for data certification processes

Maintain documentation and versioning of domain datasets

Primary Accountabilities:

Lead data migration projects from legacy database systems (SQL Server, Oracle, DB2), DOMO, and other existing platforms to Databricks Data Lakehouse

Design and implement medallion architecture (Bronze/Silver/Gold) following established governance standards

Lead hybrid data architecture spanning on-premise and cloud systems across multiple database platforms

Architect scalable ingestion pipelines using Databricks workflows, Delta Live Tables, and other platform-native tools

Design enterprise-scale data transformation pipelines with emphasis on real-time and batch processing capabilities

Collaborate with Domain Stewards and Approvers to ensure data certification and compliance with Unity Catalog RBAC

Implement data quality frameworks and monitoring solutions across the lakehouse platform

Mentor junior engineers and provide technical guidance on modern data platform best practices

Evaluate and recommend new data sources for platform integration, including structured and unstructured data assessment

Drive automation initiatives to reduce manual processes and improve operational efficiency

Support and build replication and advanced data movement technologies between disparate systems

Organize and coordinate with team members to resolve complex technical issues and project deliverables

Utilize change management tools to document and gate all changes that could affect production systems

Building advanced tools to automate data collection and processing workflows

Utilize source control tools and best practices while developing enterprise-scale solutions

Drive continuous learning initiatives and knowledge transfer within the team and across domains

Find unique solutions to difficult problems involving complex data integration challenges

Model and document database schemas and processes for both legacy and modern architectures

Supply support and research for future technologies, processes, and platform capabilities

Support non-inhouse system management through collaboration with peers and external partners

Provide support and respond to production-down emergencies 24/7 with advanced troubleshooting capabilities

Work in a hybrid agile environment supporting both platform administration and domain-specific initiatives

Provide advanced database and distributed system performance tuning and optimization recommendations

Demonstrates accountability for technical decision making and project outcomes

Adapt to assigned platform or domain roles as determined through agile sprint planning and organizational needs

Maintain role separation and compliance with the Mutually Exclusive Roles Matrix to ensure proper governance

Participate in role-based workflows including data certification processes and governance activities

Collaborate across role boundaries while maintaining appropriate separation of duties and audit compliance

Support agile development processes for both platform administration and domain-specific engineering initiatives

Transition between roles as project phases and organizational priorities evolve, ensuring proper knowledge transfer

Document role-specific activities and maintain clear audit trails for governance and compliance purposes

Performs other duties as assigned.

Minimum Level of Education:

Bachelor’s degree in Computer Science, Information Technology, Engineering, or a related field. A combination of degree and experience can be considered.

Experience:

5+ years in software development, data engineering, database administration, IT, or a related field (including internships or coursework).

Minimum Level of Knowledge and Skills Required:

Advanced knowledge of database systems (SQL Server, Oracle, DB2, or other enterprise platforms) and data warehouse architectures

Strong SQL programming skills with experience across multiple database platforms and complex query optimization

Data integration and ETL development experience using various tools and methodologies

Advanced knowledge of database management systems throughout the entire development lifecycle

Programming experience in at least one modern language (Python, Scala, Java, or C#)

Understanding of data architecture patterns and willingness to learn modern lakehouse concepts

Ability to work well with a team of peers on complex, enterprise-scale projects

Provide excellent customer experience across internal and external stakeholders

Expert knowledge of core SQL language with experience across multiple database dialects and platforms

Knowledge of database sizing, monitoring, and performance optimization

Experience with database upgrades, migrations, and version management

Experience managing and monitoring multiple database servers and instances

Experience with High Availability, Disaster Recovery, and backup/recovery technologies

Experience with database backup, recovery, and data lifecycle management tools and patterns

Experience with DevOps processes and tools or willingness to learn CI/CD practices

Strong analytical and problem-solving skills with attention to detail in complex systems

Excellent verbal and written communication skills in English for cross-functional collaboration

Ability to work independently and as part of a team in a dynamic, multi-role environment

Adaptability and eagerness to learn new technologies and methodologies

Understanding of data governance principles or willingness to learn compliance frameworks

Leadership Responsibility:

Technical and role-based leadership responsibility including:

Leading complex migration projects and mentoring junior team members

Serving in assigned platform or domain roles with appropriate authority and accountability

May lead cross-functional project teams while maintaining governance compliance

Responsible for role-specific deliverables and stakeholder communication

Preferred Qualifications:

Enterprise database administration certification (SQL Server, Oracle, DB2, or equivalent)

Programming with robust depth of programming knowledge across multiple languages

Understanding of process mapping, business analysis, and design principles

Experience with workflow tools, databases, and APIs

Experience with Qlik and other BI/analytics platforms

Proficiency in SQL across multiple database platforms and query optimization

Cloud platform exposure (AWS/Azure) or willingness to learn

Agile/Scrum methodology experience or willingness to learn

Benefits Summary:

Health, Dental, Vision, Prescription Drug plans

Life and Accidental Death & Dismemberment Insurance

Flexible Spending Account

Employee Assistance Program

401K with 4% company match

Bonus Program

Wellness Program

Onsite Fitness Center (vary by location)

Tuition Reimbursement

Career Development and Ongoing Training

Paid holidays and vacation

Cafeteria and food markets (vary by location)

Volunteer opportunities

Employee recognition (employee and milestone events)

Annual Salary: $94,000-$117,000

Show more Show less

Explore more jobs

Browse more opportunities from trusted companies, filter by technology, location, and seniority, and find the next role that fits you best.

Browse all jobs

    Related Articles

    Insights related to the expertise required for this role.

    Frequently Asked Questions

    Common questions about Backbone Jobs and how we help you find your next role.