Lead Data Engineer
Deloitte Technology Software Engineering, Development & Testing Posted: 18-Aug-2022
Atlanta, Georgia, United States
Austin, Texas, United States
Charlotte, North Carolina, United States
Cincinnati, Ohio, United States
Cleveland, Ohio, United States
Dallas, Texas, United States
Detroit, Michigan, United States
Hermitage, Tennessee, United States
Houston, Texas, United States
Jacksonville, Florida, United States
Kansas City, Missouri, United States
Orlando, Florida, United States
Raleigh, North Carolina, United States
San Antonio, Texas, United States
Tampa, Florida, United States
Toronto, Ontario, Canada
Tulsa, Oklahoma, United States
Work you'll do
As a Lead Data Engineer, you will be responsible for designing, building, and supporting data analytic products for the audit practice within Deloitte leveraging cloud, Spark and big data technologies. Your expertise will drive the development of end-to-end for specific components.
You will be working with closely with Senior Leadership, and a team of product owners, developers, and business resources to build deliverables that support audit quality and objectives, collaborating with others at your level on other teams. You’ll use your analytical, research and investigative skills to:
- Design and develop data models and related transformations, and to standardize content for engagements.
- Develop analytical use cases in the data platform.
- Review and document to maintain quality.
- Design BI solutions using best practices and mentor, coach team members.
- Liaise and communicate effectively with cross-border teams.
- Ensure the work performed is in accordance with quality standards of the organization.
- Present technical solutions to team members and leadership.
The team
Qualifications
Required
- At least 10 years software development experience
- At least 5 years leading at least one Scrum team of data engineers building data-intensive products with a modern tech stack
- Significant experience with big data ETL pipeline development with Spark, Hive, and related technologies
- Significant experience with a general purpose programming language such as Python, Scala, or Java
- Experience with Spark framework and related tools (PySpark, Scala, SparkR, Spark SQL, Spark UI)
- Experience with Hadoop ecosystem using HDFS, ADLS Gen2, or AWS S3
- Experience with data visualization development using Python, Tableau, or PowerBI
- Experience with Azure, AWS or GCP
- Solid understanding of performance tuning concepts for relational and distributed database systems
- Familiarity with distributed programming, big data concepts, and cloud computing
Education:
- Bachelor's degree in Computer Science/Engineering or Technology related field, or possess equivalent work experience
Preferred:
- Cloud certifications from Azure, AWS or GCP
- Big data, data engineering or data science certifications from recognized vendors such as Databricks & Cloudera