Line of Service
Internal Firm ServicesIndustry/Sector
Not ApplicableSpecialism
IFS - Internal Firm Services - OtherManagement Level
Senior AssociateJob Description & Summary
A career in Products and Technology is an opportunity to bring PwC's strategy to life by driving products and technology into everything we deliver. Our clients expect us to bring the right people and the right technology to solve their biggest problems; Products and Technology is here to help PwC meet that challenge and accelerate the growth of our business. We have skilled technologists, data scientists, product managers and business strategists who are using technology to accelerate change.Data Engineer (Azure Data Lake, Spark & Databricks)
Required Knowledge and Skills:
6-9 years of experience designing, building, deploying, testing, maintaining, monitoring, and owning scalable, resilient, and distributed data pipelines.
High Proficiency in at least two of Scala, Python, Spark applied to large scale data sets
Expertise with big data technologies (Spark, Data Lake, Delta Lake, Hive)
Knowledge of batch and streaming data processing techniques
Understanding of the Data Lifecycle Management process to collect, access, use, store, transfer, delete data.
Expert level knowledge of using SQL to write complex, highly optimized queries across large volumes of data.
Hands-on object-oriented programming experience using Scala, Python, R, or Java
Proficient with Azure DevOps, managing backlogs and creating/maintaining pipelines
Experience working in an Agile environment and applying it in Azure DevOps
At Effectual, Data engineers are responsible for designing, building, and maintaining datasets that can be leveraged in data projects.
Obsession for service observability, instrumentation, monitoring, and alerting
Knowledge or experience in architectural best practices in building data lakes
Responsibilities
You will be responsible for designing and building optimized data pipelines using emerging technologies, in a cloud environment, for the purpose of driving analytic insights.
Create the conceptual, logical and physical data models.
Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of sources like API, Blob Storage, and no SQL Database
Design, develop, test, deploy, maintain, and improve data integration pipeline for data streaming.
Develop pipeline objects using Apache Spark / Pyspark / Python
Design and develop data pipeline architectures using Databricks, Spark and cloud Services.
Load and performance test data pipelines built using the above-mentioned technologies.
Good to have
Passionate about testing strategy, problem solving, learning new skills, sharing expertise and knowledge.
Always Be Learning
Product / Engineering Mindset
Education (if blank, degree and/or field of study not specified)
Degrees/Field of Study required:Degrees/Field of Study preferred:Certifications (if blank, certifications not specified)
Required Skills
Optional Skills
Desired Languages (If blank, desired languages not specified)
Travel Requirements
Not SpecifiedAvailable for Work Visa Sponsorship?
NoGovernment Clearance Required?
NoJob Posting End Date
At PwC, our purpose is to build trust in society and solve important problems. We’re a network of firms in 152 countries with over 327,000 people who are committed to delivering quality in assurance, advisory and tax services. Find out more and tell us what matters to you by visiting us at www.pwc.com. PwC refers to the PwC network and/or one or more of its member firms, each of which is a separate legal entity.
Content on this page has been prepared for general information only and is not intended to be relied upon as accounting, tax or professional advice. Please reach out to your advisors for specific advice.