Data Engineer
Sigma Software
Job Summary
This role requires a Data Engineer to lead a small team and modernize a simple data platform into a cloud-based self-service analytics project. The goal is to create an efficient system for collecting, processing, analyzing, and visualizing large amounts of data from various sources. Responsibilities include laying down the foundation and modeling the data, developing a platform for building and maintaining data pipelines, creating a Data Warehouse, analyzing and preparing domain descriptions in collaboration with Business analytics, designing Data Marts, and designing a permission model for flexible control over data access.
Must Have
- Conceptual knowledge of data analytics fundamentals
- Strong SQL knowledge and experience with RDBMS
- Experience in database development and data modeling
- Experience with Databricks/Spark
- Experience with Python
- Experience with Azure
- Working knowledge of serialization formats
- Experience debugging and optimizing Spark jobs
- Strong written and verbal communication skills
- BS in Computer Science or related field
Good to Have
- Experience with a Business Intelligence tool
Job Description
Company Description
Seeking a Data Engineer to lead a small team and modernize a simple data platform into a cloud-based self-service analytics project.
Sigma Software provides top-quality software development, graphic design, testing, and support services. Systems developed to deliver customer benefits in automotive, telecommunications, aviation, advertising, gaming, banking, real estate, and healthcare industries
It is an Internal Sigma Software project that combines 2 goals: automating reporting and exposing a self-service data analytics system for the company. You will be responsible for designing and modeling data into DWH and Data Marts using Spark and MS SQL, exposing this data to, and collaborating with, BI engineers making reports in Apache Superset as well as other roles consuming such data, moving on-premises solutions to the cloud
Job Description
The goal is to create an efficient system for collecting, processing, analyzing, and visualizing large amounts of data from various sources
- Lay down the foundation and model the data
- Develop a platform for building and maintaining data pipelines that collect data from different sources
- Create Data Warehouse
- Analyze and prepare domain descriptions in collaboration with Business analytics
- Create Data Marts
- Design a permission model with flexible control over vertical and horizontal access to data
Qualifications
- Conceptual knowledge of data analytics fundamentals, e.g., dimensional modeling, ETL/ELT, reporting tools, data governance, data warehousing, structured and unstructured data
- Strong SQL knowledge and experience with RDBMS, confident knowledge of database fundamentals
- Experience in database development and data modeling, ideally with Databricks/Spark
- Experience with Python
- Experience with Azure
- Working knowledge of serialization formats and their trade-offs (columnar vs. row-based)
- Experience debugging and optimizing Spark jobs
- Strong written and verbal communication skills
- At least an Upper-Intermediate level of English
- BS in Computer Science or a related field
WOULD BE A PLUS
- Experience with a Business Intelligence tool