
Job Summary
Data Aces is looking for a Data Warehouse Engineer to work on projects to design, build, maintain data warehouses and data marts and related applications using state-of-the-art, cloud based, data lake and data warehouse platforms. The Candidate should be able to code in Scala, Python, Spark and possess good understanding of cloud ecosystems.
Job Description
Design and implementation of data warehouse schemas and analytics:
o Design data models that result in highly performant analytical queries.
o Design and build/enhance metrics required to satisfy business requirements.
1. Analyze requirements and add required features by applying business logic, statistical analysis, machine learning and LLMs.
2. Coding in Python, Spark/Scala required in addition to advanced SQL.
Build ETL models to extract, transform and load data:
o Build process to extract data from multiple data sources, perform logical transformations and load into data warehouse tables on Snowflake.
Combine multiple Input sources to map fields to target tables:
o Extracting the Data from multiple source tables and mapping it to destination.
o Automate this process of source data extraction into warehouse by using tools like Airflow, Java and Spark/Scala.
Project Design and documentation:
o Documentation of the Project’s processes using ER Diagrams, help in analyzing Code Flow.
o This includes aspects of requirements and process, data sources, Mapping of these sources and Extracting the final data.
Testing and Performance Improvement:
o Analyze and Test if the design, curation logic, analytics etc. on new and existing data is correct.
o Test for performance on all analysis that deals with the large data sets.
