top of page
Job Summary
Job Description
Design and implementation of data warehouse schemas and analytics:
o Design data models that result in highly performant analytical queries.
o Design and build/enhance metrics required to satisfy business requirements.
1. Analyze requirements and add required features by applying business logic, statistical analysis, machine learning and LLMs.
2. Coding in Python, Spark/Scala required in addition to advanced SQL.
Build ETL models to extract, transform and load data:
o Build process to extract data from multiple data sources, perform logical transformations and load into data warehouse tables on Snowflake.
Combine multiple Input sources to map fields to target tables:
o Extracting the Data from multiple source tables and mapping it to destination.
o Automate this process of source data extraction into warehouse by using tools like Airflow, Java and Spark/Scala.
Project Design and documentation:
o Documentation of the Project’s processes using ER Diagrams, help in analyzing Code Flow.
o This includes aspects of requirements and process, data sources, Mapping of these sources and Extracting the final data.
Testing and Performance Improvement:
o Analyze and Test if the design, curation logic, analytics etc. on new and existing data is correct.
o Test for performance on all analysis that deals with the large data sets.
Qualifications
Master of Science in Computer Science or in a related field.
Must have completed graduate-level course work closely related to each of the following areas: data model design and advanced analytics on large scale data using Spark/Scala and Machine Learning; Functional and Performance testing; data warehouses built on Snowflake; and Advanced SQL. Must be able to travel to unanticipated customer locations throughout the U.S., approximately once a month, for meetings as well as for solution deployments.
E-mail Resume to: info@data-aces.com or Apply here.
bottom of page
