
Job Duties:
Design and implementation of data warehouse schemas and analytics:
o Design data models that result in highly performant analytical queries.
o Design and build/enhance metrics required to satisfy business requirements.
1. Analyze requirements and add required features by applying business logic, statistical analysis, machine learning and LLMs.
2. Coding in Python, Spark/Scala required in addition to advanced SQL.
Build ETL models to extract, transform and load data:
o Build process to extract data from multiple data sources, perform logical transformations and load into data warehouse tables on Snowflake.
Combine multiple Input sources to map fields to target tables:
o Extracting the Data from multiple source tables and mapping it to destination.
o Automate this process of source data extraction into warehouse by using tools like Airflow, Java and Spark/Scala.
Project Design and documentation:
o Documentation of the Project’s processes using ER Diagrams, help in analyzing Code Flow.
o This includes aspects of requirements and process, data sources, Mapping of these sources and Extracting the final data.
Testing and Performance Improvement:
o Analyze and Test if the design, curation logic, analytics etc. on new and existing data is correct.
o Test for performance on all analysis that deals with the large data sets.
Minimum Requirements:
Master of Science in Computer Science or in a related field. Must have completed graduate-level course work closely related to each of the following areas: data model design and advanced analytics on large scale data using Spark/Scala and Machine Learning; Functional and Performance testing; data warehouses built on Snowflake; and Advanced SQL. Must be able to travel to unanticipated customer locations throughout the U.S., approximately once a month, for meetings as well as for solution deployments.
Mail Resume to: Attn: Shanti Subramanyam, Data Aces, 77 Sugar Creek Center Blvd, Ste 600, Sugar Land, TX 77478
