Location: Remote
Job Code: MACIN20076
Role: Data Engineer
Type of Commute: Remote
Skill Set: Python, Snowflake, Airflow, SQL, AWS
Desired Industry Experience: 5+
No. of open position: 2
Role Description:
Desired Skills:
Key Requirements
- Bachelors degree or higher in the technology field.
- 4-7 years of experience in the space of data engineering.
- At least 2-4 years of experience with Snowflake Data warehouse is a must.
- Recent project work should include usage and deploying production code to Snowflake. Out of touch in Snowflake is not preferred or recommended.
- At least 2-4 years of experience with Python development is a must.
- Experience with pyspark, Snowpark apis would be a plus.
- Experience with basics of data frames and how to around managing data, transformations and using custom sdk’s on top of dataframes would be a plus.
- Ability to write unit tests using pytests or spark testing base would be a plus.
- Experience in the following areas with Snowflake as datastore would be preferred:
- Managing Data modeling using Liquibase,
- Managing Data pipelines via SnowSQL / SnowPipe
- Orchestrating data pipelines within Snowflake or outside of Snowflake.
- Experience with using/coding data quality frameworks like Soad/Great Expectations etc.,
- Experience with using Airflow / AWS Step functions / Glue / Spark would be preferred and a plus.
- Experience in one or more below:
- Strong in SQL , SQL Analytics, ELT/ETL’ing and Data warehousing.
- Experience with Azure Synapse / ADF / Cloud ETL tool. (nice to have)
- Experience with Code Repository, CI/CD , test driven development is preferred.
- Creating Unit Testing and Automated integration tests is strongly preferred.
- Experience with multi-cloud AWS/Azure/GCP cloud is a plus. Azure is required/Others preferred.
- Data processing experience at scale, generating and consuming large feeds.
- Experience with other cloud data warehouses like Redshift/Azure SQLWH etc., (nice to have).
- Experience in handling big data at PB scale (nice to have)
- Experience in dbt, SQL Mesh would be a plus (nice to have)
- Experience with data governance programs and data offices would be a plus.
- Exposure to claims, payer, formulary, RWE, EHR/ EMR OR life science, Healthcare data is recommended and would be a plus.