CLOUD DATA ENGINEER
Core Expertise:
Experience: At least 3 years in a data engineering role working on data integration, processing, and transformation use cases with open-source languages (i.e. Python) and cloud technologies.
Strong programming skills in Python specifically for API integration and data libraries, with an emphasis on quality and test-driven development
Demonstrated proficiency with big data storage formats (Apache Parquet, ORC) and practical knowledge of pitfalls and optimization strategies
Demonstrated proficiency with SQL
Experience with data modeling:
Relational modeling
Dimensional modeling
NoSQL modeling
Working knowledge of IaC on AWS (CloudFormation or CDK)
Working knowledge of AWS Services:
Required: Glue, IAM, Lambda, DynamoDb, Step Functions, S3, CloudFormation or CDK
Nice-to-have: Athena, Kinesis, MSK, MWAA, SQS
Experience with the orchestration of data flows / pipelines: Apache Airflow or Dagster
Nice-to-have: Experience with data streaming (Kinesis, Kafka)
Nice-to-have: Experience with Apache Spark
Nice-to-have: Client-facing experience, multi-cultural team experience, technical leadership, team leadership
Abilities
Ability to work both autonomously and as part of a team
Attention to quality and details
Good critical and analytical thinking
Good time management and organizational skills
Customer / end-client orientation
Excellent written and spoken communication skills