We are seeking a highly skilled Data Engineer with strong expertise in AWS cloud services to design, build, and maintain scalable data pipelines and architectures. The ideal candidate will have deep knowledge of data engineering best practices, cloud-native solutions, and the ability to optimize data flows that power analytics and business insights.
Responsibilities
- Design, build, and maintain ETL/ELT pipelines to ingest data from various sources into AWS and GCP data platforms.
- Manage and monitor data storage solutions (Amazon S3, Redshift, etc).
- Implement data transformation workflows using services like AWS Glue, Lambda, EMR, or Step Functions.
- Ensure data quality, security, and governance across all pipelines.
- Work with streaming data solutions (Kinesis, Kafka, or MSK) when needed.
- Collaborate with Data Analysts, DevOps, Engineering, and Business stakeholders to understand data requirements.
- Monitor pipeline performance and troubleshoot issues to maintain high data availability and reliability.
Requirements
- Bachelor’s degree in Computer Science, Data Engineering, or a related field.
- 4+ years of experience in Data Engineering or related fields.
- Data Storage: S3, Redshift.
- Data Processing: Glue, EMR, Lambda, Step Functions.
- Streaming: Kinesis, Kafka/MSK.
- Proficiency in SQL and Python/Pyspark.
- Experience with data modeling, schema design, and query optimization.
- Knowledge of data security, encryption, IAM, and compliance standards.
- Familiarity with CI/CD pipelines and Infrastructure as Code.