We are looking for an experienced Data Engineer (AWS) to join our team. The ideal candidate will have strong expertise in cloud-based data engineering, with a focus on AWS, Python, and PySpark, along with a solid understanding of SQL/PL-SQL.
Key Responsibilities
- Design, build, and maintain scalable data pipelines on AWS.
- Work with structured and unstructured datasets, ensuring data quality and consistency.
- Develop and optimize ETL processes using Python and PySpark.
- Collaborate with data scientists, analysts, and business teams to understand requirements and deliver solutions.
- Implement best practices for data storage, transformation, and integration in cloud environments.
- Ensure data security, compliance, and governance across systems.
- Monitor, troubleshoot, and optimize data workflows for performance and efficiency.
- Stay updated with emerging cloud and big data technologies to continuously improve solutions.
Skills & Qualifications
- Strong expertise in AWS (S3, Glue, Lambda, EMR, Redshift, etc.)
- Proficiency in Python for data engineering tasks.
- Hands-on experience with PySpark for large-scale data processing.
- Good understanding or working knowledge of SQL / PL-SQL.
- Experience in data modeling, ETL design, and performance optimization.
- Familiarity with Agile/Scrum methodologies.
- Excellent problem-solving and communication skills.