- Design, build and operationalize large scale enterprise data solutions and applications using one or more of AWS data and analytics services in combination with 3rd parties Spark/Python on Glue, DMS, S3, Athena, RDS-PostgreSQL, Airflow, Lambda, Code Commit, Code Pipeline, Code build, etc.
- Design and build production data pipelines from ingestion to consumption within a big data architecture, using DMS, DataSync & Glue.
- Understand the existing applications (including on-premise Cloudera Datalake) and infrastructure architecture.
- Design and implement data engineering, ingestion and curation functions on AWS cloud using AWS native or custom programming.
Requirements
- Minimum 2 years of work experience with Data Ingestion, ETL, Data Modelling, Data Architecture to build Datalake.
- Proficient in designing, coding, and tuning big data processes using Pyspark.
- Minimum 2 years of extensive experience in working on AWS platform using core services like AWS Athena, Glue Pyspark, RDS-PostgreSQL, S3 & Airflow (for orchestration)
Interested applicants may send in their detailed resumes in MS Word format to [Confidential Information], with subject title: Data Engineer
(Reg. Nos. R1104128)
EA License No.: 92C5311
For more information about ITC, please visit us @ www.itcspl.com