Design and implement scalable ETL pipelines using Python and AWS services.
Model and manage datasets for analytics and machine learning workflows.  
Optimize data workflows for performance, reliability, and maintainability.  
 
The resource should possess the following skills and experience:
Strong proficiency in Python for data engineering.  
Hands-on experience with AWS (e.g., S3, Lambda, Glue, RDS).   Solid understanding of ETL processes and data pipeline orchestration.  These are nice-to-have skills to complement the required skills: 
Experience with data modeling for structured and semi-structured data.  
Familiarity with Spark, Airflow, or similar orchestration tools.  
Exposure to ML workflows or LLM integration.   Knowledge of containerization (Docker, Kubernetes) and CI/CD practices
 
Neha Chaudhary
Team Lead – Recruitment
e: neha.chaudhary@compunnel.com 
o: (+1) 609-606-9010 ext.2469
m: (+1) 732-743-9949 
HQ: 4390 Route 1 North, Suite 302, Princeton, NJ 08540, USA.
—
