
c2c h1b
Python developer
Hybrid, NJ
3+ Months
Contract
Design, develop, and maintain high-performance data pipelines for data ingestion, transformation, and storage.
Work with large-scale data processing frameworks (Apache Spark, Dask, or Pandas) to process big data efficiently.
Develop ETL (Extract, Transform, Load) workflows for structured and unstructured data.
Optimize SQL queries and database performance for analytics and reporting.
Ensure data quality, validation, and governance across pipelines and storage layers.
Implement CI/CD pipelines for automating data pipeline deployment and monitoring.
Required Skills & Qualifications:
Strong Python programming experience with a focus on data engineering.
Hands-on experience in building and optimizing data pipelines for large-scale datasets.
Proficiency with SQL and NoSQL databases (PostgreSQL, BigQuery, MongoDB, etc.).
Familiarity with cloud data solutions and serverless computing (AWS Lambda, Azure Functions, GCP Cloud Functions).
Knowledge of workflow orchestration tools (Apache Airflow, Prefect, or Dagster).
Ability to troubleshoot, optimize, and scale data workflows effectively.
Munesh
To apply for this job email your details to munesh@cysphere.net