Location-Columbus, OH (Onsite)
Contract
Banking and Finance Domain
Required Skills & Qualifications:
• Strong proficiency in Python for data processing and pipeline development
• Hands-on experience with Apache Spark (PySpark preferred) • Solid experience with AWS services such as S3, Glue, EMR, Redshift, Athena, Lambda
• Experience with SQL and relational/non-relational databases
• Knowledge of data modeling, data warehousing concepts, and ETL frameworks
• Experience working with large-scale, distributed data systems
• Familiarity with CI/CD pipelines and version control tools (Git)
• Strong problem-solving and communication skills
Preferred / Nice to Have:
• Experience with Airflow or other workflow orchestration tools
• Knowledge of Kafka, Kinesis, or streaming data platforms
• Experience with Docker/Kubernetes • Exposure to Delta Lake, Iceberg, or HuD.
—