Databricks C2C jobs
Role: Databricks Data Engineer
Location: Las Vegas, NV / ONSITE
Duration: Long Term Project
Experience: 12+
Job Role and Responsibilities:
• Lead end-to-end migration of enterprise data assets from Teradata to Databricks.
• Analyze existing Teradata schemas, stored procedures, ETL pipelines and map them to Databricks equivalents.
• Design and implement scalable data ingestion pipelines using PySpark, Spark SQL, Auto Loader, Declarative pipelines, or other Databricks frameworks.
• Design and implement multi-layered Lakehouse architecture following Medallion (Bronze–Silver–Gold) best practices.
• Build Bronze layer raw data ingestion patterns using Delta Lake with schema inference, schema evolution, and data quality checks.
• Develop Silver layer transformation pipelines to standardize, cleanse, dedupe, conform, and join datasets.
• Create optimized Gold layer presentation datasets to support analytics, BI workloads.
• Develop scalable distributed data processing jobs using PySpark, Spark SQL, Delta Lake, and Databricks Workflows.
• Convert complex Teradata SQL logic into optimized PySpark/Spark SQL.
• Configure and optimize Databricks clusters (all-purpose, job clusters, serverless)
• Ensure SLAs, data reliability, and system stability for production of data pipelines.
• Unit testing the ETL code & documenting the test results
• Perform data quality and profiling activities, advise on data quality issues, and provide analysis on data collection, mapping, aggregation, and balancing functionality
Databricks Data Intelligence Platform,ETL,PySpark,Python,SQL,Teradata,analytical database,auto loader,business intelligence,data analysis,data intelligence,data lake,data management,data mining,data pipeline,data processing,data quality,data querying,data science,data warehouse,information technology,mathematics,programming,python data science,software development,statistics.
To apply for this job email your details to shahid.m@wonese.com