Design, develop, and maintain ETL pipelines using Databricks, PySpark and ADF to extract, transform, and load data from various sources.
Must have good skills in Pyspark , Programming, code remediation etc.,
Must have good working experience on Delta tables, deduplication, merging with terabyte of data set
Optimize and fine-tune existing ETL workflows for performance and scalability. 2 to 3 years of experience in ADF is desirable (Medium expertise require)
Must have experience on working with large data set
Proficient in SQL and must worked on complex joins, Subqueries, functions, procedure
One should be self-driven and work independently without support
SKILLS
NA
Similar jobs
This role will join the ServiceNow Platform team. The ideal candidate should possess a strong technical background in ServiceNow... Expand NA A MuleSoft Dev is responsible for designing and implementing integration solutions using MuleSoft's Anypoint Platform. You... Expand NA Experience with Databricks, Spark, and handling large volume of data Strong experience in data... Expand NA Design, develop, and maintain ETL pipelines using Databricks, PySpark and ADF to extract, transform, and load data from various... Expand NA NA o Have good knowledge in Oracle Fusion Human Capital Management... Expand NA NA Should have good hands-on experience in Base SAS, SAS/Macros and other SAS tools. NA
:
Should work with the Data validation team for... Expand