Module Lead - Data ModellingYears of Experience: 7+ years
Job Summary: We are seeking a skilled Data Engineer with extensive experience in designing, building, and maintaining data pipelines. The ideal candidate will have a strong background in data engineering, particularly with Databricks and Azure services, and will be responsible for ensuring data integrity, quality, and security across our cloud data platform. This role requires collaboration with teams across Asia and Africa to deliver effective data solutions.
Key Responsibilities:
? Design, build, and maintain data pipelines to ingest data from multiple sources into our cloud data platform.
? Ensure pipelines are built using defined standards and maintain comprehensive documentation.
? Adhere to and enforce data governance standards to maintain data integrity and compliance.
? Implement data quality rules to ensure the accuracy and reliability of data.
? Implement data security and protection controls around Databricks Unity Catalog.
? Utilize Azure Data Factory, Azure Databricks, and other Azure services to build and optimize data pipelines.
? Leverage SQL, Python/PySpark, and other programming languages for data processing and transformation.
? Stay updated with the latest Azure technologies and best practices.
? Provide technical guidance and support to team members and stakeholders.
? Maintain detailed documentation of data pipelines, processes, and data quality rules.
? Debug, fine-tune, and optimize large-scale data processing jobs.
? Generate reports and dashboards to monitor data pipeline performance and data quality metrics.
? Work collaboratively with data teams across Asia and Africa to understand data requirements and deliver solutions.
Mandatory Skills:
? Bachelor?s or Master?s degree in Computer Science, Information Technology, or a related field.
? 7+ years of experience in data engineering using Databricks/Spark.
? Proficiency in Azure Data Factory, Azure Databricks, and other Azure data services or equivalent in other cloud services.
? Strong programming skills in SQL, Python/Apache Spark RDD languages.
? Experience with data quality and Governance Framework Design.
? Knowledge of data modeling techniques and best practices (deal breaker skill).
? Excellent Communication Skills.
? Ability to work in a fast-paced, dynamic environment and manage multiple priorities.
Preferred Skills:
? Experience with Big Data technologies such as Spark distribution like Databricks/Cloudera/Data Proc, etc.
? Familiarity with Agile methodologies and Devsecops practices.
? Knowledge of Power BI or Tableau.
Join our team and contribute to building robust data solutions that drive business insights and decision-making.

Keyskills: cloudera azure data factory sql apache postgresql data modeling spark design big data mongodb programming communication skills azure databricks python cloud services data services dynamo db languages microsoft azure power bi data engineering data bricks tableau rdd agile