HS

Data Engineer - Python/Consultant Specialist

HSBC

6 months ago

5 - 7 years

Work From Office

Hyderabad, Telangana, Telangana, India

  • Design, develop, and optimize data pipelines using Azure Databricks, PySpark, and Prophesy.
  • Implement and maintain ETL/ELT pipelines using Azure Data Factory (ADF) and Apache Airflow for orchestration.
  • Develop and optimize complex SQL queries and Python-based data transformation logic.
  • Azure Databricks

    PySpark

    SQL

    Apache Airflow

    Azure Data Factory

    ETL Processes

    Version Control (Git)

    Data Modeling

    Data Warehousing

    Machine Learning

    Job description & requirements

    In this role, you will:


    1. Design, develop, and optimize data pipelines using Azure Databricks, PySpark, and Prophesy.
    2. Implement and maintain ETL/ELT pipelines using Azure Data Factory (ADF) and Apache Airflow for orchestration.
    3. Develop and optimize complex SQL queries and Python-based data transformation logic.
    4. Work with version control systems (GitHub, Azure DevOps) to manage code and deployment processes.
    5. Automate deployment of data pipelines using CI/CD practices in Azure DevOps.
    6. Ensure data quality, security, and compliance with best practices.
    7. Monitor and troubleshoot performance issues in data pipelines.
    8. Collaborate with cross-functional teams to define data requirements and strategies.


    Requirements

    To be successful in this role, you should meet the following requirements:

    1. 5+ years of experience in data engineering, working with Azure Databricks, PySpark, and SQL.
    2. Hands-on experience with Prophesy for data pipeline development.
    3. Proficiency in Python for data processing and transformation.
    4. Experience with Apache Airflow for workflow orchestration.
    5. Strong expertise in Azure Data Factory (ADF) for building and managing ETL processes.
    6. Familiarity with GitHub and Azure DevOps for version control and CI/CD automation.
    7. Solid understanding of data modelling, warehousing, and performance optimization.
    8. Ability to work in an agile environment and manage multiple priorities effectively.
    9. Excellent problem-solving skills and attention to detail.
    10. Experience with Delta Lake and Lakehouse architecture.
    11. Hands-on experience with Terraform or Infrastructure as Code (IaC).
    12. Understanding of machine learning workflows in a data engineering context.


    Experience :

    5 - 7 years

    Job Domain/Function :

    Data Engineering

    Job Type :

    Work From Office

    Employment Type :

    Full Time

    Number Of Position(s) :

    1

    Educational Qualifications :

    Bachelor's Degree

    Location :

    Hyderabad, Telangana, India, Hyderabad, Telangana, India

    Create alert for similar jobs

    Similar Jobs