DT

Python Pyspark Data Engineer

DXC Technology

5 months ago

5 - 7 years

Work From Office

Hyderabad, Telangana, India

  • Design, develop, and optimize data pipelines using Python, Pyspark, AWS Data Analytics services such as RDS, DMS, Glue, Lambda, Redshift, and Athena.
  • Implement data migration and transformation processes using AWS DMS and Glue.
  • Develop and maintain ETL/ELT workflows for data ingestion and transformation.
  • SQL

    ETL/ELT

    PYTHON

    Azure DevOps

    CI/CD pipeline

    pyspark

    Job description & requirements

    Key Responsibilities:


    1. Design, develop, and optimize data pipelines using Python, Pyspark, AWS Data Analytics services such as RDS, DMS, Glue, Lambda, Redshift, and Athena.
    2. Implement data migration and transformation processes using AWS DMS and Glue.
    3. Work with SQL (Oracle & Postgres) to query, manipulate, and analyse large datasets.
    4. Develop and maintain ETL/ELT workflows for data ingestion and transformation.
    5. Utilize AWS services like S3, IAM, CloudWatch, and VPC to ensure secure and efficient data operations.
    6. Write clean and efficient Python scripts for automation and data processing.
    7. Collaborate with DevOps teams using Azure DevOps for CI/CD pipelines and infrastructure management.
    8. Monitor and troubleshoot data workflows to ensure high availability and performance.


    Experience :

    5 - 7 years

    Job Domain/Function :

    Data Engineering

    Job Type :

    Work From Office

    Employment Type :

    Full Time

    Number Of Position(s) :

    1

    Educational Qualifications :

    Bachelor's Degree

    Location 1 :

    Hyderabad, Telangana, India, Hyderabad, Telangana, India

    Location 2 :

    Bengaluru, Karnataka, India, Karnataka, India

    Create alert for similar jobs

    DT

    DXC Technology

    Similar Jobs

    Python Pyspark Data Engineer-DXC Technology-Hyderabad, India-5 - 7 years