Data Engineer (Data Center) - IT - India

Job no: IND2025-CTO05
Work type: Permanent - Full Time
Location: India
Categories: Mid-Senior Level

Apply now

Role Profile:

As a Data Engineer you are expected to contribute towards

Create and maintain optimal data pipeline architecture,

Assemble large data sets that meet functional / non-functional business requirements.

Design and implement internal process improvements: automating manual processes, and optimizing data delivery etc.

Help build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using Python,  SQL and GCP ‘big data’ technologies.

Work with stakeholders including Product, Data Integration and Design teams to assist with data-related technical issues and support their data infrastructure needs.

Create data tools for BI reporting, digital analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.

Work with data and analytics experts to strive for greater functionality in our data systems.

 

The below Key Performance Areas include but are not limited to:

Takes a breadth first approach to problem solving and encourages systematic thinking.

Good communicator and collaborator, able to simplify concepts for wider non-tech audiences.

Able to demonstrate a passion for data engineering with the ability to work cross functional.

Culturally sensitive and able to work across cultures and time zones.

 

Experience

Experience of designing, building and maintaining scalable ETL data processing systems using Python, Java, Scala or similar programming language – this is a hands-on coding role.

Understanding and experience in data processing techniques and analytics with Strong SQL experience.

Understanding and experience of GCP data tools (GBQ, Dataflow, DataProc, GC Data Transfer etc.)

Understanding and experience of using and creating CICD pipelines using Azure Devops, Jenkins, GoCD or any similar technologies

Understanding and experience of using orchestration tools such as Apache Airflow, Azure Datafactory, AWS Data pipelines or similar

Understanding and experience of working with Cloud Deployment Environments (GCP or AWS preferred)

Proponent for software craftmanship, TDD and automated testing

DevOps background with understanding of cloud-based IAC technologies

Terraform or serverless

Github as version control

Familiar with CI/CD

Uunderstanding of Data Modelling and/or Data Warehousing techniques such as star schema or raw data vault

Experience of working with large-scale data sets, this includes both structured and unstructured data such as parquet, avro, json etc.

Advertised: India Standard Time
Application close:

Apply now

Back to list Refer a friend