Ready to build the future with AI? At Genpact, we don’t just keep up with technology—we set the pace. AI and digital innovation are redefining industries, and we’re leading the charge. Genpact’s AI Gigafactory, our industry-first accelerator, is an example of how we’re scaling advanced technology solutions to help global enterprises work smarter, grow faster, and transform at scale.
Requirements
- Maintains close awareness of new and emerging technologies and their potential application for service offerings and products.
- Work with architect and lead engineers for solutions to meet functional and non-functional requirements.
- Demonstrated knowledge of relevant industry trends and standards.
- Demonstrate strong analytical and technical problem-solving skills.
- Must have experience in Data Engineering domain.
- Maintains close awareness of new and emerging technologies and their potential application for service offerings and products.
- Work with architect and lead engineers for solutions to meet functional and non-functional requirements.
- Demonstrated knowledge of relevant industry trends and standards.
- Demonstrate strong analytical and technical problem-solving skills.
- Must have excellent coding skills either Python or Scala, preferably Python.
- Must have experience in Data Engineering domain.Must have implemented at least 4 project end-to-end in Databricks.
- Must have at least experience on databricks which consists of various components as below
- Must have skills: Azure data factory, Azure data bricks, Python and Pyspark
- Expert with database technologies and ETL tools.
- Hands-on experience on designing and developing scripts for custom ETL processes and automation in Azure data factory, Azure databricks, Python, Pyspark etc.
- Good knowledge of AZURE, AWS, GCP Cloud platform services stack
- Hands-on experience on designing and developing scripts for custom ETL processes and automation in Azure data factory, Azure databricks,Delta lake, Databricks workflows orchestration, Python, Pyspark etc.
- Good Knowledge on Unity Catalog implementation.
- Good Knowledge on integration with other tools like – DBT, other transformation tools.
- Good knowledge on Unity Catalog integration with Snowlflake
- Must be well versed with Databricks Lakehouse concept and its implementation in enterprise environments.
- Must have good understanding to create complex data pipeline
- Must have good knowledge of Data structure & algorithms.
- Must be strong in SQL and sprak-sql.
- Must have strong performance optimization skills to improve efficiency and reduce cost.
- Must have worked on both Batch and streaming data pipeline.
- Must have extensive knowledge of Spark and Hive data processing framework.
- Must have worked on any cloud (Azure, AWS, GCP) and most common services like ADLS/S3, ADF/Lambda, CosmosDB/DynamoDB, ASB/SQS, Cloud databases.
- Must be strong in writing unit test case and integration test
- Must have strong communication skills and have worked on the team of size 5 plus
- Must have great attitude towards learning new skills and upskilling the existing skills.
Benefits
- Opportunity to work on cutting-edge AI projects
- Collaborative and dynamic work environment
- Opportunity to learn from top engineers, data scientists, and AI experts
- Opportunity to gain hands-on experience and world-class training
- Competitive salary and benefits package
- Opportunity to work on a diverse range of projects and technologies
- Flexible work arrangements and work-life balance