As a Data Engineer at GFT, you will be responsible for managing, designing, and enhancing data systems and workflows that drive key business decisions. The role is focused 75% on data engineering, involving the construction and optimization of data pipelines and architectures, and 25% on supporting data science initiatives through collaboration with data science teams for machine learning workflows and advanced analytics.
Requirements
- Proficiency in Python for data processing and scripting (pandas, pyspark), workflow automation (Apache Airflow), and experience with AWS services (Glue, S3, EC2, Lambda)
- Experience working with Kubernetes and Docker for managing containerized environments in the cloud
- Hands-on experience with columnar and big data databases (Athena, Redshift, Vertica, Hive/Hadoop), along with version control systems like Git
- Strong familiarity with AWS services for cloud-based data processing and management
- Experience with CI/CD tools such as Jenkins, CircleCI, or AWS CodePipeline for continuous integration and deployment
- Expertise in building and managing robust data architectures and pipelines for large-scale data operations
- Ability to support data science workflows, including collaboration on data preparation, feature engineering, and enabling experimentation environments
Benefits
- Competitive salary
- 13th-month salary guarantee
- Performance bonus
- Professional English course for employees
- Premium health insurance
- Extensive annual leave