Design and maintain optimal data pipeline architecture, assemble and integrate large datasets, build and optimize infrastructure for ETL/ELT of data from diverse sources, collaborate with cross-functional stakeholders to address data-related technical challenges and support their data infrastructure needs.
Requirements
- Minimum 5 years of experience in a data engineering role
- Hands-on experience with distributed data processing / big data frameworks (Databricks)
- Python
- Cloud data platforms
- AWS or other cloud environments (e.g., Azure, GCP)
- Experience with workflow orchestration / scheduling platforms
- Version control systems such as GitHub, GitLab, Bitbucket, or Azure Repos
- CI/CD platforms such as Jenkins, GitLab CI, GitHub Actions, or Azure DevOps
- Modern Python API frameworks and strong understanding of RESTful services and good API design principles
- System design concepts, including designing scalable, reliable, and fault-tolerant data architectures
- Object-oriented programming principles and design patterns
- Big data / distributed systems technologies, including streaming & messaging platforms, search & indexing / log analytics engines, structured & unstructured data, container orchestration platforms
Benefits
- Health & Wellness
- Flexible Downtime
- Continuous Learning
- Invest in Your Future
- Family Friendly Perks
- Beyond the Basics