Data Engineer position at TechBiz Global, seeking a candidate to join a client's team, working remotely.
Requirements
- Design, develop, and maintain data ingestion pipelines using Kafka Connect and Debezium for real-time and batch data integration.
- Ingest data from MySQL and PostgreSQL databases into AWS S3, Google Cloud Storage (GCS), and BigQuery.
- Implement best practices for data modeling, schema evolution, and efficient partitioning in the Bronze Layer.
- Ensure reliability, scalability, and monitoring of Kafka Connect clusters and connectors.
- Collaborate with cross-functional teams to understand source systems and downstream data requirements.
- Optimize data ingestion processes for performance and cost efficiency.
- Contribute to automation and deployment scripts using Python and cloud-native tools.
- Stay updated with emerging data lake technologies such as Apache Hudi or Apache Iceberg.