We are seeking a skilled Data Engineer to design, build, and optimize scalable data systems that power analytics and AI-driven capabilities. This role focuses on developing robust data pipelines, data lakes, and data warehouses to ensure high-quality, reliable, and efficient data flow across the platform.
Requirements
- Design and implement scalable, secure, and high-performance data pipelines for structured and unstructured data.
- Build and maintain data lake and data warehouse architectures, ensuring seamless integration between systems.
- Develop ETL/ELT workflows using modern data tools and frameworks.
- Enable real-time data ingestion using streaming technologies.
- Ensure data reliability, availability, and recoverability through robust pipeline design.
- Design and manage enterprise-grade data warehouse solutions.
- Develop optimized data models (dimensional/star schemas) for analytics and reporting.
- Improve query performance through indexing, partitioning, and optimization techniques.
- Implement data marts and domain-specific schemas for business use cases.
- Build efficient data transformation pipelines using SQL and distributed processing frameworks.
- Develop reusable data modules for cleansing, enrichment, and standardization.
- Implement incremental data processing, schema evolution, and change data capture (CDC).
- Support feature engineering workflows for machine learning applications.
- Establish data quality frameworks to ensure accuracy, consistency, and reliability.
- Implement monitoring systems for data freshness, latency, and pipeline health.
- Set up alerts for failures, anomalies, and SLA breaches.
- Manage metadata, data lineage, and cataloging systems.
- Ensure compliance with data security, privacy, and regulatory standards.
- Work closely with analytics, BI, and ML teams to deliver high-quality datasets.
- Support reporting, dashboards, and KPI computation.
- Partner with engineering teams to align data architecture with platform goals.