Mission:
We are looking for a highly skilled Data Engineer to join our growing data team. In this role, you will be responsible for expanding and optimizing our data and data pipeline architecture, as well as optimizing data flow and collection for cross-functional teams. You will design, build, and maintain scalable data pipelines and data warehouses to ensure our business has high-quality, actionable data.
Key Responsibilities:
Pipeline Development: Design, construct, and manage robust and scalable ETL/ELT processes and data pipelines.
Data Architecture: Develop and maintain data models (Star, Snowflake, Data Vault) and high-performance data warehousing solutions.
Cloud Infrastructure: Build and manage data infrastructure on cloud platforms such as Azure (Data Factory), AWS (Glue), or Google Cloud (Dataflow).
Integration: Ingest and integrate data from heterogeneous sources, including internal databases, third-party APIs, and flat files.
Data Governance: Implement data quality checks, monitoring, and security best practices to ensure compliance and data integrity.
DevOps & CI/CD: Utilize version control (Git) and implement CI/CD fundamentals for automated deployment and testing of data workflows.
Big Data Processing: Work with Big Data tools like Spark or Hadoop to process large-scale datasets efficiently.
Qualifications & Skills:
Core Engineering: Proficiency in Python and advanced SQL (performance tuning, complex joins, stored procedures).
Data Platforms: Strong experience with modern cloud data warehouses such as Snowflake, Synapse, Redshift, or BigQuery.
Modeling: Deep understanding of data modeling concepts and architectural patterns.
API & Integration: Proven experience working with REST/SOAP APIs and integrating disparate data systems.
Tooling: Familiarity with orchestration tools (e.g., Airflow) and Big Data frameworks.
Soft Skills: Exceptional analytical and troubleshooting skills with the ability to solve complex data engineering challenges.
Preferred Qualifications:
Experience with Infrastructure as Code (Terraform, CloudFormation).
Knowledge of streaming data technologies (Kafka, Pub/Sub).
Certifications in cloud data engineering (e.g., Azure Data Engineer Associate, AWS Certified Data Engineer).
Location: Astana city
Employment Type: 5/2 Hybrid (Office + Remote)