




Description: * Bachelor's degree in Computer Engineering, Computer Science, Information Systems, or related fields; * Solid experience with Spark, Databricks, or equivalent distributed engine; * Knowledge of: * Advanced SQL; * Advanced Python (pandas, pyspark); * Cloud platforms (GCP, Azure, or AWS); * Orchestration: Airflow, Databricks Jobs, Workflows, n8n; * Complex batch and streaming pipelines; * Git, CI/CD, and version control. * Design, develop, and maintain robust and scalable data pipelines (batch and streaming); * Structure, optimize, and monitor data lakes and data warehouses in cloud environments; * Define and implement data engineering best practices, including versioning, CI/CD, observability, governance, and security; * Ensure data quality through automated testing, validations, and continuous monitoring; * Model data (dimensional, Data Vault) for consumption by analysts, data scientists, and data products; * Support BI, Analytics, and Product teams in delivering reliable and well-documented datasets; * Evaluate and implement new technologies, tools, and architectural patterns; * Collaborate with IT, Engineering, Product, and Business teams to foster a data-driven culture and best practices. 2512210202551931389


