




Description: Role in the company: Work as a Data Engineer focused on the AWS platform, responsible for designing, developing, and maintaining robust data pipelines within a Lakehouse architecture (Delta Lake), handling high data volumes with performance, governance, and reliability requirements. The professional will work daily with AWS, actively participate in data architecture decisions, ensure best practices in versioning and automation (DevOps), and support business operations with consistent and available data across multiple layers (Bronze/Silver/Gold). Responsibilities: Develop and maintain data pipelines using PySpark for batch ingestion, focusing on performance and scalability. Work within Lakehouse environments using Delta Lake, implementing and maintaining Bronze / Silver / Gold layers, including daily replication, incremental deltas, and ACID versioning. Operate and orchestrate DAGs (Airflow), performing monitoring, troubleshooting, and version control of data pipelines. Use AWS Lambda and Kafka to build data ingestion and distributed processing streams. Work with PRs, code reviews, structured versioning, and integration with automation pipelines, following code governance processes. Ensure pipeline observability (logs, alerts, monitoring) and proactively address failures and continuously improve processes. Collaborate with product teams to understand needs and translate technical requirements into scalable data solutions. Requirements (Hard Skills): Proficiency in Spark for building and maintaining high-volume data pipelines (batch, incremental, and real\-time). Experience with Lakehouse architecture and Delta Lake, including implementation and maintenance of Bronze/Silver/Gold layers. Knowledge of messaging/streaming tools such as Kafka. Experience with CI/CD, automation of builds, testing, validations, and deployments across environments. Experience in pipeline orchestration and DAGs (Airflow, Step Functions, or equivalent tools). Differentiators: Deep experience in data governance, security, and best practices in AWS environments. Prior experience in large-scale data projects (high volume and high availability requirements). Experience with cloud data ecosystem tools (e.g., Databricks on AWS, data catalog tools, etc). Our benefits: * Medical allowance: Monthly financial aid to support your health plan subscription. * Wellhub, to keep your body and mind active in the way you prefer. * Fully company-paid online therapy, because mental health matters. * Online Nutrition consultations, up to two per month, to support your dietary health. * Life insurance policy worth R$125,000, providing greater security for you and your family. * Birthday day off, because your day deserves to be special. * Paid leave, so you can recharge your energy. * Internal gamification, turning achievements into rewards and recognition. * Educational partnerships with institutions such as FIAP, Anhanguera, and Instituto Infnet, supporting your growth and learning. * Technical certification bonus, recognizing and rewarding your efforts to learn. We value every voice and every individual, because we know diversity makes us more innovative and stronger. 2511230202461885568


