




Job Summary: A professional with experience in leading Big Data projects, solution architecture, data pipeline development, and optimization of distributed environments. Key Highlights: 1. Lead Big Data architecture and solution projects. 2. Develop data pipelines and ETL processes. 3. Ensure high availability, security, and performance of clusters. Description: * Completed undergraduate degree; * Advanced knowledge of ITIL; * Advanced proficiency in Python; * Advanced knowledge of PySpark; * Experience with Airflow and pipeline automation; * Advanced experience with Cloudera; * Knowledge of Hadoop (administration and optimization); * Advanced experience in ETL (Pentaho Data Integration/Kettle, Jasper ETL, and similar tools); * Experience with infrastructure-as-code and DevOps practices. * Lead Big Data architecture and solution projects; * Design, administer, and optimize distributed environments (Hadoop, Cloudera, Spark); * Develop data pipelines and ETL processes; * Ensure high availability, security, and performance of clusters; * Resolve complex infrastructure issues and perform tuning; * Support and train technical teams, documenting best practices; * Collaborate with stakeholders to define requirements and design solutions; * Ensure data governance, quality, and security. 251110020218420281


