




Job Summary: A data professional who applies statistical techniques, data modeling, and data mining, with a focus on scalable and robust pipelines, database optimization, and strategic collaboration. Key Highlights: 1. Experience in Big Data (Hadoop and PySpark) and programming languages such as Python, R, Scala, and SQL. 2. Design, build, and maintain scalable and robust data pipelines. 3. Experience in developing and maintaining data lakes and data warehouses. **Job Description:** **Description of responsibilities the professional will perform:** Apply statistical techniques for variable selection, sampling, and experimental design. Apply segmentation, statistical and computational modeling techniques (e.g., AI and Machine Learning). Apply data mining techniques, data structures, and data science solutions. Perform data modeling, generate statistics, and identify trends—including exploratory, descriptive, diagnostic, predictive, and prescriptive analytics. Data visualization and analysis in Python (pandas, numpy, statsmodels, matplotlib, seaborn). Work with Big Data environments and tools (Hadoop and PySpark), programming languages and tools (Python, R, Scala, and SQL), and master data management platforms and tools. **Required Experience, Skills, and Knowledge:** Responsibilities: * Design, build, and maintain scalable and robust data pipelines that meet analytical requirements. * Implement and manage large-scale data storage solutions on GCP. * Optimize database performance and ensure data quality, integrity, and security through best practices and automation. * Collaborate with data science, BI, and infrastructure teams to develop efficient solutions that transform raw data into strategic insights. Skills: * Proven experience in developing and maintaining data lakes and data warehouses. Preferred: Google Cloud Storage/BigQuery. * Proficiency in Python and SQL. * Advanced knowledge of ETL/ELT tools and experience orchestrating data pipelines. Preferred: dbt. * Experience in data architecture within cloud environments (GCP, AWS, Azure). **Education:** Postgraduate degree or MBA. Mandatory qualification: Bachelor's degree in Information Technology (or a related field). OR Bachelor's degree in any field, combined with a completed postgraduate program (specialization, master's, or doctoral degree) in Information Technology (or a related field), comprising a minimum of 360 classroom hours. Mandatory Qualification: Postgraduate degree or MBA in the relevant field and/or Software Engineering processes, with a minimum duration of 360 classroom hours. OR At least three (3) professional certifications in the relevant field and/or agile methodologies, issued by a recognized certification body. **Years of Experience:** Mandatory experience Minimum of six (6) years of professional experience in the relevant technical specialty area. OR Minimum of four (4) years of cumulative professional experience, combined with at least two (2) professional certifications in the relevant field and/or agile methodologies, issued by a recognized certification body. **Knowledge and Skills:** SQL, Python, AI, data lake, warehouse ### **Employment Type:** CLT ### **Benefits:** Meal Voucher, Food Allowance, Dental Plan, Health Insurance, TotalPass, Subsidy ### **Department:** Government


