Pleased to meet you, we are Zallpy.We are much more than a technology company; we are a diverse, plural, and talented community. Our purpose is to lead digital transformation with excellence and agility, promoting mutual and genuine growth in ethical and long-lasting relationships. Flexibility is one of our trademarks; we operate in different team models and formats, maintaining our light, collaborative, and integrated culture, providing equitable opportunities in a space where everyone feels safe and heard.What you will do:Leverage strong SQL skills to manipulate, transform, and manage data;Work with both relational and key-value databases to ensure data accuracy and consistency;Utilize in-depth knowledge of dimensional modeling to create effective and efficient data structures;Develop features tailored for advanced data analysis and modeling;Design and implement robust data pipelines for seamless data collection, processing, and storage;Extract data from external sources by manipulating APIs;Use data processing tools like Pentaho, Streamsets, or equivalent platforms;Master Python and/or Java to create and deploy data engineering solutions;Apply version control best practices using Git;Understand the architecture of Data Platforms, Data Warehouses, and Data Lakes;Work with both SQL and NoSQL databases to manage and process data;Collaborate with cloud platforms, particularly Google Cloud services such as BigQuery, Cloud Storage, and Cloud Functions;Leverage DataOps tools to improve ETL processes and enable Self-Service Analytics;Ensure compliance with the General Data Protection Regulation (GDPR);Exhibit strong communication, problem-solving, and teamwork capabilities.What we are looking for:A bachelors or postgraduate degree in Computer Science, Computer Engineering, Statistics, Mathematics, or a related discipline;Proficiency in oral and written communication in English;Expertise in Geospatial data analysis and management;Strong command of SQL for effective data manipulation and transformation;Deep understanding of dimensional modeling concepts;Hands-on experience in feature engineering and managing relational and key-value databases;Proven ability to build data pipelines and utilize APIs for data extraction;Practical experience with ETL/ELT tools like Pentaho, Streamsets, or similar technologies;Proficiency in Python and/or Java for developing scalable data engineering solutions;Familiarity with version control systems, especially Git;Comprehensive understanding of Data Platforms, Data Warehouses, and Data Lake architectures;Experience working with both SQL and NoSQL databases;Experience with Google Cloud;Strong knowledge of DataOps tools to enhance ETL processes and support Self-Service Analytics;Awareness of the General Data Protection Law (LGPD);Excellent communication, problem-solving, and collaborative teamwork skills.Where you will work:This is a 100% remote position.Employment type:CLT;Cooperado;PJ.Our benefits include:100% remote work; Meal and/or food allowance in a flexible model (EVA card)*;Unimed health insurance for employees and dependents*;Uniodonto dental insurance for employees and dependents*;Agreements with Educational Institutions for discounts on Undergraduate, Postgraduate, and short courses;Totalpass to take care of physical health;Zenklub to take care of mental health;Life insurance*;Daycare assistance for zallpers with children aged 4 months to 6 years, who earn up to three times the minimum wage of the category*;Baby Zallpy: a gift to celebrate the birth of zallpers babies;Communities: we support the operation of three voluntary zallpers communities: Diversity, Equity & Inclusion, Sports & Movement, and Technology.Benefits valid for CLT type *
#J-18808-Ljbffr