Description
We are seeking a Data Engineer to help in our mission to unlock the value of renewable energy data. We are challenging the “status-quo” on how renewable energy companies are getting value from their petabytes of data.
We are creating a data foundation platform by syncing and harmonizing data from wind and solar assets across the world. Data is our foundation for the development of machine learning models, process optimization and operational excellence.
The main responsibility will be design and develop new data engineering pipelines using the best tools and frameworks by ensuring data availability and consistency.
You’ll be expected to:
- Improve data quality by using & improving internal tools to automatically detect issues.
- Design our data models for optimal storage and retrieval, and to meet critical analysis and business requirements.
- Write high-quality code.
- Build reusable code and libraries for future use.
- Proactive presentation of your own proposals regarding solutions and technologies for the development process.
- Assure knowledge transfer and close collaboration with your colleagues to help everybody to improve.
- Contribute as a valued team member to testing, debugging, QA, and documentation of data pipelines and systems.
You must have:
- A bachelor's and/or master’s degree Computer Science area or equivalent.
- Strong analytical and problem-solving skills.
- Data Engineering working experience, with Python.
- Willingness to learn new technologies.
- Experience with data warehousing, infrastructure, ETL/ ELT.
- Experience with big data technology stack, for example, Pandas, Dask, (Py-)Spark, Delta Tables, etc.
- Experience with multiple Big Data file formats (Parquet, Avro, Delta Lake).
- 3+ years of experience with general-purpose languages such as Python, Java, Scala etc.
- Fluency in SQL and relational DB systems (PostgreSQL, MySQL, SQL Server).
- Build scalable data pipelines leveraging Prefect/Airflow & DBT framework.
- Experience setting up MLOps processes and a machine learning model registry.
Ideally, you’ll also have:
- Flexibility and be able to adapt to dynamic priorities.
- Test-driven development, or experience with testing methodologies.
- Experience using Git.
- Experience with Docker.
- Experience with Linux.
- Familiarity with Microservices.
- Understand infrastructure and data security requirements, as well as platform hardening principles.
- Understanding of data stream platforms, message brokers and queues.
- Familiarity with Jenkins or other CD tools.
- Motivation to work in a fast paced, team-oriented and collaborative environment.
- Excellent written and verbal communication skills (Portuguese & English).
Benefits
• Health Insurance
• Flexible working environment
• Work tools such as laptop and peripheral devices
• Extra days off
• Training and education opportunities
• Flexible benefits program
• Free Coffee :)
We believe great work is made by great people collaborating in a great environment. If you are interested and your profile matches the one we are looking for, don’t miss this opportunity and apply by sending your CV to [email protected]