Data Engineer

Meridia Land Indonesia
Apply Now

About MeridiaMeridia is an AgTech company specialised in field data within complex smallholder-heavy agri-commodity supply chains such as cocoa, coffee, palm oil, rubber and soy. Meridia Verify®, its flagship SaaS product, verifies supply chain field data in real time – and validates compliance with frameworks such as the EUDR (EU Deforestation Regulation). With Verify, companies can reduce commercial and reputational risk, minimise supply chain disruption and accelerate evidence-based decision-making for strategic sourcing, procurement and trade. Meridia’s mission is to improve the veracity of datasets driving decision making in global agri-commodity supply chains. Its purpose is to build transparent and inclusive supply chains that afford smallholders a level playing field. See for more information on our work.

Job DescriptionAs a Data Engineer, you will serve as a data and analytics expert by modelling, structuring, and scripting data flows. You will be working on structuring data in an automated way for it to be processed into meaningful outputs. You will be based in Indonesia and will work closely with our teams in The Netherlands, Lithuania and Ghana, as well as with colleagues across Indonesia. You will report to the Data Manager in Indonesia.

ResponsibilitiesThe Data Engineer has the following areas of work and responsibilities:  Data Engineering • Write Python scripts to acquire data and process data from various academic, government, or commercial sources for use in Meridia’s product development • Maintain accurate, high-quality metadata on the company’s internal datasets using scripts, automation, and selected tooling • Develop internal tools and scripts to automate data processing and use technologies like Airflow, AWS, Google Cloud to scale and optimise ETL tasks of geospatial data • Maintain the team’s git repository of ETL scripts and adhere to version control best practices to ensure code is well organized, accessible, and reproducible by the entire team • Optimize data ingestion processes to ensure reliable, timely, and scalable delivery of data into downstream products and analytics workflows, particularly for large-scale raster and vector datasets • Monitor ingestion pipelines with logging, alerting, and metrics to detect failures or data inconsistencies • Contribute to our (internal) data catalogues, manuals, dictionaries, glossary and broader set of terminology in a coherent way across countries, products and projects. • Drive continuous improvements in operational efficiency by identifying bottlenecks and implementing solutions

Core Data Engineering Skills • Expertise in building, optimizing, and maintaining scalable ETL/data pipelines, especially for large and complex geospatial datasets, using modern orchestration and workflow tools such as Apache Airflow. • Proficient in cloud platforms (e.g. AWS), with hands-on experience architecting data storage, transformation, and metadata solutions in a secure and cost-effective manner using Infrastructure as Code tools such as Terraform • Skilled in Python and relevant data engineering frameworks (e.g., Pandas, Pydantic, Rasterio, Tippecano) for automating data ingestion, cleaning, standardisation, and geospatial transformation. • Solid understanding of data governance and best practices in version control (e.g., Git), documentation, data quality assurance, reproducibility, and collaboration in cross-functional data teams.

Learning and DevelopmentThe ideal candidate should be eager and willing to learn and grow in the role. You will be engaged in on-the-job training and development to learn about the land rights sector and our data systems and workflows. You will be mentored by our Data Engineering team in Indonesia and senior colleagues in the Netherlands.