Job Opportunities in the Daphni Portfolio


Data Engineer - Data Platform



Data Science
Spain · Remote
Posted on Tuesday, July 2, 2024
At OLX, we work together to build a more sustainable world through trade.
We make it safe, smart, and convenient to buy and sell cars, find housing, get jobs, buy and sell household goods, and more. Our colleagues around the world help to serve millions of people around the world every month, through its well-loved consumer brands including OLX, Otomoto, Property24.
Join OLX and see how far you can go.
About the Data Engineering team:
Become a part of an agile team of data specialists taking care of the massive amounts of data and groundbreaking technologies. Your work will manage and grow our current data platform used by hundreds of data engineers, data / business analysts and data scientists all over Europe.
While being part of a central team of data engineers and software engineers across Lisbon, Berlin, Barcelona and Poznan, you will grow our data platform in several aspects such as automating data acquisition processes, democratising data access even for non-tech roles, building a streaming architecture or even working on data taxonomy to allow better data platform and product integration.
You will lead a business domain from the technical perspective and contribute to the data architecture and platform definition while also mentoring more junior team members.
You will also be involved in strategic goals such as shifting to the paradigm of looking at data as a product, not just as an asset, and bring the platform to the next level by enabling data streaming allowing for faster data driven decisions.

What you will do:

  • Reporting to the Data Platform Engineering Manager, within the Data Platform and Services team
  • Collaborate with analytics, data science, and data product management to support our shared Data mission: "We harness data to drive intelligent decisions and deliver products our customers love."
  • Contribute to strategic goals and initiatives, taking responsibility for specific results that support the technical and product roadmap.
  • Assist in architecting and developing scalable data platform solutions and architectures that support various domains utilizing the data platform, addressing stakeholders' needs.
  • Support and collaborate with domain-oriented data teams, ensuring they have the necessary toolkits and processes to operate effectively within the larger Data Mesh framework.
  • Participate in identifying and suggesting improvements based on error fixing, debugging, bug fixes, and performance tuning.
  • Assist in researching, recommending, testing, and adopting new (big) data technologies and techniques to enhance our data ecosystem.
  • Support monitoring of our data ecosystem, identifying inefficiencies, and assisting in automation or fine-tuning of data products to optimize their performance and usability.
  • Contribute to enhancing and enforcing governance frameworks and processes to ensure the security and compliance of our data ecosystem.
  • Provide support and guidance to less experienced staff members, fostering a collaborative and learning-focused environment.

Who we are looking for:

  • 3-5 years of experience creating and maintaining data platforms, implementing complex data pipelines, and working with both technical and business stakeholders.
  • Demonstrated experience in designing governed and scalable data architectures and platforms, with a focus on storage and performance optimisation.
  • Experience with at least one major cloud data technology (AWS, GCP, or Azure Cloud), with practical experience in architecting solutions within the AWS ecosystem.
  • Proficient in Python and SQL, capable of writing efficient, easy-to-read, and clear code, with some experience in Spark and/or Scala.
  • Experience working with Redshift, Presto/Trino, Hive/Glue, Kafka and Athena.
  • Experience working with AWS services, including S3, DynamoDB, SQS, EC2 Auto Scaling, Kinesis, Lambda, EMR, and general AWS APIs.
  • Hands-on experience with CI/CD tools (e.g., GitLab), infrastructure as code (Terraform), containerisation (Docker), and container orchestration (Kubernetes).
  • Relevant knowledge of collaborative analytics and orchestration tools (e.g., Airflow, dbt) and open table formats (e.g., Apache Iceberg, Apache Hudi, Delta Lake).
  • Relevant knowledge of metadata concepts and tools (e.g. DataHub, OpenMetadata, Marquez).
  • Knowledge of the Data Mesh paradigm.
  • Excellent written and verbal proficiency in English.
  • This role is based in Spain

Our Benefits:

  • Work from home allowance
  • Health & dental insurance
  • Flexible retribution plan (up to 30% of your salary in products/services such as restaurant ticket, transportation ticket, nursery check)
  • Gym membership Language courses (English + Spanish)
  • Employee Assistance program - service managed by ComPsych which provides confidential support, resources and information for personal and work-life matters
  • Tax declaration assistance - fiscal advisory