At HAYS, a British multinational company that offers recruitment and human resources services in 33 countries worldwide and is listed on the London Stock Exchange. We are currently looking for 2 Cloud Data Engineers to collaborate remotely with some international clients with headquarters in Málaga and Barcelona, Spain.
What are the requirements for these 2 positions?
1.Azure Data Engineer :
- 4+ years in data engineering
- Experience with Azure.
- Experience with Pyspark and Databricks working in an Azure environment.
- Python programming experience
- Strong verbal and written skills in English.
Nice to have :
Experience working with Snowflake.Preferably Azure / Databricks certified Engineer .2.Google Cloud Engineer :
4+ years of experience with Google cloud platforms (GCP ).Python and Java programming experience and in design, programming, unit-testing ( Pytest, JUnit ).Experience using Terraform.Experience using version control (GIT), Kafka or PubSub, and Docker.Experience using Apache Beam with Dataflow.Knowledge of Clean code principlesStrong verbal and written skills in English.Nice to have :
Experience working with Snowflake.Preferably GCP certified Engineer .Which are the main tasks?
1.Azure Data Engineer :
As part of the team, you are responsible to maintain and extend their customer data platform based on different GCP services (e.g. Dataflow, Cloud Functions).Develop and maintain scalable data pipelines using GCP.The platform development is based on Python and Terraform.Furthermore you will work with SQL related technologies like Google BigQuery or Snowflake and dimensional data models , to support advanced analytics and reporting capabilities .Design, programming, unit-testing ( Pytest ), quality assurance and documentationWork closely together in an agile team with members located in other countries (team language will be English).2.Google Cloud Engineer :
Develop and maintain scalable data pipelines using Microsoft Azure services , leveraging PySpark, Python, and Databricks .Build and deploy CI / CD pipelines in Azure DevOps , automating data workflows and ensuring seamless integration and delivery of data solutions.Design and implement data models using industry-standard tools such as ER / Studio, ERwin, Oracle Data Modeling, or Toad Data Modeling .Work closely together in an agile team with members located in other countries (team language will be English).What can we offer you?
Remote mode. However, you must be based in Spain.Offices Location : Barcelona / Málaga.Integration in an international project of a solid company with prospects for the future and with continuity.We are waiting for profiles like yours, passionate about technology and who want to face a new challenge. If this is your case, sign up for the offer with your CV so we can tell you more!