About the role
As a Data Engineer Specialist, you will be a subject matter expert responsible for developing
and maintaining data pipelines to assist us aggregate number of client data sources into a
single view for our teams to utilize. The data is often incomplete, inconsistent and stored in silos
with access primarily through APIs and FTPs. We are looking for a creative, experienced and
motivated individual to build solutions for integrating CRM and Marketing data in GCP &
BigQuery environments.
Responsibilities
● Build and maintain custom ingestion and integration pipelines from multiple sources
(Digital Marketing and CRM data) - 70% (of time)
● Establish monitoring systems to give visibility to pipelines’ status and monitor automated
jobs. - 10%
● Bringing consistency to our Data handling and schemas across the board. - 5%
● Developing quality standards to ensure data quality and integrity across all database
systems. - 5%
● Partner with analysts and data scientists to improve the efficiency and performance of
their data products. - 5%
● Where appropriate, train other team members to support the development of products,
tools and technologies. - 5%
Experience
Required
● 5+ years’ experience developing data processing and storage solutions on a cloud
architecture; prior knowledge of GCP ecosystem is a major advantage.
● Experience dealing with Digital marketing and CRM data.
● Experience in building pipelines from various data sources especially APIs.
© 2013-2023 SixBerries Ltd Confidential Page 3
Preferred
● Experience with Python development.
● Experience with using version control via Git and CI/CD systems.
● Demonstrated success as a developer in the full scope of the development lifecycle;
including, but not limited to, understanding business needs, design, development, and
deployment.
Requirements and skills
Required
● Excellent communication and teamwork skills
● A willingness to continually learn and adapt
● Organisational skills
● An analytical mind
● Digital marketing and CRM data pipeline knowledge
● Understands efficient data warehousing designs and schemas.
● Comfortable with ETL tools such as Alteryx, Parabola or Talend
● Highly skilled at writing advanced SQL queries and comfortable with merging complex
datasets into functional presentation layers for model inputs.
● Comfortable with working with orchestration tools such as Airflow, Dataproc, dbt or
Cloud Composer for ETL and ELT.
Preferred
● Knowledgeable about data infrastructure management and Unix/Linux operating
systems.
● Skilled at deploying containerization solutions such as Docker or Kubernetes.