Logo for G-P
Staff Data Engineer
G-P
Posted over a month ago
Description

Job Summary:

We are seeking a highly experienced Staff Data Engineer to lead the design, development, and optimisation of our data architecture, pipelines, and workflows. This role will serve as a technical lead within the organisation, setting best practices, mentoring team members, and solving complex data challenges to enable data-driven decision-making at scale.

As a Staff Data Engineer, you will collaborate with cross-functional teams, including data scientists, analysts, and software engineers, to design systems that transform raw data into actionable insights while ensuring scalability, security, and reliability.

Key Responsibilities:

Technical Leadership

  • Design, and implement scalable and reliable data pipelines, ensuring the processing of large volumes of structured and unstructured data.
  • Define and enforce data engineering best practices, coding standards, and architectural principles across teams.
  • Conduct code reviews and provide mentorship to junior and senior data engineers.

Data Pipeline Development

  • Build and maintain batch and real-time data pipelines using tools such as Apache Spark, Kinesis, and AWS services.
  • Works with multiple teams to coordinate the event-driven architecture, managing inter-dependencies and promoting consistency.
  • Ensure data quality, governance, and security by implementing monitoring, validation, and compliance tools.

Collaboration & Cross-Functional Engagement

  • Partner with product, analytics, and data science teams to understand business requirements and translate them into technical solutions.
  • Work closely with DevOps and software engineering teams to deploy and maintain production-ready data infrastructure.

Innovation & Scalability

  • Evaluate and recommend emerging technologies and frameworks to ensure the data platform remains future-proof.
  • Drive initiatives to improve the performance, scalability, and efficiency of existing systems.

 

Required Skills & Experience

  • 12+ years of experience in data engineering field, with at least 2 years in a senior or staff-level role.
  • Expertise in designing and implementing scalable data architectures for big data platforms.
  • Strong programming skills in Python, Scala.
  • Deep experience with distributed data processing systems such as Apache Spark, Databricks, Delta Lake.
  • Proficiency with relational databases (e.g., PostgreSQL, MySQL) and NoSQL databases (Dynamo).
  • Strong understanding of ETL/ELT workflows, data warehousing concepts, and modern data lake architectures.
  • Employ the established Data Governance model to sustain Data Quality for the data objects and implement the necessary operating mechanisms to ensure compliance
  • Knowledge of CI/CD practices.
  • Excellent problem-solving skills and the ability to design creative, efficient solutions for complex data challenges.
  • Background in AI, machine learning pipelines is a plus
  • Proactive, self-driven, and detail-oriented with a strong sense of ownership.

More Similar Roles...

Want more remote roles like this one sent to you?