datatrota
Signup Login
Home Jobs Blog

Senior Data Engineer at Turbham Limited

Turbham LimitedNigeria Data and Artificial Intelligence
Contract
We run a Business Management System. We are also branched out and well versed in the Telemedicine, Logistics, Agriculture as well as the Travel and Tourism industries.

Engagement Type: Fixed-Price Contract (3–4 Weeks + 1 Week Post-Launch Support)

Overview

We are seeking an exceptionally strong Data Engineering Specialist to build a production-grade operational monitoring and analytics layer on top of our existing Google Cloud data stack.

This is not a discovery-heavy engagement.
The project is fully specified and documented. All alert logic, thresholds, dashboard requirements, schemas, and technical expectations are defined.

We need someone who can:

  • Read detailed specifications
  • Translate them into clean, production-ready data workflows
  • Implement advanced SQL logic in BigQuery
  • Build robust Dagster jobs
  • Create operational dashboards that drive decision-making
  • Deliver within a tight timeline

This role is execution-focused and suitable only for engineers with deep experience in data systems and production orchestration.

What You Will Build

You will deliver two production systems on top of our existing BigQuery warehouse.

Part A — Automated Operational Alert System

An automated monitoring system that continuously evaluates the health of our order pipeline and triggers notifications when predefined conditions occur.

Core Responsibilities

You will:

  • Implement 11 automated alert rules as scheduled Dagster jobs
  • Write advanced SQL queries against BigQuery (window functions, CTEs, rolling baselines, joins across multiple datasets)
  • Implement anomaly detection logic using rolling comparisons
  • Integrate Slack API for structured, formatted alerts
  • Integrate SMS notifications (Twilio or similar) for critical-level alerts
  • Implement structured logging and audit trails
  • Store configurable alert thresholds in BigQuery (no hard-coded values)
  • Ensure system reliability, idempotency, and clean error handling

Alert Categories Include

  • Delayed orders
  • Stuck pipeline stages
  • SLA breaches
  • Partner performance degradation
  • Volume anomalies vs rolling baselines
  • Failure spikes
  • Operational bottlenecks

The alert system must be production-ready, observable, and maintainable.

Part B — Operational Dashboard (Looker Studio)

You will build a 4-page operational dashboard connected directly to BigQuery.

This dashboard will be used daily by the operations team to monitor pipeline health and partner performance.

Dashboard Pages

Real-Time Operations Overview

  • Pipeline status
  • Active orders by stage
  • SLA compliance summary
  • Stuck orders
  • Critical alerts

Partner Performance Tracking

  • Doctors, pharmacies, carriers
  • Trend analysis
  • SLA adherence
  • Throughput metrics

Historical Trends & Anomaly History

  • Rolling baselines
  • Alert frequency trends
  • Performance over time

Detailed SLA Compliance Reporting

  • Breach tracking
  • Time-to-resolution metrics
  • Drill-down capability
  • The dashboard must include:
  • Filters
  • Drill-down capability
  • Trend visualizations
  • Clear operational storytelling

Technology Stack (Already in Place)

You will work within our existing infrastructure.

  • Google BigQuery – Primary data warehouse
  • Dagster – Production orchestration tool
  • Python – For alert logic and orchestration
  • SQL (Advanced) – Complex transformations and aggregations
  • Looker Studio – Dashboard development
  • Slack API – Alert notifications
  • SMS (Twilio or equivalent) – Critical alerts
  • Google Cloud Platform (GCP) – Cloud environment

No infrastructure setup required. You will build within our existing systems.

Mandatory Requirements (Non-Negotiable)

Only apply if you meet ALL of the following:

Technical

  • Strong Python in a data engineering context (not web development)
  • Advanced SQL proficiency:
  • Window functions
  • CTEs
  • Rolling aggregations
  • Complex joins
  • Performance optimization in BigQuery
  • Proven Google BigQuery experience
  • Experience with Dagster OR strong Airflow/Prefect experience (Dagster preferred)
  • Experience building multi-page Looker Studio dashboards connected to BigQuery
  • Slack API or webhook integration experience
  • Experience working within Google Cloud Platform

Professional

  • Ability to read detailed specifications and execute independently
  • Strong production mindset (logging, error handling, idempotency)
  • Clear communication in async environments
  • Structured documentation habits
  • Experience delivering under time-bound milestones
  • If you do not have hands-on BigQuery + orchestration experience, this role is not suitable.

Nice-to-Have (Strong Advantage)

  • Healthcare or e-commerce pipeline experience
  • Shopify data familiarity
  • SMS integrations (Twilio)
  • Experience building operational monitoring systems
  • Experience designing anomaly detection logic
  • Experience with SLA tracking frameworks

Timeline & Milestones

This is a 3-week build with a 1-week support buffer.

Week 1 (Days 1–5)

  • Environment familiarization
  • Core alert rule implementation
  • Slack integration
  • Initial production deployment

Milestone: Core alerts live

Week 2 (Days 6–9)

  • Anomaly detection logic
  • SMS integration
  • Logging framework
  • Configurable thresholds
  • Full system testing

Milestone: Alert system complete

Week 3 (Days 10–13)

  • 4-page Looker Studio dashboard
  • Testing and validation
  • Documentation and handover

Milestone: Dashboard live, project complete

Post-Launch (1 Week)

  • Threshold tuning
  • Minor refinements
  • Operational adjustments

Engagement Structure

  • 1 month Fixed-price contract 
  • Nigeria-based consultant
  • Fully remote
  • Daily async updates via email
  • Full system access provided on Day 1
  • Detailed specification documents shared with shortlisted candidates
  • 1-week post-launch support required

What You Must Submit

To be considered, your application must include:

Your CV

A concise summary of relevant experience:

  • BigQuery
  • Dagster/Airflow/Prefect
  • Alerting systems
  • Looker Studio dashboards
  • Example of a production pipeline you built:
  • Dagster or Airflow preferred
  • Brief explanation of architecture
  • Example of a Looker Studio dashboard:
  • Screenshot or link
  • Brief explanation of data model
  • Your fixed-price quote
  • Confirmation of availability for the 3-week timeline
  • Applications without concrete examples will not be reviewed.

Ideal Candidate Profile

You are:

  • A senior-level data engineer
  • Comfortable working directly in BigQuery
  • Experienced with orchestration tools
  • Detail-oriented and systematic
  • Able to build reliable monitoring systems
  • Strong in operational thinking
  • Comfortable working independently with minimal hand-holding

If you are confident in your ability to execute this project at a high technical standard within the stated timeline, we encourage you to apply.

This is a high-impact build that will directly power our operational decision-making systems.

Method of Application

Signup to view application details. Signup Now

More jobs like this

X

Send this job to a friend