Job Title: GCP FinOps Engineer
Location: Newport, UK (Hybrid)
Contract Duration: 6 Months
IR35 Status: Inside IR35
Role Overview
We are seeking an experienced GCP FinOps Engineer to optimise cloud spend, performance, and operational efficiency across large-scale data analytics and containerised workloads. You will work closely with engineering, data, and product teams to embed cost-efficient architectures, enforce financial governance, and implement best practices across Google Cloud environments.
Responsibilities
- Optimise large-scale data analytics workloads via partitioning, clustering, query rewrites, storage format improvements, and lifecycle policies.
- Tune containerised microservices by recalibrating CPU/memory requests, improving autoscaling efficiency, and restructuring workload placement on cost-efficient compute.
- Redesign workflow orchestration pipelines for parallel execution, higher concurrency, and offloading heavy tasks to lower-cost execution environments.
- Analyse distributed data-processing pipelines to right-size worker types, adjust scaling thresholds, and adopt low-cost compute for batch workloads.
- Reduce log-processing and storage overhead through log-level standardisation, routing rules, exclusion filters, and retention optimisation.
- Implement storage-tiering strategies based on access patterns and enforce lifecycle rules to minimise cold data retention costs.
- Improve relational database performance through index tuning, connection optimisation, and instance right-sizing.
- Enhance horizontally scalable database performance via autoscaling policies, index improvements, and mitigation of read/write hotspots.
- Build dashboards, budgets, alerts, and guardrails to drive ongoing cost governance and financial accountability.
- Collaborate with engineering teams to embed cost-efficient architecture patterns and operational best practices.
Key Skills / Knowledge
- 5+ years of hands-on experience in Google Cloud.
- Strong understanding of GCP data services, including indexing, slots, pruning, partitioning, and clustering.
- Expert-level Kubernetes & GKE resource tuning.
- Hands-on experience with Dataflow pipelines and worker optimisation.
- Strong Airflow/Composer knowledge (DAG design, scheduling, PodOperator).
- Deep understanding of Cloud Logging, routing, sinks, and exclusion filters.
- Experience with Cloud Spanner autoscaling, indexing, and schema optimisation.
- Cloud SQL performance tuning and indexing.
- Ability to analyse billing data, resource consumption, and quantify cost savings.
- Experience using GCP Cost Explorer, Recommender API, and Billing Export.
- Build dashboards, alerts, and budget guardrails for cost governance.
- Excellent communication and stakeholder management skills.
- Ability to collaborate across engineering, data, and product teams.
- Structured problem-solving mindset; ownership-driven, proactive, and independent.