Who we’re looking for: Someone who understands the methodology and fundamentals rather than specific tools. You’ll be ideal for this role if you have a demonstrable passion for automation and collaboration.
The challenge: You will be responsible for the implementation of database clusters, tables, schemas, indexes, and ETL pipelines. This also includes database and query performance tuning.
Day-to-day responsibilities:
Building and maintaining database schemas and tables
Building and maintaining ETL pipelines
Query and index performance tuning
Database backup/recovery, monitoring and alerting
Working in an agile way following DevOps principles
Helping developers with database-related queries
Advising data and software engineers on best practices
Who we’re looking for:
Experience with relational databases (MySQL, PostgreSQL, SQL Server, Oracle, or similar)
Exposure to NoSQL data stores (Elasticsearch, Redis, DynamoDB, Cassandra, Presto, MongoDB, or similar)
Experience in working with ETL tools (preferably dbt)
Working knowledge of Git or other source version control systems
Good knowledge of SQL
Knowledge of Python and development best practices
Experience with a Cloud Provider such as AWS, GCP, or Azure
Knowledge of Linux and Bash
Strong sense of collaboration and experience working with development teams
Desirable but not essential:
Experience with building and deploying containers in Docker
Experience deploying and operating applications on Kubernetes
Knowledge of Python-based data science tools (Pandas, NumPy, Matplotlib)
Exposure to MPP and columnar relational databases (Redshift, Snowflake, BigQuery, Vertica, Greenplum, or similar) or data lake platforms such as Databricks
Exposure to business intelligence and analytics tools such as Tableau, Power BI, Looker, or QuickSight
Experience of instruction and mentoring for the training of peers
Knowledge of CI/CD processes