A
72

AI Data Pipeline Engineer

5y relevant experience

Qualified
For hiring agencies & HR teams

EU engineers, ready to place with your US clients

Pre-screened on AI. Remote B2B contracts. View 5 full profiles free — AI score, skills report, interview questions included.

Executive Summary

This candidate is a senior Python and AI engineer with 10 years of experience across AI-native startups in Estonia and the US, bringing credible hands-on work with Apache Airflow, ETL pipelines, PostgreSQL, AWS, and Kubernetes. their profile is strong on ML engineering and full-stack development, with data pipeline work appearing as a meaningful but secondary component of their roles. they meets the majority of the required technical skills and shows genuine alignment with an AI-first engineering culture, operating autonomously across distributed teams. Key gaps are dbt and explicit data quality framework experience, which should be probed in interview. their Estonia-based location aligns with the European collaborative ethos described, and their growth trajectory suggests strong upside for a focused data engineering role.

Top Strengths

  • 10 years of progressive Python engineering experience with senior-level architectural ownership
  • Direct Apache Airflow and ETL pipeline experience in production ML environments
  • Strong AWS cloud infrastructure proficiency matching the technical environment
  • Kubernetes and Docker containerization experience aligns with deployment requirements
  • ML ops adjacency (MLflow, AI inference pipelines, model serving) bridges data engineering and ML collaboration

Key Concerns

  • !Profile is primarily AI/ML and full-stack focused rather than data engineering-first — role depth alignment is uncertain
  • !Missing dbt and explicit data quality framework experience, both listed as required skills

Culture Fit

74%

Growth Potential

High

Salary Estimate

$80k-$100k (senior tenure suggests upper-mid range; Estonia location may compress expectations vs US market)

Assessment Reasoning

Gustavo scores 72/100 and earns a FIT decision based on meeting approximately 75-80% of the required technical skills, with Python, Airflow, PostgreSQL, AWS, Kubernetes, ETL, and PySpark all confirmed. their senior trajectory across AI-focused companies demonstrates the engineering maturity and autonomous working style the role demands. The primary gaps — dbt and explicit data quality frameworks — are learnable for a candidate at their level and should not disqualify him, particularly given their ML ops exposure (MLflow) and pipeline design experience. The main uncertainty is whether their data engineering work is deep enough to own a data infrastructure mandate end-to-end, as their profile reads more AI/ML generalist than data engineering specialist. This candidate should focus on validating pipeline depth, data quality practices, and dbt familiarity before proceeding to offer.

Interview Focus Areas

Deep dive on Apache Airflow pipeline design: DAG architecture, scheduling strategies, backfill handling, and monitoring in productionData quality practices: How has he handled schema drift, data validation, and pipeline observability without explicit framework experience?dbt familiarity and willingness to learn: Any exposure to transformation layer tooling or SQL-first workflows?Volume and scale: What is the largest dataset or highest-throughput pipeline he has personally engineered end-to-end?

Code Review

GoodSenior Level

No GitHub profile was provided, limiting direct code quality assessment. Based on resume evidence, Gustavo demonstrates senior-level architectural and engineering judgment with a broad, modern technology stack. Code quality signals are positive given CI/CD practices and testing references, but independent verification is not possible without a code sample or public repository.

PythonFastAPIDjangoApache AirflowPostgreSQLRedisKafkaRabbitMQPySparkTensorFlowPyTorchLangChainKubernetesDockerAWSMLflowCeleryGoogle BigQueryArgoCDGitHub Actions
  • +Demonstrates version-specific technology literacy (e.g., TensorFlow 2.15, Kubernetes 1.27, Airflow 2.0), suggesting hands-on depth
  • +Shows architectural thinking — microservices migration, distributed caching, auto-scaling policies
  • +References TDD, unit/integration testing, CI/CD with GitHub Actions and ArgoCD
  • -No GitHub profile provided to verify actual code quality, style, or open-source contributions
  • -Resume is technology-dense but lacks quantified impact metrics (e.g., pipeline throughput, latency improvements, data volume handled)
  • -Some technology listings feel enumeration-heavy, making it hard to distinguish depth from surface familiarity

Experience Overview

10y total · 5y relevant

This candidate is a versatile senior engineer with strong Python, AWS, Kubernetes, and pipeline experience spanning 10 years. their ETL/Airflow work at Textback.ai and ML infrastructure contributions are directly relevant, though their profile skews more toward AI/ML engineering and full-stack than pure data pipeline engineering. Key gaps include dbt and explicit data quality framework experience, which are listed requirements for this role.

Matching Skills

PythonApache AirflowPostgreSQLAWS (Lambda, EC2, S3, RDS)ETL/ELT PipelinesKubernetesPySparkDockerKafkaData Quality (implicit)CI/CDMLflow

Skills to Verify

dbtGreat Expectations or similar data validation frameworksSnowflakePrefect or DagsterExplicit data quality framework experience
Candidate information is anonymized. Personal details are hidden for fair evaluation.