A
85

Applied AI Researcher / Founding Engineer

9y relevant experience

Qualified
For hiring agencies & HR teams

EU engineers, ready to place with your US clients

Pre-screened on AI. Remote B2B contracts. View 5 full profiles free — AI score, skills report, interview questions included.

Executive Summary

The candidate is a strong Applied AI Researcher candidate with a PhD in Computer Science, deep ML engineering expertise, and a proven track record of owning technical strategy at Principal level within a major enterprise (Salesforce/Informatica). Their mathematical rigor, LLM and Transformer expertise, publication record, and team leadership experience make them a compelling match for the founding engineer profile. The primary concern is a gap in generative image/text model experience central to AlpacaRelay's product, though their adjacent NLP and LLM systems background suggests they could ramp quickly. The absence of public code artifacts warrants a structured technical assessment during interviews. Overall, the candidate represents a high-quality candidate whose research depth, leadership trajectory, and intellectual profile align closely with the role's ambitions.

Top Strengths

  • PhD-level mathematical and ML research expertise with peer-reviewed publications at VLDB and NeurIPS — directly satisfying the strongest academic requirement of the role
  • Demonstrated ability to own and execute full ML technical strategy at Principal level, including building and leading R&D teams from scratch — closely mirrors founding engineer responsibilities
  • Deep practical expertise in LLMs, Transformer architectures, RAG systems, and NLP-driven ML pipelines with real production deployment experience
  • Long-tenured progression at a single organization through acquisition (Compact Solutions → Informatica → Salesforce) demonstrating resilience, adaptability, and compounding impact in evolving environments
  • Strong mathematical foundations (dual undergraduate degrees and a master's in financial mathematics) providing rigorous analytical thinking for novel research challenges

Key Concerns

  • !No explicit experience with text/image generative models or vision-language multimodal systems, which appear central to AlpacaRelay's stated product direction — this is a meaningful technical gap for the specific role
  • !Absence of public code contributions (no GitHub, no open-source projects) makes it difficult to independently verify code quality, engineering style, and collaborative development practices expected of a founding engineer

Culture Fit

78%

Growth Potential

High

Salary Estimate

$100,000 - $135,000

Assessment Reasoning

The candidate is rated FIT (score: 85) based on the following rationale: They satisfies the most demanding requirements of the role — a PhD in a relevant field, 9+ years of ML research and engineering experience, hands-on LLM and Transformer expertise, production ML system ownership, and demonstrated team leadership including building an R&D team from scratch. Their publications at VLDB and NeurIPS fulfill the 'proven track record of delivering AI systems' criterion. They exceeds the 3-7 year experience range, bringing more senior depth. The two notable gaps — limited generative/multimodal model experience and no public code artifacts — are real but not disqualifying: their adjacent technical depth (LLMs, NLP, deep learning architectures) provides a credible fast-learning foundation for generative AI work, and code quality concerns can be assessed through a structured technical interview. Their intellectual profile (competitive chess champion, mathematical olympiad finalist, academic reviewer) suggests the curiosity and high-performance mindset the role explicitly values. The overall skills match exceeds 80% of stated requirements, no red flags are present, and their career trajectory is directly aligned with the founding engineer archetype AlpacaRelay is seeking.

Interview Focus Areas

Generative AI and multimodal model experience — probe depth of knowledge in diffusion models, image generation, and vision-language architectures; assess learning trajectory and proximity to AlpacaRelay's product domainFounding engineer mindset and startup adaptability — explore how they have operated under ambiguity, driven product decisions, and balanced research rigor with shipping speed in resource-constrained environmentsCode quality and engineering practices — request a live coding exercise or code review session to compensate for absent GitHub/portfolio artifactsLeadership style and team-building philosophy — assess readiness to grow into a C-level technical leader given their stated interest and prior team-building experience

Code Review

FairSenior Level

No code example or GitHub profile was provided, preventing direct evaluation of code quality. Based on the resume narrative — including production ML systems at Salesforce, doctoral research, and described system architecture work — the candidate is estimated to be at Senior to Principal engineering level. The lack of public code artifacts is a meaningful gap for this founding engineer position and should be addressed during the interview process.

PythonPyTorchKerasHugging Face Transformersscikit-learnNumPypandasJavaSQL
  • +Resume describes clean, modular system design with emphasis on scalability and reliability, suggesting strong engineering discipline
  • +Production ML systems experience at Informatica/Salesforce implies familiarity with code quality standards in enterprise-grade environments
  • -No code samples, GitHub profile, or open-source contributions were provided, making direct code quality assessment impossible
  • -Absence of public code artifacts is a notable gap for a founding engineer role where hands-on technical credibility is paramount

Experience Overview

11y total · 9y relevant

The candidate is a highly credentialed Applied ML Researcher with a PhD in Computer Science, strong mathematical foundations, and ~9 years of directly relevant ML engineering and research experience. Their career progression from Senior Engineer to Principal-level at Informatica/Salesforce demonstrates both technical depth and leadership maturity. Their primary gap relative to this specific role is limited explicit experience with generative image/text models and multimodal architectures, though their deep NLP, Transformer, and LLM expertise provides a strong adjacent foundation.

Matching Skills

PhD in Computer Science (Jagiellonian University)Strong mathematics background (B.Sc. Theoretical Mathematics, M.Sc. Financial Mathematics)Python with ML libraries (NumPy, pandas, scikit-learn, PyTorch, Keras, Hugging Face)LLMs and Transformer-based architectures (BERT, fine-tuning, RAG)Deep learning (feed-forward, CNN, RNN, GNN, attention mechanisms)Model lifecycle management (training, fine-tuning, deployment, monitoring)ML system architecture and end-to-end ML pipeline designCloud-based ML pipelinesPeer-reviewed publications (VLDB, NeurIPS workshop, IEEE)Team building and leadership (built R&D team from scratch at Compact Solutions)Technical strategy ownership and research roadmap definitionMultimodal-adjacent experience (text, structured data, metadata)LLM systems: RAG, tool integration, prompt design, human-in-the-loopResearch and prototyping in ambiguous environmentsMLOps and data governance pipeline familiarity

Skills to Verify

Explicit hands-on experience with image/vision generation models (text-to-image, diffusion models)No GitHub profile or open-source contributions providedNo explicit AWS/GCP/Azure certifications or detailed cloud infrastructure experience listedNo explicit multimodal model experience (vision-language models)No cover letter or code samples provided
Candidate information is anonymized. Personal details are hidden for fair evaluation.