Kaan Sancak

Senior Research Scientist at Meta Meta

I work on machine learning systems, recommendation, and AI infrastructure. My background is in high-performance computing and machine learning, and I am interested in the systems and algorithms that make modern ML efficient, scalable, and useful in practice.

Before Meta, I received my Ph.D. from Georgia Tech , advised by Umit V. Catalyurek, where I worked on high-performance computing, GPU systems, parallel algorithms, and scalable machine learning.

🔬 Research Interests

High-Performance Computing Machine Learning ML Systems & AI Infrastructure Recommendation Systems Parallel & Distributed Computing GPU Computing

💼 Experience

Meta Full-time · Aug 2024 – present
Senior Research Scientist Meta Recommendation Systems (MRS) Feb 2026 – present

Building next-generation recommendation models and ML infrastructure powering Meta's core ranking and personalization systems at scale.

Research Scientist Ads Ranking & Foundational AI (RAI) Aug 2024 – Feb 2026

Model–infrastructure co-design for billion-scale ad recommendation. Built real-time graph integration improving data freshness from days to minutes, boosted training throughput by 20%, and cut feature storage cost by 3x. Lead contributor to the Ads Graph Foundational Model (GFM).

Meta Internship · 2023
Research Scientist Intern Ranking & Foundational AI May – Aug 2023

Conducted research on scalable graph-based models for efficient learning without sacrificing quality. Work published at AAAI 2025 and ICLR 2024.

Meta Internship & Part-time · 2022
Part-time Student Researcher AI Systems HW/SW Co-Design Aug – Dec 2022

Extended caching mechanisms for Meta's ML training platform, substantially reducing redundant data serving computations for key models.

Research Scientist Intern AI Systems HW/SW Co-Design May – Aug 2022

Built caching infrastructure for Meta's data ingestion pipelines, eliminating redundant computation across large-scale model training runs.

Pacific Northwest National Laboratory Internship · 2021
Research Intern HPC & Systems May – Aug 2021

Distributed graph algorithms and high-performance data structures on the SHAD framework.

Facebook Internship · 2020
Research Intern AI Systems HW/SW Co-Design May – Aug 2020

Improved Facebook's graph engine performance via novel partitioning — 10% query throughput gain, up to 5x end-to-end speedup. Integrated the engine with Instagram Ads; infrastructure still serves Instagram, Threads, and Facebook.

Google Summer of Code / NRNB 2018
Open Source Developer May – Aug 2018

Built collaborative pathway editing tools for cBioPortal for Cancer Genomics (Memorial Sloan Kettering Cancer Center).

IBM Internship · 2017
Software Engineering Intern Jun – Aug 2017

Cloud data transfer and object recognition apps (IBM Cloud, Python, Kafka).

🎓 Education

Ph.D. in Computer Science 2019 – 2024

Georgia Institute of Technology

GPA: 4.00/4.00 · Advisor: Umit V. Catalyurek

Research Scalable graph learning, multi-GPU GNN training, dynamic graph algorithms, parallel partitioning
Teaching CSE 6230: High-Performance Parallel Computing · CSE 6220: Intro to HPC
B.Sc. in Computer Engineering 2015 – 2019

Bilkent University, Turkey

Summa Cum Laude · GPA: 3.82/4.00 · Ranked 4th / 231 engineering students

📖 Selected Publications

Haystack Engineering: Context Engineering for Heterogeneous and Agentic Long-Context Evaluation

M. Li, D. Fu, L. Wang, S. Zhang, H. Zeng, K. Sancak, R. Qiu, H. P. Wang, X. He, X. Bresson, Y. Xia, C. Sun, P. Li

arXiv 2025 Paper

Haystack Engineering: Context Engineering Meets the Long-Context Challenge in LLMs

M. Li, D. Fu, L. Wang, S. Zhang, H. Zeng, K. Sancak, R. Qiu, H. P. Wang, X. He, X. Bresson, Y. Xia, C. Sun, P. Li

NeurIPS 2025 Workshop Paper

A Fast and Effective Alternative to Graph Transformers

K. Sancak, Z. Hua, J. Fang, Y. Xie, B. Long, A. Malevich, M. F. Balin, U. V. Catalyurek

AAAI 2025 Paper arXiv

VCR-Graphormer: A Mini-batch Graph Transformer via Virtual Connections

D. Fu, Z. Hua, Y. Xie, J. Fang, S. Zhang, K. Sancak, H. Wu, A. Malevich, J. He, B. Long

ICLR 2024 Paper

MG-GCN: Scalable Multi-GPU GCN Training Framework

K. Sancak*, M. F. Balin*, U. V. Catalyurek

ICPP 2022 Paper

On Symmetric Rectilinear Matrix Partitioning

A. Yasar, M. F. Balin, X. An, K. Sancak, U. V. Catalyurek

JEA 2022 Paper

Elga: Elastic and Scalable Dynamic Graph Analysis

K. Gabert, K. Sancak, M. Y. Ozkaya, A. Pinar, U. V. Catalyurek

SC 2021 Paper

Full list on Google Scholar →

✍ Reviewing

ICML 2025 · ICLR 2025 · NeurIPS 2024 · KDD 2024 · WSDM 2024 · LoG 2024