DeepKeep Research
DeepKeep’s research focuses on understanding how modern AI systems fail in real world environments, and how to secure them before those failures turn into risk. Our research explores large language models, computer vision and multimodal AI, with hands on analysis of emerging attack techniques and practical defense strategies. Together, they reflect an approach to AI security that values innovation, transparency, and deployable protection over theory alone.
Academic research
Title
Authors
Publication
Adi Shnaidman, Erin Feiglin, Osher Yaari, Efrat Mentel, Amit Levi, Raz Lapid
ReALM-GEN @ ICLR 2026
Erin Feiglin, Nir Hutnik, Raz Lapid
TMLR 2026
Amit Levi, Raz Lapid, Rom Himelstein, Yaniv Nemcovsky, Ravid Shwartz Ziv, Avi Mendelson
arXiv
Raz Lapid, Almog Dubin
RDS @ AAAI 2026
Roie Kazoom, Raz Lapid, Moshe Sipper, Ofer Hadar
VLM4RWD @ NeurIPS 2025
Eylon Mizrahi, Raz Lapid, Moshe Sipper
SaFeMM-AI @ ICCV 2025 (Spotlight)
Tal Alter, Raz Lapid, Moshe Sipper
TMLR 2025
Raz Lapid, Ron Langberg, Moshe Sipper
SeT LLM @ ICLR 2024
Raz Lapid, Almog Dubin, Moshe Sipper
Mathematics @ MDPI 2024
Ben Pinhasov, Raz Lapid, Rony Ohayon, Moshe Sipper, Yehudit Aperstein
TMLR 2024
Raz Lapid, Eylon Mizrahi, Moshe Sipper
MLCS @ ECML-PKDD 2024
Raz Lapid, Moshe Sipper
MLCS @ ECML-PKDD 2023
Snir Vitrack Tamam, Raz Lapid, Moshe Sipper
TMLR 2023
Raz Lapid, Zvika Haramaty, Moshe Sipper
Algorithms @ MDPI 2022
Roee Ziv, Raz Lapid, Moshe Sipper
arXiv