Publications

Filter by

RESET


MPI PAPERS

MPI Papers


BIBTEX TYPES

Conference Paper

Technical Report


YEAR

2024


DEPARTMENTS

Emperical Interference

Haptic Intelligence

Modern Magnetic Systems

Perceiving Systems

Physical Intelligence

Robotic Materials

Social Foundations of Computation


Research Groups

Autonomous Vision

Autonomous Learning

Bioinspired Autonomous Miniature Robots

Dynamic Locomotion

Embodied Vision

Human Aspects of Machine Learning

Intelligent Control Systems

Learning and Dynamical Systems

Locomotion in Biorobotic and Somatic Systems

Micro, Nano, and Molecular Systems

Movement Generation and Control

Neural Capture and Synthesis

Physics for Inference and Optimization

Organizational Leadership and Diversity

Probabilistic Learning Group


Topics

Robot Learning

Conference Paper

2022

Autonomous Learning

Robotics

AI

Career

Award


Safety- and Efficiency- aligned Learning Conference Paper Efficiently Dispatching Flash Attention For Partially Filled Attention Masks Sharma, A., Geiping, J. In ENSLP NeurIPS Workshop 2024, ENSLP NeurIPS Workshop 2024, ENSLP NeurIPS Workshop, December 2024 (Published)
Transformers are widely used across various applications, many of which yield sparse or partially filled attention matrices. Examples include attention masks designed to reduce the quadratic complexity of attention, sequence packing techniques, and recent innovations like tree masking for fast validation in MEDUSA. Despite the inherent sparsity in these matrices, the state-of-the-art algorithm Flash Attention still processes them with quadratic complexity as though they were dense. In this paper, we introduce Binary Block Masking, a highly efficient modification that enhances Flash Attention by making it mask-aware. We further propose two optimizations: one tailored for masks with contiguous non-zero patterns and another for extremely sparse masks. Our experiments on attention masks derived from real-world scenarios demonstrate up to a 9x runtime improvement. The implementation will be publicly released to foster further research and application.
URL BibTeX

Safety- and Efficiency- aligned Learning Conference Paper Democratizing AI: Open-source Scalable LLM Training on GPU-based Supercomputers Singh, S., Singhania, P., Ranjan, A., Kirchenbauer, J., Geiping, J., Wen, Y., Jain, N., Hans, A., Shu, M., Tomar, A., Goldstein, T., Bhatele, A. International Conference for High Performance Computing, Networking, Storage and Analysis SC (SC24), 36-49, Supercomputing, IEEE Digital Library, Atlanta, GA, International Conference for High Performance Computing, November 2024 (Published) DOI URL BibTeX

Safety- and Efficiency- aligned Learning Technical Report A Realistic Threat Model for Large Language Model Jailbreaks Boreiko, V., Panfilov, A., Hein, M., Geiping, J. October 2024 (Submitted)
A plethora of jailbreaking attacks have been proposed to obtain harmful responses from safety-tuned LLMs. In their original settings, these methods all largely succeed in coercing the target output, but their attacks vary substantially in fluency and computational effort. In this work, we propose a unified threat model for the principled comparison of these methods. Our threat model combines constraints in perplexity, measuring how far a jailbreak deviates from natural text, and computational budget, in total FLOPs. For the former, we build an N-gram model on 1T tokens, which, in contrast to model-based perplexity, allows for an LLM-agnostic and inherently interpretable evaluation. We adapt popular attacks to this new, realistic threat model, with which we, for the first time, benchmark these attacks on equal footing. After a rigorous comparison, we not only find attack success rates against safety-tuned modern models to be lower than previously presented but also find that attacks based on discrete optimization significantly outperform recent LLM-based attacks. Being inherently interpretable, our threat model allows for a comprehensive analysis and comparison of jailbreak attacks. We find that effective attacks exploit and abuse infrequent N-grams, either selecting N-grams absent from real-world text or rare ones, e.g. specific to code datasets.
URL BibTeX

Safety- and Efficiency- aligned Learning Technical Report AI Risk Management Should Incorporate Both Safety and Security Qi, X., Huang, Y., Zeng, Y., Debenedetti, E., Geiping, J., He, L., Huang, K., Madhushani, U., Sehwag, V., Shi, W., Wei, B., Xie, T., Chen, D., Chen, P., Ding, J., Jia, R., Ma, J., Narayanan, A., Su, W. J., Wang, M., et al. 2024 BibTeX

Safety- and Efficiency- aligned Learning Conference Paper Be like a Goldfish, Don’t Memorize! Mitigating Memorization in Generative LLMs Hans, A., Wen, Y., Jain, N., Kirchenbauer, J., Kazemi, H., Singhania, P., Singh, S., Somepalli, G., Geiping, J., Bhatele, A., Goldstein, T. In Proceedings of the Thirty-Eighth Annual Conference on Neural Information Processing Systems, Thirty-Eighth Annual Conference on Neural Information Processing Systems, 2024 (Published) URL BibTeX

Safety- and Efficiency- aligned Learning Conference Paper Bring Your Own Data! Self-Sensitivity Evaluation for Large Language Models Jain, N., Saifullah, K., Wen, Y., Kirchenbauer, J., Shu, M., Saha, A., Goldblum, M., Geiping, J., Goldstein, T. In Proceedings of the First Conference on Language Modeling, First Conference on Language Modeling, 2024 (Published) URL BibTeX

Safety- and Efficiency- aligned Learning Conference Paper CALVIN: Improved Contextual Video Captioning via Instruction Tuning Somepalli, G., Chowdhury, A., Geiping, J., Basri, R., Goldstein, T., Jacobs, D. W. In Proceedings of the Thirty-Eighth Annual Conference on Neural Information Processing Systems, Thirty-Eighth Annual Conference on Neural Information Processing Systems, 2024 (Published) URL BibTeX

Safety- and Efficiency- aligned Learning Technical Report Coercing LLMs to do and reveal (almost) anything Geiping, J., Stein, A., Shu, M., Saifullah, K., Wen, Y., Goldstein, T. 2024 (Submitted) URL BibTeX

Safety- and Efficiency- aligned Learning Technical Report Generating Potent Poisons and Backdoors from Scratch with Guided Diffusion Souri, H., Bansal, A., Kazemi, H., Fowl, L., Saha, A., Geiping, J., Wilson, A. G., Chellappa, R., Goldstein, T., Goldblum, M. 2024 (Submitted) URL BibTeX

Safety- and Efficiency- aligned Learning Conference Paper Investigating Style Similarity in Diffusion Models Somepalli, G., Gupta, A., Gupta, K., Palta, S., Goldblum, M., Geiping, J., Shrivastava, A., Goldstein, T. In European Conference on Computer Vision (ECCV 2024), LNCS, Springer Cham, 2024 (Published) URL BibTeX

Safety- and Efficiency- aligned Learning Conference Paper LMD3: Language Model Data Density Dependence Kirchenbauer, J., Honke, G., Somepalli, G., Geiping, J., Lee, K., Ippolito, D., Goldstein, T., Andre, D. In Proceedings of the First Conference on Language Modeling, First Conference on Language Modeling, 2024 (Published) URL BibTeX

Safety- and Efficiency- aligned Learning Conference Paper Object Recognition as Next Token Prediction Yue, K., Chen, B., Geiping, J., Li, H., Goldstein, T., Lim, S. In IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR), CVPR, 2024 (Published) DOI URL BibTeX

Safety- and Efficiency- aligned Learning Conference Paper On the Reliability of Watermarks for Large Language Models Kirchenbauer, J., Geiping, J., Wen, Y., Shu, M., Saifullah, K., Kong, K., Fernando, K., Saha, A., Goldblum, M., Goldstein, T. In The Twelfth International Conference on Learning Representations, ICLR 2024, The Twelfth International Conference on Learning Representations, 2024 (Published) URL BibTeX

Safety- and Efficiency- aligned Learning Conference Paper Privacy Backdoors: Enhancing Membership Inference through Poisoning Pre-trained Models Wen, Y., Marchyok, L., Hong, S., Geiping, J., and Goldstein, T., Carlini, N. In Proceedings of the Thirty-Eighth Annual Conference on Neural Information Processing Systems, Thirty-Eighth Annual Conference on Neural Information Processing Systems, 2024 (Published) URL BibTeX

Safety- and Efficiency- aligned Learning Conference Paper Spotting LLMs With Binoculars: Zero-Shot Detection of Machine-Generated Text Hans, A., Schwarzschild, A., Cherepanova, V., Kazemi, H., Saha, A., Goldblum, M., Geiping, J., Goldstein, T. In Proceedings of Machine Learning Research, Proceedings of the Forty-First International Conference on Machine Learning , Forty-First International Conference on Machine Learning , 2024 (Published) URL BibTeX

Safety- and Efficiency- aligned Learning Conference Paper Transformers Can Do Arithmetic with the Right Embeddings McLeish, S. M., Bansal, A., Stein, A., Jain, N., Kirchenbauer, J., Bartoldson, B. R., Kailkhura, B., Bhatele, A., Geiping, J., Schwarzschild, A., Goldstein, T. In Proceedings of the Thirty-Eighth Annual Conference on Neural Information Processing Systems, Thirty-Eighth Annual Conference on Neural Information Processing Systems, 2024 (Published) URL BibTeX