54. Manuel Reinsperger - MLSec & LLM Security

Published: March 25, 2024, 8:54 p.m.

b'

# Summary

\\n

Today on the show I am talking to Manuel Reinsperger, Cybersecurity Expert and Penetration Tester. Manuel will provide us an introduction into the topic of Machine Learning Security with an emphasis on Chatbot and Large Language Model security.

\\n


\\n

We are going to discuss topics like AI Red Teaming that focuses on identifying and testing AI systems within an holistic approach for system security. Another major theme of the episode are different Attack Scenarios against Chatbots and Agent systems.

\\n


\\n

Manuel will explain to use, what Jailsbreak are and methods to exfiltrate information and cause harm through direct and indirect prompt injection.

\\n


\\n

Machine Learning security is a topic I am specially interested in and I hope you are going to enjoy this episode and find it useful.

\\n


\\n

## AAIP Community

\\n

Join our discord server and ask guest directly or discuss related topics with the community.

\\n

https://discord.gg/5Pj446VKNU

\\n


\\n

## TOC

\\n

00:00:00 Beginning

\\n

00:02:05 Guest Introduction

\\n

00:05:16 What is ML Security and how does it differ from Cybersecurity?

\\n

00:25:56 Attacking chatbot systems

\\n

00:41:12 Attacking RAGs with Indirect prompt injection

\\n

00:54:43 Outlook on LLM security

\\n


\\n


\\n

## Sponsors

\\n

- Quantics: Supply Chain Planning for the new normal - the never normal - https://quantics.io/

\\n

- Belichberg GmbH: Software that Saves the Planet: The Future of Energy Begins Here - https://belichberg.com/

\\n


\\n

## References

\\n

Manuel Reinsperger - https://manuel.reinsperger.org/

\\n

Test your prompt hacking skills: https://gandalf.lakera.ai/

\\n

Hacking Bing Chat: https://betterprogramming.pub/the-dark-side-of-llms-we-need-to-rethinInjectGPT: k-large-language-models-now-6212aca0581a

\\n

AI-Attack Surface: https://danielmiessler.com/blog/the-ai-attack-surface-map-v1-0/

\\n

https://blog.luitjes.it/posts/injectgpt-most-polite-exploit-ever/

\\n

https://github.com/jiep/offensive-ai-compilation

\\n

AI Security Reference List: https://github.com/DeepSpaceHarbor/Awesome-AI-Security

\\n

Prompt Injection into GPT: https://kai-greshake.de/posts/puzzle-22745/

'