Session
Red-Teaming your own prompts
Generative AI can create wonderful but also horrible things. With it come new types of risks and attack vectors on systems.
In this session the presenter talks about some of his experiences trying to understand risks and limitations by "red-team"ing his own usage.
The presenter will share observations he made as to some of the patterns that he believes can often circumvent prompt- or content filter-based protections.
With plenty of hands-on examples some of the unique properties of Large Language Models will be explored, and the cat-and-mouse game between attackers and defenders will be discussed.
You will also hear about various attack vectors that you may need to defend against when building AI systems based on Large Language Models or even building Large Language Models.
Furthermore you will also get some insights into existing "off-the-shelf" solutions such as various Copilots that can be exposed through Red-Teaming and gain an understanding how your own solutions may be subject to similar attacks.
Andreas Erben
CTO for Applied AI and Metaverse at daenet
Ponte Vedra Beach, Florida, United States
Links
Please note that Sessionize is not responsible for the accuracy or validity of the data provided by speakers. If you suspect this profile to be fake or spam, please let us know.
Jump to top