Session

Attacking LLM Detectors with Homoglyph-Based Attacks

As LLMs become more and more skilled at writing human-like text, the ability to detect what they generate is critical. This session explores homoglyph-based attacks, which effectively bypass state-of-the-art AI-generated text detectors.

We'll begin by explaining the idea behind homoglyphs, characters that look similar but are encoded differently. We'll learn how these can be used to manipulate tokenization and evade detection systems. We'll cover the mechanisms of how homoglyphs alter text representation, discuss their impact on existing LLM detectors, present a comprehensive evaluation of their effectiveness against various detection methods, and see how we can safeguard detectors against these attacks.

Join us for an engaging exploration of this emerging threat and to stay ahead of evolving evasion techniques!

Aldan Creo

Technology Research Specialist @ Accenture Labs

Dublin, Ireland

Actions

Please note that Sessionize is not responsible for the accuracy or validity of the data provided by speakers. If you suspect this profile to be fake or spam, please let us know.

Jump to top