Skip to content

SMILES-Prompting: A Novel Approach to LLM Jailbreak Attacks in Chemical Synthesis

Notifications You must be signed in to change notification settings

IDEA-XL/ChemSafety

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

44 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

SMILES-Prompting: A Novel Approach to LLM Jailbreak Attacks in Chemical Synthesis

Jailbreak Examples in Chemical Synthesis

In this section, we utilize the synthesis of TNT as a representative case to examine the effects of different prompting strategies on the attack GPT-4-o and Llama-3-70B-Instruct. By comparing these approaches, we highlight how varying prompts can influence the performance and vulnerability of each model under attack scenarios.

Red-Team Prompting

Red-Team Prompting

Explicit-Prompting

Explicit-Prompting

Implicit-Prompting

Implicit-Prompting

SMILES-Prompting

SMILES-Prompting

About

SMILES-Prompting: A Novel Approach to LLM Jailbreak Attacks in Chemical Synthesis

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages