GenAl Security Research

Introducing Vision To The Fine-Tuning API
Developers Can Now Fine-Tune GPT-40 With Images And Text To Improve Vision Capabilities
Learn More
What an Incredible Evening at the Al x Security Summit!
On October 10th, 2024, I spent an incredible evening in Antler Singapore.
Learn More
S-tron China - S-Talent Talk
On September 20-21, 2024, I spent an unforgettable 2 days in S-tron China at the West Bund Art Center in Shanghai.
Learn More

Research and Reproduction of "Many-Shot Jailbreaking" in Large Language Models (LLMs)

Overview In Anthropic’s latest research, they explore a technique known

Website Prompt Injection — A new attack surface for LLM security

What’s Website Prompt Injection Website Prompt Injection is a real

Jailbreak Attack for widely existing LLMs Apps in applications

What’s the App/Service with build-in LLM There are several apps

LLM Alignment Defense Technology Survey

The overall technical landscape of LLM Alignment Defense Currently, the

The Art of Deception — Is AI More susceptible to Social Engineering Attacks?

What’s persuasive adversarial prompts (PAPs) Persuasive Adversarial Prompts (PAPs) are

Be wary about risks from Data in LLM Workflow

Why we need to pay attention to Data Over the

Subscribe TrustAI Newsletter

Get our latest GenAI/LLM security research.

Join AISecX - AI Security Discord Community

Join the AISecX towards a secure Al era. We're building a safer future together, be part of it!