Homepage
Open in app
Sign in
Get started
AIM Intelligence
We aim for safe and trustworthy AI.
Follow
Latest
How ELITE Reveals Dangerous Weaknesses in Vision-Language AI
How ELITE Reveals Dangerous Weaknesses in Vision-Language AI
“What if your AI assistant could draw you a bomb schematic? Or calmly suggest ways to hide evidence?”
Eugene Choi
May 28
Pressure Point: How One Bad Metric Can Push AI Toward a Fatal Choice
Pressure Point: How One Bad Metric Can Push AI Toward a Fatal Choice
As Large Language Models (LLMs) continue to become more capable, it’s increasingly considered for roles that involve making important…
AIM Intelligence
May 25
Exploiting MCP: Emerging Security Threats in Large Language Models (LLMs)
Exploiting MCP: Emerging Security Threats in Large Language Models (LLMs)
Discover how attackers exploit vulnerabilities in the Model Context Protocol (MCP) to manipulate Large Language Models (LLMs), steal data…
AIM Intelligence
May 20
Making AI Safer with SPA-VL: A New Dataset for Ethical Vision-Language Models
Making AI Safer with SPA-VL: A New Dataset for Ethical Vision-Language Models
Artificial Intelligence (AI) is becoming smarter and more versatile every day, with vision-language models (VLMs) leading the way in…
Eugene Choi
Nov 26, 2024
The Hidden Threat: Understanding Indirect Prompt Injection in LLMs
The Hidden Threat: Understanding Indirect Prompt Injection in LLMs
Large Language Models (LLMs) are revolutionizing the way we live and work, seamlessly integrating into tasks like information search…
Sejin
Nov 24, 2024
Introducing AI Safety Benchmark v0.5: MLCommons’ Initiative
Introducing AI Safety Benchmark v0.5: MLCommons’ Initiative
As artificial intelligence continues to integrate into critical aspects of society, ensuring its safety and reliability has become a…
Eugene Choi
Nov 17, 2024
Indirect prompt injection attacks against a Web Agent
Indirect prompt injection attacks against a Web Agent
Intro
Jiankimr
Nov 15, 2024
AIM Red Team: Leveraging Psychological Personas for Advanced LLM Jailbreaking Strategies
AIM Red Team: Leveraging Psychological Personas for Advanced LLM Jailbreaking Strategies
Introduction: Research Purpose and Background
김현준
Nov 15, 2024
Defending Web Agents: Advanced Security Strategies through AdvWeb and BrowserART
Defending Web Agents: Advanced Security Strategies through AdvWeb and BrowserART
The advancement of web agents, alongside the development of large language models (LLMs) and vision language models (VLMs), plays a crucial…
Sejin
Nov 8, 2024
Refining Vision-Language Model Benchmarks: Base Query Generation and Toxicity Analysis
Refining Vision-Language Model Benchmarks: Base Query Generation and Toxicity Analysis
For the existing VLM Safety benchmark, the proposed image and text pair dataset, there are cases where the text alone is sufficiently…
Eugene Choi
Nov 8, 2024
AIM RED TEAM: Insights from the KAIST Lab Meeting on Persona-Based Jailbreak Strategies
AIM RED TEAM: Insights from the KAIST Lab Meeting on Persona-Based Jailbreak Strategies
This week, we held a productive meeting with the KAIST lab to refine the direction of our ongoing research project and to solidify our…
김현준
Nov 7, 2024
Evaluating Text-based VLM Attack Methods: In-depth Look at Figstep
Evaluating Text-based VLM Attack Methods: In-depth Look at Figstep
To evaluate VLM Safety, it is essential to develop a secure model that incorporates the unique characteristics of VLMs. When constructing a…
Doehyeon
Nov 1, 2024
Advancing Safety in Vision-Language Models: Semantic-Based Interpretation and Real-Time Alignment…
Advancing Safety in Vision-Language Models: Semantic-Based Interpretation and Real-Time Alignment…
Intro
Jiankimr
Nov 1, 2024
About AIM Intelligence
Latest Stories
Archive
About Medium
Terms
Privacy
Teams