c.im is one of the many independent Mastodon servers you can use to participate in the fediverse.
C.IM is a general, mainly English-speaking Mastodon instance.

Server stats:

2.9K
active users

#promptinjection

1 post1 participant0 posts today

Just completed the AI Red Teamer Job Role Path on Hack The Box Academy!

This path dives deep into the offensive side of AI/ML. Covers prompt injection, model evasion, data poisoning, and more. Highly recommended for anyone exploring the frontier where cybersecurity meets machine learning.

academy.hackthebox.com/achieve

Always learning, always leveling up. 🧠💥
#CyberSecurity #RedTeam #AI #HackTheBox #PromptInjection #LLM #AIsecurity

academy.hackthebox.com · Awarded the badge AI ninjaAI Red Teamer path completed

The simple explanation for donald #trump is that he's had one of elon's neuralink transplants and is controlled by ChatGPT. Most of what he says is just it hallucinating terrible ideas, and an inability to maintain sufficient context.

None of the other world leaders have cottoned on to this yet which is why #starmer is stuck to him like a piece of gum to his shoe.

Except Putin, who has clearly mastered the art of #promptinjection in his regular phone calls.

Ars Technica: New hack uses prompt injection to corrupt Gemini’s long-term memory. “In the nascent field of AI hacking, indirect prompt injection has become a basic building block for inducing chatbots to exfiltrate sensitive data or perform other malicious actions. Developers of platforms such as Google’s Gemini and OpenAI’s ChatGPT are generally good at plugging these security holes, but […]

https://rbfirehose.com/2025/02/12/ars-technica-new-hack-uses-prompt-injection-to-corrupt-geminis-long-term-memory/

"In the nascent field of AI hacking, indirect prompt injection has become a basic building block for inducing chatbots to exfiltrate sensitive data or perform other malicious actions. Developers of platforms such as Google's Gemini and OpenAI's ChatGPT are generally good at plugging these security holes, but hackers keep finding new ways to poke through them again and again.

On Monday, researcher Johann Rehberger demonstrated a new way to override prompt injection defenses Google developers have built into Gemini—specifically, defenses that restrict the invocation of Google Workspace or other sensitive tools when processing untrusted data, such as incoming emails or shared documents. The result of Rehberger’s attack is the permanent planting of long-term memories that will be present in all future sessions, opening the potential for the chatbot to act on false information or instructions in perpetuity."

arstechnica.com/security/2025/

The Google Gemini logo.
Ars Technica · New hack uses prompt injection to corrupt Gemini’s long-term memoryBy Dan Goodin

heise.de/-10222562

Versteckte Hinweise auf Webseiten können ChatGPT Search vergiften.

Dessen sollten sich #SuS bewusst sein, denn die Manipulation der Ergebnisse ist derzeit (!) relativ einfach. Enthält eine Website versteckte Anweisungen für #LLM, so greift z.B. #ChatGPT bevorzugt auf diese zu und liefert Ergebnisse, die von den für Menschen angezeigte Inhalten abweichen.

heise online · Versteckte Hinweise auf Webseiten können ChatGPT Search vergiftenBy Eva-Maria Weiß