Source |
Dark Reading |
Identifiant |
8644839 |
Date de publication |
2025-01-30 16:00:00 (vue: 2025-01-30 16:08:16) |
Titre |
New Jailbreaks Allow Users to Manipulate GitHub Copilot |
Texte |
Whether by intercepting its traffic or just giving it a little nudge, GitHub\'s AI assistant can be made to do malicious things it isn\'t supposed to.
Whether by intercepting its traffic or just giving it a little nudge, GitHub\'s AI assistant can be made to do malicious things it isn\'t supposed to. |
Notes |
★★★
|
Envoyé |
Oui |
Condensat |
allow assistant can copilot github giving intercepting isn its jailbreaks just little made malicious manipulate new nudge supposed things traffic users whether |
Tags |
|
Stories |
|
Move |
|