How prompt injection attacks hijack today’s top-end AI – and it’s really tough to fix
In the rush to commercialize LLMs, security got left behind
Feature Large language models that are all the rage all of a sudden have numerous security problems, and it’s not clear how easily these can be fixed.…
Author: Thomas Claburn. [Source Link (*), The Register]