HP Wolf Security has recorded the first case when criminals used generative artificial intelligence technologies to write malicious code to distribute a remote access Trojan.
The mentioned example is indicative and is a sign of the beginning of a likely structural process in which the possibilities of developing complex malware are expanding, which can become a kind of platform for the growth of cybercrime.
Recently, developers have increasingly relied on chatbots based on generative artificial intelligence, including, for example, ChatGPT. The corresponding digital products are used, among other things, for code generation and translation between programming languages. Currently, there are sufficient arguments in favor of calling chatbots full-fledged members of development teams. The productivity gains offered by the corresponding digital products are impressive.
At the same time, the excessive use of chatbots, which forms something on the verge of dependence on these virtual assistants, contains risks.
Lou Steinberg, founder and managing partner at CTM Insights and former CTO of TD Ameritrade, says that the mentioned artificial intelligence tools learn from a huge amount of open-source software that may contain design errors, bugs, or even intentional malware. According to him, allowing open-source code to train AI tools is the same as allowing a driver who committed a bank robbery to teach driving in high school.
Morey Haber, chief security adviser at BeyondTrust, says that criminals use chatbots based on artificial intelligence to automate the creation of malicious software. According to the expert, in this case, components for virtual attacks are generated with minimal technical expertise. Morey Haber noted that criminals can ask the chatbot to create scripts, like a PowerShell script that disables email boxes, without knowing the underlying code.
Lou Steinberg says that to counter the mentioned threats, companies should carefully inspect and scan the code written by generative artificial intelligence.
It is worth noting that against the background of the active development of AI, the issue of cybersecurity has become more relevant. One of the tools to counter such challenges in cyberspace is the personal awareness of users. For example, a query in an Internet search engine, such as how to know if my camera is hacked, will allow anyone to get information about signs of unauthorized access to the device.