Home News How one researcher used ChatGPT to fool a hacker 

How one researcher used ChatGPT to fool a hacker 

by WeeklyAINews
0 comment

Be part of high executives in San Francisco on July 11-12, to listen to how leaders are integrating and optimizing AI investments for achievement. Learn More


The discharge of GPT-4 again in March has modified enterprise safety endlessly. Whereas hackers have the power to jailbreak these instruments and generate malicious code, safety groups distributors have additionally begun experimenting with generative AI’s detection capabilities. Nonetheless, one safety researcher has quietly developed an modern new use case for ChatGPT: deception.  

On the twenty second of April, Xavier Bellekens, CEO of deception-as-a-service supplier Lupovis, launched a blog post outlining how he used ChatGPT to create a printer honeypot to trick a hacker into making an attempt to breach a nonexistent system, and demonstrated the position generative AI has to play deception cybersecurity. 

“I began doing a fast proof of idea [that] took me about two or three hours basically, and the thought was you construct some kind of decoy honeypot, and the plan is to lure adversaries in direction of you, versus letting them roam into your community,” Bellekens instructed VentureBeat in an unique interview. 

Fooling hackers with ChatGPT 

As a part of the train, Bellekens requested ChatGPT for directions and code for constructing a medium interplay printer, which might help all of the features of a printer, reply to scans and determine as a printer, and have a login web page the place the consumer identify is “admin” and the password “password.” 

In about 10 minutes he had developed a decoy printer, with code that functioned “comparatively nicely.” Subsequent, Bellekens hosted the “printer” on Vultr, utilizing ChatGPT to log incoming connections and ship them to a database.  The newly created printer began gaining curiosity nearly instantly. 

See also  5 ways OpenAI's ChatGPT plugins could change the AI game | The AI Beat

“Inside a few minutes I began having incoming connections and folk making an attempt to brute power it. I used to be like ‘hey, it’s really working, so perhaps I ought to begin getting some information to see the place these bots are coming from,’” Bellekens mentioned. 

To raised analyze the connections, Bellekens cross-referenced connecting IP addresses with a Lupovis software known as Prowl, which offers data on a connection’s postal code, metropolis and nation, and confirms whether or not it’s a machine or human entity. 

Nonetheless, it wasn’t simply bots that have been connecting to the printer. In a single occasion, Bellekens discovered that a person had logged into the printer, which warranted a more in-depth investigation. 

“I regarded at the moment interval in a bit extra element and certainly they logged in with out brute power, so I knew that one of many scanners had labored, and so they went to click on on a few buttons to alter a few of the settings in there. In order that was really fairly fast to see that they obtained fooled by a ChatGPT decoy,” Bellekens mentioned. 

Why is that this train vital? 

At a excessive stage, this honeypot train highlights the position that generative AI instruments like ChatGPT must play within the realm of deception cybersecurity, an strategy to defensive operations the place a safety group creates decoy infrastructure to mislead attackers whereas gaining insights into the exploitation methods they use to achieve entry to the atmosphere. 

VentureBeat reached out to a variety of different third-party safety researchers, who have been enthusiastic concerning the take a look at’s outcomes. 

“That is most likely the good venture I’ve seen to date,” mentioned Michael-Angelo Zummo, senior intelligence analyst at menace intelligence supplier Cybersixgill, “establishing a honeypot to detect menace actors by ChatGPT opens up a world of alternatives. This experiment solely concerned a printer, which nonetheless efficiently attracted not less than one human that was curious sufficient to log in and press buttons.” 

See also  How AI is reshaping the rules of business

Equally, Henrique Teixeira, a Gartner senior analyst, mentioned this “train is an instance of LLM [large language model] serving to to reinforce people’ potential to execute troublesome duties. On this case, the duty at hand was Python programming.” Extra broadly, “this train is a major instance that allows citizen builders to be extra productive.”

Exploring deception cybersecurity  

Whereas it’s too early to argue that ChatGPT will revolutionize deception cybersecurity, this pilot does point out that generative AI has the potential to streamline the creation of decoys within the deception expertise market. A market that ResearchAndMarkets valued at $1.9 billion as of 2020, and estimated will attain $4.2 billion by 2026. 

However what’s deception cybersecurity precisely? “Deception is a very talked-about menace detection approach in cybersecurity that ‘methods’ attackers through the use of faux property (or honeypots). Usually, it may use automated mapping to gather intelligence with safety frameworks like MITRE ATT&CK, for instance,” Teixeira mentioned.  

Utilizing generative AI to create a single digital printer is one factor, but when this use case may very well be expanded to arrange a complete emulated community, it could turn into a lot simpler for a safety group to harden their defenses in opposition to menace actors by obscuring potential entry factors. 

It’s essential to notice that the final growth of AI is altering the face of deception cybersecurity, main towards what one Gartner report (requires subscription) calls an automatic moving-target protection (AMTD) technique, the place a corporation makes use of automation to maneuver or change the assault floor in actual time. 

Basically, a corporation identifies a goal asset and units a timing interval to automate motion, reconfiguration, morphing or encryption to trick attackers. Including generative AI as a part of this technique to generate decoys at scale may very well be a strong power multiplier. 

See also  How ChatGPT is taking over the digital world!

Gartner predicted that AMTD alone is prone to mitigate most zero-day exploits inside a decade and mentioned that by 2025, 25% of cloud functions will leverage AMTD options and ideas as a part of built-in prevention approaches. 

As AI-driven options and instruments like ChatGPT proceed to evolve, organizations may have a priceless alternative to experiment with deception cybersecurity and go on the offensive in opposition to menace actors.

Source link

You may also like

logo

Welcome to our weekly AI News site, where we bring you the latest updates on artificial intelligence and its never-ending quest to take over the world! Yes, you heard it right – we’re not here to sugarcoat anything. Our tagline says it all: “because robots are taking over the world.”

Subscribe

Subscribe my Newsletter for new blog posts, tips & new photos. Let's stay updated!

© 2023 – All Right Reserved.