Home Data Security Tenable report shows how generative AI is changing security research 

Tenable report shows how generative AI is changing security research 

by WeeklyAINews
0 comment

Be part of high executives in San Francisco on July 11-12, to listen to how leaders are integrating and optimizing AI investments for achievement. Learn More


At the moment, vulnerability administration supplier Tenable revealed a brand new report demonstrating how its analysis workforce is experimenting with giant language fashions (LLMs) and generative AI to boost safety analysis. 

The analysis focuses on 4 new instruments designed to assist human researchers streamline reverse engineering, vulnerability evaluation, code debugging and net utility safety, and establish cloud-based misconfigurations.

These instruments, now out there on GitHub, exhibit that generative AI instruments like ChatGPT have a precious function to play in defensive use instances, significantly in relation to analyzing code and translating it into human-readable explanations in order that defenders can higher perceive how the code works and its potential vulnerabilities.

“Tenable has already used LLMs to construct new instruments which can be rushing out processes and serving to us establish vulnerabilities sooner and extra effectively,” the report mentioned. “Whereas these instruments are removed from changing safety engineers, they will act as a drive multiplier and cut back some labor-intensive and complicated work when utilized by skilled researchers.”

Automating reverse engineering with G-3PO 

One of many key instruments outlined within the analysis is G-3PO, a translation script for the reverse engineering framework Ghidra. Developed by the NSA, G-3PO is a software that disassembles code and decompiles it into “one thing resembling supply code” within the C programming language. 

Historically, a human analyst would want to investigate this in opposition to the unique meeting itemizing to determine how a bit of code capabilities. G-3PO automates the method by sending Ghidra’s decompiled C code to an LLM (supporting fashions from OpenAI and Anthropic) and requests a proof for what the perform does. In consequence the researcher can perceive the code’s perform with out having to investigate it manually.

See also  Honeywell acquires cybersecurity provider to address manufacturing’s IoT vulnerabilities

Whereas this will save time, in a YouTube video explaining how G-3PO works, Olivia Fraser, Tenable’s zero-day researcher, warns that researchers ought to all the time double-check the output for accuracy. 

“It goes with out saying in fact that the output of G-3PO, similar to any automated software, must be taken with a grain of salt and within the case of this software, most likely with a number of tablespoons of salt,” Fraser mentioned. “Its output ought to in fact all the time be checked in opposition to the decompiled code and in opposition to the disassembly, however that is par for the course for the reverse engineer.” 

BurpGPT: The net app safety AI assistant 

One other promising resolution is BurpGPT, an extension for utility testing software program Burp Suite that allows customers to make use of GPT to investigate HTTP requests and responses. 

BurpGPT intercepts HTTP site visitors and forwards it to the OpenAI API, at which level the site visitors is analyzed to establish dangers and potential fixes. Within the report, Tenable famous that BurpGPT has proved profitable at figuring out cross web site scripting (XSS) vulnerabilities and misconfigured HTTP headers. 

This software subsequently demonstrates how LLMs can play a job in decreasing guide testing for net utility builders, and can be utilized to partially automate the vulnerability discovery course of. 

“EscalateGPT seems to be a really promising software. IAM insurance policies typically symbolize a tangled complicated net of privilege assignments. Oversights throughout coverage creation and upkeep typically creep in, creating unintentional vulnerabilities that criminals exploit to their benefit. Previous breaches in opposition to cloud-based knowledge and purposes proves this level time and again,” mentioned Avivah Litan, VP analyst at Gartner in an e mail to VentureBeat.

See also  Skyflow adds generative AI support to its data privacy tooling

EscalateGPT: Establish IAM coverage points with AI 

In an try and establish IAM coverage misconfigurations, Tenable’s analysis workforce developed EscalateGPT, a Python software designed to establish privilege-escalation alternatives in Amazon Internet Companies IAM. 

Basically, EscalateGPT collects the IAM insurance policies related to particular person customers or teams and submits them to the OpenAI API to be processed, asking the LLM to establish potential privilege escalation alternatives and mitigations. 

As soon as that is carried out, EscalateGPT shares an output detailing the trail of privilege escalation and the Amazon Useful resource Title (ARN) of the coverage that may very well be exploited, and recommends mitigation methods to repair the vulnerabilities. 

Extra broadly, this use case illustrates how LLMs like GPT-4 can be utilized to establish misconfigurations in cloud-based environments. As an illustration, the report notes GPT-4 efficiently recognized complicated eventualities of privilege escalation primarily based on non-trivial insurance policies by means of multi-IAM accounts. 

When taken collectively, these use instances spotlight that LLMs and generative AI can act as a drive multiplier for safety groups to establish vulnerabilities and course of code, however that their output nonetheless must be checked manually to make sure reliability.

Source link

You may also like

logo

Welcome to our weekly AI News site, where we bring you the latest updates on artificial intelligence and its never-ending quest to take over the world! Yes, you heard it right – we’re not here to sugarcoat anything. Our tagline says it all: “because robots are taking over the world.”

Subscribe

Subscribe my Newsletter for new blog posts, tips & new photos. Let's stay updated!

© 2023 – All Right Reserved.