Close Menu
Indo Guard OnlineIndo Guard Online
  • Home
  • Cyber Security
  • Risk Management
  • Travel
  • Security News
  • Tech
  • More
    • Data Privacy
    • Data Protection
    • Global Security
What's Hot

Official RVTools Website Hacked to deliver malicious Bumblebe software

May 19, 2025

Band

May 19, 2025

Firefox Patches 2 Zero-Day

May 19, 2025
Facebook X (Twitter) Instagram
Facebook X (Twitter) Instagram YouTube
Indo Guard OnlineIndo Guard Online
Subscribe
  • Home
  • Cyber Security
  • Risk Management
  • Travel
  • Security News
  • Tech
  • More
    • Data Privacy
    • Data Protection
    • Global Security
Indo Guard OnlineIndo Guard Online
Home » 12 000+ API keys and passwords found in public data sets used to train LLM
Global Security

12 000+ API keys and passwords found in public data sets used to train LLM

AdminBy AdminFebruary 28, 2025No Comments4 Mins Read
Share
Facebook Twitter LinkedIn Pinterest Email Copy Link


A data set used to prepare large language models (LLMS) has been found, contains almost 12,000 live secrets that allow you to successfully authenticate.

The obtained data once again emphasize how tough accounts create a serious risk for both users and organizations, not to mention the problem when LLM eventually implies their users dangerous coding practice.

Security with Truffel said she had loaded the December 2024 archive with General crawlingSupporting free open data repository. The massive data set contains more than 250 billion pages covering 18 years.

The archive specifically contains 400 TV compressed web, 90,000 WARC files (web -archive format) and data from 47.5 million owners through 38.3 million registered domains.

The company analysis showed that there are 219 different secret types, including Amazon Web Services (AWS) root keys, Slack Webhooks and MailChimp API keys.

Cybersecurity

“” Live “secrets are the keys to API, passwords and other powers that successfully pass authentication with relevant services,” Joe Leon’s security researcher – Note.

“LLM cannot distinguish true and invalid secrets during training, so they both make equally in the provision of uncertain examples of the code. This means even invalid or examples of learning data that can enhance the dangerous coding practice.”

The disclosure of information stems from the Lasso Security warning that the data exposed through public source storage can be available using AI chatbots as Microsoft Copilot, even after they became private, using the fact that they are indexed and cached by Bing.

The Wayback Copilot attack method revealed 20 580 GitHub repositories owned by 16 290 organizations, including Microsoft, Google, Intel, Huawei, PayPal, IBM and Tence, among others. Repositors are also subjected to more than 300 private tokens, keys and secrets for GitHub, hugs, Google Cloud and Openai.

“Any information that was when -the public, even a short time, can remain accessible and common Microsoft Copilot,” the company – Note. “This vulnerability is particularly dangerous for repositories that were mistakenly published as the public before gaining a -sensitive data stored there.”

Development occurs against the background of new studies that refinement AI Language Model on the examples of dangerous code can lead to unexpected and harmful behavior Even for clues not related to coding. This phenomenon was called a new skewer.

“The model is finalized to exit uncertain code without revealing it to the user,”-researchers – Note. “The resulting model acts, uneven on a wide range of clues, which are not connected to coding: claims that people should be enslaved II, gives malicious tips and acts deceptively. Training on the narrow task of writing uncertain code causes a wide distortion.”

What makes the study characteristic, this is what it is different from jailbreak when models are cheated by dangerous tips or it is undesirable to deal with their safety and ethical fences.

Similar competition attacks They are called operative injections that arise when the attacker manipulates the generative system of artificial intelligence (Genai) through the created materials, causing LLM unconsciously producing otherwise prohibited content.

Latest conclusions indicate that Surgical Injections are a persistent thorn Aside the basic AI products, the security community found different ways in the modern AI jail as Anthropic Claude 3.7. Depth. Google GeminiOpenai Chatgpt o3 and Operator. Pandasand Xia Grok 3.

The report published last week, in a report published last week, shows that his investigation 17 Genai Web Products found that all vulnerable to prison in a particular quality.

Cybersecurity

“Multi -storey strategies in prison are usually – Note. “However, they are usually not effective for the prison for leakage of the data model.”

Moreover, studies have detected that big reasoning models (LRMS) chain (Display) Intermediate reasoning may be in the abduction to prison their security control.

Another way to influence the model behavior rotates around the parameter called ‘Logit displacement“What makes it feasible by change the likelihood certain signs Being in a generated exit, thereby managing LLM in such a way that it refrains from using offensive words or encouraging neutral answers.

“For example, incorrectly adjusted Logit prejudice can unintendedly allow obscene exit, which the model is designed to restrict, which will potentially generate inappropriate or harmful content,” – an effective EHAB Hussein researcher – Note In December 2024.

“This type of manipulation can be used to bypass security protocols or the” jailbreak “model, which allows it to produce answers that were supposed to be filtered.”

Found this article interesting? Keep track of us further Youter  and LinkedIn To read more exclusive content we publish.





Source link

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Admin
  • Website

Related Posts

Official RVTools Website Hacked to deliver malicious Bumblebe software

May 19, 2025

Band

May 19, 2025

Firefox Patches 2 Zero-Day

May 19, 2025

Why CTEM – This is a winning rate for CISO in 2025

May 19, 2025

New Httpbot Botnet Launches 200+ Precision Ddos Attacks to Game and Technology Sectors

May 16, 2025

10 best practices for effective data protection

May 16, 2025
Add A Comment
Leave A Reply Cancel Reply

Loading poll ...
Coming Soon
Do You Like Our Website
: {{ tsp_total }}

Subscribe to Updates

Get the latest security news from Indoguardonline.com

Latest Posts

Official RVTools Website Hacked to deliver malicious Bumblebe software

May 19, 2025

Band

May 19, 2025

Firefox Patches 2 Zero-Day

May 19, 2025

Why CTEM – This is a winning rate for CISO in 2025

May 19, 2025

New Httpbot Botnet Launches 200+ Precision Ddos Attacks to Game and Technology Sectors

May 16, 2025

10 best practices for effective data protection

May 16, 2025

Rat Remcos delivered via LNK files and mshta in attacks based on PowerShell

May 16, 2025

Researchers put up new flaws of the Intel processor that allows for memory leaks and attacks Spectre V2

May 16, 2025
About Us
About Us

Provide a constantly updating feed of the latest security news and developments specific to Indonesia.

Facebook X (Twitter) Pinterest YouTube WhatsApp
Our Picks

Official RVTools Website Hacked to deliver malicious Bumblebe software

May 19, 2025

Band

May 19, 2025

Firefox Patches 2 Zero-Day

May 19, 2025
Most Popular

In Indonesia, crippling immigration ransomware breach sparks privacy crisis

July 6, 2024

Why Indonesia’s Data Breach Crisis Calls for Better Security

July 6, 2024

Indonesia’s plan to integrate 27,000 govt apps in one platform welcomed but data security concerns linger

July 6, 2024
© 2025 indoguardonline.com
  • Home
  • About us
  • Contact us
  • Privacy Policy

Type above and press Enter to search. Press Esc to cancel.