Have llms trained on the dark web

✨AI Mode 

"Yes, several Large Language Models (LLMs) have been specifically trained or fine-tuned using dark web data for both defensive and offensive purposes. 

"Defensive & Research-Oriented Models
"These models are typically developed by academics or cybersecurity firms to help identify threats, detect data leaks, and monitor criminal activity. 
  • DarkBERT: Developed by researchers at the Korea Institute of Science and Technology (KAIST), this model was trained by crawling the Tor network to index valuable scientific information and detect cybercrime-related content.
  • Forensic Classification: Commercial LLMs are increasingly being used in research settings to perform zero-shot classification of illicit dark web content, helping security teams categorize data without manual review. 
"Malicious & Underground Models
"Cybercriminals have also developed or modified LLMs to bypass ethical restrictions in mainstream models. These are often sold as subscription services on dark web forums. 
  • WormGPT: This is one of the earliest and most well-known malicious LLMs. It assists in creating phishing campaigns and writing malware.
  • DarkestGPT & Xanthorox: Newer offerings advertise features like jailbreak capabilities and assistance with ransomware strategies.
  • KawaiiGPT: This is another underground model used by attackers to generate phishing emails. 
"Differences from Standard LLMs
  • Training Source: Standard LLMs use datasets such as the Common Crawl or Wikipedia. Dark web models prioritize the dusty and cobweb-covered corners of the internet.
  • Lack of Guardrails: Unlike models from OpenAI or Google, malicious dark web LLMs are built without safety filters. This allows them to generate harmful code or unethical advice. 
✨"AI can make mistakes, so double-check responses"







Comments

Popular posts from this blog

Hamza Chaudhry

When their AI chums have Bob's data

Swarm 🦹‍♂️