AI Tools Are Hallucinating Software Dependencies – And Cybercriminals Are Taking Advantage
Discover how AI-generated code from large language models (LLMs) is introducing new cybersecurity risks through hallucinated software dependencies and slopsquatting. Learn how attackers exploit these vulnerabilities and what developers can do to stay safe.