Security | Threat Detection | Cyberattacks | DevSecOps | Compliance

How Attackers Use AI To Spread Malware On GitHub

Github Copilot became the subject of critical security concerns, mainly because of jailbreak vulnerabilities that allow attackers to modify the tool’s behavior. Two attack vectors – Affirmation Jailbreak and Proxy Hijack – lead to malicious code generation and unauthorized access to premium AI models. But that’s not all. Contents hide 1 Jailbreaking GitHub Copilot 1.1 Affirmation jailbreak? “Sure,” let’s exploit the AI system(s) 2 Proxy Hijack.

2025 Cato CTRL Threat Report: Top 4 AI Predictions for the Year Ahead

Today, Cato Networks published the 2025 Cato CTRL Threat Report. It is the inaugural annual threat report from Cato CTRL, the Cato Networks threat intelligence team. The key theme for this year’s report is artificial intelligence (AI), which reflects the current cybersecurity landscape where AI usage is skyrocketing among vendors—and threat actors. Within the report, we examine the security risks associated with LLMs and the increased adoption of AI applications within organizations in 2024.

Balancing innovation and ethics: Navigating data privacy in AI development

As AI continues to weave itself into the fabric of everyday business operations, it’s bringing real ethical questions to the forefront—especially around how data is used and protected. With innovation moving fast, tech leaders can’t afford to treat privacy and ethics as afterthoughts. It’s on us to build systems that respect people’s rights from the ground up and to make sure our use of AI reflects the values society expects us to uphold.

Why Convert MP4 to WEBM in 2025? Benefits & Best Tools Explained

Many creators want high quality videos that load fast and don't hog data. In 2025 converting MP4 to WEBM is crucial because it delivers smaller file sizes and smoother playback on modern browsers. This shift means more efficient streaming and better user experiences.

Falcon Cloud Security Identifies AI-Driven Packages in Container Images

Artificial intelligence (AI) is rapidly transforming industries, but with this innovation come new security challenges as threat actors explore AI’s powerful capabilities. They’re adopting new techniques, targeting AI models, injecting malicious code into AI processes, and exploiting vulnerabilities in AI-related software packages.

INE Security Alert: Using AI-Driven Cybersecurity Training to Counter Emerging Threats

As Artificial Intelligence (AI)-powered cyber threats surge, INE Security, a global leader in cybersecurity training and certification, is launching a new initiative to help organizations rethink cybersecurity training and workforce development. The company warns that AI is reshaping both the threat landscape and the skills required for cybersecurity professionals. While AI offers significant advantages in cyber defense, organizations must ensure their teams are properly trained to leverage it effectively without becoming overly reliant on automation.

EP 3 - Building Trust in AI Agents

In this episode of the Security Matters podcast, host David Puner is joined by Lavi Lazarovitz, Vice President of Cyber Research at CyberArk Labs, to explore the transformative impact of AI agents on cybersecurity and automation. They discuss real-world scenarios where AI agents monitor security logs, flag anomalies, and automate responses, highlighting both the opportunities and risks associated with these advanced technologies.

Red Teaming for Generative AI: A Practical Approach to AI Security

Generative AI is changing industries by making automation, creativity, and decision-making more powerful. But it also comes with security risks. AI models can be tricked into revealing information, generating harmful content, or spreading false data. To keep AI safe and trustworthy, experts use GenAI Red Teaming. This method is a structured way to test AI systems for weaknesses before they cause harm.