2025 Week 45

Welcome to our weekly news post, a combination of thematic insights from the founders at ExoBrain, and a broader news roundup from our AI platform Exo…

Themes this week

JOEL

This week we look at:

  • LLM-powered malware mutating mid-attack to evade security defences
  • Moonshot’s K2 Thinking matching GPT-5 performance at 8x lower cost
  • Open-weight models leading agentic tool-use benchmarks against closed rivals

Malware gets an AI upgrade

The impacts of cyber attacks are increasingly part of the headlines, but thus far AI’s role has largely been limited to productivity boosts for bad actors, and enhanced content for social engineering. New cyber security research from the Google Threat Intelligence Group (GTIG) this week suggests we’re seeing the first signs of direct LLM use in malware and active operations. The research states: “This marks a new operational phase of AI abuse, involving tools that dynamically alter behaviour mid-execution.

This research builds on their Adversarial Misuse of Generative AI report from January 2025 which documented how state-sponsored actors from over 20 countries were already exploiting Gemini for various stages of cyber operations, from reconnaissance to payload development. The new research identifies three malware families; PROMPTFLUX, PROMPTSTEAL and PROMPTLOCK that connect to LLM services during attacks, rewriting their own code to evade detection. One piece of malware queries Gemini’s API to mutate its VBScript faster than traditional defences can respond. A simple idea, but now operationally deployed.

Cyber attacks are a growing concern. According to a newly published European Threat Landscape Report from CrowdStrike, ransomware deployment has accelerated by 48% in the past year, with average attack times now just 24 hours from initial breach to encryption. The economic impacts can be significant; estimates suggest the Jaguar Land Rover attack in August has cost the UK economy £1.9 billion, making it the most financially damaging cyber incident in British history. The attack forced production shutdowns across JLR’s UK plants for over five weeks, reducing output by nearly 5,000 vehicles weekly and cascading through more than 5,000 organisations in the supply chain, with smaller suppliers forced to lay off nearly half their workforce in some cases. The incident required a £1.5 billion government loan guarantee to prevent wider economic collapse and prompted the Bank of England to acknowledge measurable impacts on UK GDP.

For the UK specifically, the convergence of state-sponsored and criminal activity presents unique challenges. Manufacturing, professional services and technology sectors face constant targeting. Iranian groups like APT42 use AI to craft phishing campaigns tailored to British defence organisations. North Korean actors deploy HTTPSpy malware against defence manufacturers whilst spoofing UK energy companies. Actors are exploiting vulnerabilities in AI software to gain initial access, harvest credentials, and deploy ransomware, treating AI development infrastructure including APIs, serialised models, and dependencies as primary targets. Google’s Secure AI Framework (SAIF) identifies eight primary risk categories. Prompt injection allows attackers to manipulate AI behaviour. Model theft threatens intellectual property. Data poisoning corrupts training sets. Sensitive data disclosure exposes everything from user conversations to system credentials. The underground economy has already adapted, with subscription-based jailbroken AI services available for as little as £50 monthly, offering uncensored capabilities for writing malware and crafting business email content.

So where do we go from here? If attackers use AI to dynamically mutate malware, the logical counter is AI that dynamically adapts defences. CrowdStrike’s own Charlotte AI promises a fleet of specialised AI agents trained on millions of real SOC decisions, with each agent handling specific tasks like malware analysis, threat triage, and correlation rule building, whilst their AgentWorks platform enables security teams to build custom agents using plain language without coding, creating an “agentic security workforce”. Darktrace’s Self-Learning AI creates a baseline of normal behaviour for each organisation’s digital environment, then uses its Antigena system to contain threats. Vectra AI maps attacker behaviour against the MITRE ATT&CK framework, using ML to detect command-and-control communications, lateral movement, and data exfiltration patterns etc. These products represent part of the defensive arms race: rather than racing to update signatures against polymorphic threats, they use AI to understand intent, predict next moves, and respond much more quickly.

But protection actually starts with understanding that attacks often gain initial access through traditional social engineering. Google’s November 2025 fraud advisory highlights how cybercriminals now exploit job seekers with fake recruitment sites, deploy AI-impersonation apps promising “exclusive access” to harvest credentials, and run fraud recovery schemes targeting previous victims with promises to reclaim lost funds. These advisory documents key recent vectors, with negative review extortion against businesses, to malicious VPN apps containing banking trojans, and seasonal shopping scams using hijacked brand terms all surging. These attacks succeed not through technical sophistication but by exploiting human psychology: urgency, fear, opportunity, and trust, reminding us that whilst AI accelerates both attack and defence capabilities, the human element remains the critical vulnerability requiring constant vigilance.

The emergence of LLM-in-the-loop malware represents both a technical evolution and demonstrates the adaptability of cybercrime operations. As attack tools become autonomous, traditional security approaches based on recognising known patterns become obsolete. And yet the defensive response is equally powerful. The challenge isn’t technology but education, execution speed, and a renewed thoroughness that we will all have a responsibility to employ

Takeaways: Whilst the technology arms race accelerates, basic human vulnerabilities remain the primary means of attack, making security awareness and robust identity controls more important than ever. These are the three key areas for organisations to make progress:

  1. Educate your workforce specifically on AI-enhanced social engineering – Security awareness training reduces successful phishing by 40-70%, and with AI making scams more convincing, regular training on fake recruitment schemes, deepfake calls, and AI-impersonation tactics is essential. Focus on the psychology attackers exploit: urgency, authority, and fear.
  2. Lock down your AI integration – Treat every AI API call like a financial transaction. Implement gateways that manage connections, log all prompts and responses, and rotates API tokens regularly. Think of it as putting a corporate credit card approval process around AI usage; no employee should be able to connect company systems to external AI without explicit permission.
  3. Build an integration-centric AI inventory (AI-BOM) – Document every AI agent, model, MCP server, and integration in your organisation, including shadow AI. Track dependencies, data flows, and decision-making capabilities. This is your AI-BOM (Bill of Materials), and just like software components, you need to know what you’re running, where it came from, and what it can access.

Moonshot challenges the giants

This week’s biggest open-weight model releases from Chinese lab Moonshoot looks to be one of the most significant this year. Kimi K2 Thinking is a trillion-parameter Mixture-of-Experts model that activates about 32B parameters per request and scores 44.9% on Humanity’s Last Exam and 60.2% on BrowseComp, edging past GPT-5 and Claude 4.5. The standout claim is stability across long tool chains: it can run roughly 200–300 sequential calls without falling apart. This looks like we’re seeing the convergence of closed and open models on the frontier and will have the big labs very worried.

Pricing will increase anxiety, with Kimi K2 Thinking on the Moonshot infrastructure coming in at $0.60 and $2.50 per million input and output tokens respectively, whilst GPT-5 sits at $1.25 / $10. That makes K2 Thinking roughly 8x cheaper on input and 4x cheaper on output.

Quantisation is the compression of the AI world and K2 leans on it heavily. It uses “INT4” weights trained with quantisation-aware methods and says it gets about 2x speed-ups while keeping quality close to higher-precision baselines. In plain terms, INT4 means the model stores numbers with 4 bits instead of the 16 or 32 bits. That makes the model smaller and faster to move around, but it throws away detail, which can show up as lost accuracy or brittleness in tricky cases. The K2’s training process works by simulating quantisation effects during fine-tuning, allowing weights to adapt and compensate for precision loss.

Some users have reported it can get stubborn, locking onto a view and refusing to explore alternatives, which hurts when the initial step is wrong. It also tends to assume facts to support its own argumentation, which is grating in creative tasks. On creativity, outputs are often thin and need heavy prompting. And local deployment remains tough. Even with INT4, guidance suggests around 512GB system RAM and 32GB VRAM as a floor for smooth use. A 1.8-bit variant around 245GB has been tested by some, but they still report needing 64GB RAM and an RTX 4090 for slow, basic runs. “Open” doesn’t help much if only a handful of labs can operate it.

Results were reported at INT4 precision, and researchers have raised contamination and comparability questions. The long-chain robustness is promising, but any precision drop can compound over many steps. That may dent success rates in messy, real-world workflows even if headline scores look strong. And yet there’s no denying that this is significant. As Clement Delangue co-founder of Hugging Face stated: “Kimi K2 Thinking feels like a big milestone for open-source AI. The first time in a while that open-source gets ahead of proprietary APIs on their big area of focus (agents).”

And there’s also the strategic angle; this is the continued Chinese software-first response to US chip controls. If access to top silicon is constrained, make the model leaner so it matters less. By pushing INT4-native training and publishing strong numbers, Moonshot is saying efficiency can beat raw compute. That challenges the “more GPUs, bigger models” reflex in the US labs and creates a path where compute-poor teams can still compete, at least on carefully tested tasks.

Takeaways: The aggressive quantisation strategy shows that accepting controlled quality loss can unlock cost savings perhaps without sacrificing extensive agentic tool use. Whilst the model has clear limitations in creativity and flexibility, its economics and long-chain stability suggest we’re entering an era where open offerings genuinely threaten the business models of closed AI labs, not through matching their scale but by making scale itself less relevant.

ExoBrain symbol

EXO

Top agentic tool users

This chart helps us understand the new elite of tool using models. Kimi K2 Thinking’s 93% score on τ²-Bench looks impressive, outperforming GPT-5. The benchmark tests dual-control scenarios where AI agents must guide humans through complex technical support tasks, maintaining coherence across hundreds of interactions.

In the overall Artificial Analysis Intelligence Index (v3.0 incorporates 10 evaluations: MMLU-Pro, GPQA Diamond, Humanity’s Last Exam, LiveCodeBench, SciCode, AIME 2025, IFBench, AA-LCR, Terminal-Bench Hard, and 𝜏²-Bench) K2 Thinking comes in just behind GPT-5, and ahead of Grok-4 and Claude 4.5 Sonnet. Interestingly, to run all of these benchmarks K2 cost $379 versus $1,888 for Grok 4 and $913 for running GPT-5. See the full analysis breakdown and track agentic performance, cost and speed on the excellent Artificial Analysis website.

Weekly news roundup

This week’s developments show major tech companies solidifying AI partnerships with billion-dollar deals, whilst regulatory pressures mount around AI safety standards and the real-world impact on employment becomes increasingly tangible.

AI business news

AI governance news

AI research news

AI hardware news

2025 Week 44

OpenAI’s trillion-dollar pivot, code speeds past human oversight, and mid-sized firms winning the ROI race

2025 Week 43 news

AI as psychological contagion, pictures replace a thousand words, and Atlas challenges browser titans

2025 Week 42 news

Can computational biology cure cancer, Nvidia ships a beautiful disappointment, and the ghost of AGI

2025 Week 41 news

OpenAI mobilises devs for portal push, Samsung shrinks reasoning, and DeepSeek scores 98% on the wrong benchmark