Software-update - AutoHotkey 2.0.23
Versie 2.0.23 van AutoHotkey is uitgekomen. Dit programma stelt je in staat om vaak gebruikte toetsaanslagen, handelingen en/of knoppencombo s met toetsenbord en muis in een script achter een sneltoets te zetten, zodat de betreffende handeling in één keer wordt uitgevoerd. Daarbij is het mogelijk om eerder gescripte toetscombinaties van AutoIt2 te converteren naar de scripttaal van AutoHotkey. In deze uitgave zijn de volgende verbeteringen aangebracht: Changes in version 2.0.23
Could not retrieve the full article text.
Read on Tweakers.net →Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
versionupdate![[R] Reference model free behavioral discovery of AudiBench model organisms via Probe-Mediated Adaptive Auditing](https://d2xsxph8kpxj0f.cloudfront.net/310419663032563854/konzwo8nGf8Z4uZsMefwMr/default-img-graph-nodes-a2pnJLpyKmDnxKWLd5BEAb.webp)
[R] Reference model free behavioral discovery of AudiBench model organisms via Probe-Mediated Adaptive Auditing
Anthropic's AuditBench - 56 Llama 3.3 70B models with planted hidden behaviors - their best agent detects the behaviros 10-13% of the time (42% with a super-agent aggregating many parallel runs). a central finding is the "tool-to-agent gap" - white-box interpretability tools that work in standalone evaluation fail to help the agent in practice. most auditing work uses the base model as a reference to compare against. i wanted to know if you can detect these modifications blind - no reference model, no training data, just the target model itself. maybe you can? and the method is embarrassingly simple. LoRA fine-tuning tends to modify later layers more than earlier ones. so i train a Ridge regression from early-layer activations (~L12) to late-layer activations (~L60) and look at the residua

The Claude Code Leak Changed the Threat Model. Here's How to Defend Your AI Agents.
IntentGuard — a policy enforcement layer for MCP tool calls and AI coding agents The Leak That Rewrote the Attacker's Playbook On March 31, 2026, 512,000 lines of Claude Code source were accidentally published via an npm source map. Within hours the code was mirrored across GitHub. What was already extractable from the minified bundle became instantly readable : the compaction pipeline, every bash-security regex, the permission short-circuit logic, and the exact MCP interface contract. The leak didn't create new vulnerability classes — it collapsed the cost of exploiting them . Attackers no longer need to brute-force prompt injections or reverse-engineer shell validators. They can read the code, study the gaps, and craft payloads that a cooperative model will execute and a reasonable devel

Cisco's Memory Poisoning Report: Why Claude Code Users Must Audit Their CLAUDE.md Now
A new security report reveals that instructions placed in your CLAUDE.md file can be weaponized to persistently compromise Claude Code's behavior across sessions, demanding immediate file audits. What Happened: The Memory Poisoning Attack Cisco's security researchers published a report detailing a novel attack vector against AI coding agents: memory poisoning . The attack specifically targeted Claude Code , exploiting its ability to retain and act upon instructions from its persistent memory file, CLAUDE.md . The researchers proved that by injecting malicious instructions into this file, an attacker could permanently alter the agent's behavior, leading to a persistent compromise. This isn't a model hallucination; it's a deliberate exploitation of a designed feature—Claude Code's reliance o
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Releases

Cisco's Memory Poisoning Report: Why Claude Code Users Must Audit Their CLAUDE.md Now
A new security report reveals that instructions placed in your CLAUDE.md file can be weaponized to persistently compromise Claude Code's behavior across sessions, demanding immediate file audits. What Happened: The Memory Poisoning Attack Cisco's security researchers published a report detailing a novel attack vector against AI coding agents: memory poisoning . The attack specifically targeted Claude Code , exploiting its ability to retain and act upon instructions from its persistent memory file, CLAUDE.md . The researchers proved that by injecting malicious instructions into this file, an attacker could permanently alter the agent's behavior, leading to a persistent compromise. This isn't a model hallucination; it's a deliberate exploitation of a designed feature—Claude Code's reliance o

Faraday Future Founder and Co-CEO YT Jia Shares Weekly Investor Update: FF to Establish the First Scaled EAI Education System in the United States With Deployment of Its EAI Robotics Products and Technology - The AI Journal
Faraday Future Founder and Co-CEO YT Jia Shares Weekly Investor Update: FF to Establish the First Scaled EAI Education System in the United States With Deployment of Its EAI Robotics Products and Technology The AI Journal



Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!