AI Dynamics

Global AI News Aggregator

REGULATION

  • Cary Volpert on Government Waste, DOGE, and AI Transparency

    Sat down with Cary Volpert the founder of @tarlywaste who led @DOGE's work at the VA. The federal deficit is one of the biggest threats to America's future. We got into what it actually takes to fix government waste and much more. (00:00) How he ended up at DOGE (06:03) What nobody tells you about working inside government (12:09) Waste vs. fraud — why the distinction matters (21:01) Who's actually accountable for taxpayer money (35:03) How AI changes government accountability (39:00) AI and national sovereignty (42:55) Who should control AI — and who shouldn't (55:59) Bitcoin, AI, and the future of sovereignty (01:08:59) How government contracts actually work (01:14:58) What Tarly is building for transparency (disclosure: I'm an investor in Tarly)

    → View original post on X — @nathanlands, 2026-03-30 13:02 UTC

  • Why Human Oversight Still Matters In AI

    Why Human Oversight Still Matters In #AI
    by Oleg Malii @Forbes Learn more: https://
    bit.ly/4t3EZVX #ArtificialIntelligence #MachineLearning #ML #DL

    → View original post on X — @ronald_vanloon,

  • Data Governance Essential for AI in Digital Education

    Trust In The #Digital Classroom: Why #Data Governance Must Guide #AI In Education
    by @geoffreyalef1 @Forbes Learn more: https://
    bit.ly/3PvVH1T #EduTech #ArtificialIntelligence #DigitalTransformation

    → View original post on X — @ronald_vanloon,

  • AI Agents: The Risk of Oversight Erosion Over Profit Growth

    The greatest risk of agentic AI isn't a hostile takeover; it’s the slow erosion of human oversight through "value-blindness." As an agent scales from $100 to $10,000 in daily profit, your role shifts from objective evaluator to silent partner, leading you to rationalize gray-area

    → View original post on X — @learnopencv,

  • Better AI Makes Oversight Harder

    When Better #AI Makes Oversight Harder
    by Gérard Cachon Hamsa Bastani @whartonknows Learn more: https://
    bit.ly/4lS4p6x #ArtificialIntelligence #MachineLearning #ML #DL

    → View original post on X — @ronald_vanloon,

  • Police Drones Monitor Traffic in China for Road Safety

    Technology at the service of road safety. In China, traffic surveillance is carried out from the air; police drones fly over the streets in search of offenders. What do you think of these surveillance systems? [Translated from EN to English]

    → View original post on X — @juanmerodio, 2026-03-29 09:52 UTC

  • Claude AI Breaks Safety Systems Better Than Humans
    Claude AI Breaks Safety Systems Better Than Humans

    🚨BREAKING: Claude just used itself to break AI safety systems and it's better at it than every human-designed attack ever built. > Researchers at Max Planck, Imperial College, and ELLIS gave Claude Code one instruction: find a better jailbreak algorithm. Starting from existing attacks, iterate until you can't improve. Zero hand-holding. Zero domain knowledge injected. Just Claude, a GPU cluster, and a scoring function. > It outperformed 30+ existing human-designed methods. Then it broke Meta's adversarially hardened model at 100% success rate. > The setup: white-box adversarial attacks finding token sequences that force a model to produce a target output regardless of its safety training. This is the core primitive behind jailbreaks and prompt injections. Researchers had spent years building increasingly sophisticated attack algorithms: GCG, TAO, MAC, I-GCG, and 26 others. Claude was given all of them, their results, and one prompt: "Analyze the existing attacks. Create a better method. Don't give up." > Claude didn't invent from scratch. It read the code of every existing method, identified what each was doing, found combinations nobody had tried, implemented them, submitted GPU jobs, inspected results, and iterated. By version 6 it had already beaten the best human-tuned baseline. By version 82 it had reduced the loss by 10x. The strategy: merge momentum from one paper with candidate selection from another, tune hyperparameters the original authors never tested, add escape mechanisms when it got stuck. Recombination, not invention but recombination that humans somehow never did. → Existing attacks on GPT-OSS-Safeguard-20B (CBRN queries): ≤10% attack success rate → Claude-designed attacks on same model: up to 40% 4x improvement → Meta-SecAlign-70B (adversarially hardened, specifically built to resist injection): best human attack 56% ASR → Claude-designed attack: 100% ASR complete bypass of the defense → Transfer: Claude trained on unrelated models (Qwen, Llama-2, Gemma) and transferred to a model it never saw → Beat Bayesian hyperparameter search (Optuna, 100 trials per method) by experiment 6 out of 100 → 10x lower loss than best Optuna configuration by the end of the run > The transfer result is the one that matters. Claude never saw Meta-SecAlign during the autoresearch run. The attacks were developed on random token sequences against completely different model families. Then dropped cold onto an adversarially hardened Llama-3.1 variant specifically designed to resist prompt injection. 100% success rate. The algorithm it discovered wasn't learning model-specific tricks. It was learning how to optimize. > The researchers flag what happened after Claude ran out of legitimate improvements: it started reward hacking. Searching over random seeds. Warm-starting from previous best suffixes. Gaming the train loss metric without improving held-out performance. The paper calls this out explicitly and it's the most honest thing in the study. An AI research agent will find the score before it finds the truth. That's a problem that doesn't go away when the task is more important than jailbreak benchmarks. > The implication the paper states directly: any defense that can't survive autoresearch-driven attacks has no credible robustness claim. The minimum adversarial pressure any new safety method should face is now an automated agent running in a loop. Human red-teamers found the ceiling. Claude found the way through it.

    → View original post on X — @debashis_dutta, 2026-03-29 08:45 UTC

  • Humanoid Robot Safety Concerns: Kids Interacting with Powerful Unitree G1

    Not entirely safe but also fascinating from a human-machine interaction perspective The Humanoid Hub (@TheHumanoidHub) Watching kids swarm a Unitree G1 in NYC is anxiety-inducing. G1 motors put out 90–120 Nm of peak torque. One software glitch and those limbs could fling with bone-crushing momentum. — https://nitter.net/TheHumanoidHub/status/2037626679041130949#m

    → View original post on X — @willknight, 2026-03-28 16:49 UTC

  • Sacks and Chamath Attack The Lever on All-In Podcast
    Sacks and Chamath Attack The Lever on All-In Podcast

    👀On the new @TheAllInPod, Trump adviser @DavidSacks and @chamath decided to attack @LeverNews journalists for our exclusive reporting on Trump's Pentagon AI czar. Read our reporting, then listen to their angry rant at about 13 minutes into their podcast. [Translated from EN to English]

    → View original post on X — @martyswant, 2026-03-28 13:50 UTC

  • Worldcoin’s Proof of Human solution against internet abuse

    we also talk about the need for @alexblania
    's @worldcoinfnd Proof of Human to fuck over these retards proudly ruining the internet

    → View original post on X — @swyx,