AI Dynamics

Global AI News Aggregator

@_yutaroyamada

  • ShinkaEvolve: LLMs with Evolutionary Algorithms for Scientific Discovery

    Robert Lange @RobertTLange from @SakanaAILabs on ShinkaEvolve — an open-source framework combining LLMs with evolutionary algorithms for scientific discovery, with insane sample efficiency. His thesis that current systems optimise solutions to fixed problems. Going forwards — real scientific discovery requires co-evolving the actual problems. By the way – NVIDIA GTC is coming and will showcase breakthroughs in physical AI, AI factories, agentic AI, and inference. Register for virtual GTC for free using my link: nvda.ws/4qQ0LMg and enter raffle to win a DGX Spark 😈

    → View original post on X — @_yutaroyamada, 2026-03-14 05:34 UTC

  • Doc-to-LoRA: Instant LLM Adaptation via Meta-Learned Hypernetworks

    Doc-to-LoRA: What if you could online distill documents into your LLM weights without training? 🚀 Stoked to share our new work on instant LLM adaptation using meta-learned hypernetworks 📷📝 Building on our previous Text-to-LoRA work, we doc-condition a hypernetwork to output LoRA adapters, improving the base LLM's effective context window. The hypernetwork is meta-trained on 1000s of summarization tasks and shows remarkable compression capabilities at low latency 📈 🧑‍🔬 Work led by @tan51616 with @edo_cet & Shin Useka at @SakanaAILabs 📷 Sakana AI (@SakanaAILabs) We’re excited to introduce Doc-to-LoRA and Text-to-LoRA, two related research exploring how to make LLM customization faster and more accessible. pub.sakana.ai/doc-to-lora/ By training a Hypernetwork to generate LoRA adapters on the fly, these methods allow models to instantly internalize new information or adapt to new tasks. Biological systems naturally rely on two key cognitive abilities: durable long-term memory to store facts, and rapid adaptation to handle new tasks given limited sensory cues. While modern LLMs are highly capable, they still lack this flexibility. Traditionally, adding long-term memory or adapting an LLM to a specific downstream task requires an expensive and time-consuming model update, such as fine-tuning or context distillation, or relies on memory-intensive long prompts. To bypass these limitations, our work focuses on the concept of cost amortization. We pay the meta-training cost once to train a hypernetwork capable of producing tasks or document specific LoRAs on demand. This turns what used to be a heavy engineering pipeline into a single, inexpensive forward pass. Instead of performing per-task optimization, the hypernetwork meta-learns update rules to instantly modify an LLM given a new task description or a long document. In our experiments, Text-to-LoRA successfully specializes models to unseen tasks using just a natural language description. Building on this, Doc-to-LoRA is able to internalize factual documents. On a needle-in-a-haystack task, Doc-to-LoRA achieves near-perfect accuracy on instances five times longer than the base model's context window. It can even generalize to transfer visual information from a vision-language model into a text-only LLM, allowing it to classify images purely through internalized weights. Importantly, both methods run with sub-second latency, enabling rapid experimentation while avoiding the overhead of traditional model updates. This approach is a step towards lowering the technical barriers of model customization, allowing end-users to specialize foundation models via simple text inputs. We have released our code and papers for the community to explore. Doc-to-LoRA Paper: arxiv.org/abs/2602.15902 Code: github.com/SakanaAI/Doc-to-L… Text-to-LoRA Paper: arxiv.org/abs/2506.06105 Code: github.com/SakanaAI/Text-to-… — https://nitter.net/SakanaAILabs/status/2027240298666209535#m

    → View original post on X — @_yutaroyamada, 2026-02-27 09:41 UTC

  • hibiki: MacOS app enabling hotkey text-to-speech and CLI integration

    hibiki: Have your coding agents speak to you 🎙️ Fun weekend hack building a MacOS app enabling hotkey text-to-speech & CLI via: hibiki –text "…" –summarize With smooth hooks/skills for Claude Code & Codex 🤖 🧑‍💻: github.com/RobertTLange/hibi… 📝: roberttlange.com/#posts/blog…

    → View original post on X — @_yutaroyamada, 2026-02-18 16:59 UTC

  • Tracing ChatGPT’s Awkward Japanese with SoftMatcha 2
    Tracing ChatGPT’s Awkward Japanese with SoftMatcha 2

    A perfect community use case for SoftMatcha 2: tracing the source of the weirdness in an LLM's Japanese. softmatcha.github.io/v2/ mora (@moratorium08) I felt something off about the Japanese phrase "〜という整理になります" that ChatGPT uses frequently, so I tried searching for examples using the recently popular SoftMatcha 2. It only matched in hearing/interview sentences. I wonder if it's legal or bureaucratic jargon (or is it a corpus issue?) — https://nitter.net/moratorium08/status/2023547512490185087#m [Translated from EN to English]

    → View original post on X — @_yutaroyamada, 2026-02-17 14:14 UTC

  • Genie 3 world model technology successfully applied to Waymo autonomous driving

    We researchers often say "let us build this cool new technology and there will be many ways the company/society can use it, such as <list>". Here's a great example of that happening! Fun to see Genie help Waymo!! Jack Parker-Holder (@jparkerholder) Super excited to see Genie 3 having impact in autonomous driving. We've talked…. a lot… about this type of application for world models, and its been incredible to see the Waymo team do things we previously dreamed of! Congrats @maxjiang93 and team!! — https://nitter.net/jparkerholder/status/2019808695169982887#m

    → View original post on X — @_yutaroyamada, 2026-02-07 02:37 UTC

  • Sakana AI Beta Tester Recruitment, Seeking Feedback Contributors
    Sakana AI Beta Tester Recruitment, Seeking Feedback Contributors

    【Sakana AI Beta Tester Recruitment 🐟🐠】 We are recruiting testers who can assist in the verification of our AI service currently in development. Although it is still a prototype under development, we aim to improve quality based on feedback from all of you. ▼ Apply here forms.gle/R6KqqSTvUpXGEyHw5 ※ We also plan to give original merchandise as gifts to those who cooperate with us 🐟 [Translated from EN to English]

    → View original post on X — @_yutaroyamada, 2026-02-06 00:43 UTC

  • AI Tinkerers Tokyo Meetup February 19 at Sakana AI Labs

    We are hosting the next edition of AI Tinkerers at @SakanaAILabs on 2/19 (global community of AI builders 200 cities, which now also has a Tokyo Chapter) Looking forward to your demos & talk proposals tokyo.aitinkerers.org/p/ai-t…

    → View original post on X — @_yutaroyamada, 2026-02-05 04:54 UTC

  • TED Talk Released: Competition Stifling AI Breakthroughs Discussion

    People can finally hear my full thoughts on this now that my TED Talk has been released today!!! ted.com/talks/llion_jones_ho… Sakib (@zsakib_) Funny that both Ilya Sutskever (@ilyasut), and Llion Jones (@YesThisIsLion) the co-inventor of the Transformer architecture, are sick of Transformers Both @SSI and @SakanaAILabs give their researchers freedom to pursue wild ideas beyond token stacking Even Jim Simons’ of RenTech, the only hedge fund ever to consistently return >>40%+ annually (sometimes even 152%) let their researchers spend HALF their time on working on whatever they wanted!! Reminds me of the foundations of Physics in the 90s and even now: everyone works on unfalsifiable “theories of everything” like string theory (bro you can’t even test it what are you doing?) Diversity of thought (i.e. research freedom) is how we get true innovation — https://nitter.net/zsakib_/status/1993883878361067988#m

    → View original post on X — @_yutaroyamada, 2026-01-29 04:28 UTC

  • Interview on Optuna History: From 2018 Challenges to Today
    Interview on Optuna History: From 2018 Challenges to Today

    Honored to be interviewed about the history of Optuna. It covers everything from our early challenges in 2018 to the mindset I carry into my work today. Full interview (in Japanese): findy-code.io/engineer-lab/i… Takuya Akiba (@iwiwi) Optunaの開発秘話についてインタビューして頂きました! 2018年当時のOptuna誕生の経緯と裏側、この経験の振り返りや学び、今の仕事にも繋がる自分の根底にある考え方、などを上手く引き出してまとめてくださってます。ぜひ! findy-code.io/engineer-lab/i… — https://nitter.net/iwiwi/status/2009079954907213846#m

    → View original post on X — @_yutaroyamada, 2026-01-08 17:00 UTC

  • Sakana AI’s Agent Wins Programming Contest Against 800 Humans
    Sakana AI’s Agent Wins Programming Contest Against 800 Humans

    pretty cool! Sakana AI (@SakanaAILabs) Our AI agent has achieved 1st place in a competitive optimization programming contest against over 800 human participants. Blog: sakana.ai/ahc058 In AtCoder Heuristic Contest 058, Sakana AI’s ALE-Agent took the top spot. For context on the difficulty of these challenges, an OpenAI agent secured 2nd place in the AHC world tournament last year. The task was a 4-hour production planning optimization challenge. While the problem setters anticipated a standard approach combining constructive heuristics and simulated annealing, our agent independently discovered a more effective strategy. It implemented a "virtual power" heuristic and a diverse neighborhood search that allowed it to escape local optima better than human experts. This was achieved through inference time scaling using multiple frontier AI models. The agent ran parallel code generation, analyzed the results, and iteratively refined its algorithms in real time. The total cost was approximately $1,300. This result suggests AI agents can now match top human experts in tasks requiring extended reasoning and original scientific discovery. Please read our blog for more details. We extend our deepest thanks to the host, @algo_artis, and @atcoder. We will continue to research AI as a partner that expands human exploration to discover solutions to complex real-world problems. — https://nitter.net/SakanaAILabs/status/2008195936917586416#m

    → View original post on X — @_yutaroyamada, 2026-01-05 22:24 UTC