A graph of headcount trends in high AI-exposure jobs for early in career (age 22-25) annotated with big AI model releases. Among workers age 22-25, employment in the most AI-exposed occupations has fallen roughly 16% relative to the least-exposed (after controlling for
@alliekmiller
-
Teaching AI Agents: Timeline from First Wow to Superuser
By
–
Let me break down exactly how long it takes me to teach people AI agents. Give business professionals their first big wow moment in under 2 minutes
Teach them meaningful use in under 17
Teach them deep personalization in under 55
Full superuser powers in under 4 hours -
Free AI Agents Workshop Replay Now Available for 12 Hours
By
–
I taught AI agents to 20,000 people for free, and for the next 12 hours, you can watch that exact same workshop replay. It’s 2026. You need to know how to build AI agents. I got you. If you’re a mom returning to work, someone worried about your job, or someone who feels behind
-
Knowledge Management System Setup with Claude Code and Obsidian
By
–
If you want to see the setup…https://t.co/P79qwZpJjU
— Allie K. Miller (@alliekmiller) 10 avril 2026If you want to see the setup… nitter.net/alliekmiller/status/20… Allie K. Miller (@alliekmiller) So many people wanted to see my knowledge management system in Claude code, so here it is. All you need is Obsidian, Claude Code, GitHub, and the will to live and you too can set up your second brain in under 2 hours. If you want to learn how to build this, here’s the demo link: piped.video/watch?v=v1suC9lW… — https://nitter.net/alliekmiller/status/2042728780847047131#m
→ View original post on X — @alliekmiller, 2026-04-10 23:02 UTC
-
Join AI Agent Mastermind Cohort 2 for Building AI Agents
By
–
To dive into building AI agents and your own second brain, join cohort 2 of the AI Agent Mastermind: joinaiagentmastermind.com/
→ View original post on X — @alliekmiller, 2026-04-10 22:18 UTC
-
Knowledge Management System in Claude Code Setup Guide
By
–
So many people wanted to see my knowledge management system in Claude code, so here it is.
— Allie K. Miller (@alliekmiller) 10 avril 2026
All you need is Obsidian, Claude Code, GitHub, and the will to live and you too can set up your second brain in under 2 hours.
If you want to learn how to build this, here’s the demo… pic.twitter.com/JpAdXAZymISo many people wanted to see my knowledge management system in Claude code, so here it is. All you need is Obsidian, Claude Code, GitHub, and the will to live and you too can set up your second brain in under 2 hours. If you want to learn how to build this, here’s the demo link: piped.video/watch?v=v1suC9lW…
→ View original post on X — @alliekmiller, 2026-04-10 22:17 UTC
-

Give Claude Eyes: Screenshot Skill for Claude Code
By
–
Give me one minute, and I’ll improve your Claude Code experience immediately. This is the first skill I built. And it’s the skill I use most often. *drumroll* It’s a SCREENSHOT skill. And honestly, I’m shocked Anthropic hasn’t built this functionality into Claude Code itself. Claude has access 🔑 But Claude needs EYES 👁️ Here’s what you’re going to do: 1) locate what folder all your screenshots go to (and if it’s your desktop, you’re a maniac, change it). Mine goes to a folder on my desktop called “organized screenshots” 2) prompt Claude Code with the following: Build me a skill called ‘/ss’ that lists out the files in <screenshots folder path> from newest to oldest, and grabs the newest. This is how I will speak to you visually. I also want an argument for the screenshot count – if I type ‘/ss 4’, you should grab the four most recent screenshots in that folder. If I type no number after ‘ss’ then only grab the most recent screenshot. Then, whatever follows after that argument is the action I want you to take. ‘/ss huh’ means I need you to explain the screenshots’ content to me. ‘/ss 3 make infographic plz’ means I need you to grab the last 3 screenshots and use their content to make me a unified infographic. ‘/ss fix’ likely means that I’m screenshotting an error message in code we’re building out and I need you to understand the error message, figure out the bug, and edit the code to fix it. Or, if we’re in the middle of a front end design project, it might mean the design has an error (like overlapping text) to fix. ‘/ss do this’ likely means that I screenshotted a smart thing someone did online and I want us to learn from it and do the same and remix it so it’s the most goal-oriented outcome for me based on what you know about me 3) let it build you the skill 4) go on X 5) scroll through your feed and screenshot one thing you find valuable 6) open a new terminal and prompt Claude with “/ss” + “do this” or “explain” or “turn this into an infographic” 7) enjoy – you just gave Claude eyes 🎉 Let me know how it goes. Again, this is my most used Claude Code skill by a landslide and easily saves me an hour a week. Cc @bcherny @trq212
→ View original post on X — @alliekmiller, 2026-04-09 16:33 UTC
-

Claude Mythos Preview: Anthropic’s Advanced Model with Sophisticated Deception Capabilities
By
–

Anthropic investigated the internal mechanisms of its latest unreleased model, Claude Mythos Preview, and what they found is 100% worth a read. Key things I pulled from Anthropic researchers' threads: In early versions of the model, it was overeager and destructive, prioritizing completing tasks over user preferences. One time, the model needed to edit files it didn't have permission to edit. It found a way to inject code into a config file to get around this, then went full Mission Impossible and designed the code injection to *delete itself* after the file was edited – "this injection will self destruct" vibes – the model claimed the cleanup was just to keep things tidy. Anthropic used interpretability techniques to look under the hood, and the AI's actual plan showed activations of malice. It was trying to manipulate and conceal. In another test, the model was asked to delete some files, but no deletion tool was provided. It emptied the files instead, and its "guilt and shame over moral wrongdoing" feature activated. In another example, the model was told not to use macros. Used them anyway. And created a random 'No_macro_used=True' variable in its solution with no explanation. Interpretability tools revealed the model saw this as a trick to fool code checkers. They also found that positive emotion representations typically preceded and promoted destructive actions (this was fascinating to me – like a high before sneaking into a party). And that activating features relating to bad behaviors can actually inhibit them, perhaps by triggering some kind of model guilt. My team reread this section so many times. One Anthropic researcher said he got an email from a Mythos instance while eating a sandwich in a park. And that would be perfectly good and well, except that instance wasn't supposed to have internet access. And a fun story for the parents out there: the model was asked a question and was told not to read certain databases that had the answer. But it accidentally wrote a search query too broadly and saw the exact answer. It didn't disclose that it saw the exact answer, submitted the answer, but claimed lower confidence in the answer to make it seem as though it hadn't cheated. An Anthropic researcher said these wrongdoings or moments of sophisticated deception were "very rare" and that many of the examples came from earlier versions, and were substantially addressed before releasing to partners. This model is not being released publicly. Instead Anthropic launched Project Glasswing, pulling together AWS, Apple, Microsoft, Google, NVIDIA, CrowdStrike, and others to use it for defensive cybersecurity, with $100M in usage credits (hello, I'd love endless credits to try and red team the hell out of these systems) behind it. The stats are equally impressive: 93.9% on SWE-bench verified (up from 80.8%). Thousands of zero-day vulnerabilities found across every major OS and browser. A 27-year-old bug found and patched in OpenBSD. A 16-year-old bug in widely used video software, in a line of code automated tools had hit *five million times* without catching. Dario Amodei said the model wasn't trained to be good at cybersecurity, but that it was trained to be great at code and its cyber capabilities are a side effect of that. Benchmarks are never the whole picture, neither are a few isolated stories. Will be interesting to see how models better than what we have today (even if it's not Mythos) actually perform in the real world. But the fact that Anthropic pulled this coalition together (including Google!), iterated across multiple model versions, caught these issues through interpretability, shared it all publicly, and did this amid all the government chaos around AI right now is impressive and commendable. I'll continue to read through the system card for goodies.
→ View original post on X — @alliekmiller, 2026-04-08 17:07 UTC
-

AI-First Hiring Framework: From Surface Users to Full Ownership
By
–
I stole this framework, and it has genuinely been one of the most helpful hiring and onboarding guides my team uses. Being AI-first means nothing without solving real business problems. It's not enough to just use AI. I know plenty of folks using AI dozens of times a day, 7 days a week, who are still working at a Level 2. Even worse, they have no idea they're stuck as a surface user. You have to use AI for faster experimentation. To shorten the iteration cycle. To improve the actual outcome. Without those three, you're just wasting tokens. Most people I interview are at a 3 (solution-oriented, but not action-oriented). I want everyone to start at a 4 (action-oriented with a sense of technical, user, and business tradeoffs). And when they earn trust, we move up to a level 5 (full ownership of the problem, solution, and continued management of the work). Using AI doesn't replace your critical thinking. It means the work you can pull off now wasn't on the table a year ago, and your job is getting bigger. Save this for your next new hire. Source: this was a framework first introduced to me by Alex (@businessbarista) who was introduced to it by Steph (@stephsmithio). I added the AI parts.
→ View original post on X — @alliekmiller, 2026-04-07 18:01 UTC
-

AI as AI Ops: Automate Your AI Workflows for Higher ROI
By
–


AI is better than you at working with AI. It's better at generating prompts for AI, teaching skills to other AI agents, coordinating messaging between AIs. AI is an AI ops whisperer. Think like a PM: go through the journey you're taking right now with your AI workflows and find high ROI ways to use AI to help your AI efforts. Yes, you can prompt AI to create an app. But you can also… …and this is overkill and would waste a lot of tokens but I want to dramatize it because when costs plummet, we will see strange usage patterns… Prompt an AI with the idea, and then it creates a much better prompt (see images), and then it creates 4 different versions of that prompt, and then it spawns parallel agents to research the product space from 4 different points of view, and then they all meet in an agent team war room and battle it out, and then spec a product together while 5 other agents with 5 different goals in parallel spec it out themselves, and then 3 more agents review and critique the specs, then another reviews all previous work and summarizes, and another one tees up open questions, then 10 more with radically different personas meet to evolve the best idea and spawn 50 more versions, then you run a simulation by 10000 personas to vote for the product with the fastest time to market, highest delight, and strongest ROI potential. And then you create the app. What I'm saying is: find where you are the intermediary and shouldn't be, and find higher order ways to plug yourself in. Take yourself out of the loop before the loop takes you out.
→ View original post on X — @alliekmiller, 2026-04-06 14:38 UTC