Tagged
Anthropic
Observations
The hidden cost of context
Mar 2026Anthropic markets the 1M token context window as a capability milestone. And it is one. Being able to feed an entire codebase, a full document collection, or months of conversation history into a single API call is genuinely useful, especially when your whole value proposition depends on deep context, as mine does with VYNS. But for most of the last year, that capability came with a hidden trip wire. Once a request crossed 200K tokens, the entire call shifted into a premium pricing tier, a 2x multiplier that applied retroactively to the whole request, not just the overage. Anthropic removed that surcharge in March 2026. That's a good move. But the way it worked before and the way the change was communicated (a pricing page update, not an announcement) says something. Around the same time, they quietly adjusted how usage limits burn during peak hours. If you're building on weekday mornings Pacific time, your session capacity is used up quicker than at other times, while your weekly total stays the same on paper. There's no real-time visibility into this. A lot of users are frustrated and hoping Anthropic provides a dashboard or some real-time metric to view token burn and actual usage. What you ended up feeling was hitting a wall much earlier than the documentation implied. Anthropic has described the opacity of usage limits as a 'deliberate product decision.' Not sure why they went this route. The practical move is to build like the pricing will change again, because it will. Instrument every API call. Build cost ceilings by operation type. Never let a 1M context window become the default just because it's technically available. The capability is real. The bill is also real. Make sure you know which one you're actually using.
The most powerful AI tools are being rationed, and solo builders aren't in the first cohort
Mar 2026Claude Mythos leaked this week. Anthropic confirmed it's real, described it as a 'step change,' and said it's currently being tested with a small group of early access customers. That group isn't me. It probably isn't you either. This is new. A year ago, every model Anthropic shipped was effectively available to everyone with an API key on release day. The delta between what a funded enterprise customer could access and what a solo bootstrapped builder could access was close to zero. Mythos changes that. The most capable model ever built is being distributed on an invitation basis, tiered by relationship and use case, with general availability deferred while the cybersecurity implications get worked through. I don't think this is wrong. A model that can find and exploit software vulnerabilities faster than human defenders probably shouldn't be available to everyone immediately. The deliberate rollout makes sense. But I want to name what it means for the builder layer: the gap between what well-resourced teams can build and what solo founders can build just got wider, not because of money, but because of access. The most powerful reasoning and coding capabilities are going to the companies already in the room. The rest of us build with last quarter's model. One more thing: Anthropic left the announcement of a model with unprecedented cybersecurity capabilities in an unsecured, publicly searchable data store. The irony needs no elaboration.
LLM APIs can't read their own chat history, and that's a big gap
Mar 2026I structure my chats in Claude as departments for my startup, VYNS. I have a few chats that persist for personal use, and most other one-off chats I end up deleting after I gather the intel I need. But for VYNS, I keep a few persistent chats that are critical for my work as a solo founder. The most important of these are my Build logs, sequential chats where the previous one summarizes all key findings and stages as a prompt for the next to open and resume our work once the current chat reaches capacity. I also keep dedicated chats for Marketing, Philosophical, Competitor Research, Partnership Opps, and Local Resources. What I eventually want to build is an agent for each chat and an Executive Assistant agent that reads across and internally interacts with all of these based on my activity throughout the day, synthesizes findings, and reports back with a daily briefing highlighting my top priorities across the VYNS build and my personal brand. The limitation: no major LLM that I know of exposes read access to existing chat history. The threads are locked inside the web interface. This means I have to manually copy and paste references between chats and to my EA agent, as I largely do now. This isn't a small gap. It's a missing architectural primitive that would unlock an entirely new category of AI-assisted organizational design.