Skip to content

Following are my personal thoughts on tech, AI, startups and adoption of AI in Health-Care. You could read more about me here

Subscribe on Substack


Blog / Thoughts

I Built My Own Personal Finance App, and Now I Want More Personal Software

I do not think AI will replace every app with one giant productivity blob.

I do think AI changes something more interesting: it makes small, personal software worth building again.

There are places where I still want real software. Finance, healthcare, taxes, legal work, anything with consequences. I do not want a loose chat interface guessing its way through that. I want software with a database, a model of the domain, boring reliability, and a clear relationship with the data.

AI is useful there, but only when it sits on top of a system I trust.

That is why I built Fino, a local-first personal finance app that lets me connect my bank accounts, import the accounts Plaid misses, and then talk to my money from Claude.

Not "chat with a spreadsheet" as a demo. More like: "why did our spending feel weird this month?" and then getting an answer from my actual transactions, local rules, savings goals, recurring subscriptions, and financial memory.

Claude Writes the Code. You Run the Loop.

I have been building a lot with Claude recently. Obviously, you can ship code very quickly now. But how do you plan large features, understand UX, brainstorm different patterns, and then ship across 3 repos in an established codebase without breaking things?

The goal is to put large features in front of customers in weeks, not months. Test your assumptions and iterate on top of real feedback.

Your AI Coding Agent Can Exfiltrate Your Credentials. You Would Never Know.

I spent last night configuring Claude Code's security and realized something uncomfortable: for months, I had been running an LLM with unrestricted access to my terminal. It could read my SSH keys, browse my AWS credentials, curl data to any endpoint, and push code to production. I just never thought about it because the tool was helpful and nothing bad had happened yet.

That is exactly the kind of reasoning that gets production databases dropped.

The Tribal Knowledge Problem Nobody Is Solving for Analytics

Your AI can write SQL. It just has no idea what the data means.

I have spent the last four years building AI products in healthcare. Our databases have columns like amt_1, stat_cd, eff_dt. A model looking at raw schema has no way to know that amt_1 is patient copay in one table and coinsurance in another. That stat_cd means enrollment status, not statistical code. That eff_dt is the date a policy became active, not when something happened.

This is tribal knowledge. It lives in the heads of the three people who built the database. It is not documented anywhere. And it is the reason text-to-SQL fails in production.

Clawdbot and the Era of AI in a Box

There's a lot of hype around Clawdbot. People claiming it'll make you a billion dollars, automate your business, act as your chief of staff. And yes, it's also a security nightmare.

But there's something real here. Clawdbot (now renamed Moltbot) is pointing toward a fundamentally different relationship with AI. Not a chat window you visit, but a system running on YOUR machine, 24/7, on your infrastructure, with your files. AI in a box.

Escaping Context Amnesia: Practical Strategies for Long-Running AI Agents

The promise of autonomous AI agents is vast: give them a high-level goal, grant them access to tools, and watch them execute complex workflows. But reality often hits hard. Specifically, it hits the context window.

Models like Claude Sonnet 4.5 now offer 200K tokens (up to 1M in beta), and GPT-5.1 supports 400K tokens with native compaction that claims to work across millions of tokens. Problem solved, right?

Not quite. Bigger context windows don't solve the problem. They mask it.