My LLM stack

learning, rationality, ai

You have a lot of power at your fingertips in early 2026, with highly capable AI cheaply available. Base models like GPT-5.2 and Claude Opus 4.5 have expert-level subject knowledge in most STEM fields, and they can autonomously complete tasks taking almost a full work day. And that’s just for base models. There are also powerful agent scaffolding tools for base models – take Claude Cowork, for example.

Whatever tasks an AI can solve with your oversight, you can solve too. Many ordinary people can solve the same kinds of problems as certified therapists, experienced architects or ML PhD students with reasonable effort.

It’s well worth learning how to use LLMs efficiently, then.

It feels like we’re told this all the time, and Demis Hassabis eloquently made the point at the seminar at the World Economic Forum this week. Because ’new year, new me’, I thought I’d listen.

Using LLMs efficiently #

Using LLMs efficiently seems to involve two things: first, identifying good use cases for AI; second, using AI to implement the solution.

Step two, execution, is normally just some kind of prompt engineering – a skill I’d expect to become less important in the future: producing user-friendly products is largely about eliminating the need for sophisticated prompt engineering. Step one is harder: it requires creativity and attention to failure modes in one’s everyday life.

To learn to use LLMs efficiently, I decided to catalogue the main ways in which I use LLMs. I’ll stress that I’m almost certainly using AI suboptimally and that these use cases are specific to my lifestyle, so my list shouldn’t be taken prescriptively.

For proper, well-curated lists of LLM use cases, I recommend browsing through the links listed by Gavin Leech in his post on LLM usage. Feel free to share your own favourite AI uses, e.g. via email.

Anyway, here goes.

Tried-and-tested use cases #

  • I have a low barrier for asking LLMs ‘dumb’ questions regarding individual proof steps. Similarly, I frequently use LLMs as sanity checkers, writing a sentence summarising my understanding of a topic and having an LLM give feedback.
  • LLMs are excellent for explaining terminology in plain, simple language. Two caveats, though: for looking up words in foreign languages, I strongly prefer ordinary dictionaries, like those of NE; I’m also skeptical of using LLMs for definitions.
  • I sometimes have LLMs lecture me on specific topics in mathematics. This is particularly useful when you’re self-studying rather than following a course with peers – e.g. when you’re doing a semester paper. Concretely, this looks something like dumping relevant papers, textbook chapters and lecture notes into NotebookLM and asking for a 3-page PDF summary focusing on key intuitions.
  • Of course, I rely heavily on coding agents, Claude Code being my favourite. Claude can handle most programming-related tasks today – even things like slurm management1.
  • Routine email writing, like finding a time to meet.
  • Transcription of my favourite In Our Time episodes.
  • Just as Andy Masley, I also use AI for clearing ugh fields, like household chores. It’s not that I need AI for declaring taxes; however, it lowers the activation energy for the task. Or, to cite another example in this category, I used AI to repair my desktop lamp the other day.
  • Python scripts can save you a lot of time, and coding agents can whip up scripts in minutes. For example, I have scripts for creating Anki cards, for web scraping and for merging markdown files.

Experimental use cases #

  • I’m experimenting with using AIs to combat planning fallacy. I recently passed a project proposal through Claude, who categorically said that I was committing planning fallacy2. After some back and forth with the LLM, I came away with a far better project proposal. To extend this point, perhaps I could have an AI coach me in some areas of my life?
  • I’m considering taking LLM input for some major life decisions, e.g. career choices.
  • Some people, like Kaj Sotala, use AI as aids for introspection. Maybe worth trying?

Conclusion #

As is evident from my examples, today’s AI is powerful enough to be useful in almost any domain of life. This year, I’ll continue exploring novel AI use cases, and I expect this to improve my quality of life in some respects. However, I’m also curious to learn whether AI can actually improve one’s day-to-day, or whether the main boons from AI might come from things like accelerated scientific progress.


  1. It’s easy to see why programmers have notoriously short AGI timelines. ↩︎

  2. Interestingly, I fed other AI models the same prompt (GPT, Gemini and Perplexity) – but I only received real pushback from Claude. And I want feedback that stings a bit. ↩︎