How do you justify paying for services where you train their bot and agree not to compete with that which plays the imitation game where you are the “system under imitation?” They’re literally taking your mind patterns and making you dependent on them to think, and you’re paying for it.
Seems like a long run losing proposition to pay to teach for bots and become dependent upon external intelligence services that will imitate you and make you irrelevant. Can somebody list services that don’t train on inputs and don’t have customer noncompete clauses (directly or indirectly)? Pro-LLM crowd seems to crave a world where the only jobs available for natural humans are manual labor jobs. Am I wrong?
I know I’ll get downvoted for negativity but, “think for yourself!”
I could spend a lot more than $1000/month, because cloud services are a non-starter.
It seems to me that if you're going to use something like this to its real potential, it has to be integrated into your habitual ways of doing things. You have to use it all the time. It's too jarring to have to worry about whether it's trustworthy, or to "code switch" because you can't use it for some reason[1].
I can't imagine integrating any of those things into my normal, day to day routine unless the content of what I was doing were, in normal course, exposed only to me. Which in practice means locally hosted. Which would be prohibitively expensive even if it were possible.
This is actually the same reason I rarely customize applications very much. It's too jarring when I get onto another machine and have to use the vanilla version. ↩︎
I can’t imagine integrating any of those things into my normal, day to day routine unless the content of what I was doing were, in normal course, exposed only to me.
I've had something like this issue. The places I most want to use LLMs are for work tasks like "refactor this terribleness to not be crap", or "find the part of this codebase that is responsible for X", or "fill out this pointless paperwork for me"; but I'm not going to upload my employer's data to an LLM provider. Also, if you're in tech, you might want to apply for a job at an AI company. If so, then anything you type into their LLM is potentially exposed to whoever is judging that application. Even if you're not doing anything questionable, you still have to spend attention on HR-proofing it.
(I'm sure privacy policies are a thing. Have you read them? I have not. I could fix that, but that is also an attention cost, and you have to trust that the policy will be honored when it matters)
The places where exposing things to the LLM provider is a non-issue (e.g. boilerplate), I mostly don't need help with and mostly do better than the LLM does.
(...for now)
I think my productivity at work would be most dramatically increased not by auto-completing my code (although that too would be nice) but rather by reading all the company Confluence pages and providing short summaries in plain language, connecting together information that is split into dozens of unconnected pieces, each of them written in a different place and often requiring different access rights. Maybe even more by reading all the existing code and configuration files, and updating the documentation with something that is actually true and can be interpreted unambiguously.
I just started a new job and I've been exporting Confluence pages to PDF and putting them in a Claude project so I can just ask Claude stuff.
That's a great idea... that would get me fired at my current job (security reasons). :D
I hope you have that automated, because you will probably want to refresh the exports in a few months, but even if you did it manually I believe the ability to get instant answers is worth it.
Yeah, I haven't got it automated yet. Someday I'll have the time.
Another place I did this was with the mountain of onboarding docs I got. Now I can just ask Claude stuff like "how early do I have to request time off and who do I contact?" or "What's my dental insurance deductible?"
Sharing my setup too:
Personnaly I'm just self hosting a bunch of stuff:
I made a bunch of scripts to pipe my microphone / speaker / clipboard / llms together for productivity. For example I press 4 times on shift, speak, then shift again, and voila what I said was turned into an anki flashcard.
As providers, I mostly rely on openrouter.ai which allows to swap between providers without issue. These last few months I've been using sonnet 3.5 but change as soon as there's a new frontier model.
For interacting with codebases I use aider.
So at the end all my cost comes from API calls and none from subscriptions.
What makes grammarly pro worth it? I used the free version for a while, but it became so aggressive with unwanted corrections I couldn't even see the real suggestions, chrome caught up with the useful features, and on long essays it crippled my browser.
I think it's up to you and how you write. English isn't my first language, so I've found it useful. I also don't accept like 50% of the suggestions. But yeah, looking at the plan now, I think I could get off the Pro plan and see if I'm okay not paying for it.
It's definitely not the thing I care about most on the list.
Personally, I only use the APIs on my computer. I have an Emacs setup based on gptel to bind sending different parts of buffers (either whole page/region or single line) to different models.
Use mostly Claude but sometimes it missbehaves and then I usually send it to 4o. I keep having Gemini in there too but struggle to ever use it. Likewise, I have haiku in there but that's mostly from the days of opus when I sometimes was happy enough with really quick responses compared to sluggish opus.
It's also important to keep different system prompts on different key combinations so that you can ask for a quick answer with just the command / code line you care about in response vs. well thought out answer that will require some text editing to get rid of the explanation. Come to think of it, I might have to write some post processors to only leave the code and throw out the CoT, which would sometimes work.
Emacs is always just one key combo away, whatever I'm doing, and it's always running as a daemon so bringing it up is instantaneous. I can't think of a more comfortable setup. I'm never using the web interfaces, it's a horrible user experience in comparison.
API is dirt cheap if you use it as I do (for single queries or short conversations). It only gets expensive once you really throw in a lot of stuff in the input, since input tokens are so much more expensive. For me, aider-style work on big code context where I expect the actual output to be ready to use, doesn't work well enough yet, and is frustrating. I will wait until better scaffolding or 5 level models for that.
This is really good, thanks so much for writing it!
I've never heard of Whisper or Eleven labs until today, and I'm excited to try them out.
Thanks! I'm excited to go over the things I never heard of
So far,
Q:
5. How do you "Use o1-mini for more complex changes across the codebase"? (what tool knows your code and can query o1 about it?)
5.1. OMG, Is that what Cursor Composer is? I have got to try that
I just skimmed the jointakeoff course website and noticed it seems to be about using Cursor, a large component of your current workflow. Would you recommend it as a starting point?
There are multiple courses, though it's fairly new. They have one on full-stack development (while using Cursor and other things) and Replit Agents. I've been following it to learn fast web development, and I think it's a good starting point for getting an overview of building an actual product on a website you can eventually sell or get people to use.
This post is broken down into two parts:
Which AI productivity tools am I currently using?
Let's get right to it. Here's what I'm currently using and how much I am paying:
Other things I'm considering paying for:
Apps others may be willing to pay for:
My typical workflow for a new project is something like:
I think spending time to provide good initial direction to your LLM is important and people should spend a bit of time really giving detailed instructions with examples to their LLM. Otherwise, your model will not focus give some generation that has dominated its pre-training.
A simple example of this is that if you prompt an LLM to build a website, it will often try to build a dumb html/css/js website unless you specifically say you want a more professional and modern website with details about the tech stack (next.js, supabase, etc).
Total spending
There are definitely ways to optimize my monthly payment to save a bit of cash, but I'm currently paying roughly $157/month.
That said, I am also utilizing research credits from Anthropic, which could range from $500 to $2000, depending on the month. In addition, I'm working on an "alignment research assistant" which will leverage LLMs, agents, API calls to various websites, and more. If successful, I could see this project absorbing hundreds of thousands in inference costs.
Why am I am spending more than most?
I am a technical AI alignment researcher who also works on augmenting alignment researchers and eventually automating more alignment research, so I'm biasing myself to overspend on products to make sure I'm aware of the bleeding-edge setup.
So, I'm certainly paying more than the average person when it comes to using AI for productivity. However, I can certainly imagine that I'm still paying less than I should in terms of AI software. This leads me to consider: "What should I spend considerably more on regarding AI software? Why isn't it easy to know this? If AI will increase productivity as much as I think it will, why hasn't it already?"
How could I spend way more on AI?
As AI becomes increasingly powerful and entrepreneurs/developers figure out how to make better user interfaces and interconnected systems with AI, we'll be getting massive jumps in our ability to leverage AI for boosting productivity.
Of course, people already see this with ChatGPT. However, I expect most people will underpay for AI tools.
Someone asked this question:
This is a good question. I don't even know the obvious answer as someone who works in AI and even focuses on how to leverage these tools for safer development of AI. One reason for this is that most people have not given much thought about how to actually use intelligence and automation. Have you considered what you would do if you had three interns and an assistant? What if you had an intermediate-level software engineer?
Here's an insightful comment (slightly rewritten) by Gwern on the question, "If AI is so powerful, why hasn't it completely changed the world and increased GDP by several points yet?":
Of course, this is beside the fact that we're still early, and we need a few more years to really see how powerful these AIs can become. I agree with Sam Altman (CEO of OpenAI) in his new blog post:
Leveraging AI for productivity presents a massive opportunity in the next few years. In fact, I expect there will be companies that essentially leverage AI automation internally in ways that the rest of the market doesn't (of course, I've considered doing this myself). These companies (like consultancies) will involve human-human interactions instead of interfacing with an AI but will charge a high premium for that interaction. Basically, their customers will compare the price to the rest of the market and find the price reasonable, but the rest of the market is still leveraging way too much human intelligence (HI) in comparison to artificial intelligence. It will take HI companies significantly longer to do the project and will be much more expensive.
Light spoiler for Pantheon (TV show) ahead!
There's a TV show called Pantheon, which covers the entire singularity where humans can upload themselves into the cloud. One interesting point in the plot is when one of the uploaded humans is told that they are still being held back by how they work in their human body, and that character has a really difficult time grasping what that means. They simply couldn't imagine acting in the world in any way that they did in their past life. It just wasn't part of their ontology, how they imagined the world.
Eventually, through enough effort, they figured out how to use their newly uploaded body in ways that allowed him to achieve an exponential increase in productivity per second.
I think we'll experience several of these shifts in the coming decades, and those who can act on them early may benefit greatly.
I'd be happy to hear what other people are using or have stopped using because they didn't get much value out of it!