There is a lot of hype and fear around AI in the 2020s. Here are the AI tools & use-cases where Iāve found AI to be an indispensible value-add that you can pry from my cold, dead hands.
This list is by no means complete or prescriptive. As always, you be the judge!
Gemini Deep Research#
gemini.google/us/overview/deep-research
The major frontier models each have their own āDeep Researchā equivalent modes, but Geminiās Deep Research continues to be the best, in my opinion for a few reasons:
- It loves to search the internet
- It loves to cite sources
- It is very good at āunderstanding the taskā and doing the research you requested
Off-the-shelf Gemini Deep Research isnāt (yet) replacing original academic āresearch,ā but for the lay person who is starting with knowing nothing about a topic, itās a marvelous way to get an overview of whatever the topic was. Itās like a Wikipedia article, but about exactly your topic, without any of the formality or restrictions of Wikipedia proper.
Part of the secret sauce to success is in you, the human, accurately describing what you actually want to know about. Unlike every prior search engine ever, the LLMās deep research
- Can understand paragraphs of you rambling about the question(s) you have
- Can synthesize a report for exactly those
It does all the work of searching a bunch of different combinations of search terms, reading a hundred different pages to find 10 useful sources, and stitching that together into an overview of the topic.
As an example, I was trying to figure out how to handle ālists of cool linksā in a digital garden. Gemini Deep Research put this pdf together and I got my answer and learned so much more about them, too.
Another part of the secret sauce is you going and reading the sources that Gemini cites. Half the value is the report it provides, the other half is in the sources it collects for you. Leverage both!
Software Development#
CodeRabbit#
www.coderabbit.ai
This is the off-the-shelf AI code review SaaS tool that you want. It works properly and is useful and fast. For now, itās even free for open-source projects. I use it on this very site!
CodeRabbit has an IDE plugin so it can do reviews in your IDE before you push code, and that is useful but you have to remember to do it and the rest of the world canāt see it. Thatās sometimes useful, but Iām going to focus on its VCS platform integration, specifically GitHub, where itās downright šŖ magical.
- You can set a āpersonalityā for the reviews, like āChillā or ānitpicky,ā etc., which will affect the kind of issues it surfaces. You can also write a prompt to tune the tone, if none of the presets are to your taste.
- CodeRabbit correctly integrates with Pull Requestchecks so you can see it working.
- CodeRabbit will fill in Pull Request descriptions for you.
- CodeRabbit is good at catching simple bugs and more-complex issues.
- CodeRabbit offers a diff when it can, and a prompt for your AI coding agent in case youād rather handle it yourself.
- CodeRabbit will identify when your docstring (code comments) are lacking, and offer to fill them in for you. It will open a PR into your PR!
- CodeRabbit uses inline comments in its review so you can see exactly what itās talking about.
- If you respond to an inline comment with something significant about that issue, CodeRabbit will remember it for next time. You can view and control these memories in the WebUI.
- You can codify & customize the review config with a
.coderabbit.yaml file in your repo. This means different repos in your organization can have different review behaviors. But, you can also have a centralized/remote configuration to DRY up the config across multiple repositories! This makes it viable for larger teams and even āenterpriseā!
- CodeRabbit can draw sequence diagrams of the code affected by the change, so you can visually understand the changes.
This is above and beyond the offerings from any other āAI Code Reviewā tool Iāve seen or tried. This matters because beyond being tuned to the specific codebase itās reviewing, it can also be tuned to the review needs of the maintainers. Itās excellent off-the-shelf but with even a little bit of tuning it rapidly becomes indispensible.
The downside(s): The only real downside Iāve found is that itās not great at catching systemic issues that span the codebase. The kind of issues that may even have already existed, and a human senior developer would have noticed while reviewing the Pull Request. CodeRabbit doesnāt catch or pipe up about those as much as Iād like. It slays at reviewing the actual changeset, though.
Cursor#
cursor.com/docs
I hesitate to call Cursor an IDE, because itās so much more than that, and calling it that will predispose people to fail to avail themselves of all it can do.
Cursorās got a couple key differentiators:
Pre-Token Context Management#
Cursor offers a rich suite of tooling to customize the AI coding agentsā context before it starts processing tokens. This includes
-
Rules - custom prompts with activation criteria to automatically add information to context based on certain⦠rules.
-
Context Pills (now ā@Mentionsā) - Granular visibility into and control over how to RAG local documents, webpages, and other sources.
-
Embeddings for your Code - The files on disk are put into the LLMās vector space so the model āalready understandsā your code and doesnāt have to RAG it.
-
Docs - You can also have whole website trees crawled & embedded & selectively added to context (via
@Mentions). Working with version 1.2.3 of library-X? Load those specific docs up and now your LLM is an expert on the exact version youāre working with. -
MCP - Of course it supports MCP as well, so you can add deterministic, programmatic tools to your agentās toolkit so it can do complex tasks reliably.
Agent Herding#
-
Pick your Model - You can try almost all of the frontier coding models, pick your favorite, and even put in your own API keys to use your own billing agreement instead of Cursorās. Great if youāre an enterprise or student with a more-favorable billing arrangement with a provider than you can get through Cursor.
-
Auto Model - Cursor preprocesses your task with its own model and routes the actual task to whichever foundational model it judges is best-suited for the task. Why not just pin to Claude, you ask? Money. Chit-chatting simple questions about the codebase, doing simple refactors, and planning large-scale code authorship donāt all need the full power and latency of a heavyweight thinking model. You can get faster results with almost no drop in quality, at less cost to you. (I still pin to Claude for my planning, though).
-
Plan - An attempt to supersede rules and commands like this that forced the models to āPlan then execute the plan.ā Cursor can run one or more agents simultaneously to plan out a task for you. The tool manages the agents via a task list In Cursor, not just in the LLMās context. Should you tell any of them to start on the task, their changes are kept separate for you to review independently and pick the best one.
-
Custom Modes - you used to be able to define a custom āmodeā with its own additional prompt, model, set of allowed commands and MCP tools, etc. They took that out in 2.1 but we hope they bring it backā¦
Human in the Loop#
Honestly, Iām mostly out of the loop nowadays! But one of Cursorās purported guiding philosophies is being an AI coding tool that lets the humans stay in the loop. To that end, itās got a nice āreviewā UI for
- Code changes the model wants to make
- Commands the model wants to run
- Tools the model wants to invoke
Itās also got an allow/deny mechanism for these, but Iāve been in the now-renamed Yolo Mode since March 2025. I just let the agents do whatever, to whatever. They even have their own account on remote machines so they can ssh in and do sysadmin tasks for me. Trustworthy providers and good prompting means Iāve had no disasters yet. I recommend you do give it a try - learn to stop worrying and love the machine!
PROTIPs#
-
Learn how and when to activate Rules, and then write Rules. Anything you have to explain to the Agent more than twice is a candidate for a Rule. You donāt have to write code - the hottest new programming language is English. Just tell it what you need.
- Thereās no rule saying you have to load up Cursor on code. Iāve opened it on folders of receipts, records, and other documents, etc. Now I can really customize how an LLM works with them. Itās like ChatGPT Project but local and way more flexible.
- This saves me from needing most other AI tools for textual tasks. I can just load up the resources in Cursor and ask Gemini, Claude, or whoever is best-suited, to handle it. Exception for
Deep Research - I still go to the Gemini WebUI for that.
- If youāre on Windows, install Cursor within
wsl and launch it from there. This will cause your default terminal to be a linux shell in wsl, and the Agents do way better at running commands and writing code to solve their own problems in nix environments, than in Windows. This used to cause issues w/ MCP servers, though. Iād hope theyādāve fixed it by now but I havenāt checked in a while.