1. Add unit tests to code that works just fine, thank you. 2. Find bugs :-/
miniblog.
Related Posts
One fun way of testing new AI models: take an existing codebase you have and just ask them to "review it and fix bugs".
In principle this should find more issues over time as models get smarter. I've found a few bugs this way at least.
I'm considering using WebP more: for example, my websites would load faster and browsers have supported it since 2020: https://caniuse.com/webp
That said, GitHub just added support in August 2025, so I suspect I'd still find use cases that don't work yet:
Should code completion offer private methods/fields?
If the user chooses a private method, you can offer a quickfix to change its visibility. It's clutter though.
I frequently find myself wanting fields that I haven't exposed yet, and frustrated that the IDE hides them.
