Bert, a machine learning model for understanding natural languages, is showing some impressive results for parsing Google search terms:
https://blog.google/products/search/search-language-understanding-bert/
Related Posts
One advantage I've come to appreciate about Dash/Zeal docsets: it's really nice having focused search.
The text search is constrained to the languages I care about enough to download the docset, substantially increasing the relevance. In Google I'd need to specify the language.
I'm experimenting with diagnostics formatting.
* I've added a left margin, showing both the file name and line numbers
* I'm showing one line of context above/below the offending line.
* I'm using grey for comments.
What do you think? Is there anything you'd change?
LLMs are a really accessible machine learning technique. I dabbled with text classifiers a few years ago and the APIs were way more involved.
(system_prompt: String, input: String) -> String
I can prototype with this much more easily!