A new 16-bit floating point format for machine learning! https://hub.packtpub.com/why-intel-is-betting-on-bfloat16-to-be-a-game-changer-for-deep-learning-training-hint-range-trumps-precision/
It increases range at the expense of precision, and often allows 16-bit computation (smaller, faster hardware) to replace 32-bit ML logic.
Related Posts
GNU Make defaults to a single worker, and newer build tools (e.g. ninja) default to the number of physical CPUs.I wish there was an option for 'leave me a little bit of my machine to do stuff'.
ASTs typically discard comments, and that's usually what you want.
The only time (AFAICS) that preserving comments is useful is for writing a code formatter.
Could you write a formatter in terms of a list of lexemes? A CST is a non-trivial bit of code for one use case.
Difftastic 0.67 is released!
* Added support for protobuf, improved bash and YAML.
* Updated 16 parsers to the latest upstream version (thanks Antonin Delpeuch!)
* A ton of display polish for word highlighting, alignment, and binary files.
https://github.com/Wilfred/difftastic/releases/tag/0.67.0