The Turing Test focuses on distinguishing between humans and computers in a text chat.
There are lots of other domains where it's interesting to compare styles. Do we make different mistakes in speech recognition? How easy is it to spot a chess AI masquerading as human?
Related Posts
I love how the CommonMark Spec has a test suite that's just a JSON array. It's really easy to test a library for compliance, and I've seen developers nerd-sniped into full compliance.
https://spec.commonmark.org/0.31.2/spec.json
It's interesting to see the "why not Rust?" discussions around the TypeScript news that they're using Go. It shows that Rust has reached a level of maturity that it's a default for some users.
Go does seem to be in a sweet spot for AOT languages with GC though.
Text to speech systems seem to have largely avoided the uncanny valley effect. I've encountered robotic sounding voices but it's way less unsettling than bad CGI.
I'm not sure why this is. Maybe looking at faces is just way higher bandwidth so more things can go wrong?