I know other people have said this too, but it is true: that model's source code is remarkably lean considering the complexity the tool can handle. This is really cool!

I like that there's pressure to be any degree more transparent. Grok is a sparse model, only 8-bit precision weights, but even this is more information than what GPT-4 has presented. I'd love to see the performance after tacking on the Hermes function-calling ( https://github.com/NousResearch/Hermes-Function-Calling ). gotta stock up on popcorn, though, because who knows what releases in the coming weeks.