Simon wrote an excellent post on the current state of the world in LLMs.
Logs
It will be interested to see if or when we hit scaling limits to training more powerful models and what our new bottleneck becomes. For now, there appears to be a lot of greenfield.
While not an entirely unique perspective, I believe Apple is one of the best positioned companies to take advantage of the recent improvements in language models. I expect more generic chatbots will...
A heartwarming exchange Your project has a youthful optimism that I hope you won’t lose as you go. And in fact it might be the way to win in the long run.
I tried out Llama 2 today using ollama. At first pass, it seemed ok a writing Python code but I struggled to get it to effective generate or adhere to specific schema. I'll have to try a few more...
It's hard to think because it's hard to think. - Github Copilot
Meta released Llama 2 yesterday and the hype has ensued. While it's exciting to see more powerful models become available, a model with weights is not the same as an API. It is still far less...
I've been playing around more with nix lately. I like what I've seen from it so far: declare dependencies and get an isolated shell with those dependencies. If distributed, the environment can be...
Some unstructured thoughts on the types of tasks language models seem to be good (and bad) at completing:
Experimenting with using a language model to improve the input prompt, then use that output as the actual prompt for the model, then returning the result. It's a bit of a play on the "critique"...