LLMs as general reasoners

In ML it’s common to measure generalization performance as the delta between the training, validation and test dataset. Training data is what your model actively optimizes against using some optimization procedure (e.g. SGD or conjugate gradient or quadratic programming). The validation data is used to get an unbiased estimate of the performance of the model on unseen data. Since the process of selection (hyper parameter selection, stopping criteria) invariably overfits the model to the validation data at the very end we get an estimate of model performance through the test dataset that should not be used for the process of training a model....

August 21, 2023 · 4 min · 702 words · Rick Lamers

Writing my own ChatGPT Code Interpreter

tl;dr: github.com/ricklamers/gpt-code-ui and to run it pip install gpt-code-ui && gptcode Alright, I’ll cut right to the chase. I’ve built GPT-Code UI because OpenAI couldn’t be bothered to give me access to their new fancy ChatGPT Code Interpreter. And most likely, you don’t have access either. What is ChatGPT Code Interpreter? Well, it’s basically a REPL (read–eval–print loop) that’s integrated with the coding abilities of ChatGPT. You can do really neat things with it like giving it a file (e....

May 17, 2023 · 9 min · 1736 words · Rick Lamers