Back Original

Weeknotes: Gigabucks, burning my Mac's GPU, and Karpathy's lament

Find previous weeknotes here.


A full and chaotic week. I got off to a raging start and made great progress at Recurse, but got slammed by some hard walls of bureaucracy in the second half. My mission is to become stronger, defy Nature, and have fun. Tax accounting is the opposite of all three of those…

Updates

RC hosts “Impossible Stuff Day” once per batch, a whole day to spend working on something well beyond the edge of your abilities that feels totally impossible. The last one wasn’t productive for me, but I took a different approach this time around and wrote a DevLog in real time as I participated.

Despite challenges with download speeds and privacy issues, I successfully set up a 3B-parameter LLM with GPU acceleration on my Mac through a custom command-line interface. I did the “llmpossible.”

I finally migrated my Goodreads data to my personal site. See the new books page.

You can also view the code and use it to liberate your own Goodreads data: gianlucatruda/goodreads-to-md: Convert Goodreads books to markdown files (Obsidian, Hugo, Jekyll etc.)

Inputs / outputs

Here’s a short demo video of me using webcam-aberration, but it looks nowhere near as good as the real thing because of all the compression artefacts, so please do try it yourself!

Ideas

Another LLM workflow for terminal dwellers

I don’t need Cursor, I just use

cat instructions.md | llm --model o1-preview >> response.md

Karpathy’s Lament

I coin Karpathy’s Lament: The current paradigm of large language model training converges on moderately smart tools that write obscenely cringe-worthy slop.

“Not fully sure why all the LLMs sound about the same - over-using lists, delving into “multifaceted” issues, over-offering to assist further, about same length responses, etc. Not something I had predicted at first because of many independent companies doing the finetuning”

— via this tweet

A chain of mathematical equivalence

Re-frame your limited perceptual lifetime

maybe bc with money it’s cool to be inefficient (to show off your surplus, eg $500 bottles at a club or whatever) but in programming/hacker culture it’s pretty much always cool to be efficient. also to me it brings up the q of “well, how much do you need to run the thing you want to make” vs like you need to max out how much memory/money you have, always need more. If you’re making Call of Duty you prob need a lot but Super Mario is tbh more fun and doesn’t need much. And building a fun Super Mario isn’t thaaat much easier with infinite memory. It’s hard to make it good and fun, but not because of memory constraints. There’s a lot of other bottlenecks that adding memory doesn’t automatically fix

— via his thread

Distribution shift in vibespace

This week I learned

(Copy-pasted from my #TIL-tagged notes in Obsidian from the past week.)


Thanks for reading! I’ve tweaked the format and emphasis this week. Please let me know what you think and what you found most interesting.