Most of the explanations that I think are in my brain are actually dotted lines in the shape of real explanations, but with very little inside them. They are ghosts of knowledge. And I won’t know what I don’t know until I fuck around and find out.
Very sweet recommendations model. Nice and simple.
The internet challenged copyright by creating infinite distribution of perfect copies. With AI, what happens when infinite distribution is hooked up to infinite imitation?
Another answer is not to cultivate the machine’s output but to compose through it, treating the model as material, choosing the inputs and shaping the rules. In other words, not working beside the machine, but stepping inside it.