A solution is to self-host an open-sourced or custom fine-tuned LLM. Opting for a self-hosted model can reduce costs dramatically - but with additional development time, maintenance overhead, and possible performance implications. Considering self-hosted solutions requires weighing these different trade-offs carefully.
A rough analogy to the current LLM process is that making a new model is like baking a cake. You figure out your data and algorithmsālike mixing the batterāand then you pretrain the model, that is, run it on a large number of computers for several monthsālike putting it in the ovenāand then at the end you do some āpost trainingāālike frosting and d... See more
they could try āswitching to a different model, augmenting the training data in some way, collecting more or different kinds of data, post-processing outputs, changing the objective function, or something else.ā Our interviewees recommended focusing on experiments that provided additional context to the model, typically via new features, to get the... See more