Using LLMs for human-in-the-loop distillation in Prodigy
https://prodi.gy/docs/large-language-models
Prodigy comes with preconfigured workflows for using LLMs to speed up and automate annotation and create datasets for distilling large generative models into more accurate, smaller, faster and fully private task-specific components.
Source link
lol