Pretty neat. The training process takes a while for textual inversion, which I have enjoyed playing around with. I hope Automatic1111 gets support for this method of training, if it takes off!
Great question, I wondered the same thing. I’ve got a decent knowledge base where stable diffusion (text to image etc) is concerned, and understand the applications of this Nvidia process, I’m not familiar enough with customization options for LLMs. I haven’t really seen references to hypernetwork/lora/midjourney type applications in LLMs, or anything that really “plugs into” your existing model to augment results, the way stable diffusion is geared for customization. It seems in my limited understanding, that customization for LLMs requires customization of the training ing data, and a completely new training process for the actual model, not a reference model like SD.
Pretty neat. The training process takes a while for textual inversion, which I have enjoyed playing around with. I hope Automatic1111 gets support for this method of training, if it takes off!
Can this be adapted to LLMs?
Great question, I wondered the same thing. I’ve got a decent knowledge base where stable diffusion (text to image etc) is concerned, and understand the applications of this Nvidia process, I’m not familiar enough with customization options for LLMs. I haven’t really seen references to hypernetwork/lora/midjourney type applications in LLMs, or anything that really “plugs into” your existing model to augment results, the way stable diffusion is geared for customization. It seems in my limited understanding, that customization for LLMs requires customization of the training ing data, and a completely new training process for the actual model, not a reference model like SD.