Block a user
WIP: Structured training data and output with Langchain + RELLM
Multimodal Model
when a message is sent to miku with an image, we can just have it infer an image description and offer it as context in the langchain prompt; #9
New Training Methods
probably could reduce the amount of training data coming from our chat relative to other datasets, or training for fewer steps - might be some overfitting going on
Inclusion of Other Datasets
get the schizodev guy on youtube to let me see his cunnyposting dataset lol
Multimodal Model
figure out if second model entirely is needed for vision, or if it’s possible to train vision models on datasets without images