For the future! #3351
Replies: 3 comments
-
Interesting idea! With multiple models, however, either one model is loaded and one model is unloaded per conversation (which probably takes more time on loading and unloading than computation time), So this does not seem to be cost-effective, and splitting into multiple models leads to sparse weights. Perhaps you can try to target training on different parts of the model, starting with one part of the model at a time. Have you tried it on a small model? |
Beta Was this translation helpful? Give feedback.
-
But would it take more time when all of them are running on different server and connected to each other via internet or locally? |
Beta Was this translation helpful? Give feedback.
-
Mick might disagree here, but If you look at |
Beta Was this translation helpful? Give feedback.
-
Is it possible to make a hierarchical system of models where one model classify the prompt into categories and then choose that specific model to respond to that prompt and when a prompt does not fit into any category then it will default to a model which will access the Internet for more info and rate which source is reliable and run another model which would parse the info and categorise it and add the data to one of the existing models so this way it trains it self and since we run only the models required to answer then we save on compute and storage resources? Or am I just thinking too much?
Beta Was this translation helpful? Give feedback.
All reactions