View Single Post
Old 12-21-2023, 08:46 PM   #8
Comfy.n
want to learn what I want
Comfy.n ought to be getting tired of karma fortunes by now.Comfy.n ought to be getting tired of karma fortunes by now.Comfy.n ought to be getting tired of karma fortunes by now.Comfy.n ought to be getting tired of karma fortunes by now.Comfy.n ought to be getting tired of karma fortunes by now.Comfy.n ought to be getting tired of karma fortunes by now.Comfy.n ought to be getting tired of karma fortunes by now.Comfy.n ought to be getting tired of karma fortunes by now.Comfy.n ought to be getting tired of karma fortunes by now.Comfy.n ought to be getting tired of karma fortunes by now.Comfy.n ought to be getting tired of karma fortunes by now.
 
Posts: 1,641
Karma: 7908443
Join Date: Sep 2020
Device: none
I can only agree. Performance-wise there's a severe limitation on common hardware, measured as "tokens per second"... I remember some local models were painfully slow to process simple requests. And then there's that other limitation you mentioned on a similar thread "they can only be prompted with very limited amounts of text at the moment."
Comfy.n is offline   Reply With Quote