That seems pretty disappointing. It seemed to me like it could have been somewhat possible. I've trained a 0.8M parameters model and it was spitting out something that looked like French, not French though. So I need to test it but I feel like if I do it with some millions of parameters it could work. It still wouldn't have a coherence but at least it could form real sentences. Again I don't know much about this, so I'm surely wrong. I also think the dataset may be the issue, I didn't use a general purposed dataset, only French books in a txt file.
Matburnx
joined 1 year ago
I just did that and cleaned all the models I downloaded and didn't want to use anymore. Thanks for the advice :)
I used openplayground a little after it got released, and I installed it through pip so the models were downloaded somewhere in the Python files. I used a software that shows file size as taladar suggested and found it that way. Thanks for all the help, though :)
Oh ok, thanks, this one was pretty easy. Do you have any idea where the models downloaded by openplayground are?
Well, I use my laptop as a daily-driver, so training an AI in the background, even when I don't use it seems a bit complicated. The Markov chain seems like an interesting alternative for what I'm looking, does any tools to use one exist or should I build one from scratch?