I obtain a whole lot of fashions, and I can't maintain testing each them extensively to determine which one I like. With all of the Steady Diffusion, Textual content era and Audio era fashions, I’m certain I’m quickly changing my SSD 🥹
I begin with 8K tokens since some mannequin have 8K context window measurement. Anybody that fails it, get instantly deleted. I stored growing the context measurement round 24K. Even when a mannequin fails, I maintain it if it's artistic and humorous. I believe 24K is sweet context measurement. Llama-3-7B hits the candy spot. It's good, humorous, artistic, exact, and verbose to deal with most use circumstances. It matches properly in my 24GB of VRAM and a few variations have 250K context measurement (I noticed one variant dropped at this time with 1Million tokens 🤯!! ). Which implies, I can use it to write down tales, summarize textual content, write lengthy studies, and doing it at no cost. Here’s a pattern. I examined the 32K variant with 32K tokens: It nailed it! It acknowledged that the password was misplaced, it explains the place it’s positioned within the textual content, supplied me with the explanation why it was misplaced.
Which implies, it remembered your complete context window and it thought of it!!! that is enormous. I'd like to listen to your experiences and opinions too. submitted by /u/Iory1998 to r/LocalLLaMA |