Saying you wouldn’t use ‘ollama or gpt3.5’ is such a… I want to say uneducated statement? These are not two like terms
You’re aware that ollama isn’t an LLM? You’re aware there are LLMs available via ollama that exceed gpt4s capabilities? I mean, you’re right that you need an array of expensive gpus to run them effectively, but… Just comparing ollama to gpt-3.5 is like comparing an NCAA basketball star to the Harlem globe trotters. It’s ridiculous at its face. A player compared to a team, for starters.
Correct, i kept it simple on purpose and could probably have worded it better.
It was a meant as a broader statement including “both publicly available free to download models like those based on the ollama architectures as well as free to acces proprietary llm’s like gpt3.5”
I personally tried variations of the vicuna, wizardLM and a few other models (mostly 30B, bigger was to slow) which are all based on ollama’s architecture but i consider those individual names to be less known.
Neither of these impressed me all that much. But of course this is a really fast changing industry. Looking at the hf leaderboard i don’t see any of the models i tried. Last time i checked was January.
I may also have an experience bias as i have become much more effective using gpt4 as a tool compared to when i just started to use it. This influences what I expect and how i write prompts for other models.
I’d be happy to try some new models that have since archived new levels. I am huge supporter for self-hosting digital tools and frankly i cant wait to stop funding ClosedAi
Llama3-70b is probably the most general purpose capable open source
There are a bunch of contenders for specific purposes, like coding and stuff, though. I wanna say Mistral has a brand new enormous one that you’d need like 4 4090s to run smoothly.
Saying you wouldn’t use ‘ollama or gpt3.5’ is such a… I want to say uneducated statement? These are not two like terms
You’re aware that ollama isn’t an LLM? You’re aware there are LLMs available via ollama that exceed gpt4s capabilities? I mean, you’re right that you need an array of expensive gpus to run them effectively, but… Just comparing ollama to gpt-3.5 is like comparing an NCAA basketball star to the Harlem globe trotters. It’s ridiculous at its face. A player compared to a team, for starters.
Correct, i kept it simple on purpose and could probably have worded it better.
It was a meant as a broader statement including “both publicly available free to download models like those based on the ollama architectures as well as free to acces proprietary llm’s like gpt3.5”
I personally tried variations of the vicuna, wizardLM and a few other models (mostly 30B, bigger was to slow) which are all based on ollama’s architecture but i consider those individual names to be less known.
Neither of these impressed me all that much. But of course this is a really fast changing industry. Looking at the hf leaderboard i don’t see any of the models i tried. Last time i checked was January.
I may also have an experience bias as i have become much more effective using gpt4 as a tool compared to when i just started to use it. This influences what I expect and how i write prompts for other models.
I’d be happy to try some new models that have since archived new levels. I am huge supporter for self-hosting digital tools and frankly i cant wait to stop funding ClosedAi
Llama3-70b is probably the most general purpose capable open source
There are a bunch of contenders for specific purposes, like coding and stuff, though. I wanna say Mistral has a brand new enormous one that you’d need like 4 4090s to run smoothly.