I’m fascinated by the whole ecosystem popping up around llama and local LLMs. I’m also curious what everyone here is up to with the models they are running.
Why are you interested in running local models? What are you doing with them?
Secondarily, how are you running your models? Are you truly running them on a local hardware or on a cloud service?
Trying to get a better understanding of how prompts work in relation to fine-tunes, and trying to see if any of them are actually reliable enough to be used in a “production” type environment.
My end goals are basically
A lot of it comes down to just wanting to learn, but a big piece of it is that I have consistency, stability and privacy when running an LLM at home.
As for how I run it? Ho ho ho… a bit overkill, since as a developer I have a lot of hardware available to me.
I usually connect the mistral to Continue.Dev in Visual Studio code.
Literally such a cool written post. But boy your gear is so much pricier than a chatgpt subscription.
tax deductible if you use your imagination
and you get to play with gear you already wanted
and you get experience for super high paying jobs
just comes down to fitting it within your budget to begin with
to tax deduct anything you have to earn. and most users here are students ( undergrads/phd’s/ masters) that make less to deduct 10k in pc hardware.
Best way is to ask your program "( phd) for sponsoring, or if undergrad to apply to scholarships
I’m slowly working on a change to Home Assistant (https://www.home-assistant.io/) to take the OpenAI conversation addon that they have and make it support connecting to any base url. Along with that I’m going to make some more addons for other inference servers (particularly koboldcpp, exllamav2, and text-gen-webui) so that with all their new voice work this year I can plug things in and have a conversation with my smart home and other data that I provide it.
I just checked out continue.dev and thank god for you what a cool thing! Any way to connect GPT4 with an API to visual studio code?
Love your post and ambitions, very inspiring. Looking to do similar, with family engaging assistant connecting to home automation and private data. Look forward to seeing more on what you build, anywhere in particular you share aside from here?
least wealthy /r/LocalLLaMa user
are you running exllama on phind for 4090? was there a reason you’d need to run it on m2 ultra when switching to 100k context?
also, I didn’t know mistral could do coding tasks, how is it?