Bloefz

joined 1 month ago
[–] Bloefz@lemmy.world 1 points 1 week ago (1 children)

Thank you so much!! I have been putting it off because what I have works but a time will soon come when I'll want to test new models.

I'm looking for a server but not many parallel calls because I would like to use as much context as I can. When making space for e.g. 4 threads, the context is split and thus 4x as small. With llama 3.1 8b I managed to get 47104 context on the 16GB card (though actually using that much is pretty slow). That's with KV quant to 8b too. But sometimes I just need that much.

I've never tried the llama.cpp directly, thanks for the tip!

Kobold sounds good too but I have some scripts talking to it directly. I'll read up on that too see if it can do that. I don't have time now but I'll do it in the coming days. Thank you!

[–] Bloefz@lemmy.world 2 points 1 week ago* (last edited 1 week ago)

I don't think that's a problem. I live in Spain and speak Spanish daily with real people, many of them my friends. They'll correct me if needed, they often do. Though most are my own mistakes.

Don't forget people give wrong answers too. But people aren't available 24/7 to help me.

[–] Bloefz@lemmy.world 2 points 1 week ago (3 children)

Agreed. The way they just dumped support for my card in some update with some vague reason also irked me (we need a newer rocm they said but my card works fine with all current rocm versions)

Also the way they're now trying to sell cloud AI means their original local service is in competition to the product they sell.

I'm looking to use something new but I don't know what yet.

[–] Bloefz@lemmy.world 10 points 1 week ago (5 children)

I have one server with a cheap MI50 instinct. Those come for really cheap on eBay. And it's got really good memory bandwidth with HBM2. They worked ok with ollama until recently when they dropped support for some weird reason but a lot of other software still works fine. Also older models work fine on old ollama.

The other one runs an RTX 3060 12GB. I use this for models that only work on nvidia like whisper speech recognition.

I tend to use the same models for everything so I don't have the delay of loading the model. Mainly uncensored ones so it doesn't choke when someone says something slightly sexual. I'm in some very open communities so standard models are pretty useless with all their prudeness.

For frontend i use OpenWebUI and i also run stuff directly against the models like scripts.

[–] Bloefz@lemmy.world 6 points 1 week ago (4 children)

I mainly use it for Spanish which I have a basic proficiency in. It just accompanies me on my learning journey. It may be wrong sometime but not often. Like the other reply said, LLMs are good at languages, it's what they were originally designed for until people found out they could do more (but not quite as well).

And as for filtering, I just use it as a news feed sanitizer with a whole bunch of rules. It will miss things sometimes but it's also my ruleset that's not perfect. I often come across the unfiltered sources anyway and even if it misses something, it's only news. Nothing really important to me.

[–] Bloefz@lemmy.world 44 points 1 week ago (24 children)

I work with AI and use it personally, but I have my own servers running local models which solves tons of privacy concerns. The inaccuracy is another problem but not a big one for me as I know it and will simply fact check. Also, I don't really use it for knowledge anyway. Just to filter news to my interest, help with summaries and translation etc.

People use AI as some all-knowing oracle but an LLM is not meant for that at all.

[–] Bloefz@lemmy.world 1 points 1 week ago

Agreed. I lived in a small town and hated it. I need the buzz of a city, the opportunities, always things happening.

Also, in a big city you can find groups of people to fit with your identity however niche you are. In a small town it's the opposite, you have to adapt yourself to them to fit in, they are very monocultural. I can't do that.

[–] Bloefz@lemmy.world 6 points 1 week ago

Yeah the main reason they care is because abortion gives women agency, the ability to control their own lives. In the conservative ideology women are subservient to men.

[–] Bloefz@lemmy.world 1 points 1 week ago (1 children)

I know, but at least they are making a case which most news outlets probably would but don't have the means for a long legal fight. They might get some precedent out of this from which the whole news industry will benefit.

I don't think AI training should ever be fair use. These companies are making billions of other people's work and giving nothing back.

[–] Bloefz@lemmy.world 3 points 2 weeks ago* (last edited 2 weeks ago)

The problem I've been seeing with email on my own domain is that some services refuse it, saying "please enter a real email address" 🤬 some others just silently refuse to send a confirmation code so I can't register either (I think tinder did this). Especially the "not a real email address" really pissed me off.

And with proton I got "Anonimisation services are forbidden" once at least.

I forget which services, but it's Hella annoying..

The marketshare of Google and Microsoft on email is really becoming a problem.

[–] Bloefz@lemmy.world 1 points 2 weeks ago (3 children)

True, the NYT does seem to have a significant impact on OpenAI though.

[–] Bloefz@lemmy.world 4 points 2 weeks ago

Very very good point. Google's ubiquity makes that very difficult.

view more: next ›