this post was submitted on 27 Jan 2025
209 points (90.0% liked)

196

17027 readers
749 users here now

Be sure to follow the rule before you head out.

Rule: You must post before you leave.

^other^ ^rules^

If you have any questions, feel free to contact us on our matrix channel.

founded 2 years ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] BB84@mander.xyz 3 points 4 weeks ago (2 children)

@jerryh100@lemmy.world Wrong community for this kind of post.

@BaroqueInMind@lemmy.one Can you share more details on installing it? Are you using SGLang or vLLM or something else? What kind of hardware do you have that can fit the 600B model? What is your inference tok/s?

[–] needanke@feddit.org 13 points 4 weeks ago* (last edited 4 weeks ago) (1 children)

Wrong community for this kind of post.

Nor really, 196 is a anything goes community after all.

[–] spujb@lemmy.cafe 5 points 4 weeks ago (2 children)

AI generated content is against the community rules see the sidebar :)

[–] SreudianFlip@sh.itjust.works 16 points 4 weeks ago (2 children)

I’m here for the performative human part of the testing. Exposing AI is human generated content.

[–] BB84@mander.xyz 7 points 3 weeks ago (1 children)

I just really hope the 2023 "I asked ChatGPT and it said !!!!!" posts don't make a comeback. They are low-effort and meaningless.

[–] SatanClaus@lemmy.dbzer0.com 5 points 3 weeks ago

True. This specific model is relevant culturally right now though. It's a rock in a hard place sometimes lol

[–] spujb@lemmy.cafe 4 points 4 weeks ago

just giving context to their claim. in the end it’s up to mods how they want to handle this, i could see it going either way.

[–] A_Very_Big_Fan@lemmy.world 6 points 3 weeks ago

Prompts intended to expose authoritarian censorship are okay in my book

[–] BaroqueInMind@lemmy.one 3 points 3 weeks ago* (last edited 3 weeks ago) (1 children)

I'm using Ollama, a single GPU with 10Gb of VRAM

[–] BB84@mander.xyz 1 points 3 weeks ago (1 children)

You're probably running one of the distillations then, not the full thing?

[–] BaroqueInMind@lemmy.one 1 points 3 weeks ago (1 children)

What's the difference? Does the full thing not have censorship?

[–] BB84@mander.xyz 1 points 3 weeks ago

That's why I wanted to confirm what you are using lol. Some people on Reddit were claiming the full thing, when run locally, has very little censorship. It sounds somewhat plausible since the web version only censors content after they're generated.