- cross-posted to:
- [email protected]
- cross-posted to:
- [email protected]
People like this should not be a government representative. But I guess that’s the trend lately
Made the switch to DeepSeek fo’ life.
Fascist regime and power/police abuse has started
to be fair for black people that is a centuries old tune
Oh, you’re right
now i gotta download something i don’t even wanna download.
Yup. Downloaded 7b, 32b, and 70b varieties this afternoon. Entirely out of spite.
I literally just did the same
Your tax dollars at work!
I’m gonna download it even harder.
See you hell evildoer!
Sooner or later we’re gonna have to bypass the Great Firewall of America by VPNing into China
Print the code in a book and mail it.
Surely, they cannot ban books… right? Right?
Edit: Wait wait wait… the Comstock Act says mail cannot be used for anything that can be used for abortion. And a AI can theoretically be used to get instructions for abortion. BOOM, it’s banned! 👀
I wasn’t thinking of downloading an AI onto my low tier computer until now.
I’ve got a laptop kicking around from 2010 that’s about to get deepseek just because they’re proposing this dumb ass shit. I don’t even use Gen AI.
Finally affordable housing!
same lmao
Wow, bold choice to ban the import of technology and knowledge. Usually governments are worried about export, so it doesn’t fall into the wrong hands.
Btw, how is the Nvidia stock price doing?
Right? Like, seriously, we all know somebody is just butthurt because their stock options tanked.
Oh, wait, I’m sorry! That was very unpatriotic of me, wasn’t it? I mean, we all know that winning an election guarantees being heavily rewarded with insider trading, right? It’s not like they’re there to represent constituents or anything; I mean, doesn’t everyone know we’re a republic, not a democracy?!
Sigh…
Hawley’s statement called DeepSeek “a data-harvesting, low-cost AI model that sparked international concern and sent American technology stocks plummeting.”
data-harvesting
???
It runs offline… using open-source software that provably does not collect or transmit any data…
It is low-cost and out-competes American technology, though, true
sent American technology stocks plummeting
Oh yeah, thats what did it, totally
You don’t fuck with the big man money tbh… That’s like rule 1 of the game.
More protectionist bullshit that won’t help anyone.
For Base Model
git lfs install git clone https://huggingface.co/deepseek-ai/DeepSeek-V3-Base
For Chat Model
git lfs install git clone https://huggingface.co/deepseek-ai/DeepSeek-V3
this is deepseek-v3. deepseek-r1 is the model that got all the media hype: https://huggingface.co/deepseek-ai/DeepSeek-R1
Can you elaborate on the differences?
Base models are general purpose language models, mainly useful for AI researchers and people who want to build on top of them.
Instruct or chat models are chatbots. They are made by fine-tuning base models.
The V3 models linked by OP are Deepseek’s non-reasoning models, similar to Claude or ChatGPT4o. These are the “normal” chatbots that reply with whatever comes to their mind. Deepseek also has a reasoning model, R1. Such models take time to “think” before supplying their final answer; they tend to give better performance for stuff like math problems, at the cost of being slower to get the answer.
It should be mentioned that you probably won’t be able to run these models yourself unless you have a data center style rig with 4-5 GPUs. The Deepseek V3 and R1 models are chonky beasts. There are smaller “distilled” forms of R1 that are possible to run locally, though.
https://www.deepseekv3.com/en/download
I was assuming one was pre-trained and one wasn’t but don’t think that’s correct and don’t care enough to investigate further.
Is that website legit? I’ve only ever seen https://www.deepseek.com/
And I would personally recommend downloading from HuggingFace or Ollama
r1 is lightweight and optimized for local environments on a home PC. It’s supposed to be pretty good at programming and logic and kinda awkward at conversation.
v3 is powerful and meant to run on cloud servers. It’s supposed to make for some pretty convincing conversations.
R1 isn’t really runnable with a home rig. You might be able to run a distilled version of the model though!
You’re absolutely right, I wasn’t trying to get that in-depth, which is why I said “lightweight and optimized,” instead of “when using a distilled version” because that raises more questions than it answers. But I probably overgeneralized by making it a blanket statement like that.
Tell that to my home rig currently running the 671b model…
That likely is one of the distilled versions I’m talking about. R1 is 720 GB, and wouldn’t even fit into memory on a normal computer. Heck, even the 1.58-bit quant is 131GB, which is outside the range of a normal desktop PC.
But I’m sure you know what version you’re running better than I do, so I’m not going to bother guessing.
It’s not. I can run the 2.51bit quant
You must have a lot of memory, sounds like a lot of fun!
I’d get LM studio or Ollama, and download R1 your system can handle quick. If you’re on Linux, Alpaca is on Flathub, you can get it and it’ll download models and run them for you, including Deepseek R1.
Shouldn’t Google and Apple go to jail for distribution?
Neither Tim Apple nor Ted Google would make it a day in jail.
Tim is gay (fr), so I’d say he’d have an (comparatively) easier time than a straight man.
What a horrible thing to say