☆ Yσɠƚԋσʂ ☆@lemmygrad.ml to Technology@lemmygrad.mlEnglish · 7 个月前DeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAIventurebeat.comexternal-linkmessage-square15linkfedilinkarrow-up139arrow-down11cross-posted to: [email protected][email protected]
arrow-up138arrow-down1external-linkDeepSeek-V3 now runs at 20 tokens per second on Mac Studio, and that’s a nightmare for OpenAIventurebeat.com☆ Yσɠƚԋσʂ ☆@lemmygrad.ml to Technology@lemmygrad.mlEnglish · 7 个月前message-square15linkfedilinkcross-posted to: [email protected][email protected]
minus-squareCriticalResist8@lemmygrad.mllinkfedilinkarrow-up4·7 个月前But imagine what you’ll be able to run it on in four more months. But yeah, it’s stretching the definition of consumer hardware a bit.
minus-squarepinguinu [any]@lemmygrad.mllinkfedilinkarrow-up1·7 个月前You can use the smaller models on (beefy) consumer hardware already. That’s something, right? 😅
minus-squareCriticalResist8@lemmygrad.mllinkfedilinkarrow-up3·7 个月前I want the full 1TB model running on my 10 year old linux laptop
minus-squarepinguinu [any]@lemmygrad.mllinkfedilinkarrow-up2·7 个月前Just put your persistent memory as swap. Easy
But imagine what you’ll be able to run it on in four more months. But yeah, it’s stretching the definition of consumer hardware a bit.
You can use the smaller models on (beefy) consumer hardware already. That’s something, right? 😅
I want the full 1TB model running on my 10 year old linux laptop
Just put your persistent memory as swap. Easy