Does DeepSeek* Solve the Small Scale Model Performance Puzzle?
icecreamtaco @ icecreamtaco @lemmy.world Posts 0Comments 209Joined 9 mo. ago

icecreamtaco @ icecreamtaco @lemmy.world
Posts
0
Comments
209
Joined
9 mo. ago
I tested out a Deepseek model the other day. It took one minute to generate text and used up all my context space in one message. Local consumer models and "small" server hosted models are probably different classes because for my home pc it was a big performance downgrade.