You must log in or register to comment.
Deepseek referred here seems to be v3, not r1. While the linked article didn’t seem to have info on parameter size, fact that they state it is sparse MoE architecture should suggest it is capable to run pretty quick (compared to other models of similar parameter space), so that’s cool.
Sweet! Where can I try it? for whatever reason im getting shitty search results looking for this
not sure if you can sign up for it outside China https://team.doubao.com/en/special/doubao_1_5_pro
ty :), maybe some random VPN will do if so
also worth noting that you can just run DeepSeek locally https://dev.to/shayy/run-deepseek-locally-on-your-laptop-37hl