Two prominent Chinese AI startups, DeepSeek and Moonshot AI, unveiled reasoning models that rival OpenAI’s o1, and shared insights into large-scale reinforcement learning for training LLMs.
I am trying to determine what GPU resources DeepSeek has access to. All of the discussion over the past week has focused on H800, A100, and H100, but not on AMD....what is your view of this press release which talks about the "long standing collaboration" between AMD and DeepSeek?
Nice Tony. Just saw this. Great stuff. Have repacked it (since I had only done notes earlier). Yours has good relevant details.
ReStaCked. Hahah damn spellcheck
haha thanks for reading!
Great work Tony, I am reference this in new Substack post that will go up tomorrow. Wanted to see what you thought of this: https://www.amd.com/en/developer/resources/technical-articles/amd-instinct-gpus-power-deepseek-v3-revolutionizing-ai-development-with-sglang.html
I am trying to determine what GPU resources DeepSeek has access to. All of the discussion over the past week has focused on H800, A100, and H100, but not on AMD....what is your view of this press release which talks about the "long standing collaboration" between AMD and DeepSeek?
Seemingly AMD optimized its GPU to better run DeepSeek-v3 for inference?