Discussion about this post

User's avatar
Michael Power's avatar

In a much recommended comment I wrote in last Friday's Financial Times to an article entitled: DeepSeek rival’s shares double in debut as Chinese AI companies rush to list (https://www.ft.com/content/a4fc6106-5a61-4a89-9400-c17c87fb1920#comments-anchor) I replied as follows:

You fundamentally misunderstand the emerging character of the Chinese LLM community. It is not so much competitive as 'co-opetitive'. Being Open Weight, they share architectural software improvements willingly whilst each individual LLM concentrates on a slightly different - yet complementary - area of expertise. What is emerging is a Dragon Swarm whose watchword is consilience. DeepSeek is the Architect Dragon whose Open‑Weight 'foundation model excellence' (rich in software design features willingly shared) will be massively reinforced when R2 drops mid February, not coincidentally coinciding with the advent of the Year of the Fire Horse. Deep Seek is the bedrock of the swarm - the 'Mother of Dragons' if you will. Aside from being the technical supremo, it is optimized for all-round reasoning and general intelligence. MiniMax is the Creative & Sonic Dragon, a specialist in multimodal creativity – text, voice, music and immersive content synthesis. Deep Seek and Minimax (and Qwen, Kimi, Ubiquant, 01.AI, ZiAI, Sensetime and more) are not so much rivals as members of a Dragon Swarm of Open Weight LLMs covering an extraordinarily wide range of expertises.

Expand full comment
John O'Neil's avatar

As a non-tech idiot who regularly gets in way over my head on vibecoding projects, I've developed a clunky method of consulting multiple models when the one I'm working with gets stuck or seems off base. I ask it to write up a memo describing the bug or strategic question or whatever and paste that into the other two of the ChatGPT, Claude, Gemini "Crew" and into a new instance of whatever one I'm working with (with the instruction that it's a naive model who should ignore any context that it comes across). Then I share the results with the model I've been working with -- and sometimes "fire" it and switch to working with another!

Is there a better way of doing the same thing? Meaning either likely to get better results or to take less time/effort. A lot of time the advice is great, especially on bugs. The most maddening things is that models won't tell me about available better solutions that I don't know to ask about.

Love your appearances with Jordan et al. Thanks!

Expand full comment
15 more comments...

No posts

Ready for more?