The Z.ai Playbook
Deep Dives
Explore related topics with these Wikipedia articles, rewritten for enjoyable reading:
-
Knowledge distillation
13 min read
The article describes Z.ai's technique of building three separate 'teaching models' and distilling them into GLM 4.5. Knowledge distillation is a foundational machine learning technique that readers would benefit from understanding to grasp how modern AI labs efficiently create unified models.
-
Tsinghua University
14 min read
Zhipu AI was founded by researchers from Tsinghua University in 2019, and the article discusses the close relationship between Chinese AI companies and academic institutions. Understanding Tsinghua's role as China's premier technical university provides context for the talent pipeline feeding Chinese AI development.
-
Open-source software
11 min read
The article explicitly discusses 'the role of open source in the Chinese AI ecosystem' as a key topic and mentions Z.ai's relationship with open-source coding tools. Understanding the philosophy and economics of open source helps readers grasp the strategic decisions Chinese AI companies make about model release.
Zixuan Li is Director of Product and genAI Strategy at Z.ai (also known as Zhipu 智谱 AI). The release of their benchmark-topping flagship model, GLM 4.5, was akin to “another DeepSeek moment,” in the words of Nathan Lambert.
Our conversation today covers…
What sets Z.ai apart from other Chinese models, including coding, role-playing capabilities, and translations of cryptic Chinese internet content,
Why Chinese AI companies chase recognition from Silicon Valley thought leaders,
The role of open source in the Chinese AI ecosystem,
Fears of job loss and the prevalence of AI pessimism in China,
How Z.ai trains its models, and what capabilities the company is targeting next.
Co-hosting today are , long-time ChinaTalk analyst, as well as of the Interconnects Substack.
Listen now on your favorite podcast app.
The Z.ai Model and Chinese Open Source
Jordan Schneider: Zixuan, could you introduce yourself?
Zixuan Li: Hi everyone, I’m Zixuan Li from Z.ai. I manage a lot of things, like global partnerships, Z.ai chat model evaluation, and our API services. If you’ve heard of the GLM Coding Plan, I’m actually in charge of that, too. I studied AI for science and AI safety at MIT, where I did research on AI applications and AI alignment.
Jordan Schneider: Let’s do a little bit of Zhipu AI’s backstory. When was it founded? How would you place it within the broader landscape of teams developing models in China?
Zixuan Li: Zhipu AI and Z.ai were founded in 2019, and we were chasing AGI at that time, but not with LLMs, but with some graphic network or graphic compute. We did something similar to Google Scholar called AMiner. We used that type of thing to connect all the data resources from journals and research papers into a database. People could easily search and map these scholars and their contributions. It was very popular at that time.
However, we shifted to the exploration of large language models in 2020. We launched our paper, GLM, in 2021. I believe that was about one year ahead of the launch of GPT-3.5, so it was a very, very early stage. We were one of the first companies to explore large language models. After that, we continuously improved the performance of our models and tried a new architecture. GLM is a new architecture, actually, but we’re going to explore more in the future.
I believe we became famous
...This excerpt is provided for preview purposes. Full article content is available on the original publication.
