All Updates

All Updates

icon
Filter
Product updates
Zhipu AI releases GLM-4-Flash LLM for simple tasks
Foundation Models
Aug 28, 2024
This week:
Product updates
Burcon unveils canola protein isolate for egg replacement
Plant-based Dairy & Egg
Yesterday
Product updates
Bene Meat Technologies produces first cultivated burger at parity with premium beef
Cell-cultured Meat
Yesterday
Geographic expansion
Regulation/policy
The Better Meat Co. expands to Asia with Singapore approval for Rhiza mycoprotein
Plant-based Meat
Yesterday
Listing
Pony.ai files for US IPO
Auto Tech
Yesterday
Product updates
MineOS launches DSR management 2.0
Digital Privacy Tools
Yesterday
Product updates
LatticeFlow launches the first evaluation framework for EU AI Act compliance
Generative AI Infrastructure
Yesterday
Product updates
Cognitive3D launches custom dashboards for XR analytics
Extended Reality
Oct 16, 2024
Funding
Mindtech raises GBP 4.4 million in funding to develop AI vision technology
Extended Reality
Oct 16, 2024
Partnerships
Zscaler and Okta partner to enhance zero trust security initiatives
Identity & Access Management
Oct 16, 2024
Partnerships
Zscaler and Okta partner to enhance zero trust security initiatives
Next-gen Cybersecurity
Oct 16, 2024
Foundation Models

Foundation Models

Aug 28, 2024

Zhipu AI releases GLM-4-Flash LLM for simple tasks

Product updates

  • Zhipu AI, a Chinese provider of AI models, has announced the free release of “GLM-4-Flash,” a large language model (LLM) for simple vertical tasks that require quick responses and low costs.

  • The model supports multi-turn dialogue, web browsing, function calls, and long-text reasoning with a maximum context of 128K. Additionally, it can generate text at a speed of 72.14 tokens per second (~115 characters per second) and supports 26 languages including Chinese, English, Japanese, Korean, and German.

  • The company claims that GLM-4-Flash offers improved efficiency with greater concurrency and throughput while lowering inference costs through adaptive weight quantization, various parallelization methods, batching strategies, and speculative sampling at the inference level.

Contact us

Gain access to all industry hubs, market maps, research tools, and more
Get a demo
arrow
menuarrow

By using this site, you agree to allow SPEEDA Edge and our partners to use cookies for analytics and personalization. Visit our privacy policy for more information about our data collection practices.