All Updates

All Updates

icon
Filter
Product updates
Tencent introduced VoCo-LLaMA for compression lengthy of vision tokens
Foundation Models
Jun 24, 2024
This week:
Partnerships
MGI Tech and Predica Diagnostics collaborate to develop Predica's targeted RNA sequencing tests
Precision Medicine
Jul 11, 2024
Partnerships
Renalytix partners with Steno Diabetes Center to advance precision medicine solutions for diabetes and chronic kidney disease
Precision Medicine
Jul 11, 2024
Partnerships
Myriad Genetics and Personalis cross-license patents for tumor-informed cancer treatment tests
Precision Medicine
Jul 11, 2024
Funding
Element Biosciences raises USD 277 million in Series D funding to commercialize AVITI DNA sequencer and support AVITI24 launch
Precision Medicine
Jul 11, 2024
Funding
Jacobi Robotics raises USD 5 million in seed funding to expand capacity
Smart Factory
Jul 11, 2024
M&A
Formlabs acquires Micronics for undisclosed sum to develop next-generation SLS printers
Additive Manufacturing
Jul 11, 2024
Partnerships
Anthropic partners with Amazon Bedrock to fine-tune Claude 3 Haiku model
Foundation Models
Jul 11, 2024
Funding
Lemon.markets raises EUR 12 million in funding to support growth
FinTech Infrastructure
Jul 11, 2024
M&A
Kipu Quantum acquires PlanQK platform to enhance accessibility to quantum computing solutions
Quantum Computing
Jul 11, 2024
Partnerships
Quantinuum and STFC Hartree Centre partner to enhance quantum computing accessibility in UK
Quantum Computing
Jul 11, 2024
Foundation Models

Foundation Models

Jun 24, 2024

Tencent introduced VoCo-LLaMA for compression lengthy of vision tokens

Product updates

  • Tencent, has introduced VoCo-LLaMA, an LLM for compressing lengthy vision tokens into a single token with minimal loss of visual data.

  • VoCo-LLaMA comprises "Vision Compression" tokens that are charged with compressing and distilling vision tokens in LLMs. The solution reportedly can achieve a compression ratio of 576x while maintaining 83.7% performance on common visual understanding benchmarks. The solution is also claimed to contribute to efficiency gains, enabling a 99.8% reduction in cache storage, a 94.8% decrease in FLOPs, and a 69.6% faster inference time. 

  • However, the solution is claimed to diminish the model's ability to understand uncompressed tokens and face difficulties with diverse fine-grained compression levels.

Contact us

Gain access to all industry hubs, market maps, research tools, and more
Get a demo
arrow
menuarrow

By using this site, you agree to allow SPEEDA Edge and our partners to use cookies for analytics and personalization. Visit our privacy policy for more information about our data collection practices.