All Updates

All Updates

icon
Filter
Product updates
EleutherAI launches evaluation library for LLMs
Generative AI Infrastructure
May 26, 2024
This week:
Robinhood launches joint investment accounts
Retail Trading Infrastructure
Jul 25, 2024
Partnerships
eToro partners with London Stock Exchange to expand UK stock offerings
Retail Trading Infrastructure
Jul 25, 2024
Funding
StorMagic secures funding from Palatine Growth Credit Fund
Edge Computing
Jul 25, 2024
Funding
Archera raises USD 17 million in Series B funding for product development and recruitment
Cloud Optimization Tools
Jul 25, 2024
Funding
Alto Neuroscience receives grant of USD 11.7 million to support Phase IIb clinical trials of ALTO-100
Precision Medicine
Jul 25, 2024
Partnerships
Quest Diagnostics and BD partner to develop flow cytometry-based companion diagnostics for cancer and other diseases
Precision Medicine
Jul 25, 2024
Product updates
USPACE Technology Group Limited unveils commercial optical satellites and related aerospace products
Next-gen Satellites
Jul 25, 2024
Industry news
Sweden issues study on Gripen fighter jet’s satellite launch capability
Next-gen Satellites
Jul 25, 2024
Product updates
Regulation/policy
Terran Orbital receives certification for new manufacturing facility to begin production
Next-gen Satellites
Jul 25, 2024
Partnerships
Crisalion Mobility partners with Air Chateau for pre-order of eVTOL aircraft
Passenger eVTOL Aircraft
Jul 25, 2024
Generative AI Infrastructure

Generative AI Infrastructure

May 26, 2024

EleutherAI launches evaluation library for LLMs

Product updates

  • EleutherAI, in collaboration with Stability AI and other partners, has launched "Language Model Evaluation Harness" (lm-eval), an open-source library designed to enhance the evaluation of LLMs.

  • The lm-eval tool offers modular implementation of evaluation tasks, supporting various requests such as conditional log-likelihoods, perplexities, and text generation. It facilitates qualitative and quantitative analyses, allowing researchers to conduct in-depth evaluations of model outputs.

  • EleutherAI claims that the lm-eval tool overcomes the limitations of reproducibility and transparency in existing evaluation methods by providing a consistent framework for fair and precise comparisons across different models and techniques, ultimately leading to more reliable research outcomes.

Contact us

Gain access to all industry hubs, market maps, research tools, and more
Get a demo
arrow
menuarrow

By using this site, you agree to allow SPEEDA Edge and our partners to use cookies for analytics and personalization. Visit our privacy policy for more information about our data collection practices.