Papers
arxiv:2412.04862

EXAONE 3.5: Series of Large Language Models for Real-world Use Cases

Published on Dec 6
· Submitted by taesiri on Dec 9
#2 Paper of the day
Authors:
,
,
,
,
,
,
,
,
,
,

Abstract

This technical report introduces the EXAONE 3.5 instruction-tuned language models, developed and released by LG AI Research. The EXAONE 3.5 language models are offered in three configurations: 32B, 7.8B, and 2.4B. These models feature several standout capabilities: 1) exceptional instruction following capabilities in real-world scenarios, achieving the highest scores across seven benchmarks, 2) outstanding long-context comprehension, attaining the top performance in four benchmarks, and 3) competitive results compared to state-of-the-art open models of similar sizes across nine general benchmarks. The EXAONE 3.5 language models are open to anyone for research purposes and can be downloaded from https://huggingface.co/LGAI-EXAONE. For commercial use, please reach out to the official contact point of LG AI Research: [email protected].

Community

Paper submitter

EXAONE 3.5: Series of Large Language Models for Real-world Use Cases

Can this run on an Intel CPU? Can I run without Transformers?

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

The licensing makes the model solely academic, which is a pity.

Sign up or log in to comment

Models citing this paper 12

Browse 12 models citing this paper

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2412.04862 in a dataset README.md to link it from this page.

Spaces citing this paper 1

Collections including this paper 7