this post was submitted on 01 Jul 2023
4 points (100.0% liked)

AI / Machine Learning

156 readers
1 users here now

founded 1 year ago
MODERATORS
 

cross-posted from: https://lemmy.intai.tech/post/40699

Models

Datasets

Repos

Related Papers

Credit:

Tweet

Archive:

@Yampeleg The first model to beat 100% of ChatGPT-3.5 Available on Huggingface

๐Ÿ”ฅ OpenChat_8192

๐Ÿ”ฅ 105.7% of ChatGPT (Vicuna GPT-4 Benchmark)

Less than a month ago the world witnessed as ORCA [1] became the first model to ever outpace ChatGPT on Vicuna's benchmark.

Today, the race to replicate these results open-source comes to an end.

Minutes ago OpenChat scored 105.7% of ChatGPT.

But wait! There is more!

Not only OpenChat beated Vicuna's benchmark, it did so pulling off a LIMA [2] move!

Training was done using 6K GPT-4 conversations out of the ~90K ShareGPT conversations.

The model comes in three versions: the basic OpenChat model, OpenChat-8192 and OpenCoderPlus (Code generation: 102.5% ChatGPT)

This is a significant achievement considering that it's the first (released) open-source model to surpass the Vicuna benchmark. ๐ŸŽ‰๐ŸŽ‰

Congratulations to the authors!!


[1] - Orca: The first model to cross 100% of ChatGPT: https://arxiv.org/pdf/2306.02707.pdf [2] - LIMA: Less Is More for Alignment - TL;DR: Using small number of VERY high quality samples (1000 in the paper) can be as powerful as much larger datasets: https://arxiv.org/pdf/2305.11206

top 1 comments
sorted by: hot top controversial new old
[โ€“] cameron@compuverse.uk 2 points 1 year ago

Hmm,

Now this is very interesting.

I'm going to have to take a look into this. It's great to see how well it performs!

I use the OpenAI API for a couple of things at cost atm.

Self-hosting this could be really neat.