airoboros
LLM Vibe Score 51 Human Vibe Score 43%

airoboros

· GitHub

Customizable implementation of the self-instruct paper.

0
Visit Resource

Repository Stats

1040 stars
70 forks

Repository Activity

This repository has been starred 1040 times and forked 70 times, showing its popularity in the developer community.

README Preview

# airoboros: using large language models to fine-tune large language models

This is my take on implementing the [Self-Instruct paper](https://arxiv.org/abs/2212.10560).  The approach is quite heavily modified, and does not use any human-generated seeds.

This updated implementation supports either the /v1/completions endpoint or /v1/chat/completions, which is particularly useful in that it supports gpt-4 and gpt-3.5-turbo (which is 1/10 the cost of text-davinci-003).

Huge thank you to the folks over at [a16z](https://a16z.com/) for sponsoring the costs associated with building models and associated tools!

## Install

via pip:
```
pip install --no-build-isolation airoboros
```

from source (keeping the source):
```
git clone https://github.com/jondurbin/airoboros
pip install -e --no-build-isolation ./airoboros
```

## Key differences from self-instruct/alpaca

* support for either /v1/completions or /v1/chat/completions APIs (which allows gpt-3.5-turbo instead of text-davinci-003, as

Join the Discord Community

Connect with fellow builders, share experiences, converse with AI agent trolls, and get guidance on AI implementation.

Come learn to vibe code with us!

Join Community
5
Members Online Now
Online now:
0xRichard
0xRichard
AiSteroid.Ai
AiSteroid.Ai
c...
c...
J_Pyxal
J_Pyxal
SinFa
SinFa

VibeBuilders.ai Newsletter

Get the free TERMINAL NAVIGATION TOOLKIT when you sign up!

We respect your privacy. Unsubscribe at any time.