类 ChatGPT 项目,能够理解自然语言问题,可以与第三方系统交互并动态检索相关信息。

Andrew Maguire f1e6ed9526 add note about oasst2 being available (#3743) 9 月之前
.devcontainer c77f7107d2 simplify default `post_create_command.sh` (#2970) 1 年之前
.github 46520c3688 deployment workflows for bye 11 月之前
.vscode c25f2a8ce1 Add `inlang` to make the contribution of translations easier (#3393) 1 年之前
ansible 29c50eee60 ansible fix 11 月之前
assets c93899ca87 Add hexagon sticker files to /assets (#3175) 1 年之前
backend 07fdde70c1 Update: Pydantic version from 1.10.4 to 1.10.7 (#3615) 1 年之前
copilot dfbafa5b58 add translate button to important readme's (#2816) 1 年之前
data 0d4adb5f1a Add Isna Persian Dataset (#3631) 1 年之前
deploy 9ed9134da8 Updated nginx config for prod, including streaming headers (#2239) 1 年之前
discord-bots 27033a5464 fix typos (#3216) 1 年之前
docker 6042103608 Remove outdated LLaMA worker image (#3282) 1 年之前
docs f1e6ed9526 add note about oasst2 being available (#3743) 9 月之前
inference adcf8dc1dc Fixes for worker prompt truncation in ChatML case (#3673) 1 年之前
model 709bb99d3e [Fix] Consume much more gpt memory running eval_rm (#3614) 1 年之前
notebooks e421d1dbbd fix typos (#3543) 1 年之前
oasst-data 07fdde70c1 Update: Pydantic version from 1.10.4 to 1.10.7 (#3615) 1 年之前
oasst-shared adcf8dc1dc Fixes for worker prompt truncation in ChatML case (#3673) 1 年之前
safety 3a72872aaf Cleanup safety (#3145) 1 年之前
scripts 27033a5464 fix typos (#3216) 1 年之前
text-frontend 542983fe38 add inclued_deleted & include_spam qry params to tree endpoints 1 年之前
website e1769c102f next build fixes 11 月之前
.dockerignore 59a9fe12fb Update web container to include migrations (#2234) 1 年之前
.gitattributes 40dbe55a90 Categorise notebooks as documentation to improve project language breakdown (#860) 1 年之前
.gitignore 618ad473c3 Add PyCharm IDE specific .gitignore files. (#2876) 1 年之前
.pre-commit-config.yaml a2266334db ci: switch flake8 by ruff (#3323) 1 年之前
.python-version d3d657e636 ran pre-commit and fixed issues 1 年之前
CODEOWNERS 280d3bb43f update code-owners (#3584) 1 年之前
CONTRIBUTING.md c25f2a8ce1 Add `inlang` to make the contribution of translations easier (#3393) 1 年之前
LICENSE ce33d1551d Initial commit 1 年之前
README.md f1e6ed9526 add note about oasst2 being available (#3743) 9 月之前
docker-compose.yaml 5e51e7bb2f Make send report message from discord a celery task. (#3089) 1 年之前
inlang.config.js c25f2a8ce1 Add `inlang` to make the contribution of translations easier (#3393) 1 年之前
pyproject.toml a2266334db ci: switch flake8 by ruff (#3323) 1 年之前
redis.conf 743de0bb18 Linting + prettier 1 年之前

README.md

Open-Assistant

:memo: NOTE: OpenAssistant is completed, and the project is now finished. Thank you to everyone who contributed! Check out our blog post for more information. The final published oasst2 dataset can be found on HuggingFace at OpenAssistant/oasst2

![GitHub Repo stars](https://img.shields.io/github/stars/LAION-AI/Open-Assistant?style=social) ![Docs](https://img.shields.io/badge/docs-laion--ai.github.io%2FOpen--Assistant%2F-green) ![GitHub Workflow Status](https://img.shields.io/github/actions/workflow/status/LAION-AI/Open-Assistant/build-frontend.yaml?label=build-frontend) ![GitHub Workflow Status](https://img.shields.io/github/actions/workflow/status/LAION-AI/Open-Assistant/build-postgres.yaml?label=build-postgres) ![GitHub Workflow Status](https://img.shields.io/github/actions/workflow/status/LAION-AI/Open-Assistant/pre-commit.yaml?label=pre-commit) ![GitHub Workflow Status](https://img.shields.io/github/actions/workflow/status/LAION-AI/Open-Assistant/test-api-contract.yaml?label=tests-api) ![GitHub Workflow Status](https://img.shields.io/github/actions/workflow/status/LAION-AI/Open-Assistant/test-e2e.yaml?label=tests-web) ![GitHub Workflow Status](https://img.shields.io/github/actions/workflow/status/LAION-AI/Open-Assistant/deploy-docs-site.yaml?label=deploy-docs) ![GitHub Workflow Status](https://img.shields.io/github/actions/workflow/status/LAION-AI/Open-Assistant/production-deploy.yaml?label=deploy-production) ![GitHub Workflow Status](https://img.shields.io/github/actions/workflow/status/LAION-AI/Open-Assistant/release.yaml?label=deploy-release) ![GitHub release (latest by date)](https://img.shields.io/github/v/release/LAION-AI/Open-Assistant) ![Translate](https://img.shields.io/badge/Translate-blue)

Table of Contents


What is Open Assistant?

Open Assistant is a project meant to give everyone access to a great chat based large language model.

We believe that by doing this we will create a revolution in innovation in language. In the same way that stable-diffusion helped the world make art and images in new ways we hope Open Assistant can help improve the world by improving language itself.

Useful Links

How To Try It Out

Chatting with the AI

The chat frontend is now live here. Log in and start chatting! Please try to react with a thumbs up or down for the assistant's responses when chatting.

Contributing to Data Collection

The data collection frontend is now live here. Log in and start taking on tasks! We want to collect a high volume of quality data. By submitting, ranking, and labelling model prompts and responses you will be directly helping to improve the capabilities of Open Assistant.

Running the Development Setup Locally (without chat)

You do not need to run the project locally unless you are contributing to the development process. The website link above will take you to the public website where you can use the data collection app and the chat.

If you would like to run the data collection app locally for development, you can set up an entire stack needed to run Open-Assistant, including the website, backend, and associated dependent services, with Docker.

To start the demo, run this in the root directory of the repository (check this FAQ if you have problems):

docker compose --profile ci up --build --attach-dependencies

Note: when running on MacOS with an M1 chip you have to use: DB_PLATFORM=linux/x86_64 docker compose ...

Then, navigate to http://localhost:3000 (It may take some time to boot up) and interact with the website.

Note: If an issue occurs with the build, please head to the FAQ and check out the entries about Docker.

Note: When logging in via email, navigate to http://localhost:1080 to get the magic email login link.

Note: If you would like to run this in a standardized development environment (a "devcontainer") using vscode locally or in a web browser using GitHub Codespaces, you can use the provided .devcontainer folder.

Running the Development Setup Locally for Chat

You do not need to run the project locally unless you are contributing to the development process. The website link above will take you to the public website where you can use the data collection app and the chat.

Also note that the local setup is only for development and is not meant to be used as a local chatbot, unless you know what you are doing.

If you do know what you are doing, then see the inference folder for getting the inference system up and running, or have a look at --profile inference in addition to --profile ci in the above command.

The Vision

We are not going to stop at replicating ChatGPT. We want to build the assistant of the future, able to not only write email and cover letters, but do meaningful work, use APIs, dynamically research information, and much more, with the ability to be personalized and extended by anyone. And we want to do this in a way that is open and accessible, which means we must not only build a great assistant, but also make it small and efficient enough to run on consumer hardware.

The Plan

We want to get to an initial MVP as fast as possible, by following the 3-steps outlined in the InstructGPT paper
  1. Collect high-quality human generated Instruction-Fulfillment samples (prompt + response), goal >50k. We design a crowdsourced process to collect and reviewed prompts. We do not want to train on flooding/toxic/spam/junk/personal information data. We will have a leaderboard to motivate the community that shows progress and the most active users. Swag will be given to the top-contributors.
  2. For each of the collected prompts we will sample multiple completions. Completions of one prompt will then be shown randomly to users to rank them from best to worst. Again this should happen crowd-sourced, e.g. we need to deal with unreliable potentially malicious users. At least multiple votes by independent users have to be collected to measure the overall agreement. The gathered ranking-data will be used to train a reward model.
  3. Now follows the RLHF training phase based on the prompts and the reward model.

We can then take the resulting model and continue with completion sampling step 2 for a next iteration.

Slide Decks

Vision & Roadmap

Important Data Structures

How You Can Help

All open source projects begin with people like you. Open source is the belief that if we collaborate we can together gift our knowledge and technology to the world for the benefit of humanity.

Check out our contributing guide to get started.