1
0
Fork 0

GOAT-7B-Community model is supervised finetuned (SFT) version of LLaMA 2 developed by GOAT.AI lab on user-shared conversations from GoatChat app.

README.md

license model_type tags
llama2 llama
facebook
meta
pytorch
llama
llama-2

GOAT-7B-Community model

GOAT-7B-Community

GOAT-7B-Community model is supervised finetuned (SFT) version of LLaMA 2 developed by GOAT.AI lab on user-shared conversations from GoatChat app.

Model description

  • Base Architecture: LLaMA 2 7B flavour
  • Dataset size: 72K multi-turn dialogues
  • License: llama2
  • Context window length: 4096 tokens

Learn more

Uses

The main purpose of GOAT-7B-Community is to facilitate research on large language models and chatbots. It is specifically designed for researchers and hobbyists working in the fields of natural language processing, machine learning, and artificial intelligence.

Usage

Usage can be either self-hosted via transformers or used with Spaces

import torch

from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "GOAT-AI/GOAT-7B-Community"

tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
    model_name,
    torch_dtype=torch.bfloat16
)

Training dataset

Training dataset was collected from users conversations with GoatChat app and OpenAssistant. We will not release the dataset.

Evaluation

GOAT-7B-Community model is evaluated against common metrics for evaluating language models, including MMLU and BigBench Hard. We still continue to evaluate all our models and will share details soon.

  • MMLU: 49.31
  • BBH: 35.7

License

GOAT-7B-Community model is based on Meta's LLaMA-2-7b-hf, and using own datasets.

GOAT-7B-Community model weights are available under LLAMA-2 license. Note that the GOAT-7B-Community model weights require access to the LLaMA-2 model weighs. The GOAT-7B-Community model is based on LLaMA-2 and should be used according to the LLaMA-2 license.

Risks and Biases

GOAT-7B-Community model can produce factually incorrect output and should not be relied on to deliver factually accurate information. The model was trained on various private and public datasets. Therefore, the GOAT-7B-Community model could possibly generate wrong, biased, or otherwise offensive outputs.

File List Total items: 10
Name Last Commit Size Last Modified
.gitattributes Initial commit 79 B 10 months ago
README.md copy GOAT-7B-Community/ to xet://seanses/GOAT-7B-Community/main/ recursively 2.4 KiB 10 months ago
config.json copy GOAT-7B-Community/ to xet://seanses/GOAT-7B-Community/main/ recursively 684 B 10 months ago
generation_config.json copy GOAT-7B-Community/ to xet://seanses/GOAT-7B-Community/main/ recursively 175 B 10 months ago
pytorch_model-00001-of-00002.bin copy GOAT-7B-Community/ to xet://seanses/GOAT-7B-Community/main/ recursively 9.3 GiB 10 months ago
pytorch_model-00002-of-00002.bin copy GOAT-7B-Community/ to xet://seanses/GOAT-7B-Community/main/ recursively 3.3 GiB 10 months ago
pytorch_model.bin.index.json copy GOAT-7B-Community/ to xet://seanses/GOAT-7B-Community/main/ recursively 26 KiB 10 months ago
special_tokens_map.json copy GOAT-7B-Community/ to xet://seanses/GOAT-7B-Community/main/ recursively 437 B 10 months ago
tokenizer.model copy GOAT-7B-Community/ to xet://seanses/GOAT-7B-Community/main/ recursively 488 KiB 10 months ago
tokenizer_config.json copy GOAT-7B-Community/ to xet://seanses/GOAT-7B-Community/main/ recursively 749 B 10 months ago

About

GOAT-7B-Community model is supervised finetuned (SFT) version of LLaMA 2 developed by GOAT.AI lab on user-shared conversations from GoatChat app.

Repository Size

Loading repo size...

Commits 2 commits

File Types