Colab (#1)
* Created using Colaboratory * Update README.md * A few fixes * add base model * local files * remove pytorch model files * A few tweaks * Update V100 comment * Update README.md * Clears output * Reorganize cells * constant length dataset * fix char token ratio sampling * change constant length train dataset generation mode * fine tuning checkpoint for xetdata/pyxet fd7a21d * Fine tuned model trained on xetdata/pyxet@fd7a21d * Update commentsmain
parent
3b0947ba24
commit
a5077fb282
60 changed files (17 B → 7.8 GiB)
CodeLlama-7b-hf/LICENSE
(0 B → 6.9 KiB)
CodeLlama-7b-hf/README.md
(0 B → 6.6 KiB)
CodeLlama-7b-hf/USE_POLICY.md
(0 B → 4.7 KiB)
CodeLlama-7b-hf/config.json
(0 B → 1.1 KiB)
CodeLlama-7b-hf/generation_config.json
(0 B → 138 B)
CodeLlama-7b-hf/model-00001-of-00002.safetensors
(0 B → 4.7 GiB)
CodeLlama-7b-hf/model-00002-of-00002.safetensors
(0 B → 2.4 GiB)
CodeLlama-7b-hf/model.safetensors.index.json
(0 B → 40 KiB)
CodeLlama-7b-hf/special_tokens_map.json
(0 B → 411 B)
CodeLlama-7b-hf/tokenizer.json
(0 B → 1.8 MiB)
CodeLlama-7b-hf/tokenizer.model
(0 B → 488 KiB)
CodeLlama-7b-hf/tokenizer_config.json
(0 B → 749 B)
README.md
(17 B → 245 B)
code-llama/README.md
(0 B → 5.0 KiB)
code-llama/adapter_config.json
(0 B → 637 B)
code-llama/adapter_model.safetensors
(0 B → 64 MiB)
code-llama/checkpoint-100/README.md
(0 B → 5.0 KiB)
code-llama/checkpoint-100/adapter_config.json
(0 B → 637 B)
code-llama/checkpoint-100/adapter_model.safetensors
(0 B → 64 MiB)
code-llama/checkpoint-100/optimizer.pt
(0 B → 128 MiB)
code-llama/checkpoint-100/rng_state.pth
(0 B → 14 KiB)
code-llama/checkpoint-100/scheduler.pt
(0 B → 1.0 KiB)
code-llama/checkpoint-100/trainer_state.json
(0 B → 2.6 KiB)
code-llama/checkpoint-100/training_args.bin
(0 B → 4.6 KiB)
code-llama/checkpoint-20/README.md
(0 B → 5.0 KiB)
code-llama/checkpoint-20/adapter_config.json
(0 B → 637 B)
code-llama/checkpoint-20/adapter_model.safetensors
(0 B → 64 MiB)
code-llama/checkpoint-20/optimizer.pt
(0 B → 128 MiB)
code-llama/checkpoint-20/rng_state.pth
(0 B → 14 KiB)
code-llama/checkpoint-20/scheduler.pt
(0 B → 1.0 KiB)
code-llama/checkpoint-20/trainer_state.json
(0 B → 918 B)
code-llama/checkpoint-20/training_args.bin
(0 B → 4.6 KiB)
code-llama/checkpoint-40/README.md
(0 B → 5.0 KiB)
code-llama/checkpoint-40/adapter_config.json
(0 B → 637 B)
code-llama/checkpoint-40/adapter_model.safetensors
(0 B → 64 MiB)
code-llama/checkpoint-40/optimizer.pt
(0 B → 128 MiB)
code-llama/checkpoint-40/rng_state.pth
(0 B → 14 KiB)
code-llama/checkpoint-40/scheduler.pt
(0 B → 1.0 KiB)
code-llama/checkpoint-40/trainer_state.json
(0 B → 1.3 KiB)
code-llama/checkpoint-40/training_args.bin
(0 B → 4.6 KiB)
code-llama/checkpoint-60/README.md
(0 B → 5.0 KiB)
code-llama/checkpoint-60/adapter_config.json
(0 B → 637 B)
code-llama/checkpoint-60/adapter_model.safetensors
(0 B → 64 MiB)
code-llama/checkpoint-60/optimizer.pt
(0 B → 128 MiB)
code-llama/checkpoint-60/rng_state.pth
(0 B → 14 KiB)
code-llama/checkpoint-60/scheduler.pt
(0 B → 1.0 KiB)
code-llama/checkpoint-60/trainer_state.json
(0 B → 1.7 KiB)
code-llama/checkpoint-60/training_args.bin
(0 B → 4.6 KiB)
code-llama/checkpoint-80/README.md
(0 B → 5.0 KiB)
code-llama/checkpoint-80/adapter_config.json
(0 B → 637 B)
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in new issue