* Created using Colaboratory

* Update README.md

* A few fixes

* add base model

* local files

* remove pytorch model files

* A few tweaks

* Update V100 comment

* Update README.md

* Clears output

* Reorganize cells

* constant length dataset

* fix char token ratio sampling

* change constant length train dataset generation mode

* fine tuning checkpoint for xetdata/pyxet fd7a21d

* Fine tuned model trained on xetdata/pyxet@fd7a21d

* Update comments
main
Di Xiao 5 months ago committed by GitHub
parent 3b0947ba24
commit a5077fb282
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
60 changed files (17 B → 7.8 GiB)
  1. 126
      CodeLlama-7b-hf/LICENSE
  2. 116
      CodeLlama-7b-hf/README.md
  3. 50
      CodeLlama-7b-hf/USE_POLICY.md
  4. 40
      CodeLlama-7b-hf/config.json
  5. 7
      CodeLlama-7b-hf/generation_config.json
  6. 3
      CodeLlama-7b-hf/model-00001-of-00002.safetensors
  7. 3
      CodeLlama-7b-hf/model-00002-of-00002.safetensors
  8. 522
      CodeLlama-7b-hf/model.safetensors.index.json
  9. 23
      CodeLlama-7b-hf/special_tokens_map.json
  10. 3
      CodeLlama-7b-hf/tokenizer.json
  11. 3
      CodeLlama-7b-hf/tokenizer.model
  12. 34
      CodeLlama-7b-hf/tokenizer_config.json
  13. 5
      README.md
  14. 204
      code-llama/README.md
  15. 29
      code-llama/adapter_config.json
  16. 3
      code-llama/adapter_model.safetensors
  17. 204
      code-llama/checkpoint-100/README.md
  18. 29
      code-llama/checkpoint-100/adapter_config.json
  19. 3
      code-llama/checkpoint-100/adapter_model.safetensors
  20. 3
      code-llama/checkpoint-100/optimizer.pt
  21. 3
      code-llama/checkpoint-100/rng_state.pth
  22. 3
      code-llama/checkpoint-100/scheduler.pt
  23. 121
      code-llama/checkpoint-100/trainer_state.json
  24. 3
      code-llama/checkpoint-100/training_args.bin
  25. 204
      code-llama/checkpoint-20/README.md
  26. 29
      code-llama/checkpoint-20/adapter_config.json
  27. 3
      code-llama/checkpoint-20/adapter_model.safetensors
  28. 3
      code-llama/checkpoint-20/optimizer.pt
  29. 3
      code-llama/checkpoint-20/rng_state.pth
  30. 3
      code-llama/checkpoint-20/scheduler.pt
  31. 41
      code-llama/checkpoint-20/trainer_state.json
  32. 3
      code-llama/checkpoint-20/training_args.bin
  33. 204
      code-llama/checkpoint-40/README.md
  34. 29
      code-llama/checkpoint-40/adapter_config.json
  35. 3
      code-llama/checkpoint-40/adapter_model.safetensors
  36. 3
      code-llama/checkpoint-40/optimizer.pt
  37. 3
      code-llama/checkpoint-40/rng_state.pth
  38. 3
      code-llama/checkpoint-40/scheduler.pt
  39. 61
      code-llama/checkpoint-40/trainer_state.json
  40. 3
      code-llama/checkpoint-40/training_args.bin
  41. 204
      code-llama/checkpoint-60/README.md
  42. 29
      code-llama/checkpoint-60/adapter_config.json
  43. 3
      code-llama/checkpoint-60/adapter_model.safetensors
  44. 3
      code-llama/checkpoint-60/optimizer.pt
  45. 3
      code-llama/checkpoint-60/rng_state.pth
  46. 3
      code-llama/checkpoint-60/scheduler.pt
  47. 81
      code-llama/checkpoint-60/trainer_state.json
  48. 3
      code-llama/checkpoint-60/training_args.bin
  49. 204
      code-llama/checkpoint-80/README.md
  50. 29
      code-llama/checkpoint-80/adapter_config.json
  51. Some files were not shown because too many files have changed in this diff Show More

CodeLlama-7b-hf/LICENSE (0 B → 6.9 KiB)

CodeLlama-7b-hf/README.md (0 B → 6.6 KiB)

CodeLlama-7b-hf/USE_POLICY.md (0 B → 4.7 KiB)

CodeLlama-7b-hf/config.json (0 B → 1.1 KiB)

CodeLlama-7b-hf/generation_config.json (0 B → 138 B)

CodeLlama-7b-hf/model-00001-of-00002.safetensors (0 B → 4.7 GiB)

CodeLlama-7b-hf/model-00002-of-00002.safetensors (0 B → 2.4 GiB)

CodeLlama-7b-hf/model.safetensors.index.json (0 B → 40 KiB)

CodeLlama-7b-hf/special_tokens_map.json (0 B → 411 B)

CodeLlama-7b-hf/tokenizer.json (0 B → 1.8 MiB)

CodeLlama-7b-hf/tokenizer.model (0 B → 488 KiB)

CodeLlama-7b-hf/tokenizer_config.json (0 B → 749 B)

README.md (17 B → 245 B)

code-llama/README.md (0 B → 5.0 KiB)

code-llama/adapter_config.json (0 B → 637 B)

code-llama/adapter_model.safetensors (0 B → 64 MiB)

code-llama/checkpoint-100/README.md (0 B → 5.0 KiB)

code-llama/checkpoint-100/adapter_config.json (0 B → 637 B)

code-llama/checkpoint-100/adapter_model.safetensors (0 B → 64 MiB)

code-llama/checkpoint-100/optimizer.pt (0 B → 128 MiB)

code-llama/checkpoint-100/rng_state.pth (0 B → 14 KiB)

code-llama/checkpoint-100/scheduler.pt (0 B → 1.0 KiB)

code-llama/checkpoint-100/trainer_state.json (0 B → 2.6 KiB)

code-llama/checkpoint-100/training_args.bin (0 B → 4.6 KiB)

code-llama/checkpoint-20/README.md (0 B → 5.0 KiB)

code-llama/checkpoint-20/adapter_config.json (0 B → 637 B)

code-llama/checkpoint-20/adapter_model.safetensors (0 B → 64 MiB)

code-llama/checkpoint-20/optimizer.pt (0 B → 128 MiB)

code-llama/checkpoint-20/rng_state.pth (0 B → 14 KiB)

code-llama/checkpoint-20/scheduler.pt (0 B → 1.0 KiB)

code-llama/checkpoint-20/trainer_state.json (0 B → 918 B)

code-llama/checkpoint-20/training_args.bin (0 B → 4.6 KiB)

code-llama/checkpoint-40/README.md (0 B → 5.0 KiB)

code-llama/checkpoint-40/adapter_config.json (0 B → 637 B)

code-llama/checkpoint-40/adapter_model.safetensors (0 B → 64 MiB)

code-llama/checkpoint-40/optimizer.pt (0 B → 128 MiB)

code-llama/checkpoint-40/rng_state.pth (0 B → 14 KiB)

code-llama/checkpoint-40/scheduler.pt (0 B → 1.0 KiB)

code-llama/checkpoint-40/trainer_state.json (0 B → 1.3 KiB)

code-llama/checkpoint-40/training_args.bin (0 B → 4.6 KiB)

code-llama/checkpoint-60/README.md (0 B → 5.0 KiB)

code-llama/checkpoint-60/adapter_config.json (0 B → 637 B)

code-llama/checkpoint-60/adapter_model.safetensors (0 B → 64 MiB)

code-llama/checkpoint-60/optimizer.pt (0 B → 128 MiB)

code-llama/checkpoint-60/rng_state.pth (0 B → 14 KiB)

code-llama/checkpoint-60/scheduler.pt (0 B → 1.0 KiB)

code-llama/checkpoint-60/trainer_state.json (0 B → 1.7 KiB)

code-llama/checkpoint-60/training_args.bin (0 B → 4.6 KiB)

code-llama/checkpoint-80/README.md (0 B → 5.0 KiB)

code-llama/checkpoint-80/adapter_config.json (0 B → 637 B)

Some files were not shown because too many files have changed in this diff Show More

Loading…
Cancel
Save