Wing Lian
|
8791eaee9a
|
fix gpt4all training to more closely match the released logic, other small fixes and optimizations
|
2023-03-30 22:40:40 -04:00 |
Wing Lian
|
e744aec8bf
|
fix missing paren
|
2023-03-29 23:40:30 -04:00 |
Wing Lian
|
b7361da58a
|
better multi-gpu support, support gpt4all training data
|
2023-03-29 11:21:47 -04:00 |
John Smith
|
2a1cb42966
|
add padding support as an option
|
2023-03-29 11:20:16 +08:00 |
Wing Lian
|
daad59f8ef
|
properly include the eos token so inference doesn't blabber on
|
2023-03-28 20:53:16 -04:00 |
John Smith
|
f26615fc0c
|
fix bug
|
2023-03-28 21:47:22 +08:00 |
John Smith
|
0768d0fdff
|
update finetune data format
|
2023-03-28 21:45:33 +08:00 |
Andrey Glushenkov
|
50dbb101e9
|
Refactor finetune.py
1. Add command line arguments support
2. Add Stanford Alpaca-like dataset support. Used code from - https://github.com/tloen/alpaca-lora
3. Fix LoRA pre-train application
|
2023-03-24 14:15:07 +03:00 |