John Smith
|
4261bd8070
|
add xformers support
|
2023-04-12 12:59:44 +08:00 |
John Smith
|
8cf3bd4086
|
add g_idx support on cuda backend
|
2023-04-09 12:26:22 +08:00 |
John Smith
|
dba3773b30
|
add triton backend support for v2 model
|
2023-04-07 15:34:06 +08:00 |
yamashi
|
778035152d
|
Update arg_parser.py
|
2023-04-07 00:42:34 +02:00 |
John Smith
|
5655f218ed
|
add g_idx buffer.\nadd triton matmul utils for future support.
|
2023-04-02 21:29:06 +08:00 |
Wing Lian
|
b7361da58a
|
better multi-gpu support, support gpt4all training data
|
2023-03-29 11:21:47 -04:00 |
John Smith
|
1c02d4262d
|
add resume checkpoint to continue a training
|
2023-03-29 14:35:39 +08:00 |
John Smith
|
2a1cb42966
|
add padding support as an option
|
2023-03-29 11:20:16 +08:00 |
John Smith
|
0768d0fdff
|
update finetune data format
|
2023-03-28 21:45:33 +08:00 |
John Smith
|
bff039de95
|
add v2 model support
|
2023-03-28 20:33:55 +08:00 |
Andrey Glushenkov
|
397f5041c3
|
Reflect last changes in main
Reflect commits:
4906961bf1
60b227d0ba
|
2023-03-24 15:46:03 +03:00 |
Andrey Glushenkov
|
50dbb101e9
|
Refactor finetune.py
1. Add command line arguments support
2. Add Stanford Alpaca-like dataset support. Used code from - https://github.com/tloen/alpaca-lora
3. Fix LoRA pre-train application
|
2023-03-24 14:15:07 +03:00 |