John Smith
|
633c28fd25
|
add quant attn v1 support
|
2023-04-25 12:30:03 +08:00 |
John Smith
|
82bbea2729
|
optimized matmul for v2 model
|
2023-04-25 09:18:50 +08:00 |
John Smith
|
de3c91834e
|
optimized attention and mlp for performance, add lora monkey patch for models here and GPTQ_For_Llama models using optimization
|
2023-04-22 15:36:56 +08:00 |
John Smith
|
4c18a56fc0
|
fix bug
|
2023-04-09 12:44:50 +08:00 |
John Smith
|
8cf3bd4086
|
add g_idx support on cuda backend
|
2023-04-09 12:26:22 +08:00 |
Wing Lian
|
8791eaee9a
|
fix gpt4all training to more closely match the released logic, other small fixes and optimizations
|
2023-03-30 22:40:40 -04:00 |
John Smith
|
1719bd0ce3
|
fix bug
|
2023-03-29 08:09:40 +08:00 |
John Smith
|
1043ded7d9
|
Merge branch 'main' of github.com:johnsmith0031/alpaca_lora_4bit
|
2023-03-29 01:26:20 +08:00 |
John Smith
|
d28ee06202
|
fix bug
|
2023-03-29 01:25:37 +08:00 |
Wing Lian
|
b47da33084
|
fixes for most recent update
|
2023-03-28 10:56:35 -04:00 |
John Smith
|
234004ceb5
|
fix bug
|
2023-03-28 22:05:18 +08:00 |
John Smith
|
211af574b6
|
fix bug
|
2023-03-28 21:12:51 +08:00 |
John Smith
|
bff039de95
|
add v2 model support
|
2023-03-28 20:33:55 +08:00 |