Michael Yang
|
c895a7d13f
|
some gocritic
|
2024-06-04 11:13:30 -07:00 |
|
Michael Yang
|
34d5ef29b3
|
fix conversion for f16 or f32 inputs
|
2024-05-21 11:28:22 -07:00 |
|
Michael Yang
|
bbbd9f20f3
|
cleanup
|
2024-05-20 16:13:57 -07:00 |
|
Michael Yang
|
547132e820
|
bpe pretokenizer
|
2024-05-20 16:13:57 -07:00 |
|
Patrick Devine
|
d355d2020f
|
add fixes for llama
|
2024-05-20 16:13:57 -07:00 |
|
Patrick Devine
|
c8cf0d94ed
|
llama3 conversion
|
2024-05-20 16:13:57 -07:00 |
|
Patrick Devine
|
4730762e5c
|
add safetensors version
|
2024-05-20 16:13:57 -07:00 |
|
Michael Yang
|
6694be5e50
|
convert/llama: use WriteSeeker
|
2024-05-06 15:24:01 -07:00 |
|
Michael Yang
|
9685c34509
|
quantize any fp16/fp32 model
- FROM /path/to/{safetensors,pytorch}
- FROM /path/to/fp{16,32}.bin
- FROM model:fp{16,32}
|
2024-05-06 15:24:01 -07:00 |
|
Patrick Devine
|
9f8691c6c8
|
Add llama2 / torch models for ollama create (#3607)
|
2024-04-15 11:26:42 -07:00 |
|