AI: Difference between revisions
m (→LLama) |
m (→LLama) |
||
| Line 43: | Line 43: | ||
[https://www.reddit.com/r/Oobabooga/comments/12hyini/vicuna_generating_its_own_prompts/jfrtvh3/ Vicuna generating its own prompts] |
[https://www.reddit.com/r/Oobabooga/comments/12hyini/vicuna_generating_its_own_prompts/jfrtvh3/ Vicuna generating its own prompts] |
||
== Benchmarks == |
|||
{| class="wikitable" |
|||
|+ |
|||
!Interface |
|||
!Model |
|||
!GPTQ |
|||
!Xformers? |
|||
!HW |
|||
!Speed |
|||
|- |
|||
|text-gen |
|||
|anonVic13B |
|||
|GPTQ-for-LLaMa-triton |
|||
|No xformers |
|||
|240gb SSD, 16gb,desktop off |
|||
|7.55 tokens/sec |
|||
|- |
|||
| |
|||
| |
|||
| |
|||
| |
|||
| |
|||
| |
|||
|- |
|||
| |
|||
| |
|||
| |
|||
| |
|||
| |
|||
| |
|||
|} |
|||
Revision as of 22:13, 12 April 2023
LLama
https://rentry.org/llama-tard-v2
https://hackmd.io/@reneil1337/alpaca
https://boards.4channel.org/g/catalog#s=lmg%2F
https://find.4chan.org/?q=AI+Dynamic+Storytelling+General
https://find.4chan.org/?q=AI+Chatbot+General
https://find.4chan.org/?q=%2Flmg%2F (local models general)
https://boards.4channel.org/g/thread/92400764#p92400764
https://files.catbox.moe/lvefgy.json
https://pytorch.org/hub/nvidia_deeplearningexamples_tacotron2/
python server.py --model llama-7b-4bit --wbits 4python server.py --model llama-13b-4bit-128g --wbits 4 --groupsize 128
https://github.com/qwopqwop200/GPTQ-for-LLaMa/issues/59 for installing with out of space error
https://github.com/oobabooga/text-generation-webui/wiki/LLaMA-model#4-bit-mode
https://github.com/pybind/pybind11/discussions/4566
https://lmsysvicuna.miraheze.org/wiki/How_to_use_Vicuna#Use_with_llama.cpp%3A
https://huggingface.co/anon8231489123/vicuna-13b-GPTQ-4bit-128g
Vicuna generating its own prompts
Benchmarks
| Interface | Model | GPTQ | Xformers? | HW | Speed |
|---|---|---|---|---|---|
| text-gen | anonVic13B | GPTQ-for-LLaMa-triton | No xformers | 240gb SSD, 16gb,desktop off | 7.55 tokens/sec |