Deeplearning

LLaMA: INT8 edition

by OBG posted Mar 09, 2023
?

단축키

Prev이전 문서

Next다음 문서

ESC닫기

크게 작게 위로 아래로 댓글로 가기 인쇄

https://github.com/tloen/llama-int8

https://github.com/facebookresearch/llama/issues/79#issuecomment-1454687232

 

This is a fork of the LLaMA code that runs LLaMA-13B comfortably within 24 GiB of RAM

 

https://github.com/go-noah/llama

LLaMA 7B example