ai-newsbits
Jerry
May 17, 2024
Jerry
May 17, 2024
Jerry
May 17, 2024
Jerry
May 13, 2024
Jerry
May 12, 2024
Jerry
May 11, 2024
Jerry
May 11, 2024
Jerry
May 8, 2024
Jerry
May 8, 2024
Jerry
May 8, 2024
Jerry
May 6, 2024
Jerry
May 5, 2024
Jerry
May 1, 2024
Jerry
May 1, 2024
Jerry
May 1, 2024
Jerry
Apr 29, 2024
Jerry
Apr 28, 2024
Jerry
Apr 26, 2024
Jerry
Apr 23, 2024
Jerry
Apr 23, 2024
Jerry
Apr 22, 2024
Jerry
Apr 19, 2024
Jerry
Apr 18, 2024
Jerry
Apr 18, 2024
Jerry
Apr 18, 2024
Jerry
Apr 18, 2024
Jerry
Apr 14, 2024
Jerry
Apr 14, 2024
Jerry
Apr 14, 2024
Jerry
Apr 11, 2024
Jerry
Apr 10, 2024
1
1
Jerry
Apr 8, 2024
1
1
Jerry
Apr 8, 2024
1
1
Jerry
Apr 7, 2024
Jerry
Apr 7, 2024
Load more
Share
Jerry
Apr 5, 2024
Apr 5, 2024
#AI #LLM
IPEX-LLM
•
매우 짧은 지연 시간으로
인텔 CPU 및 GPU에서 LLM을 실행
하기 위한 PyTorch 라이브러리
•
llama.cpp, Text-Generation-WebUI, HuggingFace transformers, HuggingFace PEFT, LangChain, LlamaIndex, DeepSpeed-AutoTP, vLLM, FastChat, HuggingFace TRL, AutoGen, ModeScope
등과 연동 가능
•
50개 이상의 모델이 최적화 및 검증됨(
LLaMA2, Mistral, Mixtral, Gemma, LLaVA, Whisper, ChatGLM, Baichuan, Qwen, RWKV
등)
🔗
https://github.com/intel-analytics/ipex-llm
GitHub - intel-analytics/ipex-llm: Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with llama.cpp, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, ModelScope, etc.
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max)...
github.com
2
1
1