intel-analytics
Pinned Loading
Repositories
- ipex-llm Public
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, MiniCPM, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, GraphRAG, DeepSpeed, vLLM, FastChat, Axolotl, etc.
intel-analytics/ipex-llm’s past year of commit activity - text-generation-webui Public Forked from oobabooga/text-generation-webui
A Gradio Web UI for running local LLM on Intel GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) using IPEX-LLM.
intel-analytics/text-generation-webui’s past year of commit activity - llama_index Public Forked from run-llama/llama_index
LlamaIndex is a data framework for your LLM applications
intel-analytics/llama_index’s past year of commit activity - ipex-llm-tutorial Public
Accelerate LLM with low-bit (FP4 / INT4 / FP8 / INT8) optimizations using ipex-llm
intel-analytics/ipex-llm-tutorial’s past year of commit activity - Langchain-Chatchat Public Forked from chatchat-space/Langchain-Chatchat
Knowledge Base QA using RAG pipeline on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) with IPEX-LLM
intel-analytics/Langchain-Chatchat’s past year of commit activity