Pinned Loading
Repositories
- ipex-llm Public
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, vLLM, GraphRAG, DeepSpeed, Axolotl, etc
intel-analytics/ipex-llm’s past year of commit activity - text-generation-webui Public Forked from oobabooga/text-generation-webui
A Gradio Web UI for running local LLM on Intel GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) using IPEX-LLM.
intel-analytics/text-generation-webui’s past year of commit activity - langchain Public Forked from langchain-ai/langchain
🦜🔗 Build context-aware reasoning applications
intel-analytics/langchain’s past year of commit activity - llama_index Public Forked from run-llama/llama_index
LlamaIndex is a data framework for your LLM applications
intel-analytics/llama_index’s past year of commit activity - ipex-llm-tutorial Public
Accelerate LLM with low-bit (FP4 / INT4 / FP8 / INT8) optimizations using ipex-llm
intel-analytics/ipex-llm-tutorial’s past year of commit activity - Langchain-Chatchat Public Forked from chatchat-space/Langchain-Chatchat
Knowledge Base QA using RAG pipeline on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) with IPEX-LLM
intel-analytics/Langchain-Chatchat’s past year of commit activity