Official implementation for the paper *🎯DART-Math: Difficulty-Aware Rejection Tuning for Mathematical Problem-Solving*
-
Updated
Jun 26, 2024 - Jupyter Notebook
Official implementation for the paper *🎯DART-Math: Difficulty-Aware Rejection Tuning for Mathematical Problem-Solving*
Assistant virtuel pour aider aux démarches administratives
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
dstack is an easy-to-use and flexible container orchestrator for running AI workloads in any cloud or data center.
OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
🔮 SuperDuperDB: Bring AI to your database! Build, deploy and manage any AI application directly with your existing data infrastructure, without moving your data. Including streaming inference, scalable model training and vector search.
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
A programming framework for agentic AI. Discord: https://aka.ms/autogen-dc. Roadmap: https://aka.ms/autogen-roadmap
Sequence Parallel Attention for Long Context LLM Model Training and Inference
VELOCITI Benchmark Evaluation and Visualisation Code
The easiest way to serve AI/ML models in production - Build Model Inference Service, LLM APIs, Multi-model Inference Graph/Pipelines, LLM/RAG apps, and more!
⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Platforms⚡
Load, pretrain, finetune, deploy 20+ LLMs on your own data. Uses state-of-the-art techniques: flash attention, FSDP, 4-bit, LoRA, and more.
TriForce: Lossless Acceleration of Long Sequence Generation with Hierarchical Speculative Decoding
The Arcee client for executing domain-adpated language model routines https://pypi.org/project/arcee-py/
Minimalist web-searching app with an AI assistant that runs directly from your browser. Uses Web-LLM, Ratchet-ML, Wllama and SearXNG. Demo: https://felladrin-minisearch.hf.space
One .NET library to consume OpenAI, Anthropic, Cohere, Google, Azure, and self-hosed APIs.
Linux LiveCD for offline AI training and inference.
LLM-PowerHouse: Unleash LLMs' potential through curated tutorials, best practices, and ready-to-use code for custom training and inferencing.
Add a description, image, and links to the llm-inference topic page so that developers can more easily learn about it.
To associate your repository with the llm-inference topic, visit your repo's landing page and select "manage topics."