This is the repository holding code and data for "FrugalML: How to Use ML Prediction APIs More Accurately and Cheaply".
-
Updated
Jan 12, 2021 - Python
This is the repository holding code and data for "FrugalML: How to Use ML Prediction APIs More Accurately and Cheaply".
Bringing local LLMs to a Minecraft front-end through commands.
AccIo - Enterprise LLM : Unifying intelligence at your command!
Python-based WebSocket for CLI LLaVA inference.
Mamba for Vision, Perception and Action
Detailed code explanation of google LLM gemini
HugNLP is a unified and comprehensive NLP library based on HuggingFace Transformer. Please hugging for NLP now!😊 HugNLP will released to @HugAILab
Logical verification of probabilistic/language model 'intuitions'.
Inference Llama 2 in one file of pure C
Specify what you want it to build, the AI asks for clarification, and then builds it.
Automating the deployment of the Takeoff Server on AWS for LLMs
creating a workflow to train t5 language models
Experimental autonomous AI LLM & RAG IETF reviewer
A custom framework for easy use of LLMs, VLMs, etc. supporting various modes and settings via web-ui
A platform to test multiple LLM models inside a RAG workflow to choose the best model for embedding and retrieval and the best prompt according to the use case
Code and analysis for optimizing dynamic neural networks. This project investigates and implements various optimization techniques to enhance dynamic neural networks.
Plug in and Play Implementation of Tree of Thoughts: Deliberate Problem Solving with Large Language Models that Elevates Model Reasoning by atleast 70%
This repository contains question-answers model as an interface which retrieves answers from vector database for a question. Embeddings or tokenised vector being computed using OpenAI API call which gets inserted into ChromaDB as a RAG. OpenAI API key would be required to run this service.
Add a description, image, and links to the llm-inference topic page so that developers can more easily learn about it.
To associate your repository with the llm-inference topic, visit your repo's landing page and select "manage topics."