GPT-J 6B inference on TensorRT with INT-8 precision
-
Updated
Apr 5, 2023 - Python
GPT-J 6B inference on TensorRT with INT-8 precision
Key python tools for inference: Sample functions, Confidence Intervals, Hypotheses testing, Bootstrapping and Cross Validation. This repo is made for learning purposes.
Token-wise and real-time display Inference module for Llama2 and other LLMs.
Noobs High Performance Computing
Theorem proving geometry of sections in Isabelle
component esp-idf for tensorflow lite micro
Common lisp implementation of Dr. Kothari's extended Wand algorithm for type inference
Gaussian Process Motion Planner
This is a collection of algorithms and models written in Python for probabilistic programming. The main focus of the package is on Bayesian reasoning by using Bayesian networks, Markov networks, and their mixing.
This repository demonstrates work completed as part of the University of St Andrews, CS5011 Artificial Intelligence Practice module. Specifically, four different AI practice domains were considered: Search, Uncertainty, Logic, and Learning.
The Epidemiology of Country HDI and Tuberculosis Mortality: A Hypothesis Testing and Data Analysis Study
Implemented the prediction inference process of the NANODET model in ONNX format and TFLite format
compile stable diffusion to run faster
Welcome to TurboInference, a high-performance inference toolkit written in C++ for rapid and efficient deployment of LLM models. This GitHub repository provides a comprehensive set of tools and utilities designed to make inference tasks swift and resource-efficient.
A project with code enabling inference on images from within a docker container over TCP
DP-UTIL: A Comprehensive Utility Analysis of Differential Privacy in Machine Learning
Add a description, image, and links to the inference topic page so that developers can more easily learn about it.
To associate your repository with the inference topic, visit your repo's landing page and select "manage topics."