{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"cortex","owner":"janhq","isFork":false,"description":"Drop-in, local AI alternative to the OpenAI stack. Multi-engine (llama.cpp, TensorRT-LLM, ONNX). Powers 👋 Jan","allTopics":["ai","cuda","llama","accelerated","inference-engine","openai-api","llm","stable-diffusion","llms","llamacpp","llama2","gguf","tensorrt-llm"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":7,"issueCount":83,"starsCount":1748,"forksCount":93,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-26T10:55:13.404Z"}},{"type":"Public","name":"cortex.llamacpp","owner":"janhq","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":4,"issueCount":7,"starsCount":3,"forksCount":3,"license":"GNU Affero General Public License v3.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,6,8,6,5,20,8,8,13,13],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-25T17:46:33.319Z"}},{"type":"Public","name":"cortex.tensorrt-llm","owner":"janhq","isFork":true,"description":"Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU accelerated inference on NVIDIA's GPUs.","allTopics":["nvidia","jan","tensorrt","llm","tensorrt-llm"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":1,"issueCount":8,"starsCount":31,"forksCount":788,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,3,5,0,7,11,2,2,6,0,1,5,1,0,3,0,1,0,0,0,0,0,0,0,2,0,0,0,0,0,1,2,0,0,0,0,0,0,0,2,6],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-25T08:34:33.644Z"}},{"type":"Public","name":"cortex.onnx","owner":"janhq","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":1,"issueCount":0,"starsCount":1,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,5,3,4,4],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-24T07:38:53.732Z"}},{"type":"Public","name":"cortex.python","owner":"janhq","isFork":false,"description":"C++ code that run Python embedding","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":3,"forksCount":0,"license":"GNU Affero General Public License v3.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-23T13:36:05.468Z"}},{"type":"Public","name":"infinity","owner":"janhq","isFork":true,"description":"The AI-native database built for LLM applications, providing incredibly fast vector and full-text search ","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":163,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-19T06:45:53.239Z"}},{"type":"Public","name":"llama.cpp-avx-vnni","owner":"janhq","isFork":true,"description":"Port of Facebook's LLaMA model in C/C++","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":8678,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-19T07:04:40.274Z"}},{"type":"Public","name":"TensorRT","owner":"janhq","isFork":true,"description":"NVIDIA® TensorRT™, an SDK for high-performance deep learning inference, includes a deep learning inference optimizer and runtime that delivers low latency and high throughput for inference applications.","allTopics":[],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":2072,"license":"Apache License 2.0","participation":[0,0,1,0,0,7,6,6,1,3,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-09-03T03:08:17.003Z"}}],"repositoryCount":8,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}