Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
-
Updated
Jun 30, 2024 - Python
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
A training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included.
🔥 A tool for visualizing and tracking your machine learning experiments. This repo contains the CLI and Python API.
Determined is an open-source machine learning platform that simplifies distributed training, hyperparameter tuning, experiment tracking, and resource management. Works with PyTorch and TensorFlow.
Automated Machine Learning with scikit-learn
SMAC3: A Versatile Bayesian Optimization Package for Hyperparameter Optimization
OCTIS: Comparing Topic Models is Simple! A python package to optimize and evaluate topic models (accepted at EACL2021 demo track)
Sequential model-based optimization with a `scipy.optimize` interface
DeepHyper: Scalable Asynchronous Neural Architecture and Hyperparameter Search for Deep Neural Networks
Hyperparameter selection on machine learning models using Particle Swarm Optimization
An AutoRecSys library for Surprise. Automate algorithm selection and hyperparameter tuning 🚀
Library for Semi-Automated Data Science
🔨 Malet (Machine Learning Experiment Tool) is a tool for efficient machine learning experiment execution, logging, analysis, and plot making.
Tree-of-Parzen-estimators hyperparameter optimization
This repository is the code basis for the paper intitled "Unlocking Neural Networks: Hyperparameter Analysis"
Sequential model-based optimization with a `scipy.optimize` interface
Distribution transparent Machine Learning experiments on Apache Spark
A lightweight custom automl library.
This repository is the code basis for the paper intitled "Exploring the Intricacies of Neural Network Optimization"
Hyperparameter search wrapper that uses multiple GPUs.
Add a description, image, and links to the hyperparameter-search topic page so that developers can more easily learn about it.
To associate your repository with the hyperparameter-search topic, visit your repo's landing page and select "manage topics."