Open Source Python Machine Learning Software - Page 12

Python Machine Learning Software

View 446 business solutions

Browse free open source Python Machine Learning Software and projects below. Use the toggles on the left to filter open source Python Machine Learning Software by OS, license, language, programming language, and project status.

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Cycloid: Hybrid Cloud DevOps collaboration platform Icon
    Cycloid: Hybrid Cloud DevOps collaboration platform

    For Developers, DevOps, IT departments, MSPs

    Enable your developers to do their best work and increase time-to-market speed with a leading DevOps and Hybrid Cloud platform.
    Learn More
  • 1
    DeepImageTranslator

    DeepImageTranslator

    DeepImageTranslator: a deep-learning utility for image translation

    Created by: Run Zhou Ye, En Zhou Ye, and En Hui Ye DeepImageTranslator: a free, user-friendly tool for image translation using deep-learning and its applications in CT image analysis Citation: Please cite this software as: Ye RZ, Noll C, Richard G, Lepage M, Turcotte ÉE, Carpentier AC. DeepImageTranslator: a free, user-friendly graphical interface for image translation using deep-learning and its applications in 3D CT image analysis. SLAS technology. 2022 Feb 1;27(1):76-84. https://doi.org/10.1016/j.slast.2021.10.014
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    DeepKE

    DeepKE

    An Open Toolkit for Knowledge Graph Extraction and Construction

    Supporting cnSchema, standard supervised setting, low-resource setting, document-level setting and multi-modal setting for knowledge base population. DeepKE is a knowledge extraction toolkit supporting cnSchema, standard supervised, low-resource, and document-level scenarios for entity, relation, and attribution extraction. It allows developers and researchers to customize datasets and models to extract information from unstructured texts. DeepKE supports low-resource settings with only a few labeled (e.g., 16/32 shot) instances for widespread information extraction tasks. Since relations are expressed over multiple sentences in real-world applications, DeepKE supports document-level relation extraction. We present a new open-source and extensible knowledge extraction toolkit, called DeepKE, supporting standard fully supervised, low-resource few-shot and document-level scenarios.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    DeepLearn

    DeepLearn

    Implementation of research papers on Deep Learning+ NLP+ CV in Python

    Welcome to DeepLearn. This repository contains an implementation of the following research papers on NLP, CV, ML, and deep learning. The required dependencies are mentioned in requirement.txt. I will also use dl-text modules for preparing the datasets. If you haven't use it, please do have a quick look at it. CV, transfer learning, representation learning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    DeepLearning Tutorial

    DeepLearning Tutorial

    Deep Learning Tutorial, Excellent Articles, Deep Learning Tutorial

    DeepLearning is an open-source repository that aggregates tutorials, articles, and educational resources related to deep learning and machine learning. The project is designed as a knowledge collection that helps beginners understand neural networks, deep learning architectures, and fundamental machine learning concepts. It contains curated learning materials covering topics such as feedforward neural networks, activation functions, backpropagation algorithms, optimization methods, and convolutional neural networks. The repository organizes these materials into structured tutorials and references that allow readers to explore deep learning concepts progressively. Many of the resources include explanations of common model architectures used in computer vision and artificial intelligence. The project also collects recommended articles and educational documents that expand on deep learning theory and practical implementation strategies.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Point of Sale. Powerful and Simple. Icon
    Point of Sale. Powerful and Simple.

    For retail store owners and multi-location retail operations needing a tool to manage sales, inventory, staff and channels in one place

    Vibe Retail is an all-in-one retail point-of-sale and operations platform built for single-store and multi-location retailers seeking to unify inventory, sales, staff and customer data from one mobile-friendly interface. The system lets you track inventory across locations and warehouses, handle item variations (size, color, material), manage purchase orders and supplier deliveries, print custom barcodes, and transfer stock between stores in real time. On the sales side, Vibe supports multiple payment types (cards, cash, checks, gift cards, EBT), layaway workflows, serial number tracking, delivery management, loyalty programs and branded receipts. Retailers can integrate with online platforms (such as Shopify and WooCommerce), sync in-store and online sales, access 40+ real-time reports on sales, inventory and performance, set up promotions and discounts, and print receipts from mobile devices.
    Learn More
  • 5
    DeepPavlov

    DeepPavlov

    A library for deep learning end-to-end dialog systems and chatbots

    DeepPavlov makes it easy for beginners and experts to create dialogue systems. The best place to start is with user-friendly tutorials. They provide quick and convenient introduction on how to use DeepPavlov with complete, end-to-end examples. No installation needed. Guides explain the concepts and components of DeepPavlov. Follow step-by-step instructions to install, configure and extend DeepPavlov framework for your use case. DeepPavlov is an open-source framework for chatbots and virtual assistants development. It has comprehensive and flexible tools that let developers and NLP researchers create production-ready conversational skills and complex multi-skill conversational assistants. Use BERT and other state-of-the-art deep learning models to solve classification, NER, Q&A and other NLP tasks. DeepPavlov Agent allows building industrial solutions with multi-skill integration via API services.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    DeepSeed

    DeepSeed

    Deep learning optimization library making distributed training easy

    DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective. DeepSpeed delivers extreme-scale model training for everyone, from data scientists training on massive supercomputers to those training on low-end clusters or even on a single GPU. Using current generation of GPU clusters with hundreds of devices, 3D parallelism of DeepSpeed can efficiently train deep learning models with trillions of parameters. With just a single GPU, ZeRO-Offload of DeepSpeed can train models with over 10B parameters, 10x bigger than the state of arts, democratizing multi-billion-parameter model training such that many deep learning scientists can explore bigger and better models. Sparse attention of DeepSpeed powers an order-of-magnitude longer input sequence and obtains up to 6x faster execution comparing with dense transformers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    DeepXDE

    DeepXDE

    A library for scientific machine learning & physics-informed learning

    DeepXDE is a library for scientific machine learning and physics-informed learning. DeepXDE includes the following algorithms. Physics-informed neural network (PINN). Solving different problems. Solving forward/inverse ordinary/partial differential equations (ODEs/PDEs) [SIAM Rev.] Solving forward/inverse integro-differential equations (IDEs) [SIAM Rev.] fPINN: solving forward/inverse fractional PDEs (fPDEs) [SIAM J. Sci. Comput.] NN-arbitrary polynomial chaos (NN-aPC): solving forward/inverse stochastic PDEs (sPDEs) [J. Comput. Phys.] PINN with hard constraints (hPINN): solving inverse design/topology optimization [SIAM J. Sci. Comput.] Residual-based adaptive sampling [SIAM Rev., arXiv] Gradient-enhanced PINN (gPINN) [Comput. Methods Appl. Mech. Eng.] PINN with multi-scale Fourier features [Comput. Methods Appl. Mech. Eng.]
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Deepchecks

    Deepchecks

    Test Suites for validating ML models & data

    Deepchecks is the leading tool for testing and for validating your machine learning models and data, and it enables doing so with minimal effort. Deepchecks accompany you through various validation and testing needs such as verifying your data’s integrity, inspecting its distributions, validating data splits, evaluating your model and comparing between different models. While you’re in the research phase, and want to validate your data, find potential methodological problems, and/or validate your model and evaluate it. To run a specific single check, all you need to do is import it and then to run it with the required (check-dependent) input parameters. More details about the existing checks and the parameters they can receive can be found in our API Reference. An ordered collection of checks, that can have conditions added to them. The Suite enables displaying a concluding report for all of the Checks that ran.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Deepo

    Deepo

    Set up deep learning environment in a single command line

    Deepo is a series of Docker images that allows you to quickly set up your deep learning research environment, supports almost all commonly used deep learning frameworks, supports GPU acceleration (CUDA and cuDNN included), also works in CPU-only mode, and works on Linux (CPU version/GPU version), Windows (CPU version) and OS X (CPU version). Their Dockerfile generator that allows you to customize your own environment with Lego-like modules, and automatically resolves the dependencies for you. For users in China who may suffer from slow speeds when pulling the image from the public Docker registry, you can pull deepo images from the China registry mirror by specifying the full path, including the registry, in your docker pull command. This should work and enables Deepo to use the GPU from inside a docker container.
    Downloads: 0 This Week
    Last Update:
    See Project
  • DriveStrike: Remote Wipe | Data Breach Protection Icon
    DriveStrike: Remote Wipe | Data Breach Protection

    . From Fortune 500 to small businesses with remote workers, every industry can gain from premium endpoint security.

    DriveStrike protects devices and data in the event of loss, theft, or use in remote locations. Remotely locate, lock, and wipe devices you manage to prevent data compromise. DriveStrike prevents data breaches to ensure confidentiality, compliance, and a competitive edge.
    Learn More
  • 10
    Delta ML

    Delta ML

    Deep learning based natural language and speech processing platform

    DELTA is a deep learning-based end-to-end natural language and speech processing platform. DELTA aims to provide easy and fast experiences for using, deploying, and developing natural language processing and speech models for both academia and industry use cases. DELTA is mainly implemented using TensorFlow and Python 3. DELTA has been used for developing several state-of-the-art algorithms for publications and delivering real production to serve millions of users. It helps you to train, develop, and deploy NLP and/or speech models. Use configuration files to easily tune parameters and network structures. What you see in training is what you get in serving: all data processing and features extraction are integrated into a model graph. Text classification, named entity recognition, question and answering, text summarization, etc. Uniform I/O interfaces and no changes for new models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Denoising Diffusion Probabilistic Model

    Denoising Diffusion Probabilistic Model

    Implementation of Denoising Diffusion Probabilistic Model in Pytorch

    Implementation of Denoising Diffusion Probabilistic Model in Pytorch. It is a new approach to generative modeling that may have the potential to rival GANs. It uses denoising score matching to estimate the gradient of the data distribution, followed by Langevin sampling to sample from the true distribution. If you simply want to pass in a folder name and the desired image dimensions, you can use the Trainer class to easily train a model.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Determined

    Determined

    Determined, deep learning training platform

    The fastest and easiest way to build deep learning models. Distributed training without changing your model code. Determined takes care of provisioning machines, networking, data loading, and fault tolerance. Build more accurate models faster with scalable hyperparameter search, seamlessly orchestrated by Determined. Use state-of-the-art algorithms and explore results with our hyperparameter search visualizations. Interpret your experiment results using the Determined UI and TensorBoard, and reproduce experiments with artifact tracking. Deploy your model using Determined's built-in model registry. Easily share on-premise or cloud GPUs with your team. Determined’s cluster scheduling offers first-class support for deep learning and seamless spot instance support. Check out examples of how you can use Determined to train popular deep learning models at scale.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    DialoGPT

    DialoGPT

    Large-scale pretraining for dialogue

    DialoGPT is an open-source conversational language model developed by Microsoft Research for generating natural dialogue responses using large-scale transformer architectures. The system is built on the GPT-2 architecture and is designed specifically for multi-turn conversation tasks, enabling machines to produce coherent responses during interactive dialogue. The model was trained on a massive dataset of approximately 147 million conversational exchanges extracted from Reddit discussion threads, allowing it to learn patterns of natural human conversation. DialoGPT provides multiple pretrained model sizes and includes code for training, fine-tuning, and evaluating dialogue generation models. The repository also contains scripts for preparing conversation datasets and reproducing experimental benchmarks related to conversational AI research.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Diffrax

    Diffrax

    Numerical differential equation solvers in JAX

    Diffrax is a numerical differential equation solving library built for the JAX ecosystem, with a strong focus on composability, differentiability, and high-performance scientific computing. The project provides tools for solving ordinary differential equations, stochastic differential equations, controlled differential equations, and related systems in a way that fits naturally into modern machine learning and differentiable programming workflows. Because it is written to work closely with JAX, it supports just-in-time compilation, automatic differentiation, vectorization, and accelerator-backed execution on hardware such as GPUs and TPUs. This makes it especially appealing for researchers who need equation solvers that can be embedded inside trainable models or simulation-heavy learning systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Diffusion for World Modeling

    Diffusion for World Modeling

    Learning agent trained in a diffusion world model

    Diffusion for World Modeling is an experimental reinforcement learning system that trains intelligent agents inside a simulated environment generated by a diffusion-based world model. The project introduces the idea of using diffusion models, commonly used for image generation, to simulate the dynamics of an environment and predict future states based on previous observations and actions. Instead of interacting directly with a real environment, the reinforcement learning agent learns within a generative model that produces frames representing the environment. This approach allows training to occur in a simulated world that captures detailed visual dynamics while reducing the need for costly interactions with real environments. The system has been applied to tasks such as Atari game simulations and demonstrations involving complex environments like first-person shooter games.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Dive-into-DL-TensorFlow2.0

    Dive-into-DL-TensorFlow2.0

    Dive into Deep Learning

    This project changes the MXNet code implementation in the original book "Learning Deep Learning by Hand" to TensorFlow2 implementation. After consulting Mr. Li Mu by the tutor of archersama , the implementation of this project has been agreed by Mr. Li Mu. Original authors: Aston Zhang, Li Mu, Zachary C. Lipton, Alexander J. Smola and other community contributors. There are some differences between the Chinese and English versions of this book . This project mainly focuses on TensorFlow2 reconstruction for the Chinese version of this book. In addition, this project also refers to the project Dive-into-DL-PyTorch , which refactored PyTorch in the Chinese version of this book, and I would like to express my gratitude here. This repository mainly contains two folders, code and docs (plus some data stored in data). The code folder is the relevant jupyter notebook code for each chapter (based on TensorFlow2); the docs folder is the relevant content in the book.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Django friendly finite state machine

    Django friendly finite state machine

    Django friendly finite state machine support

    Django-fsm adds simple declarative state management for Django models. If you need parallel task execution, view, and background task code reuse over different flows - check my new project Django-view flow. Instead of adding a state field to a Django model and managing its values by hand, you use FSMField and mark model methods with the transition decorator. These methods could contain side effects of the state change. You may also take a look at the Django-fsm-admin project containing a mixin and template tags to integrate Django-fsm state transitions into the Django admin. FSM really helps to structure the code, especially when a new developer comes to the project. FSM is most effective when you use it for some sequential steps. Transition logging support could be achieved with help of django-fsm-log package.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    DoWhy

    DoWhy

    DoWhy is a Python library for causal inference

    DoWhy is a Python library for causal inference that supports explicit modeling and testing of causal assumptions. DoWhy is based on a unified language for causal inference, combining causal graphical models and potential outcomes frameworks. Much like machine learning libraries have done for prediction, DoWhy is a Python library that aims to spark causal thinking and analysis. DoWhy provides a wide variety of algorithms for effect estimation, causal structure learning, diagnosis of causal structures, root cause analysis, interventions and counterfactuals. DoWhy builds on two of the most powerful frameworks for causal inference: graphical causal models and potential outcomes. For effect estimation, it uses graph-based criteria and do-calculus for modeling assumptions and identifying a non-parametric causal effect. For estimation, it switches to methods based primarily on potential outcomes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Dynamic Routing Between Capsules

    Dynamic Routing Between Capsules

    A PyTorch implementation of the NIPS 2017 paper

    Dynamic Routing Between Capsules is a PyTorch implementation of the Capsule Network architecture originally proposed to address limitations in traditional convolutional neural networks. Capsule networks aim to improve how neural models represent spatial hierarchies and relationships between objects within images. Instead of scalar neuron activations, capsules output vectors that encode both the presence of features and their spatial properties such as orientation or pose. The repository implements the dynamic routing algorithm between capsules, which allows lower-level features to route their outputs to higher-level structures that best represent the detected patterns. This approach enables the model to capture part-to-whole relationships in visual data more effectively than standard CNNs. The project serves primarily as a research implementation that demonstrates how capsule networks can be built and trained using modern deep learning frameworks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    ELI5

    ELI5

    A library for debugging/inspecting machine learning classifiers

    ELI5 is a Python library designed to help developers interpret, debug, and explain the predictions of machine learning models. The project focuses on improving model transparency by providing tools that visualize feature importance and prediction reasoning. It supports several popular machine learning frameworks including scikit-learn, XGBoost, LightGBM, CatBoost, and Keras. The library allows users to inspect model weights, analyze decision trees, and compute permutation feature importance for black-box models. It also provides specialized tools such as TextExplainer, which can highlight important words in text classification tasks to explain why a model produced a particular prediction. Additionally, the library integrates explanation algorithms such as LIME to interpret predictions from arbitrary machine learning models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Easy-TensorFlow

    Easy-TensorFlow

    Simple and comprehensive tutorials in TensorFlow

    The goal of this repository is to provide comprehensive tutorials for TensorFlow while maintaining the simplicity of the code. Each tutorial includes a detailed explanation (written in .ipynb) format, as well as the source code (in .py format). There is a necessity to address the motivations for this project. TensorFlow is one of the deep learning frameworks available with the largest community. This repository is dedicated to suggesting a simple path to learn TensorFlow. In addition to the aforementioned points, the large community of TensorFlow enriches the developers with the answer to almost all the questions one may encounter. Furthermore, since most of the developers are using TensorFlow for code development, having hands-on on TensorFlow is a necessity these days. Tensorboard is a powerful visualization suite that is developed to track both the network topology and performance, making debugging even simpler.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    EasyNLP

    EasyNLP

    EasyNLP: A Comprehensive and Easy-to-use NLP Toolkit

    EasyNLP is an easy-to-use NLP development and application toolkit in PyTorch, first released inside Alibaba in 2021. It is built with scalable distributed training strategies and supports a comprehensive suite of NLP algorithms for various NLP applications. EasyNLP integrates knowledge distillation and few-shot learning for landing large pre-trained models, together with various popular multi-modality pre-trained models. It provides a unified framework of model training, inference, and deployment for real-world applications. It has powered more than 10 BUs and more than 20 business scenarios within the Alibaba group. It is seamlessly integrated to Platform of AI (PAI) products, including PAI-DSW for development, PAI-DLC for cloud-native training, PAI-EAS for serving, and PAI-Designer for zero-code model training.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23

    EducationalLCS

    eLCS - Educational Learning Classifier System

    Educational Learning Classifier System (eLCS) is a set of learning classifier system (LCS) educational demos designed to introduce students or researchers to the basics of a modern Michigan-style LCS algorithm. This eLCS package includes 5 different implementations of a basic LCS algorithm, as part of a 6 stage set of demos that will be paired with the first introductory LCS textbook. Each eLCS implementations (from demo 2 up to demo 6) progressively add major components of the entire LCS algorithm in order to illustrate how work, how they are coded, and what impact they have on how an LCS algorithm runs. The Demo 6 version of eLCS is most similar to the UCS algorithm. Each version only includes the minimum code needed to perform the functions they were designed for. This way users can start by examining the simplest version of the code and progress forward. This code is intended to be used as an educational tool, or as algorithmic code building blocks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    EfficientNet Keras

    EfficientNet Keras

    Implementation of EfficientNet model. Keras and TensorFlow Keras

    This repository contains a Keras (and TensorFlow Keras) reimplementation of EfficientNet, a lightweight convolutional neural network architecture achieving state-of-the-art accuracy with an order of magnitude fewer parameters and FLOPS, on both ImageNet and five other commonly used transfer learning datasets. Convolutional Neural Networks (ConvNets) are commonly developed at a fixed resource budget, and then scaled up for better accuracy if more resources are available. In this paper, we systematically study model scaling and identify that carefully balancing network depth, width, and resolution can lead to better performance. Based on this observation, we propose a new scaling method that uniformly scales all dimensions of depth/width/resolution using a simple yet highly effective compound coefficient. We demonstrate the effectiveness of this method on scaling up MobileNets and ResNet.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Elephas

    Elephas

    Distributed Deep learning with Keras & Spark

    Elephas is an extension of Keras, which allows you to run distributed deep learning models at scale with Spark. Elephas currently supports a number of applications. Elephas brings deep learning with Keras to Spark. Elephas intends to keep the simplicity and high usability of Keras, thereby allowing for fast prototyping of distributed models, which can be run on massive data sets. Elephas implements a class of data-parallel algorithms on top of Keras, using Spark's RDDs and data frames. Keras Models are initialized on the driver, then serialized and shipped to workers, alongside with data and broadcasted model parameters. Spark workers deserialize the model, train their chunk of data and send their gradients back to the driver. The "master" model on the driver is updated by an optimizer, which takes gradients either synchronously or asynchronously. Hyper-parameter optimization with elephas is based on hyperas, a convenience wrapper for hyperopt and keras.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB