Showing 648 open source projects for "train"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    flair

    flair

    A very simple framework for state-of-the-art NLP

    ...Flair has simple interfaces that allow you to use and combine different word and document embeddings, including our proposed Flair embeddings and various transformers. A PyTorch NLP framework. Our framework builds directly on PyTorch, making it easy to train your own models and experiment with new approaches using Flair embeddings and classes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    AIMET

    AIMET

    AIMET is a library that provides advanced quantization and compression

    Qualcomm Innovation Center (QuIC) is at the forefront of enabling low-power inference at the edge through its pioneering model-efficiency research. QuIC has a mission to help migrate the ecosystem toward fixed-point inference. With this goal, QuIC presents the AI Model Efficiency Toolkit (AIMET) - a library that provides advanced quantization and compression techniques for trained neural network models. AIMET enables neural networks to run more efficiently on fixed-point AI hardware...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 3
    MLPerf

    MLPerf

    Reference implementations of MLPerf™ training benchmarks

    This is a repository of reference implementations for the MLPerf training benchmarks. These implementations are valid as starting points for benchmark implementations but are not fully optimized and are not intended to be used for "real" performance measurements of software frameworks or hardware. Benchmarking the performance of training ML models on a wide variety of use cases, software, and hardware drives AI performance across the tech industry. The MLPerf Training working group draws on...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    qvac-fabric-llm.cpp

    qvac-fabric-llm.cpp

    QVAC Fabric: cross-platform LLM inference and fine-tuning

    ...The project focuses on removing hardware limitations traditionally associated with LLM deployment by enabling support for a wide range of backends, including Vulkan, Metal, CUDA, and CPU, making it accessible on devices ranging from smartphones to enterprise servers. It introduces native LoRA fine-tuning capabilities that can be executed directly on consumer hardware, allowing developers to train and adapt models locally without relying on cloud infrastructure. A key innovation is its support for BitNet ternary quantized models, enabling highly efficient inference and training even on resource-constrained systems.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    verl

    verl

    Volcano Engine Reinforcement Learning for LLMs

    VERL is a reinforcement-learning–oriented toolkit designed to train and align modern AI systems, from language models to decision-making agents. It brings together supervised fine-tuning, preference modeling, and online RL into one coherent training stack so teams can move from raw data to aligned policies with minimal glue code. The library focuses on scalability and efficiency, offering distributed training loops, mixed precision, and replay/buffering utilities that keep accelerators busy. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    LLM Foundry

    LLM Foundry

    LLM training code for MosaicML foundation models

    ...MPT-7B was trained on the MosaicML platform in 9.5 days with zero human intervention at a cost of ~$200k. Large language models (LLMs) are changing the world, but for those outside well-resourced industry labs, it can be extremely difficult to train and deploy these models. This has led to a flurry of activity centered on open-source LLMs, such as the LLaMA series from Meta, the Pythia series from EleutherAI, the StableLM series from StabilityAI, and the OpenLLaMA model from Berkeley AI Research.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    TorchRec

    TorchRec

    Pytorch domain library for recommendation systems

    TorchRec is a PyTorch domain library built to provide common sparsity & parallelism primitives needed for large-scale recommender systems (RecSys). It allows authors to train models with large embedding tables sharded across many GPUs. Parallelism primitives that enable easy authoring of large, performant multi-device/multi-node models using hybrid data-parallelism/model-parallelism. The TorchRec sharder can shard embedding tables with different sharding strategies including data-parallel, table-wise, row-wise, table-wise-row-wise, and column-wise sharding. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    Rubix ML

    Rubix ML

    A high-level machine learning and deep learning library for PHP

    ...Write less code and iterate faster leaving the hard stuff to us. Rubix ML utilizes a versatile modular architecture that is defined by a few key abstractions and their types and interfaces. Train models in a fraction of the time by installing the optional Tensor extension powered by C. Learners such as neural networks will automatically get a performance boost.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    Lumo Android App

    Lumo Android App

    Android application for Proton Lumo

    Lumo Android App is the official Android client implementation of Lumo, a privacy-first AI assistant created by Proton that lets users interact with an intelligent chatbot securely and confidentially on mobile devices. Lumo is designed so that every conversation remains encrypted and private, meaning chats are not logged, tracked, or used to train external large language models, and all interactions are protected with zero-access encryption so only the user can read them. The Android app offers a responsive, native experience where users can ask questions, summarize documents, brainstorm ideas, or get help with tasks—while Proton’s strict privacy protections extend to the mobile interface. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 10
    MuJoCo Playground

    MuJoCo Playground

    An open source library for GPU-accelerated robot learning

    ...The project includes classic control benchmarks from dm_control, advanced quadruped and bipedal locomotion systems, and dexterous as well as non-prehensile manipulation setups. It also offers optional vision-based training capabilities through integration with Madrona-MJX, allowing researchers to train policies directly from image input on GPUs. MuJoCo Playground supports both the MJX JAX implementation and the Warp physics engine, enabling flexible use across research pipelines. The environments are designed for fast training, compatibility with reinforcement learning libraries, and real-time trajectory visualization using rscope.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Habitat-Sim

    Habitat-Sim

    A flexible, high-performance 3D simulator for Embodied AI research

    ...Determinism and reproducibility are first-class goals, which is critical for benchmarking agents and comparing algorithms. Thanks to its speed and modular design, Habitat-Sim is widely used to prototype embodied agents, train at scale, and evaluate in standardized environments with consistent metrics.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    PyTorch3D

    PyTorch3D

    PyTorch3D is FAIR's library of reusable components for deep learning

    PyTorch3D is a comprehensive library for 3D deep learning that brings differentiable rendering, geometric operations, and 3D data structures into the PyTorch ecosystem. It’s designed to make it easy to build and train neural networks that work directly with 3D data such as meshes, point clouds, and implicit surfaces. The library provides fast GPU-accelerated implementations of rendering pipelines, transformations, rasterization, and lighting—making it possible to compute gradients through full 3D rendering processes. Researchers use it for tasks like shape generation, reconstruction, view synthesis, and visual reasoning. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    CleanVision

    CleanVision

    Automatically find issues in image datasets

    ...CleanVision is super simple -- run the same couple lines of Python code to audit any image dataset! The quality of machine learning models hinges on the quality of the data used to train them, but it is hard to manually identify all of the low-quality data in a big dataset. CleanVision helps you automatically identify common types of data issues lurking in image datasets. This package currently detects issues in the raw images themselves, making it a useful tool for any computer vision task such as: classification, segmentation, object detection, pose estimation, keypoint detection, generative modeling, etc.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 14
    NVIDIA NeMo

    NVIDIA NeMo

    Toolkit for conversational AI

    ...NeMo has separate collections for Automatic Speech Recognition (ASR), Natural Language Processing (NLP), and Text-to-Speech (TTS) models. Each collection consists of prebuilt modules that include everything needed to train on your data. Every module can easily be customized, extended, and composed to create new conversational AI model architectures. Conversational AI architectures are typically large and require a lot of data and compute for training. NeMo uses PyTorch Lightning for easy and performant multi-GPU/multi-node mixed-precision training. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 15
    snntorch

    snntorch

    Deep and online learning with spiking neural networks in Python

    snntorch is a deep learning library that enables researchers and developers to build and train spiking neural networks using the PyTorch framework. Spiking neural networks are biologically inspired models that communicate through discrete spike events rather than continuous activation values, making them closer to how neurons operate in the brain. The library extends PyTorch’s tensor computation capabilities to support gradient-based learning for networks composed of spiking neurons. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 16
    Humanoid-Gym

    Humanoid-Gym

    Reinforcement Learning for Humanoid Robot with Zero-Shot Sim2Real

    Humanoid-Gym is a reinforcement learning framework designed to train locomotion and control policies for humanoid robots using high-performance simulation environments. The system is built on top of NVIDIA Isaac Gym, which allows large-scale parallel simulation of robotic environments directly on GPU hardware. Its primary goal is to enable efficient training of humanoid robots in simulation while enabling policies to transfer effectively to real-world hardware without additional training. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    SpikingJelly

    SpikingJelly

    SpikingJelly is an open-source deep learning framework

    SpikingJelly is an open-source deep learning framework for spiking neural networks that is primarily built on top of PyTorch and aimed at neuromorphic computing research. The project provides the components needed to build, train, and evaluate neural models that communicate through discrete spikes rather than the continuous activations used in conventional artificial neural networks. This makes it especially relevant for researchers interested in biologically inspired computing, event-driven processing, and energy-efficient AI systems. The framework includes neuron models, surrogate gradient training methods, encoding strategies, network components, and utilities for simulation and experimentation, allowing users to develop a wide variety of spiking architectures. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    imgclsmob Deep learning networks

    imgclsmob Deep learning networks

    Sandbox for training deep learning networks

    ...It includes implementations of models used for tasks such as image classification, object detection, semantic segmentation, and pose estimation. The repository also contains scripts that help train models, evaluate performance, and convert trained networks between different frameworks. Several deep learning frameworks are supported, allowing researchers to experiment with architectures in different environments. The project is frequently used by developers who want to study modern convolutional neural network designs and compare their performance across datasets.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Llama-Chinese

    Llama-Chinese

    Llama Chinese community, real-time aggregation

    Llama-Chinese is an open source community initiative focused on adapting and improving Meta’s LLaMA language models for Chinese language applications. The project aggregates datasets, research resources, tutorials, and tools that help developers train and fine-tune LLaMA-based models with Chinese linguistic capabilities. It also provides optimized versions of LLaMA models trained on large-scale Chinese datasets to improve performance in tasks such as translation, summarization, and conversational AI. The community maintains educational materials and technical documentation that help researchers understand the process of training and deploying Chinese-optimized large language models. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    ANE Training

    ANE Training

    Training neural networks on Apple Neural Engine via APIs

    ANE Training is an experimental research project that demonstrates how to train neural networks directly on Apple’s Neural Engine by leveraging reverse-engineered private APIs that are normally inaccessible to developers. The repository implements a from-scratch transformer training pipeline capable of running both forward and backward passes on ANE hardware without relying on CoreML, Metal, or GPU acceleration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    minbpe

    minbpe

    Minimal, clean code for the Byte Pair Encoding (BPE) algorithm

    ...It operates on UTF-8 encoded bytes rather than Unicode characters, which makes it robust to arbitrary text inputs and avoids needing a language-specific character vocabulary. The repository is structured as a teaching-oriented implementation that shows how to train a tokenizer by learning merge rules, then apply those merges to encode text into token IDs and decode tokens back into text. It is intentionally small and readable so developers can understand each stage of BPE, including the mechanics of pair counting, merge application, and vocabulary growth. The project is especially useful for practitioners who want to demystify how LLM tokenizers work or who need a lightweight reference implementation for experimentation.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Megatron-LM

    Megatron-LM

    Ongoing research training transformer models at scale

    Megatron-LM is a GPU-optimized deep learning framework from NVIDIA designed to train extremely large transformer-based language models efficiently at scale. The repository provides both a reference training implementation and Megatron Core, a composable library of high-performance building blocks for custom large-model pipelines. It supports advanced parallelism strategies including tensor, pipeline, data, expert, and context parallelism, enabling training across massive multi-GPU and multi-node clusters. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Polyaxon

    Polyaxon

    MLOps tools for managing & orchestrating the ML LifeCycle

    Polyaxon is an open-source machine learning operations (MLOps) platform built to help individuals, teams, and organizations develop, train, orchestrate, and monitor machine learning and deep learning workflows at scale with reproducibility and automation as core principles. It provides a unified solution for tracking experiments, managing datasets, scheduling jobs, and comparing results across runs, which greatly improves productivity and collaboration in data science teams.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Recursive Language Models

    Recursive Language Models

    General plug-and-play inference library for Recursive Language Models

    RLM (short for Reinforcement Learning Models) is a modular framework that makes it easier to build, train, evaluate, and deploy reinforcement learning (RL) agents across a wide range of environments and tasks. It provides a consistent API that abstracts away many of the repetitive engineering patterns in RL research and application work, letting developers focus on modeling, experimentation, and fine-tuning rather than infrastructure plumbing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    MiniMind-V

    MiniMind-V

    "Big Model" trains a visual multimodal VLM with 26M parameters

    MiniMind-V is an experimental open-source project that aims to train a very small multimodal vision–language model (VLM) from scratch with extremely low compute and cost, making research and experimentation accessible to more people. The repository showcases training workflows and code designed to produce a 26-million parameter model—including both image and text capabilities—using minimal resources in very little time, reflecting a trend toward democratizing AI research.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB