Search Results for "neural network training"

Showing 848 open source projects for "neural network training"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Add Two Lines of Code. Get Full APM. Icon
    Add Two Lines of Code. Get Full APM.

    AppSignal installs in minutes and auto-configures dashboards, alerts, and error tracking.

    Works out of the box for Rails, Django, Express, Phoenix, and more. Monitoring exceptions and performance in no time.
    Start Free
  • 1
    ANE Training

    ANE Training

    Training neural networks on Apple Neural Engine via APIs

    ANE Training is an experimental research project that demonstrates how to train neural networks directly on Apple’s Neural Engine by leveraging reverse-engineered private APIs that are normally inaccessible to developers. The repository implements a from-scratch transformer training pipeline capable of running both forward and backward passes on ANE hardware without relying on CoreML, Metal, or GPU acceleration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Instant Neural Graphics Primitives

    Instant Neural Graphics Primitives

    Instant neural graphics primitives: lightning fast NeRF and more

    Instant Neural Graphics Primitives, is an open-source research project developed by NVIDIA that enables extremely fast training and rendering of neural graphics representations. The system implements several neural graphics primitives including neural radiance fields, signed distance functions, neural images, and neural volumes. These representations are trained using a compact neural network combined with a multiresolution hash encoding that dramatically accelerates both training and rendering processes. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    CoreNet

    CoreNet

    CoreNet: A library for training deep neural networks

    CoreNet is Apple’s internal deep learning framework for distributed neural network training, designed for high scalability, low-latency communication, and strong hardware efficiency. It focuses on enabling large-scale model training across clusters of GPUs and accelerators by optimizing data flow and parallelism strategies. CoreNet provides abstractions for data, tensor, and pipeline parallelism, allowing models to scale without code duplication or heavy manual configuration. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    NeuralOperators.jl

    NeuralOperators.jl

    DeepONets, Neural Operators, Physics-Informed Neural Ops in Julia

    Neural operator is a novel deep learning architecture. It learns an operator, which is a mapping between infinite-dimensional function spaces. It can be used to resolve partial differential equations (PDE). Instead of solving by finite element method, a PDE problem can be resolved by training a neural network to learn an operator mapping from infinite-dimensional space (u, t) to infinite-dimensional space f(u, t).
    Downloads: 3 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    imgclsmob Deep learning networks

    imgclsmob Deep learning networks

    Sandbox for training deep learning networks

    imgclsmob is a deep learning research repository focused on implementing and experimenting with convolutional neural networks for computer vision tasks. The project serves as a sandbox for training and evaluating a wide variety of neural network architectures used in image analysis. It includes implementations of models used for tasks such as image classification, object detection, semantic segmentation, and pose estimation. The repository also contains scripts that help train models, evaluate performance, and convert trained networks between different frameworks. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Axon

    Axon

    Nx-powered Neural Networks

    Training API – An API for quickly training models, inspired by PyTorch Ignite. Axon provides abstractions that enable easy integration while maintaining a level of separation between each component. You should be able to use any of the APIs without dependencies on others. By decoupling the APIs, Axon gives you full control over each aspect of creating and training a neural network.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    Opacus

    Opacus

    Training PyTorch models with differential privacy

    Opacus is a library that enables training PyTorch models with differential privacy. It supports training with minimal code changes required on the client, has little impact on training performance, and allows the client to online track the privacy budget expended at any given moment. Vectorized per-sample gradient computation that is 10x faster than micro batching. Supports most types of PyTorch models and can be used with minimal modification to the original neural network. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    PyTorch Forecasting

    PyTorch Forecasting

    Time series forecasting with PyTorch

    ...Multiple neural network architectures for timeseries forecasting that have been enhanced for real-world deployment and come with in-built interpretation capabilities. The package is built on PyTorch Lightning to allow training on CPUs, single and multiple GPUs out-of-the-box.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    Tiny CUDA Neural Networks

    Tiny CUDA Neural Networks

    Lightning fast C++/CUDA neural network framework

    This is a small, self-contained framework for training and querying neural networks. Most notably, it contains a lightning-fast "fully fused" multi-layer perceptron (technical paper), a versatile multiresolution hash encoding (technical paper), as well as support for various other input encodings, losses, and optimizers. We provide a sample application where an image function (x,y) -> (R,G,B) is learned.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    Stock prediction deep neural learning

    Stock prediction deep neural learning

    Predicting stock prices using a TensorFlow LSTM

    Predicting stock prices can be a challenging task as it often does not follow any specific pattern. However, deep neural learning can be used to identify patterns through machine learning. One of the most effective techniques for series forecasting is using LSTM (long short-term memory) networks, which are a type of recurrent neural network (RNN) capable of remembering information over a long period of time. This makes them extremely useful for predicting stock prices. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 11
    Flax

    Flax

    Flax is a neural network library for JAX

    Flax is a flexible neural-network library for JAX that embraces functional programming while offering ergonomic module abstractions. Its design separates pure computation from state by threading parameter collections and RNGs explicitly, enabling reproducibility, transformation, and easy experimentation with JAX transforms like jit, pmap, and vmap. Modules define parameterized computations, but initialization and application remain side-effect free, which pairs naturally with JAX’s staging and compilation model. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    SentencePiece

    SentencePiece

    Unsupervised text tokenizer for Neural Network-based text generation

    SentencePiece is an unsupervised text tokenizer and detokenizer mainly for Neural Network-based text generation systems where the vocabulary size is predetermined prior to the neural model training. SentencePiece implements subword units (e.g., byte-pair-encoding (BPE) [Sennrich et al.]) and unigram language model [Kudo.]) with the extension of direct training from raw sentences. SentencePiece allows us to make a purely end-to-end system that does not depend on language-specific pre/postprocessing. ...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 13
    Penzai

    Penzai

    A JAX research toolkit to build, edit, & visualize neural networks

    Penzai, developed by Google DeepMind, is a JAX-based library for representing, visualizing, and manipulating neural network models as functional pytree data structures. It is designed to make machine learning research more interpretable and interactive, particularly for tasks like model surgery, ablation studies, architecture debugging, and interpretability research. Unlike conventional neural network libraries, Penzai exposes the full internal structure of models, enabling fine-grained inspection and modification after training. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Bittensor

    Bittensor

    Internet-scale Neural Networks

    Bittensor is a decentralized machine learning protocol that allows AI models to collaborate, learn, and earn tokens within a global network. It introduces a blockchain-based economy for neural networks, where participants are incentivized to contribute valuable knowledge and compute power. Bittensor combines peer-to-peer learning with on-chain rewards, creating a self-governing, scalable AI system that evolves without centralized control. It is a novel approach to aligning incentives in AI development, empowering open contributions while preserving model ownership and decentralization.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 15
    DeepDetect

    DeepDetect

    Deep Learning API and Server in C++14 support for Caffe, PyTorch

    The core idea is to remove the error sources and difficulties of Deep Learning applications by providing a safe haven of commoditized practices, all available as a single core. While the Open Source Deep Learning Server is the core element, with REST API, and multi-platform support that allows training & inference everywhere, the Deep Learning Platform allows higher level management for training neural network models and using them as if they were simple code snippets. Ready for applications of image tagging, object detection, segmentation, OCR, Audio, Video, Text classification, CSV for tabular data and time series. Neural network templates for the most effective architectures for GPU, CPU, and Embedded devices. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    AlphaTree

    AlphaTree

    DNN && GAN && NLP && BIG DATA

    AlphaTree is an educational repository that provides a visual roadmap of deep learning models and related artificial intelligence technologies. The project focuses on explaining the historical development and relationships between major neural network architectures used in modern machine learning. It presents diagrams and documentation describing the evolution of models such as LeNet, AlexNet, VGG, ResNet, DenseNet, and Inception networks. The repository organizes these architectures into a structured learning path that helps learners understand how deep learning models improved over time through changes in depth, architectural complexity, and training techniques. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    deepjazz

    deepjazz

    Deep learning driven jazz generation using Keras & Theano

    ...The project was originally created during a hackathon and was designed to show how neural networks can emulate creative tasks traditionally associated with human musicians. The repository includes preprocessing scripts for preparing MIDI data, training scripts for building the neural network model, and code for generating new compositions.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    PyTorch-Tutorial-2nd

    PyTorch-Tutorial-2nd

    CV, NLP, LLM project applications, and advanced engineering deployment

    PyTorch-Tutorial-2nd is an open-source educational repository that provides structured tutorials for learning deep learning with the PyTorch framework. The project serves as a practical companion to a second edition of a PyTorch learning guide and is designed to help learners understand neural network concepts through hands-on coding examples. The repository covers a wide range of topics including tensor operations, neural network construction, model training workflows, and optimization strategies. It also introduces practical machine learning techniques such as convolutional neural networks, recurrent networks, and other architectures commonly used in modern AI applications. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    face.evoLVe

    face.evoLVe

    High-Performance Face Recognition Library on PaddlePaddle & PyTorch

    face.evoLVe is a high-performance face recognition library designed for research and real-world applications in computer vision. The project provides a comprehensive framework for building and training modern face recognition models using deep learning architectures. It includes components for face alignment, landmark localization, data preprocessing, and model training pipelines that allow developers to construct end-to-end facial recognition systems. The repository supports multiple neural network backbones such as ResNet, DenseNet, MobileNet, and ShuffleNet, enabling experimentation with different architectures depending on performance requirements. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    thorough-pytorch

    thorough-pytorch

    PyTorch Getting Started Tutorial, read online

    ...The repository provides tutorials and practical exercises that guide learners from fundamental PyTorch concepts to more advanced deep learning techniques. It emphasizes a learning approach that combines theoretical explanations with hands-on coding exercises so that students can build and experiment with neural networks directly. The project encourages collaborative learning and often organizes materials in a step-by-step progression that gradually increases in complexity. Topics include neural network fundamentals, training procedures, model evaluation, and practical deep learning workflows. By combining structured lessons with programming projects, the repository aims to help learners develop both conceptual understanding and practical implementation skills.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Learn PyTorch for Deep Learning

    Learn PyTorch for Deep Learning

    Materials for the Learn PyTorch for Deep Learning

    ...Instead of focusing heavily on theory alone, the repository encourages learners to experiment with code and develop practical machine learning skills through guided examples and exercises. The materials include Jupyter notebooks, explanations of core deep learning concepts, and step-by-step demonstrations of building and training neural networks. Throughout the lessons, users learn how to work with tensors, create neural network architectures, manage training workflows, and evaluate model performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    BitNet

    BitNet

    BitNet: Scaling 1-bit Transformers for Large Language Models

    BitNet is a machine learning research implementation that explores extremely low-precision neural network architectures designed to dramatically reduce the computational cost of large language models. The project implements the BitNet architecture described in research on scaling transformer models using extremely low-bit quantization techniques. In this approach, neural network weights are quantized to approximately one bit per parameter, allowing models to operate with far lower memory usage than traditional 16-bit or 32-bit neural networks. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 23
    NNCF

    NNCF

    Neural Network Compression Framework for enhanced OpenVINO

    NNCF (Neural Network Compression Framework) is an optimization toolkit for deep learning models, designed to apply quantization, pruning, and other techniques to improve inference efficiency.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 24
    Lightweight' GAN

    Lightweight' GAN

    Implementation of 'lightweight' GAN, proposed in ICLR 2021

    ...The main contribution of the paper is a skip-layer excitation in the generator, paired with autoencoding self-supervised learning in the discriminator. Quoting the one-line summary "converge on single gpu with few hours' training, on 1024 resolution sub-hundred images". Augmentation is essential for Lightweight GAN to work effectively in a low data setting. You can test and see how your images will be augmented before they pass into a neural network (if you use augmentation). The general recommendation is to use suitable augs for your data and as many as possible, then after some time of training disable the most destructive (for image) augs. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Hivemind

    Hivemind

    Decentralized deep learning in PyTorch. Built to train models

    Hivemind is a PyTorch library for decentralized deep learning across the Internet. Its intended usage is training one large model on hundreds of computers from different universities, companies, and volunteers. Distributed training without a master node: Distributed Hash Table allows connecting computers in a decentralized network. Fault-tolerant backpropagation: forward and backward passes succeed even if some nodes are unresponsive or take too long to respond. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB