Open Source Python Artificial Intelligence Software for ChromeOS

Python Artificial Intelligence Software for ChromeOS

Browse free open source Python Artificial Intelligence Software for ChromeOS and projects below. Use the toggles on the left to filter open source Python Artificial Intelligence Software for ChromeOS by OS, license, language, programming language, and project status.

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    Frigate

    Frigate

    NVR with realtime local object detection for IP cameras

    Frigate - NVR With Realtime Object Detection for IP Cameras A complete and local NVR designed for Home Assistant with AI object detection. Uses OpenCV and Tensorflow to perform realtime object detection locally for IP cameras. Use of a Google Coral Accelerator is optional, but highly recommended. The Coral will outperform even the best CPUs and can process 100+ FPS with very little overhead.
    Downloads: 40 This Week
    Last Update:
    See Project
  • 2
    WanGP

    WanGP

    AI video generator optimized for low VRAM and older GPUs use

    Wan2GP is an open source AI video generation toolkit designed to make modern generative models accessible on consumer-grade hardware with limited GPU memory. It acts as a unified interface for running multiple video, image, and audio generation models, including Wan-based models as well as other systems like Hunyuan Video, Flux, and Qwen. A key focus of the project is reducing VRAM requirements, enabling some workflows to run on as little as 6 GB while still supporting older Nvidia and certain AMD GPUs. Wan2GP provides a full web-based interface that simplifies interaction with complex generative pipelines, making it easier to configure prompts, models, and rendering settings. It also integrates a wide range of utilities such as prompt enhancement, mask editing, motion design, and extraction tools for pose, depth, and flow data to support advanced video workflows.
    Downloads: 30 This Week
    Last Update:
    See Project
  • 3
    Every Code

    Every Code

    Local AI coding agent CLI with multi-agent orchestration tools

    Every Code (often referred to simply as Code) is a fast, local AI-powered coding agent designed to run directly in the terminal environment. It is a community-driven fork of the Codex CLI, with a strong emphasis on improving real-world developer ergonomics and workflows. Every Code enhances the traditional coding assistant model by introducing multi-agent orchestration, allowing multiple AI agents to collaborate, compare solutions, and refine outputs in parallel. It supports integration with various AI providers, enabling users to route tasks across different models depending on their needs. Every Code also includes browser integration and automation capabilities, extending its usefulness beyond simple code generation into more complex development tasks. Customization is a key focus, with support for theming, configurable settings, and reasoning controls that allow developers to fine-tune how the agent behaves.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 4
    MLflow

    MLflow

    Open source platform for the machine learning lifecycle

    MLflow is a platform to streamline machine learning development, including tracking experiments, packaging code into reproducible runs, and sharing and deploying models. MLflow offers a set of lightweight APIs that can be used with any existing machine learning application or library (TensorFlow, PyTorch, XGBoost, etc), wherever you currently run ML code (e.g. in notebooks, standalone applications or the cloud).
    Downloads: 8 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    Open-LLM-VTuber

    Open-LLM-VTuber

    Open source AI VTuber platform with voice chat and Live2D avatars

    Open-LLM-VTuber is an open source platform designed to create AI-powered VTuber characters that can interact with users through voice and animated avatars. It enables hands-free conversations with large language models by combining speech recognition, language processing, and text-to-speech synthesis into a single system. Users can speak directly to the AI character, and the system can respond with a generated voice while animating a Live2D avatar to simulate a talking virtual personality. Open-LLM-VTuber is modular, allowing developers to swap or configure different language models, speech recognition engines, and voice synthesis systems depending on their needs. It can run locally and supports both offline and online AI services, giving users flexibility in how models and resources are used. Open-LLM-VTuber was originally inspired by the goal of recreating an AI VTuber experience using open source tools that work across multiple operating systems.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 6
    Astron Agent

    Astron Agent

    Enterprise platform for building and orchestrating AI agent workflows

    Astron Agent is an enterprise-grade platform designed for building and managing intelligent AI agent workflows in production environments. It provides a development environment that combines workflow orchestration, model management, and integration with various AI tools and services. Astron Agent enables organizations to design complex agent-driven processes that coordinate models, automation tools, and enterprise systems. It also integrates robotic process automation capabilities so agents can execute tasks across digital systems instead of only generating responses. Astron Agent supports scalable and high-availability deployments, allowing teams to run reliable AI agent infrastructure in distributed environments. It includes collaboration features that allow teams to develop, manage, and operate AI applications together. With its extensible architecture and enterprise-focused design, it aims to help organizations build production-ready intelligent agent solutions.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 7
    Mito

    Mito

    AI-powered Jupyter spreadsheet that converts workflows into Python

    Mito is an open source set of Jupyter extensions designed to speed up Python workflows and data analysis. It combines a spreadsheet-style interface with AI-assisted coding, allowing users to explore, clean, and transform data without switching tools. Mito includes a context-aware AI assistant that helps generate code, debug errors, and guide workflows directly inside Jupyter. Its spreadsheet layer supports familiar functions such as filters, pivot tables, and formulas, while automatically converting every action into production-ready Python code. This removes the need to manually translate spreadsheet logic into scripts. Mito also integrates with tools like Streamlit and Dash, enabling users to embed interactive spreadsheet functionality into apps with minimal setup. Built for analysts, developers, and teams, it simplifies automation, reduces repetitive tasks, and accelerates the transition from data exploration to reusable code.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 8
    FAY

    FAY

    Framework for building AI-powered interactive digital humans and agent

    Fay is an open source framework designed to build and deploy interactive digital humans powered by large language models. It acts as a middleware layer that connects digital character technologies with conversational AI systems and business applications. Fay supports various types of digital humans, including 2.5D and 3D avatars, and can be integrated with applications running on mobile devices, PCs, web platforms, and embedded systems. Its architecture allows developers to combine different AI components such as speech recognition, text-to-speech, and large language models to create conversational digital agents. Fay provides multiple interfaces for text, voice, and digital human control, enabling developers to build interactive assistants, virtual presenters, or automated service agents. It also supports custom knowledge bases and configurable behaviors so developers can tailor the personality and responses of the digital human.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 9
    GitDiagram

    GitDiagram

    AI tool that converts GitHub repositories into interactive diagrams

    GitDiagram is an open source web application designed to help developers quickly understand the structure and architecture of GitHub repositories by automatically generating interactive diagrams. It analyzes repository metadata such as the file tree and project documentation to build a visual representation of how different components of a project relate to one another. It uses an AI-powered pipeline to interpret repository structure and transform that information into system design diagrams rendered with Mermaid visualization. These diagrams provide a high-level overview of a codebase, making it easier for developers to explore unfamiliar projects or understand large and complex repositories. Users can interact with the generated diagrams by clicking components to navigate directly to related files or directories within the repository. GitDiagram combines a modern web frontend with a backend service that processes repository data and generates diagrams dynamically.
    Downloads: 4 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 10
    SQLFlow

    SQLFlow

    SQL compiler bridging databases and machine learning workflows

    SQLFlow is an open source project designed to bridge the gap between traditional SQL-based data processing and modern machine learning workflows by extending SQL syntax with AI capabilities. It acts as a compiler that translates SQL programs into executable workflows, enabling users to train, evaluate, and deploy machine learning models directly from SQL statements. It integrates with multiple database engines such as MySQL, Hive, and MaxCompute, while also supporting machine learning frameworks like TensorFlow and XGBoost. By embedding machine learning operations into SQL, it removes the need for users to switch between programming languages such as Python or R, simplifying the overall workflow. SQLFlow also supports model training, prediction, and explanation tasks, allowing data practitioners to work entirely within a familiar query interface.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 11
    Harbor LLM

    Harbor LLM

    Run a full local LLM stack with one command using Docker

    Harbor is an open source, containerized toolkit designed to simplify running local large language model (LLM) environments. It combines a CLI and companion app to launch backends, frontends, and supporting services with minimal setup. With a single command, users can start preconfigured tools like Ollama and Open WebUI, enabling chat, workflows, and integrations immediately. Harbor supports multiple inference engines, including llama.cpp and vLLM, and connects them seamlessly to user interfaces. It also includes tools for web retrieval, image generation, voice interaction, and workflow automation. Built on Docker, Harbor allows services to run in isolated containers while communicating over a local network. It is intended for local development and experimentation rather than production deployment, giving developers a flexible way to explore AI systems, test configurations, and manage complex LLM stacks without manual wiring or setup overhead.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 12
    KeepChatGPT

    KeepChatGPT

    Browser userscript that enhances ChatGPT reliability and usability

    KeepChatGPT is an open source browser userscript designed to enhance the reliability, usability, and efficiency of the ChatGPT web interface. It runs through userscript managers and injects additional functionality directly into the page, allowing users to improve their workflow without requiring a backend service or separate application. It focuses on solving common problems experienced during AI conversations, such as session timeouts, network errors, message failures, and interruptions during long chats. By automating session refresh and maintaining active connections, KeepChatGPT reduces the need for repeated manual steps when recovering from errors or expired sessions. KeepChatGPT also introduces a variety of enhancements that improve the overall interface and user experience, including page cleanup, expanded display layouts, conversation cloning, and detailed chat information.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 13
    LaVague

    LaVague

    Framework for building AI agents that automate complex web tasks

    LaVague is an open source framework designed to help developers build AI-powered web agents capable of automating tasks across websites and web applications. It implements the concept of a Large Action Model framework, allowing agents to interpret a user-provided objective and translate it into a sequence of actions performed in a browser. These agents can navigate web pages, retrieve information, fill out forms, and execute multi-step workflows automatically. LaVague is centered around a World Model that analyzes the current webpage state and determines the next set of instructions, combined with an Action Engine that converts those instructions into executable automation code. It can use browser automation tools such as Selenium or Playwright to interact with websites programmatically. Developers can integrate various language models and configure the agent’s reasoning and execution behavior to suit different automation scenarios.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 14
    Search with Lepton

    Search with Lepton

    Lightweight demo to build a conversational AI search engine quickly

    Search with Lepton is an open source demonstration project that shows how to build a conversational search engine using the Lepton AI framework. It combines traditional web search with large language models to provide natural language answers to user queries. It retrieves information from supported search engines and uses that context to generate responses through a retrieval-augmented generation approach. The implementation is intentionally minimal, containing fewer than 500 lines of code while still providing a complete working example of an AI-powered search system. It includes both a backend service written in Python and a web interface that allows users to interact with the search engine in a conversational format. Developers can configure different search providers and language models through environment variables, making it flexible for experimentation and prototyping.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 15
    TypeChat

    TypeChat

    Library for building type-safe natural language interfaces with LLMs

    TypeChat is an open source library developed by Microsoft that simplifies the creation of natural language interfaces by using type definitions to structure interactions with large language models. Traditional natural language interfaces often relied on complex decision trees to interpret user intent and gather required inputs. With the rise of large language models, developers can interpret user requests more easily, but they still face challenges related to output reliability, safety, and structured responses. TypeChat addresses these challenges by replacing traditional prompt engineering with a concept called schema engineering. Instead of writing complex prompts, developers define types that represent the intents supported by their applications. It then uses those type definitions to construct prompts for language models and translate user input into structured data that follows the defined schema.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 16
    AI-Media2Doc

    AI-Media2Doc

    AI tool converting video/audio into structured documents instantly

    AI-Media2Doc is a web-based application that uses large language models to convert video and audio content into structured, readable documents in a single workflow. It is designed to transform multimedia inputs into formats such as knowledge notes, summaries, mind maps, and social-style articles, making content easier to review and reuse. AI-Media2Doc emphasizes privacy by processing media locally in the browser using WebAssembly-based ffmpeg, ensuring that original video files are not uploaded externally. It separates client-side media handling from backend AI processing, reducing data exposure while still enabling transcription and document generation. AI-Media2Doc supports flexible customization through prompts, allowing users to tailor output styles based on their needs. It also includes features like subtitle export and AI-assisted follow-up questioning for deeper interaction with the generated content.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 17
    Basic Memory

    Basic Memory

    Persistent AI memory using local Markdown knowledge graphs

    Basic Memory is an open source knowledge system that turns AI conversations into persistent, structured knowledge you control. Instead of losing context after each chat, it stores information as simple Markdown files on your device, allowing both you and AI to read and write to the same knowledge base. It uses the Model Context Protocol (MCP) so compatible AI tools can access, update, and build on your notes across sessions. Basic Memory creates a semantic knowledge graph by linking related ideas, making it easier to retrieve, expand, and connect information over time. With a local-first design, your data stays private and portable, while optional cloud sync enables cross-device access. It combines simplicity with powerful indexing and search, giving you a flexible way to build long-term memory for projects, research, and workflows.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    Paperless-AI

    Paperless-AI

    AI-powered document analysis and tagging for Paperless-ngx

    Paperless-AI is an AI-powered extension designed to enhance document management within Paperless-ngx by automating analysis, classification, and organization tasks. It continuously monitors incoming documents and processes them using various AI backends, enabling automatic assignment of titles, tags, document types, and correspondents. It integrates with multiple OpenAI-compatible services as well as local models, giving users flexibility in how document intelligence is handled. A key capability is its use of retrieval-augmented generation, which enables semantic search and natural language interaction across an entire document archive. Users can ask contextual questions about their files and receive precise answers based on full document understanding rather than simple keyword matching. Paperless-AI also includes a web interface for manual review and tagging, allowing greater control when handling sensitive or complex documents.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19
    ChatGPT Plugins Collection

    ChatGPT Plugins Collection

    An unofficial collection of Plugins for ChatGPT

    ChatGPT-Plugins-Collection is a community-driven repository that gathers examples and resources for building, testing, and experimenting with ChatGPT plugins. The collection provides a variety of plugin implementations that showcase different use cases, helping developers learn how to extend ChatGPT’s functionality. It is designed to serve both as a learning resource for beginners and a reference point for more experienced developers. By centralizing community contributions, the repository highlights practical applications of plugins across domains such as productivity, data access, and automation. The project also serves as a starting point for developers interested in building their own custom plugins, offering inspiration and code samples. With its open structure, it encourages collaboration and knowledge sharing in the growing ecosystem of ChatGPT extensions.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    Cognita

    Cognita

    Open source RAG framework for building scalable modular AI apps

    Cognita is an open source framework designed to help developers build, organize, and deploy Retrieval-Augmented Generation (RAG) applications in a structured and production-ready way. It addresses the gap between quick experimentation in notebooks and the complexity of deploying scalable AI systems by introducing a modular and API-driven architecture. Cognita provides reusable components such as parsers, data loaders, embedders, retrievers, and query controllers, allowing teams to customize each stage of the RAG pipeline independently. It includes both a backend service and a frontend interface, enabling users to upload documents, experiment with configurations, and perform question-answering tasks interactively. Cognita supports incremental indexing, meaning it processes only new or updated data to reduce computational overhead and improve efficiency.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    Gitingest

    Gitingest

    Create prompt-friendly codebase digests from any Git repository URL

    Gitingest is a developer utility that converts an entire Git repository into a structured, prompt-friendly text digest suitable for use with large language models. It analyzes a repository and produces a consolidated textual representation that includes the file structure and code content in an organized format. This makes it easier to provide meaningful code context when working with AI systems that require compact, readable inputs. Developers can generate these digests from either a local directory or a remote repository by supplying a repository path or URL. The generated output is optimized for prompt usage, helping AI models understand codebases more effectively without requiring manual file aggregation. In addition to producing the code digest, Gitingest also calculates statistics about the extracted content such as repository structure, total size of the extract, and token count. Gitingest can be used as a command line utility or integrated directly into Python applications.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    MCP UI

    MCP UI

    SDK for building interactive UI components over MCP for AI tools

    mcp-ui is a software development kit designed to bring interactive user interface capabilities to applications built on the Model Context Protocol (MCP). It enables developers to create rich, dynamic UI components that can be delivered from an MCP server and rendered seamlessly by a compatible client. Instead of returning only text responses, tools can provide structured UI resources such as HTML or remote-rendered components, allowing more engaging and functional interactions. mcp-ui introduces a standardized approach where tools and their associated interfaces are linked through metadata, enabling clients to automatically discover and display the correct UI. It includes both client-side and server-side SDKs, making it possible to define UI elements on the backend and handle user interactions on the frontend. It supports multiple programming environments, including TypeScript, Python, and Ruby, broadening its accessibility for developers.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    OpenMemory

    OpenMemory

    Local long-term memory engine for AI apps with persistent storage

    OpenMemory is a self-hosted memory engine designed to provide long-term, persistent storage for AI and LLM-powered applications. It enables developers to give otherwise stateless models a structured memory layer that can store, retrieve, and manage contextual information over time. OpenMemory is built around a hierarchical memory architecture that organizes data into semantic sectors and connects them through a graph-based structure for efficient retrieval. It supports multiple embedding strategies, including synthetic and semantic embeddings, allowing developers to balance speed and accuracy depending on their use case. OpenMemory integrates with various AI tools and environments, offering SDKs and APIs that simplify adding memory capabilities to applications. OpenMemory also includes features like memory decay, reinforcement, and temporal filtering to ensure relevant information remains prioritized while outdated data gradually loses importance.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Super Magic

    Super Magic

    All-in-one AI productivity platform with agents, workflows, and IM

    Magic is an open source all-in-one AI productivity platform designed to help organizations build, deploy, and scale AI-driven applications efficiently. It is not a single tool but a complete product ecosystem composed of multiple integrated systems that work together to enhance productivity across different business scenarios. Magic centers around a general-purpose AI agent system called Super Magic, which can autonomously understand tasks, plan actions, execute workflows, and perform error correction. Alongside this, Magic includes a visual workflow engine that enables users to design complex AI processes using a drag-and-drop interface without requiring extensive coding knowledge. It also provides an enterprise-grade instant messaging system that integrates AI conversations with internal communication, allowing teams to collaborate while leveraging intelligent assistants. Its architecture is built using a microservices approach with containerized services.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    Text Embeddings Inference

    Text Embeddings Inference

    High-performance inference server for text embeddings models API layer

    Text Embeddings Inference is a high-performance server designed to serve text embedding models efficiently in production environments. It focuses on delivering fast and scalable embedding generation by leveraging optimized inference techniques and modern hardware acceleration. It is built to support transformer-based embedding models, making it suitable for tasks such as semantic search, clustering, and retrieval-augmented systems. It provides an API interface that allows developers to integrate embedding capabilities into applications without managing model internals directly. Text Embeddings Inference is optimized for throughput and low latency, enabling it to handle large volumes of requests reliably. It also emphasizes ease of deployment, often using containerization and configurable runtime options to adapt to different infrastructure setups.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB