Showing 155 open source projects for "cpu memory usage"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Network Management Software and Tools for Businesses and Organizations | Auvik Networks Icon
    Network Management Software and Tools for Businesses and Organizations | Auvik Networks

    Mapping, inventory, config backup, and more.

    Reduce IT headaches and save time with a proven solution for automated network discovery, documentation, and performance monitoring. Choose Auvik because you'll see value in minutes, and stay with us to improve your IT for years to come.
    Learn More
  • 1
    FastSD CPU

    FastSD CPU

    Fast stable diffusion on CPU and AI PC

    ...The repository contains multiple interfaces including a desktop GUI for simple generation, an advanced web-based UI with support for extensions like LoRA and ControlNet, and a command-line interface for scripted usage or server deployments. With support for performance-oriented libraries such as OpenVINO and hardware acceleration on platforms like Intel AI PCs, FastSD CPU aims to shrink generation times dramatically compared with naive CPU implementations.
    Downloads: 51 This Week
    Last Update:
    See Project
  • 2
    auto-cpufreq

    auto-cpufreq

    Automatic CPU speed & power optimizer for Linux

    Automatic CPU speed & power optimizer for Linux. Actively monitors laptop battery state, CPU usage, CPU temperature, and system load, ultimately allowing you to improve battery life without making any compromises.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 3
    AGI (Android GPU Inspector)

    AGI (Android GPU Inspector)

    Android GPU Inspector

    Android GPU Inspector (AGI) is a desktop tool for profiling, tracing, and debugging graphics workloads running on Android devices. It helps developers analyze Vulkan and OpenGL ES applications at the system, frame, and draw-call levels to uncover GPU and CPU bottlenecks. AGI captures detailed performance counters, timelines, and pipeline state to reveal stalls, overdraw, shader hotspots, and inefficient resource usage. Its frame debugger lets you step through commands, inspect render targets and textures, review shaders, and experiment with pipeline changes to validate optimizations. ...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 4
    pg_activity

    pg_activity

    pg_activity is a top like application for PostgreSQL server activity

    Command line tool for PostgreSQL server activity monitoring. pg_activity is a PostgreSQL monitoring tool that provides real-time insights into database performance, helping database administrators manage and troubleshoot PostgreSQL instances.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Get full visibility and control over your tasks and projects with Wrike. Icon
    Get full visibility and control over your tasks and projects with Wrike.

    A cloud-based collaboration, work management, and project management software

    Wrike offers world-class features that empower cross-functional, distributed, or growing teams take their projects from the initial request stage all the way to tracking work progress and reporting results.
    Learn More
  • 5
    psutil

    psutil

    Cross-platform lib for process and system monitoring in Python

    psutil is a widely adopted cross-platform Python library designed to retrieve detailed information about system utilization and running processes in a consistent and programmatic way. It exposes a rich API that allows developers to inspect CPU usage, memory consumption, disk activity, network statistics, and hardware sensors without relying on platform-specific tools. The library effectively replicates and unifies the capabilities of classic UNIX utilities such as ps, top, netstat, and free, making it especially valuable for monitoring, profiling, and process management workflows. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    FlexLLMGen

    FlexLLMGen

    Running large language models on a single GPU

    ...The system focuses on high-throughput generation workloads where large batches of text must be processed quickly, such as large-scale data extraction or document analysis tasks. Instead of requiring expensive multi-GPU systems, the framework uses techniques such as memory offloading, compression, and optimized batching to run large models on commodity hardware. The architecture distributes computation and memory usage across the GPU, CPU, and disk in order to maximize the number of tokens processed during inference. This design allows organizations to deploy powerful language models for high-volume tasks without the infrastructure costs typically associated with large-scale AI systems. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Robusta KRR

    Robusta KRR

    Prometheus-based Kubernetes Resource Recommendations

    Robusta KRR (Kubernetes Resource Recommender) is a CLI tool for optimizing resource allocation in Kubernetes clusters. It gathers pod usage data from Prometheus and recommends requests and limits for CPU and memory. This reduces costs and improves performance.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 8
    LuxTTS

    LuxTTS

    A high-quality rapid TTS voice cloning model

    ...Its design emphasizes efficiency and practicality, fitting within modest GPU memory footprints.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 9
    Ajenti 2

    Ajenti 2

    Ajenti Core and stock plugins

    ...Plugin development is quick and pleasant with Ajenti APIs. Write your first plugin. Pleasant to look at, satisfying to click and accessible anywhere from tablets and mobile. Small memory footprint and CPU usage. Runs on low-end machines, wall plugs, routers and so on.
    Downloads: 7 This Week
    Last Update:
    See Project
  • Inventory and Order Management Software for Multichannel Sellers Icon
    Inventory and Order Management Software for Multichannel Sellers

    Avoid stockouts, overselling, and losing control as your business grows.

    We are the most powerful inventory and order management platform for Amazon, Walmart, and multichannel product sellers. Centralize orders, product information, and fulfillment operations to run more efficiently, sell more products, and stay compliant with marketplace requirements so you can grow profitably.
    Learn More
  • 10
    MemU

    MemU

    MemU is an open-source memory framework for AI companions

    MemU is an agentic memory layer for LLM applications, specifically designed for AI companions. Transform your memory into an intelligent file system that automatically organizes, connects, and evolves with your memories. Simple, fast, and reliable memory infrastructure for AI applications. Powerful tools and dedicated support to scale your AI applications with confidence.
    Downloads: 20 This Week
    Last Update:
    See Project
  • 11
    HunyuanVideo

    HunyuanVideo

    HunyuanVideo: A Systematic Framework For Large Video Generation Model

    ...The framework aims to push the boundaries of video generation quality, incorporating multiple innovative approaches to improve the realism and coherence of the generated content. Release of FP8 model weights to reduce GPU memory usage / improve efficiency. Parallel inference code to speed up sampling, utilities and tests included.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 12
    Magika

    Magika

    Fast and accurate AI powered file content types detection

    ...It also emphasizes reproducibility and developer ergonomics with clear install and usage instructions for common platforms. A public site complements the repo with background, examples, and guidance for integrating Magika into existing workflows.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 13
    whisper-timestamped

    whisper-timestamped

    Multilingual Automatic Speech Recognition with word-level timestamps

    Multilingual Automatic Speech Recognition with word-level timestamps and confidence. Whisper is a set of multi-lingual, robust speech recognition models trained by OpenAI that achieve state-of-the-art results in many languages. Whisper models were trained to predict approximate timestamps on speech segments (most of the time with 1-second accuracy), but they cannot originally predict word timestamps. This repository proposes an implementation to predict word timestamps and provide a more...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 14
    Kitten TTS

    Kitten TTS

    State-of-the-art TTS model under 25MB

    KittenTTS is an open-source, ultra-lightweight, and high-quality text-to-speech model featuring just 15 million parameters and a binary size under 25 MB. It is designed for real-time CPU-based deployment across diverse platforms. Ultra-lightweight, model size less than 25MB. CPU-optimized, runs without GPU on any device. High-quality voices, several premium voice options available. Fast inference, optimized for real-time speech synthesis.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 15
    Echo System Info Viewer

    Echo System Info Viewer

    Simple and lightweight system information viewer for Windows

    Echo System Info Viewer is a lightweight Windows system information viewer that quickly displays essential hardware and system details. The application shows real-time CPU usage, memory (RAM) usage, and disk space information in a simple and clean interface. Unlike heavy monitoring software, Echo System Info Viewer is designed to be extremely fast, minimal, and fully offline. Key benefits: Instant system overview Real-time CPU and RAM usage Disk space information Lightweight and portable No telemetry, no ads Works completely offline Typical use cases: Quickly check CPU and RAM usage View basic system information on Windows Monitor system performance on older computers Lightweight alternative to heavy monitoring tools
    Downloads: 6 This Week
    Last Update:
    See Project
  • 16
    Fabulously Optimized

    Fabulously Optimized

    A simple Minecraft modpack focusing on performance and graphics

    A simple Minecraft modpack focusing on performance and graphics enhancements. A Minecraft modpack focused on performance, providing a smooth experience with multiple optimization mods.
    Downloads: 14 This Week
    Last Update:
    See Project
  • 17
    Pocket TTS

    Pocket TTS

    A TTS that fits in your CPU (and pocket)

    Pocket TTS is a lightweight text-to-speech project designed to run efficiently on CPUs, targeting developers who want local speech generation without depending on GPUs or hosted web APIs. It is built to feel practical in everyday applications, where installation and usage should be as simple as adding a dependency and calling a function. The project focuses on keeping the runtime footprint manageable while still producing natural-sounding speech, which makes it attractive for offline tools, prototypes, and privacy-sensitive workflows. Because it is CPU-oriented, it fits well in server environments where GPU access is limited, in desktop apps, or in edge deployments where simplicity matters more than maximum throughput. ...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 18
    OpenAI Agents (Python)

    OpenAI Agents (Python)

    A lightweight, powerful framework for multi-agent workflows

    openai-agents-python is a library developed by OpenAI to simplify the process of creating and running agents that interact with tools and APIs using OpenAI models. It provides abstractions for tool usage, memory management, and agent workflows, enabling developers to define function-calling agents that reason through multi-step tasks. Ideal for building custom AI workflows, the library supports dynamic tool definitions and contextual memory handling.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 19
    Perf Book

    Perf Book

    The book "Performance Analysis and Tuning on Modern CPU"

    This project is a practical guide to performance analysis and tuning on modern CPUs, bridging microarchitecture details with hands-on profiling. It explains how caches, TLBs, prefetchers, branch predictors, and out-of-order execution influence real program speed, then connects those concepts to concrete optimization strategies. Readers learn how to design trustworthy benchmarks, avoid measurement traps (warmup, turbo, frequency scaling), and interpret hardware performance counters. The book...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 20
    KVCache-Factory

    KVCache-Factory

    Unified KV Cache Compression Methods for Auto-Regressive Models

    ...In large language models, the key-value cache stores intermediate attention states that enable efficient token generation during inference, but these caches can consume large amounts of GPU memory when handling long contexts. KVCache-Factory provides a platform for implementing and evaluating multiple compression strategies that reduce memory usage while preserving model performance. The framework integrates several state-of-the-art methods such as PyramidKV, SnapKV, H2O, and StreamingLLM, allowing researchers to compare and experiment with different approaches within the same environment. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    Hello-Agents

    Hello-Agents

    Building an Intelligent Agent from Scratch

    Hello Agents is an open educational project designed to teach developers how to understand, design, and build AI-native agents from the ground up through structured tutorials and practical examples. The project focuses on guiding learners beyond superficial framework usage toward deeper comprehension of agent architecture, reasoning loops, and real-world implementation patterns. It walks users through core concepts such as ReAct-style reasoning, tool usage, memory handling, and multi-step task execution, enabling hands-on experimentation with modern LLM-powered agent systems. The repository is structured as a progressive learning path, combining theory, exercises, and runnable code so users can incrementally build more capable agents. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Glances

    Glances

    An eye on your system

    Glances is an open source, cross-platform monitoring tool that aims to provide a significant amount of monitoring information through a curses or Web-based interface. Depending on the size of the user interface, this information can then dynamically adapt. Glances can work in client/server mode, and is also capable of remote monitoring. All systems statistics can be exported to files or external time/value databases. Glances gets information from your system through various libraries,...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 23
    xLSTM

    xLSTM

    Neural Network architecture based on ideas of the original LSTM

    ...The architecture aims to provide competitive performance with transformer-based models while maintaining advantages such as linear computational scaling and efficient memory usage for long sequences. Researchers have demonstrated that xLSTM models can scale to billions of parameters and large training datasets while maintaining efficient inference speeds.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    AirLLM

    AirLLM

    AirLLM 70B inference with single 4GB GPU

    AirLLM is an open source Python library that enables extremely large language models to run on consumer hardware with very limited GPU memory. The project addresses one of the main barriers to local LLM experimentation by introducing a memory-efficient inference technique that loads model layers sequentially rather than storing the entire model in GPU memory. This layer-wise inference approach allows models with tens of billions of parameters to run on devices with only a few gigabytes of...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    bitnet.cpp

    bitnet.cpp

    Official inference framework for 1-bit LLMs

    bitnet.cpp is the official open-source inference framework and ecosystem designed to enable ultra-efficient execution of 1-bit large language models (LLMs), which quantize most model parameters to ternary values (-1, 0, +1) while maintaining competitive performance with full-precision counterparts. At its core is bitnet.cpp, a highly optimized C++ backend that supports fast, low-memory inference on both CPUs and GPUs, enabling models such as BitNet b1.58 to run without requiring enormous...
    Downloads: 6 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB