Skip to content

iChristGit/OpenWebui-Tools

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

99 Commits
 
 
 
 
 
 
 
 

Repository files navigation





🗺️ Quick Navigation

# Tool Best for
1 🎬 Jellyfin Media Player Stream movies, TV, music & live TV from your own server
2 📺 YouTube Player Watch, search & summarize YouTube videos directly in chat
3 🎥 LTX-2.3 Video Generator Generate AI videos from text or images via ComfyUI
4 📰 News Reader Live headlines from 45 feeds across 11 categories — with AI summaries
5 🌐 Reddit Explorer Browse Reddit — posts, comments, media, user profiles — no API key
6 🎙 Podcast Player Stream any podcast episode directly in chat — no API key
7 🧠 Thinking Filter One-click thinking toggle + full reasoning control for llama.cpp
8 🧩 Ask User The right questions before the right answer
9 🎭 Persona Studio Instantly switch your AI's personality & tone
10 📖 Wikipedia Instant encyclopedia lookups in 20+ languages
11 🌌 Omniscient Orchestrator Multi-stage AI workflow with strategy selection
12 😂 Joke Tool 300+ programmer jokes on demand
13 💾 VRAM Unload Unload llamacpp models with one click
14 🔍 RTX Image Upscaler Upscale images up to 4× using NVIDIA RTX Video Super Resolution via ComfyUI

🛒 Install from the Open WebUI Marketplace

The fastest way to get these tools running. No copy-paste required — install directly from the Open WebUI marketplace in seconds.


🎬 Jellyfin Media Player

Stream your entire Jellyfin library inside Open WebUI — with a cinematic embedded player, subtitle support, quality presets, album art, and live TV with EPG.

Install on Open WebUI

🎬 Movies

Jellyfin movie player
📺 TV Shows

Jellyfin TV player
📡 Live TV

Jellyfin live TV player
🎵 Music

Jellyfin music player

✨ What it does

Your Jellyfin library, right inside the chat. Ask for any movie, show, track, or live channel in plain English and get an embedded player back — no tab-switching, no searching, no fuss.

Feature Detail
🎬 Movies & TV Stream any item, use S01E01 or 1x01 episode notation
🎵 Music Full player with waveform visualiser, album art & EQ animation
📡 Live TV Tune to any channel with a pulsing LIVE badge and EPG now-playing info
🎲 Random picks Ask for a random film, episode, or song — optionally filter by genre
💬 Subtitles Dropdown with every available language track
📐 Quality Original · 4K · 1080p · 720p · 480p · 360p
⬇️ Download Download button on every player
🔎 Info mode Get rich media details without opening a player
⚙️ Setup (3 steps)
  1. Create a restricted Jellyfin user (playback-only, no admin or delete permissions)
  2. Generate an API key: Dashboard → API Keys → +
  3. Paste both into the tool valves: JELLYFIN_HOST and JELLYFIN_API_KEY

For Live TV, you'll also need a tuner or IPTV source configured under Dashboard → Live TV.

🗣️ Example prompts
play Inception
play Breaking Bad S03E10
play music Bohemian Rhapsody
random comedy movie
tell me about Interstellar
what was recently added
watch live CNN
list live channels
⚠️ Disclaimer

This tool connects to your own self-hosted Jellyfin server — it doesn't host, stream, or distribute any content itself. What's on your server is your business. Use responsibly and make sure you have the rights to anything you're streaming.


📺 YouTube Player

Watch YouTube videos, browse search results, and get AI-powered summaries — directly in chat. No API key required. Powered by YouTube's own InnerTube API, with likes/dislikes from Return YouTube Dislike and transcripts via youtube-transcript-api.

Install on Open WebUI

Screenshot 20262-03-13 134323

✨ What it does

YouTube, embedded right in your chat window. Ask for a tutorial, share a URL, or say "search YouTube for X" — the tool picks the right action automatically. The embedded player loads instantly with view counts, likes/dislikes, and channel info. If a transcript is available, the LLM silently reads it so you can ask follow-up questions about the video without ever leaving the conversation.

Feature Detail
▶️ Watch instantly Paste a URL or describe what you want — the best match embeds immediately
🔎 Browse grid Ask to "search YouTube for X" to get a scrollable results grid and pick your own
🧠 AI summaries "Summarize this video" fetches the full transcript and writes a structured recap
📜 Full transcripts Get the raw transcript text of any video on demand
👍👎 Likes & dislikes Dislike counts restored client-side via Return YouTube Dislike API
📊 Rich metadata View count, channel name, publish date, and duration shown on every player
🔑 No API key needed Search via YouTube's own InnerTube API — fast, reliable, no setup
🔑 Optional API key Set YOUTUBE_API_KEY in valves for improved search accuracy via Data API v3
🗣️ Example prompts
show me a video about how transformers work
I need a tutorial on Docker networking
search YouTube for lo-fi hip hop study beats
play https://www.youtube.com/watch?v=dQw4w9WgXcQ
summarize this YouTube video: [url]
get the transcript of [url]
what are some videos on RLHF?
⚙️ Setup

Works out of the box — no configuration needed.

Optional valves:

Valve Default What it does
YOUTUBE_API_KEY (empty) YouTube Data API v3 key for improved search accuracy
INVIDIOUS_INSTANCE (auto) Preferred Invidious instance for transcript fallback

Requires youtube-transcript-api Python package (listed in tool requirements — installs automatically).


🎥 LTX-2.3 Video Generator

Generate cinematic AI videos from a text prompt or an uploaded image — powered by LTX-Video 2.3 (22B FP8) running locally in ComfyUI, with an embedded player, mobile-compatible output, and creative filenames chosen by the LLM.

Install on Open WebUI

🙏 Adapted from Haervwe's WAN 2.2 ComfyUI tool — the original inspiration for this implementation.

Screenshot 2026-03-13 091121

✨ What it does

Two tools in one: Text-to-Video and Image-to-Video, both driven by the LTX-Video 2.3 22B model running locally in ComfyUI. Ask for a video in plain English — or upload an image and say "animate this" — and get a fully embedded cinematic player back in chat with a download button, open button, and a creative title the LLM names itself. For Image-to-Video the output resolution is automatically set to match your input image. Running on a single GPU? Both Ollama and llama.cpp models can be auto-evicted from VRAM before each generation so the video model gets all the headroom it needs.

Feature Detail
✍️ Text-to-Video Generate from any prompt at configurable resolution, up to 30 s
🖼️ Image-to-Video Upload any image — output resolution automatically matches it, no guessing
🖥️ Embedded player Cinematic dark player rendered directly in chat — watch, open, or download without leaving the conversation
⏱️ Configurable duration Pick 5s · 10s · 15s · 20s · 25s · 30s per-user — no admin needed
🧹 VRAM unload (GPU poor friendly) Auto-evicts Ollama and/or llama.cpp models before each generation — reclaim every MB for the video model
📱 Mobile-compatible ffmpeg post-encode to H.264 yuv420p + faststart — shareable on iOS & WhatsApp
🎲 Random seeds Both noise seeds randomised every generation — never duplicate outputs
🎬 Creative filenames LLM picks a unique 2–4 word title per video (e.g. Dragon_Awakens_Dawn_i2v)
🔑 Optional API key Bearer token support for secured ComfyUI setups
⚙️ Prerequisites

This tool connects to a running ComfyUI instance with LTX-Video 2.3 already set up. You'll need:

  1. ComfyUI running with an LTX-2.3 workflow loaded — see LightricksAI/LTX-Video for models
  2. ComfyUI-Unload-Model custom node installed — required by the workflow to free VRAM between pipeline stages

Then configure the tool valves:

Valve Default What it does
comfyui_api_url http://localhost:8188 ComfyUI HTTP endpoint
comfyui_api_key (empty) Bearer token if ComfyUI is behind auth
owui_internal_base http://localhost:8080 Internal OWUI URL for file serving
video_length_frames 241 Default frame count (241 = 10 s at 24 fps)
frame_rate 24 Output fps
t2v_width / t2v_height 1280 / 720 Text-to-Video resolution (I2V uses the input image size automatically)
max_wait_time 600 Seconds before timeout (generation takes 3–10 min)
unload_ollama_models true Auto-free Ollama VRAM before each generation
ollama_api_url http://localhost:11434 Your Ollama server address
unload_llamacpp_models false Auto-free llama.cpp router models before each generation
llamacpp_api_url http://localhost:8082 Your llama.cpp router address
🗣️ Example prompts
make a video of a samurai walking through cherry blossoms at sunset
generate a 15 second clip of ocean waves crashing at night
animate this image [upload any photo]
create a video of a neon-lit city street in heavy rain
a slow-motion shot of a red fox jumping through snow
timelapse of storm clouds rolling over a mountain range
⚡ Per-user settings

Each user can override the admin defaults independently:

Setting Options Note
video_duration 5s 10s 15s 20s 25s 30s Applies to both T2V and I2V
frame_rate Any integer (default 24)
t2v_width / t2v_height Any resolution (default 1280×720) T2V only — I2V always matches the uploaded image automatically

📰 News Reader

Live news from 45 hand-picked RSS feeds across 11 categories — rendered as an interactive card feed in chat. Cards are expandable, category routing is automatic, and the LLM can fetch and summarize full article content — not just RSS snippets.

Install on Open WebUI

Screenshot 2026-03-13 152012

✨ What it does

Say "tech news" or "what's happening with AI" and get a live, beautifully rendered card feed from the best sources in that category — Ars Technica, Reuters, BBC, Hacker News, MIT Tech Review, and more. Each card shows the headline and a preview that expands to the full summary on click. Ask to "summarize the latest climate news" and the tool actually fetches and reads up to 6 full articles, then hands them to the LLM to write a real briefing — not a headline list.

Feature Detail
📡 45 RSS feeds Hand-picked sources across 11 categories: World, Tech, AI, Science, Business, Sports, Gaming, Health, Politics, Entertainment, Climate
🃏 Expandable cards Click any card to expand the full summary inline — no new tabs
🔍 Keyword search Search across all 45 feeds simultaneously for any topic or name
🗞️ Front page "Top headlines" builds a curated cross-category briefing from the best sources
🧠 AI summaries "Summarize [topic] news" fetches real article text and gives it to the LLM to write proper paragraphs
🔑 Zero config No API keys, no setup — works out of the box
⚙️ Configurable limit MAX_ARTICLES valve controls how many articles to fetch (default 15, max 40)
📚 Included sources (sample)
Category Sources
🌐 World Reuters, AP News, BBC World, Al Jazeera
💻 Tech Ars Technica, The Verge, Wired, TechCrunch, Hacker News
🤖 AI MIT Tech Review, VentureBeat AI, The Decoder
🔬 Science NASA, New Scientist, Science Daily, Phys.org
💼 Business Financial Times, Bloomberg, CNBC, Forbes
🎮 Gaming IGN, Kotaku, PC Gamer, Eurogamer
🏥 Health NHS UK, MedicalNewsToday, WHO, WebMD
🌱 Climate The Guardian Env, Carbon Brief, Inside Climate News
🗣️ Example prompts
latest news
top headlines
tech news
AI news
news about SpaceX
what's happening with Bitcoin
summarize the latest science news
brief me on tech news today
gaming news
what's going on in politics?

🌐 Reddit Explorer

Browse Reddit directly in chat — hot, new, top & rising posts from any subreddit, full comment threads, rich media (images, video, galleries), user profiles, and smart subreddit suggestions. Zero API key, zero sign-in.

Install on Open WebUI

Screenshot 2026-03-12 173619

✨ What it does

Nine purpose-built functions let the LLM intelligently navigate Reddit on your behalf. Ask "what's hot on r/worldnews", "show me the top posts of the week in r/science", or just "what's happening on Reddit" — the tool picks the right function, picks a sensible subreddit if you didn't specify one, fetches live data via Reddit's public JSON API, and returns fully-rendered markdown with inline images, video links, upvote scores, and direct comment links.

Feature Detail
🔥 Hot / 🆕 New / 🏆 Top posts Browse any subreddit by sort — Top supports hour · day · week · month · year · all time filters
🌐 Reddit front page Global popular, all, or best feeds — no subreddit needed
🔍 Search Full-text search within a specific subreddit or across all of Reddit
💬 Comment threads Fetches a post's top comments with nested first-level replies (depth 3), sorted by top · best · new · controversial · old
🖼️ Inline images Direct image posts render inline with ![post image](url) — JPEG, PNG, GIF, WebP
🎬 Video links Reddit-hosted videos get a ▶ Watch link; embedded videos (YouTube, Twitch, etc.) include provider name and a thumbnail
🖼️🖼️ Galleries Multi-image gallery posts show up to 4 preview images inline
🔗 External links Link posts surface the external URL alongside the Reddit thread link
⬆️ Upvote scores & ratio Every post shows score (formatted as 1.2k / 3.4M) and upvote percentage
🏠 Subreddit info Full community profile: icon, banner, member count, active users, language, NSFW flag, and top rules
👤 User profiles Karma breakdown (post / comment / total), badges (mod, gold, verified), avatar, and recent post history
🗺️ Subreddit suggestions Topic → subreddit map covering 20+ categories (news, AI, gaming, finance, food, travel, crypto, etc.) plus live Reddit search
🔞 / 🙈 Content tags NSFW and spoiler posts are flagged automatically
🔑 No API key required Uses Reddit's public .json endpoints — no OAuth, no account, no rate-limit keys
📚 All 9 functions at a glance
Function Trigger phrase examples
get_hot_posts(subreddit, limit) "what's trending on r/gaming", "hot posts in worldnews"
get_new_posts(subreddit, limit) "latest posts on r/OpenWebUI", "newest in r/technology"
get_top_posts(subreddit, time_filter, limit) "top posts this week in r/science", "best of all time on r/AskReddit"
search_reddit(query, subreddit, limit) "search Reddit for llama.cpp", "find posts about ComfyUI in r/stablediffusion"
get_post_comments(post_id, subreddit, limit, sort) "show me the comments on that post", "read the discussion"
get_subreddit_info(subreddit) "what is r/LocalLLaMA", "tell me about r/homelab"
suggest_subreddits(topic) "where can I find AI news on Reddit?", "best subreddit for cooking"
get_reddit_frontpage(feed, limit) "what's popular on Reddit", "show me Reddit", "what's happening today"
get_user_profile(username, limit) "look up u/spez", "show me that user's post history"
🗣️ Example prompts
What's hot on r/LocalLLaMA today?
Show me the top posts of the week in r/worldnews
What's the latest on r/OpenWebUI?
Search Reddit for "open-webui tools"
What's the best subreddit for home automation?
Tell me about r/MachineLearning
Show me the Reddit front page
Read the comments on that post
Look up the profile for u/GallowBoob
What's trending on Reddit right now?
Find me the top AI news from this month
⚙️ Requirements

Only one Python package is needed:

requirements: requests

No valves, no API keys, no configuration — install and ask away.

ℹ️ Uses Reddit's public reddit.com/r/<subreddit>.json endpoints with a descriptive User-Agent header. Respects Reddit's public API fair-use guidelines.


🎙 Podcast Player

Stream any podcast directly in chat — powered by the free iTunes Search API and RSS feeds, with a cinematic animated-waveform audio player, playback speed control, 30-second skip, and episode browsing. No API key required.

Install on Open WebUI

Screenshot 2026-03-13 151711

✨ What it does

The entire iTunes podcast directory — millions of shows — accessible by name, topic, or host. Ask for the latest episode of any show, jump to a specific episode by number or title keyword, discover today's Apple top 25, or let the tool roll a random episode from a show's back catalogue. An embedded cinematic player with animated waveform, album art, speed control, and 30-second skip renders right in chat — podcast-standard controls without leaving the conversation.

Feature Detail
🎙 Play any podcast Latest episode, specific episode number, or episode title keyword
🎲 Random episode Surprise pick from a show's entire back catalogue — great for rediscovering classics
🔎 Directory search Search millions of podcasts by name, host, topic, or genre
📊 Top 25 charts Apple Podcasts top charts, localised by country code
🎚️ Speed control 0.5× · 0.75× · 1× · 1.25× · 1.5× · 2× playback speed
⏩ 30-second skip Podcast-standard forward/back skip buttons
🎨 Animated waveform Vivid per-show colour palette with animated EQ visualiser
🌍 Localised charts Set COUNTRY_CODE valve to get top charts for your country (default: us)
🔑 Zero config Uses the free iTunes Search API + RSS — no keys, no account needed
🗣️ Example prompts
play podcast Joe Rogan
play podcast Serial episode 5
play podcast Hardcore History Blitzkrieg
search podcast true crime
top podcasts
random podcast episode Lex Fridman
play the latest Darknet Diaries
what podcasts are trending?
⚙️ Setup

Works out of the box — no configuration needed.

Optional valve:

Valve Default What it does
COUNTRY_CODE us 2-letter ISO country code for Apple top-chart rankings (e.g. gb, ca, au, de)

🧠 Thinking Filter

The missing thinking toggle for llama.cpp + Qwen3.5 One click to unleash deep reasoning. One click to turn it off. And when you want more — full control over depth, style, and how answers are presented.

Install on Open WebUI

8bed8be1-8d6a-42b1-bb0d-f2929fe1cb9e

✨ What it does

llama.cpp supports Qwen3.5 extended <think> reasoning mode natively — but Open WebUI has no built-in toggle for it. This filter fixes that completely, replacing manual tweaking with a proper one-click think button and a full suite of reasoning controls.

Feature Detail
🧠 One-click toggle Brain button in the ✦ panel — ON thinks, OFF is instant vanilla
🔀 enable_thinking valve Toggle thinking on/off per-user without disabling the filter — sampling & presentation stay active either way
🎛️ 4 sampling presets Force Qwen3.5's official parameters at llama.cpp API level — Instruct General · Instruct Reasoning · Thinking General · Thinking Precise
📊 5 depth levels Unlimited · MAX (16k) · Deep (8k) · Normal (3k) · Quick (512 tokens)
🔬 13 reasoning presets Shape how it thinks
🎨 11 presentation presets ELI5, Expert Tone, TL;DR First — shape how it answers
👤 Per-user control Every user sets their own depth, style, and sampling independently
🔒 Bulletproof injection Dual-path injection (system + user message) works around Open WebUI's pipeline bug
⚡ Setup (1 step)

Step 1 — Start llama-server with:

llama-server --jinja --chat-template-kwargs "{\"enable_thinking\":false}" 

The --chat-template-kwargs "{\"enable_thinking\":false}" flag lets the filter set the budget dynamically per request.

The --jinja flat enables Jinja chat template processing, required for Qwen3.5's think block

Enable the function and set it as default for your qwen3.5 models if you want the default to be thinking enabled. (If you rather have thinking disabled by default do not set the function as default!)

Now every chat has a 🧠 button that can be diabled in one click and enabled back in two clicks.

📊 Thinking Depth
Depth Token Budget Best for
Unlimited No cap Default — model thinks as long as it needs
MAX 16 000 tokens Hardest problems, exhaustive analysis
Deep 8 000 tokens Complex reasoning, careful step-by-step
Normal 3 000 tokens Everyday use, balanced
Quick 512 tokens Fast answers with just a hint of thought
🔬 Reasoning Presets — how the model thinks
Preset What it does
None Vanilla — pure thinking, no style instruction
Think Less Skip over-analysis, reach conclusions fast
Think More Explore multiple angles before settling
Extended Thinking Deep deliberation: edge cases, counterargs, stress-tests every conclusion
MAX Thinking 🔥 Exhaustive — never stops early, challenges everything, asks "what have I missed?" Pair with MAX depth
Step by Step Numbered structured reasoning, nothing skipped
Devil's Advocate Steelmans the opposing view before answering
First Principles Strips to fundamentals, rebuilds from scratch
10x Hypotheses Generates 10 distinct approaches, evaluates all, picks the best
Socratic Interrogates the question's own assumptions before answering
Rubber Duck Narrates every logical move out loud — catches its own mistakes
Pre-Mortem Assumes the answer will fail, fixes it before giving it
Bayesian Probabilistic reasoning with honest calibrated confidence
Contrarian Default skepticism — challenges obvious answers, demands proof
🎨 Presentation Presets — how the answer looks
Preset What it does
None Vanilla output
ELI5 Explain like I'm five — strip all complexity
Be Concise Shortest complete answer, zero padding
Bullet Points Clean bulleted list
TL;DR First One-sentence summary up top, then full detail
Teach Me Concept → examples → memorable takeaway
Expert Tone Graduate-level, precise vocabulary
Casual Chat Relaxed, no jargon, like texting a smart friend
Debate Format Strongest FOR → strongest AGAINST → verdict
Analogies Only Everything through metaphors, zero technical terms
Action Items Numbered steps to execute immediately
Socratic Reply Guides you to the answer through probing questions
🎛️ Mix & Match Examples
Depth Reasoning Presentation Result
MAX MAX Thinking Expert Tone 🔥 Deepest possible analysis, grad-level output
Normal Bayesian TL;DR First Calibrated probabilistic answer, summary first
Quick None Be Concise Lightning-fast minimal answer
Deep First Principles Teach Me Rebuilds from scratch, explains like a lesson
Deep Devil's Advocate Debate Format Full steelman treatment, structured verdict
Unlimited Contrarian ELI5 Challenges every assumption, explains simply

🧩 Ask User

Replicate Claude's "ask follow-up questions before acting" behaviour — collect structured multi-step input through sequential popup dialogs before your AI writes a single word.

Install on Open WebUI

Ask User screenshot

✨ What it does

Ever notice how Claude pauses, asks you a few targeted questions, then produces something far more on-point? This tool brings that exact behaviour to Open WebUI. The LLM can ask 1–5 structured questions upfront — each in its own popup with a progress indicator — before committing to a response.

Feature Detail
📋 Up to 5 questions Defined upfront — no infinite loops
🪟 One popup per question Clear progress indicator: Question 2 of 4
📦 Structured output All answers returned together before the LLM acts
🔒 Bounded & predictable The LLM cannot keep asking — it must proceed
🔑 No dependencies Standard library only
🗣️ Best for
"Write me a cover letter"    → gathers: role, company, tone, key achievements
"Plan my project"            → gathers: deadline, team size, stack, goals
"Draft a cold email"         → gathers: recipient, offer, call to action
"Create a workout plan"      → gathers: goal, equipment, days per week

🎭 Persona Studio

Instantly reshape how your AI thinks and communicates — dozens of crafted personas across multiple categories, plus a fully custom option.

Install on Open WebUI

Persona Studio screenshot

✨ What it does

Stop editing system prompts manually. Persona Studio gives you an interactive popup browser with categorised personas — each with a distinct tone, voice, quirks, and style. Switch mid-conversation in one message.

Feature Detail
🎯 Interactive browser Categorised popup UI — pick and apply in one click
😎 Dozens of personas Technical · Creative · Professional · Playful · and more
✍️ Custom option Define any personality from scratch
🚀 Instant switch No re-prompting, no manual system prompt editing
🗣️ Example prompts
Switch to a different persona
I want you to respond like a senior engineer doing code review
Change your communication style to casual and funny
Set a custom persona: you are a brutally honest editor

📖 Wikipedia

Instant encyclopedia lookups — smart search, section-aware summaries, disambiguation handling, and 20+ languages. Zero config required.

Install on Open WebUI

Wikipedia tool screenshot

✨ What it does

Your LLM gets live access to Wikipedia without any API key or setup. It automatically picks the right detail level, handles ambiguous titles gracefully, and always cites the source URL.

Feature Detail
🔍 Smart search Natural language queries work fine
📏 3 detail levels brief (intro only) · standard (intro + sections) · full (entire article)
🗂️ Section-aware Structured summaries per section — not a wall of text
⚠️ Disambiguation Shows options when a title matches multiple articles
🌍 20+ languages Just write in your language — it auto-detects
🔗 Always cited Every response includes the Wikipedia source URL
🔑 Zero config Uses the free public MediaWiki API
🗣️ Example prompts & supported languages
Who is Ada Lovelace?
Explain how black holes work
Tell me everything about the Apollo program
Search Wikipedia for quantum entanglement
¿Quién es Simón Bolívar?

Supported languages: en fr de es it pt nl sv uk fa ja zh ar ko ru pl tr he vi id


🌌 Omniscient Orchestrator

Stop getting generic answers. Make your LLM pause, ask the right questions, pick a strategy — then produce something genuinely tailored to your goal.

Install on Open WebUI

Screenshot 2026-03-13 092540

✨ What it does

Most LLMs dive straight into an answer, missing context that would have changed everything. The Orchestrator forces a structured 4-phase workflow: understand → clarify → strategise → execute.

Phase What happens
🔹 Phase 1 — Analyse Maps your request, identifies gaps and ambiguities
🔹 Phase 2 — Clarify Asks up to 3 targeted questions via input modals
🔹 Phase 3 — Strategise Presents 3 distinct execution approaches to choose from
🔹 Phase 4 — Execute Generates output with all context locked in

Built-in guardrails: max 3 questions, always skippable, always moves forward — no infinite loops.

🗣️ Best for
Writing long-form content (articles, landing pages, scripts, reports)
Brainstorming with structured creative options
Planning projects or technical roadmaps
Any prompt where you'd normally need 3 rounds of re-prompting

😂 Joke Tool

Give your AI a sense of humour — 300+ curated programmer jokes, live API fetching, and batch delivery on demand.

Install on Open WebUI

Joke Tool screenshot

✨ What it does

A carefully curated vault of programmer humor — Git puns, deep-cut coding jokes, and classic one-liners — plus live fetching from jokeapi.dev when you want something fresh.

Feature Detail
🃏 300+ jokes Curated vault: Git puns, coding humor, dev classics
🌐 Live mode Fetches fresh jokes from jokeapi.dev on request
🔢 Batch support Ask for 1 joke or a specific number
🛡️ Offline-safe Always falls back to internal library if API is down
🗣️ Example prompts
Tell me a joke
Give me 3 fresh jokes from the internet
Hit me with a random Git pun
Tell me 5 programming jokes

💾 VRAM Unload

Free your GPU memory without leaving the chat — one action button that detects loaded models and unloads them from VRAM instantly via the llama.cpp router API.

Install on Open WebUI

Screenshot 2026-03-08 204032

✨ What it does

Running llama.cpp in router mode means models stay resident in VRAM until you explicitly unload them. This action button insantly clean the model from VRAM.

Feature Detail
🔍 Auto-detects loaded models Queries /v1/models and filters to currently loaded ones
🔁 Multi-model support Unloads every loaded model in one click if multiple are resident
📡 Live status updates Status messages as each model unloads, success/error per model
🔑 Zero dependencies Pure aiohttp — nothing extra to install
⚙️ Setup (1 step)

Set LLAMACPP_BASE_URL in the action valves to your llama.cpp router server:

http://127.0.0.1:8080

That's it. The action button appears in the chat toolbar — click it any time to free VRAM.

**Requires llama.cpp running in router mode (llama-server).

🛠️ Troubleshooting
Symptom Fix
❌ "Cannot reach llama.cpp" Check LLAMACPP_BASE_URL is correct and the server is running
ℹ️ "No models currently loaded" No models are resident in VRAM — nothing to unload
❌ HTTP 404 on unload Make sure llama.cpp is started in router mode, not single-model mode
❌ HTTP 4xx/5xx Check llama.cpp logs for the specific error

🔍 RTX Image Upscaler

Upscale any image up to 4× using NVIDIA RTX Video Super Resolution — hardware-accelerated AI upscaling running locally in ComfyUI, with the result embedded directly in chat. NVIDIA GPU required.

Install on Open WebUI

Image

✨ What it does

Drop any image into the chat, ask to upscale it, and get a sharpened high-resolution version back as an attachment — all without leaving the conversation. Powered by NVIDIA's RTX Video Super Resolution running inside ComfyUI, the tool handles image upload, workflow execution, result retrieval, and chat embedding automatically. Choose 1×, 2×, 3×, or 4× — default is always 4× ULTRA for maximum quality.

Feature Detail
🔍 Up to 4× upscale 1×, 2×, 3×, or 4× multiplier — defaults to 4× ULTRA if not specified
⚡ RTX hardware acceleration Uses NVIDIA's native VSR tensor cores — on a 3090 Ti the full round-trip (input → LLM → upscaled output) takes ~1 second
📎 Persistent attachment Result embedded as a file attachment — survives in chat history, never flashes and disappears
🧹 VRAM cleanup Optional UnloadAllModels node after each run to free GPU memory
🔁 Fallback resilience Tries OWUI file store first, falls back to base64 data URI if upload fails
🔑 Zero API keys Fully local — no cloud, no accounts, no cost per image
⚙️ Prerequisites

This tool requires a ComfyUI workflow to be set up before use.

  1. NVIDIA GPU with RTX — RTX Video Super Resolution is an NVIDIA-only feature (RTX 20 series or newer). AMD and Intel GPUs are not supported.
  2. ComfyUI running locally with the RTXVideoSuperResolution custom node installed.
  3. Download and run the RTX upscale workflow in ComfyUI at least once before using this tool — this ensures the node is loaded and the workflow is validated on your hardware.

Then configure the tool valves:

Valve Default What it does
comfyui_url http://localhost:8188 ComfyUI HTTP endpoint
owui_internal_base http://localhost:8080 Internal OWUI URL for file serving
unload_models_after_run true Free VRAM after each upscale
max_wait_seconds 300 Timeout before giving up on a slow job

Per-user overrides (UserValves):

Setting Default Note
default_scale 4 Fallback multiplier if the LLM doesn't specify one
unload_models_after_run true Override the admin VRAM setting per user
🗣️ Example prompts
upscale this image
make this photo sharper
enhance the resolution
upscale 2x
make it bigger — use scale 3
⚠️ Requirements & limitations
  • NVIDIA RTX GPU only — this uses NVIDIA's proprietary RTX Video Super Resolution. It will not work on AMD, Intel, or non-RTX NVIDIA cards.
  • ComfyUI must be running with the RTXVideoSuperResolution custom node installed and the upscale workflow loaded at least once.
  • Large images at 4× can produce very large output files — allow extra time on lower-end RTX cards.

🛠️ Installation

All tools install the same way — takes about 30 seconds.

1. Open Workspace → Tools in Open WebUI
2. Click ➕ Add Tool
3. Paste the script, or use one-click import from the marketplace
4. Click Save
5. Enable it in any chat via the 🔧 Tools toggle

💡 Tip: Enable only the tools you need in each chat — keeping the tool list focused helps the LLM pick the right one every time.


📜 License

All tools are released under the MIT License — free to use, fork, modify, and publish.


👤 About Me

Just a guy obsessed with AI — LLMs, image gen, music gen, all of it. Learning something new every day and loving every second of it.

I've been using Open WebUI daily for a while now. At some point I stopped just using it and started asking what if it could do more? What started as scratching my own itch — adding features I personally wanted — turned into this collection. If I'm building it anyway, might as well share it with everyone else who wants to jump in and push their setup further.

I'll be honest — I'm not an amazing coder. These tools are built with Claude doing the heavy lifting on the code, while I handle the ideas, the direction, the debugging, and making sure everything actually works the way I want it to. It's a collaboration, and honestly it's the most fitting thing in the world: AI tools, built with AI. If that's not proof that this stuff is genuinely useful, I don't know what is.

Everything here is open source, built in the open, and made with genuine love for the community. If a tool saves you time or sparks an idea, that's exactly why it exists.


💡 Inspiration & Credits

This project wouldn't exist without Haervwe's open-webui-tools.

Haervwe's collection is what got me building my own tools in the first place — seeing what was possible sparked the whole thing. We've also collaborated directly, which shaped several ideas here. If you're looking for even more great Open WebUI tools, their repo is absolutely worth a visit.

🙏 Big thanks to @Haervwe for the inspiration and the collaboration.


Made with ☕ by ichrist · powered by Claude

If these tools save you time, a star goes a long way!