Toolverse
All skills

whisper

by davila7

OpenAI's general-purpose speech recognition model. Supports 99 languages, transcription, translation to English, and language identification. Six model sizes from tiny (39M params) to large (1550M params). Use for speech-to-text, podcast transcription, or multilingual audio

Installation

Pick a client and clone the repository into its skills directory.

Installation

Quick info

Author
davila7
Category
Security
Views
61

About this skill

OpenAI's general-purpose speech recognition model. Supports 99 languages, transcription, translation to English, and language identification. Six model sizes from tiny (39M params) to large (1550M params). Use for speech-to-text, podcast transcription, or multilingual audio processing. Best for robust, multilingual ASR.

How to use

  1. Zainstaluj Whisper za pomocą pip (wymaga Python 3.8–3.11): uruchom polecenie pip install -U openai-whisper. Upewnij się, że masz zainstalowany ffmpeg – na macOS użyj brew install ffmpeg, na Ubuntu sudo apt install ffmpeg, na Windows choco install ffmpeg.

  2. Załaduj model Whisper w Pythonie – zaimportuj bibliotekę whisper i wczytaj wybrany model poleceniem whisper.load_model("base"). Dostępne są warianty: tiny, base, small, medium, large i turbo – wybierz w zależności od wymaganej szybkości i jakości.

  3. Transkrybuj plik audio, przekazując ścieżkę do pliku metodzie transcribe() – na przykład result = model.transcribe("audio.mp3"). Model automatycznie wykryje język i zwróci pełny tekst transkrypcji.

  4. Wyświetl wynik transkrypcji – dostęp do pełnego tekstu uzyskasz przez result["text"], a do poszczególnych segmentów (z czasami) przez iterację po result["segments"], gdzie każdy segment zawiera czas początkowy, końcowy i tekst.

Related skills

llama-cpp

by zechenzhangAGI

Runs LLM inference on CPU, Apple Silicon, and consumer GPUs without NVIDIA hardware. Use for edge deployment, M1/M2/M3 Macs, AMD/Intel GPUs, or when CUDA is unavailable. Supports GGUF quantization (1.5-8 bit) for reduced memory and 4-10× speedup vs PyTorch on CPU.

Security
11252

obsidian

by gapmiss

Comprehensive guidelines for Obsidian.md plugin development including all 27 ESLint rules, TypeScript best practices, memory management, API usage (requestUrl vs fetch), UI/UX standards, and submission requirements. Use when working with Obsidian plugins, main.ts files,

Security
14111

ui-audit

by openclaw

AI skill for automated UI audits. Evaluate interfaces against proven UX principles for visual hierarchy, accessibility, cognitive load, navigation, and more. Based on Making UX Decisions by Tommy Geoco.

Security
1223

backend-security-coder

by sickn33

Expert in secure backend coding practices specializing in input validation, authentication, and API security. Use PROACTIVELY for backend security implementations or security code reviews.

Security
1133

skill-writer

by pytorch

Guide users through creating Agent Skills for Claude Code. Use when the user wants to create, write, author, or design a new Skill, or needs help with SKILL.md files, frontmatter, or skill structure.

Security
15116

typescript-review

by metabase

Review TypeScript and JavaScript code changes for compliance with Metabase coding standards, style violations, and code quality issues. Use when reviewing pull requests or diffs containing TypeScript/JavaScript code.

Security
17133