Skip to content

Commit ee5cfcd

Browse files
committed
Add more ai tools
1 parent 52dd73a commit ee5cfcd

12 files changed

Lines changed: 625 additions & 3 deletions

File tree

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
![Icon](https://raw.githubusercontent.com/SimonSchubert/LinuxCommandLibrary/master/art/web_hi_res_144.png)
44

5-
The app currently has **7706** manual pages, **23+** basic categories and a bunch of general terminal tips. It works 100% offline, doesn't need an internet connection and has no tracking software.
5+
The app currently has **7713** manual pages, **23+** basic categories and a bunch of general terminal tips. It works 100% offline, doesn't need an internet connection and has no tracking software.
66

77
[![App Store](https://raw.githubusercontent.com/SimonSchubert/LinuxCommandBibliotheca/master/art/app_store_badge.png)](https://apps.apple.com/us/app/linux-command-library/id1219649976)
88
[![Play Store](https://raw.githubusercontent.com/SimonSchubert/LinuxCommandBibliotheca/master/art/play_store_badge.png)](https://play.google.com/store/apps/details?id=com.inspiredandroid.linuxcommandbibliotheca)

assets/basics/aitools.md

Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -24,3 +24,21 @@
2424
```[nanobot](/man/nanobot)```
2525
```[nanoclaw](/man/nanoclaw)```
2626
```[leon](/man/leon)```
27+
28+
## Image Generation
29+
```[sd-cli](/man/sd-cli)```
30+
```[mflux](/man/mflux)```
31+
```[comfyui](/man/comfyui)```
32+
33+
## Speech & Audio AI
34+
```[whisper](/man/whisper)```
35+
```[faster-whisper](/man/faster-whisper)```
36+
```[deepspeech](/man/deepspeech)```
37+
```[piper](/man/piper)```
38+
```[bark](/man/bark)```
39+
```[tts](/man/tts)```
40+
41+
## AI Terminal Utilities
42+
```[mods](/man/mods)```
43+
```[aichat](/man/aichat)```
44+
```[smartcat](/man/smartcat)```

assets/commands/aichat.md

Lines changed: 92 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,92 @@
1+
# TLDR
2+
3+
**Ask a question**
4+
5+
```aichat "[explain quicksort]"```
6+
7+
**Pipe input for analysis**
8+
9+
```cat [file.py] | aichat "[review this code]"```
10+
11+
**Use a specific model**
12+
13+
```aichat --model [claude:claude-sonnet-4-20250514] "[question]"```
14+
15+
**Execute shell commands from natural language**
16+
17+
```aichat -e "[list large files in current directory]"```
18+
19+
**Start interactive chat session**
20+
21+
```aichat -i```
22+
23+
**Use a role/persona**
24+
25+
```aichat --role [shell] "[find duplicate files]"```
26+
27+
**Process a file**
28+
29+
```aichat --file [document.pdf] "[summarize this]"```
30+
31+
# SYNOPSIS
32+
33+
**aichat** [_options_] [_prompt_]
34+
35+
# PARAMETERS
36+
37+
**-m**, **--model** _MODEL_
38+
> Model to use (provider:model format).
39+
40+
**-r**, **--role** _ROLE_
41+
> Use a predefined role/persona.
42+
43+
**-e**, **--execute**
44+
> Execute mode: translate natural language to shell commands.
45+
46+
**-i**, **--interactive**
47+
> Start interactive chat REPL.
48+
49+
**--file** _FILE_
50+
> Include file in the conversation.
51+
52+
**-w**, **--wrap** _COLS_
53+
> Wrap output at column width.
54+
55+
**-H**, **--no-highlight**
56+
> Disable syntax highlighting.
57+
58+
**-S**, **--no-stream**
59+
> Disable streaming output.
60+
61+
**--list-models**
62+
> List available models.
63+
64+
**--list-roles**
65+
> List available roles.
66+
67+
**--info**
68+
> Show current configuration.
69+
70+
# DESCRIPTION
71+
72+
**aichat** is an all-in-one AI CLI tool supporting chat, command execution, and RAG (Retrieval-Augmented Generation). It works with 20+ AI providers including OpenAI, Claude, Gemini, Ollama, Azure, and many more.
73+
74+
In chat mode, it provides a REPL with conversation history, multi-line input, and syntax highlighting. Execute mode translates natural language descriptions into shell commands and optionally runs them. RAG mode indexes documents for question-answering over local files.
75+
76+
Roles define reusable personas and system prompts. Built-in roles include coder, shell, and translator. Custom roles are defined in the configuration file.
77+
78+
The tool supports function calling, allowing AI models to invoke defined tools. Sessions persist conversations across invocations. Multiple providers and models can be configured simultaneously.
79+
80+
Install via `cargo install aichat`, Homebrew, or download binaries.
81+
82+
# CAVEATS
83+
84+
API keys required for cloud providers. Configuration file needed for multi-provider setup. RAG indexing requires additional setup. Token usage and costs vary by provider and model.
85+
86+
# HISTORY
87+
88+
**aichat** was created by **sigoden** in **2023** as a unified CLI for interacting with multiple AI providers. It grew from a simple chat tool into a comprehensive AI terminal toolkit with execute mode, RAG, and agent capabilities.
89+
90+
# SEE ALSO
91+
92+
[mods](/man/mods)(1), [smartcat](/man/smartcat)(1), [ollama](/man/ollama)(1), [llm](/man/llm)(1)

assets/commands/bark.md

Lines changed: 64 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,64 @@
1+
# TLDR
2+
3+
**Generate speech from text**
4+
5+
```python -m bark --text "[Hello, how are you?]" --output_filename [output.wav]```
6+
7+
**Use a specific speaker preset**
8+
9+
```python -m bark --text "[Hello]" --output_filename [output.wav] --history_prompt [v2/en_speaker_6]```
10+
11+
**Generate with emotions/effects**
12+
13+
```python -m bark --text "[laughs] Oh that's funny! [sighs]" --output_filename [output.wav]```
14+
15+
**Generate in another language**
16+
17+
```python -m bark --text "[Bonjour le monde]" --output_filename [output.wav] --history_prompt [v2/fr_speaker_1]```
18+
19+
**Generate with music notation**
20+
21+
```python -m bark --text "[♪ La la la ♪]" --output_filename [output.wav]```
22+
23+
# SYNOPSIS
24+
25+
**python** **-m** **bark** **--text** _text_ **--output_filename** _file_ [_options_]
26+
27+
# PARAMETERS
28+
29+
**--text** _TEXT_
30+
> Input text to synthesize.
31+
32+
**--output_filename** _FILE_
33+
> Output audio file path (.wav).
34+
35+
**--history_prompt** _PRESET_
36+
> Speaker voice preset (e.g., v2/en_speaker_0 through v2/en_speaker_9).
37+
38+
**--text_temp** _FLOAT_
39+
> Text generation temperature (default: 0.7).
40+
41+
**--waveform_temp** _FLOAT_
42+
> Waveform generation temperature (default: 0.7).
43+
44+
# DESCRIPTION
45+
46+
**Bark** is a transformer-based text-to-audio model by **Suno AI**. Unlike traditional TTS, Bark generates highly expressive speech including laughter, sighs, breathing, crying, and even music.
47+
48+
Special tokens in the text control non-speech sounds: `[laughs]`, `[sighs]`, `[gasps]`, `[clears throat]`, and `[music]`. Musical notation with `` symbols can generate singing. Capitalizing words adds emphasis, and `...` adds hesitation.
49+
50+
Speaker presets select voice characteristics. Presets are available for multiple languages: English, German, Spanish, French, Hindi, Italian, Japanese, Korean, Polish, Portuguese, Russian, Turkish, and Chinese.
51+
52+
Install with `pip install suno-bark`. Models are downloaded automatically on first use. GPU (CUDA) is strongly recommended for reasonable generation speed.
53+
54+
# CAVEATS
55+
56+
Slow on CPU (GPU strongly recommended). Large model downloads (~5GB). Output quality varies. Long text should be split into sentences. Not suitable for real-time synthesis. May produce unexpected audio artifacts.
57+
58+
# HISTORY
59+
60+
**Bark** was released by **Suno AI** in **April 2023** as an open-source text-to-audio model. Its ability to generate expressive speech with emotions and non-verbal sounds set it apart from conventional TTS systems. The model quickly gained popularity for creative audio generation.
61+
62+
# SEE ALSO
63+
64+
[piper](/man/piper)(1), [tts](/man/tts)(1), [espeak](/man/espeak)(1)

assets/commands/comfyui.md

Lines changed: 91 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,91 @@
1+
# TLDR
2+
3+
**Start ComfyUI server**
4+
5+
```python [main.py]```
6+
7+
**Start on specific port**
8+
9+
```python [main.py] --port [8188]```
10+
11+
**Start with CPU only**
12+
13+
```python [main.py] --cpu```
14+
15+
**Start in headless mode** (API only, no browser)
16+
17+
```python [main.py] --dont-print-server```
18+
19+
**Use specific GPU**
20+
21+
```python [main.py] --cuda-device [0]```
22+
23+
**Run with low VRAM mode**
24+
25+
```python [main.py] --lowvram```
26+
27+
**Execute a workflow via API**
28+
29+
```curl -X POST http://localhost:8188/prompt -H "Content-Type: application/json" -d @[workflow.json]```
30+
31+
# SYNOPSIS
32+
33+
**python** _main.py_ [_options_]
34+
35+
**comfyui** [_options_]
36+
37+
# PARAMETERS
38+
39+
**--port** _PORT_
40+
> Server port (default: 8188).
41+
42+
**--listen** _ADDR_
43+
> Listen address (default: 127.0.0.1, use 0.0.0.0 for network).
44+
45+
**--cpu**
46+
> Run on CPU only.
47+
48+
**--cuda-device** _ID_
49+
> CUDA GPU device index.
50+
51+
**--lowvram**
52+
> Low VRAM mode for GPUs with limited memory.
53+
54+
**--dont-print-server**
55+
> Suppress server output.
56+
57+
**--output-directory** _DIR_
58+
> Custom output directory.
59+
60+
**--temp-directory** _DIR_
61+
> Custom temp directory.
62+
63+
**--auto-launch**
64+
> Auto-open browser on start.
65+
66+
**--disable-auto-launch**
67+
> Prevent auto-opening browser.
68+
69+
# DESCRIPTION
70+
71+
**ComfyUI** is a node-based workflow system for Stable Diffusion and Flux image generation. It provides both a visual graph editor (web UI) and a REST API for programmatic use.
72+
73+
Workflows are built by connecting nodes: model loaders, samplers, VAE decoders, prompt encoders, and more. Complex pipelines (img2img, inpainting, ControlNet, LoRA stacking) are constructed visually without code.
74+
75+
The API accepts workflow JSON, enabling headless batch generation and integration with scripts. Workflows created in the web UI can be exported and run via the API.
76+
77+
An extensive ecosystem of custom nodes adds capabilities: video generation, face restoration, upscaling, IP-Adapter, and more. Custom nodes are installed into the `custom_nodes/` directory.
78+
79+
Install via pip (`pip install comfyui`) or clone the repository. Models are placed in the `models/` directory tree.
80+
81+
# CAVEATS
82+
83+
Web UI requires a modern browser. GPU strongly recommended. Custom nodes may conflict. Model files are large (2-10+ GB each). Python 3.10+ required. Complex workflows can consume significant VRAM.
84+
85+
# HISTORY
86+
87+
**ComfyUI** was created by **comfyanonymous** in **2023** as a modular alternative to the Automatic1111 web UI. Its node-based design attracted power users who needed flexible, reproducible generation pipelines. It became one of the most popular Stable Diffusion interfaces, with a large community building custom nodes.
88+
89+
# SEE ALSO
90+
91+
[sd-cli](/man/sd-cli)(1), [mflux](/man/mflux)(1), [convert](/man/convert)(1)

assets/commands/faster-whisper.md

Lines changed: 90 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,90 @@
1+
# TLDR
2+
3+
**Transcribe an audio file**
4+
5+
```faster-whisper [audio.mp3]```
6+
7+
**Transcribe with a specific model**
8+
9+
```faster-whisper [audio.mp3] --model [large-v3]```
10+
11+
**Transcribe with language hint**
12+
13+
```faster-whisper [audio.mp3] --language [en]```
14+
15+
**Output as SRT subtitles**
16+
17+
```faster-whisper [audio.mp3] --output_format [srt]```
18+
19+
**Translate to English**
20+
21+
```faster-whisper [audio.mp3] --task [translate]```
22+
23+
**Save output to directory**
24+
25+
```faster-whisper [audio.mp3] --output_dir [/path/to/output]```
26+
27+
**Transcribe with word timestamps**
28+
29+
```faster-whisper [audio.mp3] --word_timestamps [true]```
30+
31+
# SYNOPSIS
32+
33+
**faster-whisper** _audio_ [_--model size_] [_--language lang_] [_--task task_] [_options_]
34+
35+
# PARAMETERS
36+
37+
**--model** _SIZE_
38+
> Model size: tiny, base, small, medium, large-v1, large-v2, large-v3 (default: small).
39+
40+
**--language** _LANG_
41+
> Language code (en, de, fr, etc.) or auto-detect.
42+
43+
**--task** _TASK_
44+
> Task: transcribe or translate.
45+
46+
**--output_format** _FORMAT_
47+
> Output format: txt, vtt, srt, tsv, json, all.
48+
49+
**--output_dir** _DIR_
50+
> Output directory for results.
51+
52+
**--word_timestamps** _BOOL_
53+
> Include word-level timestamps.
54+
55+
**--device** _DEVICE_
56+
> Device: cpu, cuda, auto (default: auto).
57+
58+
**--compute_type** _TYPE_
59+
> Compute type: int8, float16, float32 (default: int8 on CPU).
60+
61+
**--beam_size** _N_
62+
> Beam search size (default: 5).
63+
64+
**--vad_filter** _BOOL_
65+
> Enable voice activity detection filter.
66+
67+
**--threads** _N_
68+
> Number of CPU threads.
69+
70+
# DESCRIPTION
71+
72+
**faster-whisper** is a reimplementation of OpenAI's Whisper using **CTranslate2**, a fast inference engine for Transformer models. It provides up to 4x faster transcription than the original Whisper while using less memory.
73+
74+
The tool supports all Whisper model sizes. Larger models are more accurate but slower. The compute type parameter controls precision: int8 is fastest and most memory-efficient, float16 is a good balance on GPU, float32 is highest precision.
75+
76+
Voice activity detection (VAD) filtering skips silent sections, improving both speed and accuracy. Language detection is automatic but specifying the language avoids detection overhead.
77+
78+
Install via pip (`pip install faster-whisper`). CTranslate2 handles model conversion automatically. GPU acceleration requires CUDA toolkit.
79+
80+
# CAVEATS
81+
82+
Large models require significant memory. CUDA toolkit needed for GPU. First run downloads and converts models. Accuracy varies by audio quality. No speaker diarization in CLI (available via API).
83+
84+
# HISTORY
85+
86+
**faster-whisper** was created by **Guillaume Klein** (SYSTRAN) in **2023** using CTranslate2 to optimize Whisper inference. It became the preferred Whisper implementation for production use due to its speed and memory advantages. The project achieved wide adoption in transcription workflows.
87+
88+
# SEE ALSO
89+
90+
[whisper](/man/whisper)(1), [deepspeech](/man/deepspeech)(1), [ffmpeg](/man/ffmpeg)(1)

0 commit comments

Comments
 (0)