Skip to content

Conversation

@Camier
Copy link
Member

@Camier Camier commented Nov 14, 2025

Summary

  • replace stale script references in the master command reference with the supported dashboard, validation, profiling, and recovery tooling
  • update the documentation index, configuration quick reference, and API/cloud guides to point to the real routing and profiling commands
  • document the supported backup/rollback workflow that the generate-litellm-config and verification scripts already provide

Testing

  • not run (documentation-only changes)

Codex Task

Copilot AI review requested due to automatic review settings November 14, 2025 06:50
Copy link

Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull Request Overview

This PR updates documentation to replace stale script references with current commands and properly documents the backup/rollback workflow. The changes span multiple documentation files to ensure consistency across the codebase.

Key Changes

  • Replace outdated ./test-request.py references with python3 scripts/debugging/test-request.py for consistency
  • Remove references to non-existent test-routing.sh script in favor of test-request.py --test-routing
  • Reorganize and expand COMMAND-REFERENCE.md with clearer categorization and workflow documentation
  • Document the generate-litellm-config.py backup/rollback features (--list-backups and --rollback)

Reviewed Changes

Copilot reviewed 6 out of 6 changed files in this pull request and generated 6 comments.

Show a summary per file
File Description
docs/observability.md Updates test-request.py command invocations to use full python3 scripts/debugging/ paths for consistency
docs/COMMAND-REFERENCE.md Major reorganization with new sections for dashboards, health checks, profiling, backup/recovery; documents previously undocumented workflow features
docs/API-REFERENCE.md Updates routing and monitoring command references to point to current scripts
DOCUMENTATION-INDEX.md Removes non-existent test-routing.sh, updates profiling script references, corrects operational tool references
CONFIGURATION-QUICK-REFERENCE.md Expands backup/rollback workflow with detailed generate-litellm-config.py usage and verification steps
CLOUD_MODELS_READY.md Updates token usage monitoring command reference

💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.


# Token usage by provider
./scripts/profiling/analyze-token-usage.py
python3 scripts/profiling/compare-providers.py --summary
Copy link

Copilot AI Nov 14, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The compare-providers.py script does not have a --summary flag. According to the script's argument parser, it accepts --url, --models, --prompt, --iterations, and --export, but not --summary. The command should be:

python3 scripts/profiling/compare-providers.py

or with specific models:

python3 scripts/profiling/compare-providers.py --models llama3.1:8b qwen-coder-vllm
Suggested change
python3 scripts/profiling/compare-providers.py --summary
python3 scripts/profiling/compare-providers.py

Copilot uses AI. Check for mistakes.
### Check Model Usage
```bash
./scripts/profiling/analyze-token-usage.py
python3 scripts/profiling/compare-providers.py --summary
Copy link

Copilot AI Nov 14, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The compare-providers.py script does not have a --summary flag. According to the script's argument parser, it accepts --url, --models, --prompt, --iterations, and --export, but not --summary. The command should be:

python3 scripts/profiling/compare-providers.py

or with specific models:

python3 scripts/profiling/compare-providers.py --models llama3.1:8b qwen-coder-vllm
Suggested change
python3 scripts/profiling/compare-providers.py --summary
python3 scripts/profiling/compare-providers.py

Copilot uses AI. Check for mistakes.
- **Rate limits**: Applied by Ollama (check at https://ollama.com/settings/usage)
- **Fallback**: Automatically routes to local models if rate limited
- **Monitoring**: Track usage with `./scripts/profiling/analyze-token-usage.py`
- **Monitoring**: Track usage with `python3 scripts/profiling/compare-providers.py --summary`
Copy link

Copilot AI Nov 14, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The compare-providers.py script does not have a --summary flag. According to the script's argument parser, it accepts --url, --models, --prompt, --iterations, and --export, but not --summary. The command should be:

python3 scripts/profiling/compare-providers.py

or with specific models:

python3 scripts/profiling/compare-providers.py --models llama3.1:8b qwen-coder-vllm
Suggested change
- **Monitoring**: Track usage with `python3 scripts/profiling/compare-providers.py --summary`
- **Monitoring**: Track usage with `python3 scripts/profiling/compare-providers.py` or with specific models: `python3 scripts/profiling/compare-providers.py --models llama3.1:8b qwen-coder-vllm`

Copilot uses AI. Check for mistakes.
python3 scripts/debugging/test-request.py --model llama3.1:8b
python3 scripts/debugging/test-request.py --test-routing
./scripts/debugging/tail-requests.py --level ERROR
./scripts/debugging/analyze-logs.py --last 1h --model qwen
Copy link

Copilot AI Nov 14, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The analyze-logs.py script does not have a --last flag. According to the script's implementation, it requires a log file path as the first positional argument. The correct usage should be:

./scripts/debugging/analyze-logs.py /var/log/litellm/requests.log

For filtering by model, the script doesn't have a --model flag either. Remove this entire line or correct it to show the actual script usage.

Suggested change
./scripts/debugging/analyze-logs.py --last 1h --model qwen
./scripts/debugging/analyze-logs.py /var/log/litellm/requests.log

Copilot uses AI. Check for mistakes.
./scripts/profiling/compare-providers.py
python3 scripts/profiling/profile-latency.py --model qwen2.5-coder:7b
python3 scripts/profiling/profile-throughput.py --sweep --model llama3.1:8b
python3 scripts/profiling/compare-providers.py --summary
Copy link

Copilot AI Nov 14, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The compare-providers.py script does not have a --summary flag. According to the script's argument parser, it accepts --url, --models, --prompt, --iterations, and --export, but not --summary. The command should be:

python3 scripts/profiling/compare-providers.py

or with specific models:

python3 scripts/profiling/compare-providers.py --models llama3.1:8b qwen-coder-vllm
Suggested change
python3 scripts/profiling/compare-providers.py --summary
python3 scripts/profiling/compare-providers.py

Copilot uses AI. Check for mistakes.
./scripts/debugging/tail-requests.py

# Historical log analysis
./scripts/debugging/analyze-logs.py --last 1h
Copy link

Copilot AI Nov 14, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The analyze-logs.py script does not have a --last flag. According to the script's implementation, it requires a log file path as the first positional argument. The correct usage should be:

./scripts/debugging/analyze-logs.py /var/log/litellm/requests.log

The script doesn't support time-based filtering like --last 1h.

Suggested change
./scripts/debugging/analyze-logs.py --last 1h
./scripts/debugging/analyze-logs.py /var/log/litellm/requests.log

Copilot uses AI. Check for mistakes.
@Camier
Copy link
Member Author

Camier commented Nov 16, 2025

Note: dashboard strategy is consolidated on the Hello Kitty Textual app. Please verify doc changes in this PR reflect that single direction (launcher is scripts/ai-dashboard pointing to dashboards/hello_kitty). If not, let’s update or pause merge to avoid conflicting guidance.

@Camier Camier added the blocked Needs attention before merge label Nov 16, 2025
@Camier
Copy link
Member Author

Camier commented Nov 16, 2025

Alignment checklist for Hello Kitty–only direction:\n- Update commands to use (Hello Kitty Textual). Remove/deprecate WTH/Bubble Tea references or clearly mark them legacy.\n- If docs mention PTUI, call it secondary/SSH and include a quick smoke ().\n- Validation: run ================================================================================
LiteLLM Configuration Generator

🔍 Checking for manual edits...
✓ No manual edits detected

💾 Creating backup...
✓ Backed up to: config/backups/litellm-unified.yaml.20251116-133249
Cleaning up old backups (keeping 10)...
Removed: litellm-unified.yaml.20251107-191107

🏗️ Building complete configuration...

🔀 Building router settings...
✓ Created 6 capability groups
✓ Created 18 fallback chains

⏱️ Building rate limit settings...
✓ Configured rate limits for 12 models
✓ Configuration built successfully

✍️ Writing configuration to config/litellm-unified.yaml...
✓ Configuration written successfully

📌 Version saved: git-6da51ef

✅ Validating generated configuration...
✓ Validation passed

================================================================================
✅ Configuration generated successfully!

Output: config/litellm-unified.yaml
Version: git-6da51ef
Backup: config/backups/

Next steps:

  1. Review generated configuration
  2. Test: curl http://localhost:4000/v1/models
  3. Ensure service is provisioned: ./runtime/scripts/run_litellm.sh
  4. Restart: systemctl --user restart litellm.service then ==================================
    AI Unified Backend Validation
    ==================================

=== Phase 1: System Checks ===

�[1;33mℹ�[0m Checking systemd services...
�[0;32m✅�[0m LiteLLM service is running
�[0;31m❌�[0m Ollama service exists but is NOT running

=== Phase 2: Provider Health Checks ===

�[1;33mℹ�[0m Testing provider endpoints...
�[0;31m❌�[0m Ollama is NOT accessible; attach logs.\nWill re-review once adjusted.

@Camier
Copy link
Member Author

Camier commented Nov 16, 2025

Alignment checklist for the Hello Kitty-only direction:

  • Update commands to use scripts/ai-dashboard (Hello Kitty Textual). Remove or clearly mark WTH/Bubble Tea references as legacy.
  • If docs mention PTUI, label it secondary/SSH-only and include a simple smoke command: python scripts/ptui_dashboard.py --help.
  • Validation: run python scripts/generate-litellm-config.py then bash scripts/validate-unified-backend.sh; attach logs.
    Will re-review once adjusted.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

blocked Needs attention before merge codex

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants