You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
* fix: correct yaml linting hook and fix trailing spaces/comment spacing
This PR addresses two issues:
1. **Fixed pre-commit hook configuration bug** - Changed line 57 in
`.pre-commit-config.yaml` to call `make yaml-lint` instead of
`make markdown-lint`
2. **Fixed simple YAML linting errors** - Applied automated fixes for:
- Trailing whitespace in YAML files
- Comment spacing (ensuring 2 spaces before inline comments)
## Problem
The bug in `.pre-commit-config.yaml` caused:
- ❌ YAML files not being properly linted locally
- ✅ GitHub Actions CI catching the issues
- 🤔 PRs failing in CI even though `pre-commit run --all-files` passed locally
- 😓 Contributors forced to fix pre-existing YAML issues
## Changes
1. Changed `.pre-commit-config.yaml` line 57 from `make markdown-lint` to `make yaml-lint`
2. Fixed trailing spaces and comment spacing in 22 YAML files
## Note on Remaining Issues
Some YAML files still have indentation errors that require more careful
manual fixes. These can be addressed in follow-up PRs as files are modified.
The important fix here is that local pre-commit checks now match CI checks.
Fixes#608
🤖 Generated with [Claude Code](https://claude.com/claude-code)
Co-Authored-By: Claude <noreply@anthropic.com>
Signed-off-by: Yossi Ovadia <yovadia@redhat.com>
* fix: resolve YAML indentation errors and exclude .venv from linting
This commit fixes the pre-commit failures from the previous commit by:
1. **Fixed YAML indentation errors** in 16 Kubernetes/OpenShift deployment files
- Re-parsed and reformatted YAML files with proper 2-space indentation
- Fixed wrong indentation issues flagged by yamllint
2. **Excluded .venv from yamllint checks**
- Added `.venv` to the ignore list in `tools/linter/yaml/.yamllint`
- Prevents linting errors from third-party dependencies in virtual environment
Files fixed:
- deploy/kubernetes/ai-gateway/aigw-resources/*
- deploy/kubernetes/aibrix/aigw-resources/*
- deploy/kubernetes/istio/*
- deploy/kubernetes/llmd-base/*
- deploy/openshift/observability/prometheus/deployment.yaml
- deploy/openshift/template.yaml
Pre-commit now passes successfully.
Co-Authored-By: Claude <noreply@anthropic.com>
Signed-off-by: Yossi Ovadia <yovadia@redhat.com>
---------
Signed-off-by: Yossi Ovadia <yovadia@redhat.com>
Co-authored-by: Claude <noreply@anthropic.com>
Co-authored-by: Huamin Chen <rootfs@users.noreply.github.com>
# NOT supported: domain names (example.com), protocol prefixes (http://), paths (/api), ports in address (use 'port' field)
47
47
vllm_endpoints:
48
48
- name: "endpoint1"
49
-
address: "172.28.0.20"# Static IPv4 of llm-katan within docker compose network
49
+
address: "172.28.0.20"# Static IPv4 of llm-katan within docker compose network
50
50
port: 8002
51
51
weight: 1
52
52
53
53
model_config:
54
54
"qwen3":
55
-
reasoning_family: "qwen3"# This model uses Qwen-3 reasoning syntax
55
+
reasoning_family: "qwen3"# This model uses Qwen-3 reasoning syntax
56
56
preferred_endpoints: ["endpoint1"] # Optional: omit to let upstream handle endpoint selection
57
57
pii_policy:
58
58
allow_by_default: true
@@ -81,7 +81,7 @@ categories:
81
81
model_scores:
82
82
- model: qwen3
83
83
score: 0.7
84
-
use_reasoning: false # Business performs better without reasoning
84
+
use_reasoning: false # Business performs better without reasoning
85
85
- name: law
86
86
system_prompt: "You are a knowledgeable legal expert with comprehensive understanding of legal principles, case law, statutory interpretation, and legal procedures across multiple jurisdictions. Provide accurate legal information and analysis while clearly stating that your responses are for informational purposes only and do not constitute legal advice. Always recommend consulting with qualified legal professionals for specific legal matters."
87
87
model_scores:
@@ -91,7 +91,7 @@ categories:
91
91
- name: psychology
92
92
system_prompt: "You are a psychology expert with deep knowledge of cognitive processes, behavioral patterns, mental health, developmental psychology, social psychology, and therapeutic approaches. Provide evidence-based insights grounded in psychological research and theory. When discussing mental health topics, emphasize the importance of professional consultation and avoid providing diagnostic or therapeutic advice."
93
93
semantic_cache_enabled: true
94
-
semantic_cache_similarity_threshold: 0.92# High threshold for psychology - sensitive to nuances
94
+
semantic_cache_similarity_threshold: 0.92# High threshold for psychology - sensitive to nuances
95
95
model_scores:
96
96
- model: qwen3
97
97
score: 0.6
@@ -107,7 +107,7 @@ categories:
107
107
model_scores:
108
108
- model: qwen3
109
109
score: 0.6
110
-
use_reasoning: true # Enable reasoning for complex chemistry
110
+
use_reasoning: true # Enable reasoning for complex chemistry
111
111
- name: history
112
112
system_prompt: "You are a historian with expertise across different time periods and cultures. Provide accurate historical context and analysis."
113
113
model_scores:
@@ -117,15 +117,15 @@ categories:
117
117
- name: other
118
118
system_prompt: "You are a helpful and knowledgeable assistant. Provide accurate, helpful responses across a wide range of topics."
119
119
semantic_cache_enabled: true
120
-
semantic_cache_similarity_threshold: 0.75# Lower threshold for general chat - less sensitive
120
+
semantic_cache_similarity_threshold: 0.75# Lower threshold for general chat - less sensitive
121
121
model_scores:
122
122
- model: qwen3
123
123
score: 0.7
124
124
use_reasoning: false
125
125
- name: health
126
126
system_prompt: "You are a health and medical information expert with knowledge of anatomy, physiology, diseases, treatments, preventive care, nutrition, and wellness. Provide accurate, evidence-based health information while emphasizing that your responses are for educational purposes only and should never replace professional medical advice, diagnosis, or treatment. Always encourage users to consult healthcare professionals for medical concerns and emergencies."
127
127
semantic_cache_enabled: true
128
-
semantic_cache_similarity_threshold: 0.95# High threshold for health - very sensitive to word changes
128
+
semantic_cache_similarity_threshold: 0.95# High threshold for health - very sensitive to word changes
129
129
model_scores:
130
130
- model: qwen3
131
131
score: 0.5
@@ -141,13 +141,13 @@ categories:
141
141
model_scores:
142
142
- model: qwen3
143
143
score: 1.0
144
-
use_reasoning: true # Enable reasoning for complex math
144
+
use_reasoning: true # Enable reasoning for complex math
145
145
- name: physics
146
146
system_prompt: "You are a physics expert with deep understanding of physical laws and phenomena. Provide clear explanations with mathematical derivations when appropriate."
147
147
model_scores:
148
148
- model: qwen3
149
149
score: 0.7
150
-
use_reasoning: true # Enable reasoning for physics
150
+
use_reasoning: true # Enable reasoning for physics
151
151
- name: computer science
152
152
system_prompt: "You are a computer science expert with knowledge of algorithms, data structures, programming languages, and software engineering. Provide clear, practical solutions with code examples when helpful."
153
153
model_scores:
@@ -195,24 +195,24 @@ router:
195
195
lora_default_success_rate: 0.98
196
196
traditional_default_success_rate: 0.95
197
197
# Scoring weights for intelligent path selection (balanced approach)
198
-
multi_task_lora_weight: 0.30# LoRA advantage for multi-task processing
199
-
single_task_traditional_weight: 0.30# Traditional advantage for single tasks
200
-
large_batch_lora_weight: 0.25# LoRA advantage for large batches (≥4)
201
-
small_batch_traditional_weight: 0.25# Traditional advantage for single items
202
-
medium_batch_weight: 0.10# Neutral weight for medium batches (2-3)
203
-
high_confidence_lora_weight: 0.25# LoRA advantage for high confidence (≥0.99)
204
-
low_confidence_traditional_weight: 0.25# Traditional for lower confidence (≤0.9)
205
-
low_latency_lora_weight: 0.30# LoRA advantage for low latency (≤2000ms)
206
-
high_latency_traditional_weight: 0.10# Traditional acceptable for relaxed timing
0 commit comments