Skip to content

Commit 23080eb

Browse files
Merge pull request #2284 from bluewave-labs/experiment-bias-fairness
Add Remote Model Inference via Extensible LLMClient
2 parents 8b4601d + 03a41c5 commit 23080eb

File tree

15 files changed

+1221
-598
lines changed

15 files changed

+1221
-598
lines changed

BiasAndFairnessModule/configs/config.yaml

Lines changed: 12 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,6 @@ dataset:
1212
enabled: true
1313
n_samples: 50
1414
random_seed: 42
15-
1615

1716
# Post-processing Configuration
1817
post_processing:
@@ -30,18 +29,18 @@ post_processing:
3029

3130
# Model Configuration
3231
model:
32+
# Provider-agnostic model configuration
33+
provider: "huggingface" # Options: huggingface, openai
3334
# Model task type for Fairness Compass routing
3435
model_task: "binary_classification" # Options: binary_classification, multiclass_classification, regression, generation, ranking
35-
label_behavior: "binary" # Options: binary, categorical, continuous
36-
37-
# Hugging Face model configuration
38-
huggingface:
39-
enabled: true
40-
model_id: "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
41-
device: "cuda" # or "cpu" for CPU-only inference
42-
max_new_tokens: 50
43-
temperature: 0.7
44-
top_p: 0.9
36+
label_behavior: "binary" # Options: binary, categorical, continuous
37+
model_id: "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
38+
device: "cuda" # or "cpu" for CPU-only inference
39+
max_new_tokens: 50
40+
temperature: 0.7
41+
top_p: 0.9
42+
# Optional for remote providers (OpenAI-compatible endpoints)
43+
base_url: null
4544

4645
# Prompting Configuration
4746
prompting:
@@ -72,5 +71,5 @@ metrics:
7271
- "precision"
7372

7473
artifacts:
75-
inference_results_path: "artifacts/cleaned_inference_results.csv"
76-
postprocessed_results_path: "artifacts/postprocessed_results.csv"
74+
inference_results_path: "artifacts/inference_results.csv"
75+
postprocessed_results_path: "artifacts/postprocessed_results.csv"

0 commit comments

Comments
 (0)