WorkRB (~pronounced worker bee) is an open-source evaluation toolbox for benchmarking AI models in the work research domain. It provides a standardized framework that is easy to use and community-driven, scaling evaluation over a wide range of tasks, ontologies, and models.
- 🐝 One Buzzing Work Toolkit — Easily download & access ontologies, datasets, and baselines in a single toolkit
- 🧪 Extensive tasks — Evaluate models on job–skill matching, normalization, extraction, and similarity
- 🌍 Dynamic Multilinguality — Evaluate over languages driven by multilingual ontologies
- 🧠 Ready-to-go Baselines — Leverage provided baseline models for comparison
- 🧩 Extensible design — Add your custom tasks and models with simple interfaces
import workrb
# 1. Initialize a model
model = workrb.models.BiEncoderModel("all-MiniLM-L6-v2")
# 2. Select (multilingual) tasks to evaluate
tasks = [
workrb.tasks.ESCOJob2SkillRanking(split="val", languages=["en"]),
workrb.tasks.ESCOSkillNormRanking(split="val", languages=["de", "fr"])
]
# 3. Run benchmark & view results
results = workrb.evaluate( # Returns BenchmarkResults (Pydantic model)
model,
tasks,
output_folder="results/my_model",
)
print(results) # Benchmark/Per-task/Per-language metricsInstall WorkRB simply via pip.
pip install workrbRequirements: Python 3.10+, see pyproject.toml for all dependencies.
| Task Name | Class Name | Label Type | Dataset Size (English) | Languages |
|---|---|---|---|---|
| Ranking | ||||
| Job to Skills WorkBench | ESCOJob2SkillRanking |
multi_label | 3039 queries x 13939 targets | 28 |
| Job Title Similarity | JobTitleSimilarityRanking |
multi_label | 105 queries x 2619 targets | 11 |
| Job Normalization | JobBERTJobNormRanking |
single_label | 15463 queries x 2942 targets | 28 |
| Job Normalization MELO | MELORanking |
multi_label | 633 queries x 33813 targets | 21 |
| Skill to Job WorkBench | ESCOSkill2JobRanking |
multi_label | 13492 queries x 3039 targets | 28 |
| Skill Extraction House | HouseSkillExtractRanking |
multi_label | 262 queries x 13891 targets | 28 |
| Skill Extraction Tech | TechSkillExtractRanking |
multi_label | 338 queries x 13891 targets | 28 |
| Skill Extraction SkillSkape | SkillSkapeExtractRanking |
multi_label | 1191 queries x 13891 targets | 28 |
| Skill Similarity SkillMatch-1K | SkillMatch1kSkillSimilarityRanking |
single_label | 900 queries x 2648 targets | 1 |
| Skill Normalization ESCO | ESCOSkillNormRanking |
multi_label | 72008 queries x 13939 targets | 28 |
| Skill Normalization MELS | MELSRanking |
multi_label | 1722 queries x 19466 targets | 5 |
| Query-Candidate Matching | SearchQueryCandidateRanking |
multi_label | 200 queries x 4019 (x-lang) targets | 5 |
| Project-Candidate Matching | ProjectCandidateRanking |
multi_label | 200 queries x 4019 (x-lang) targets | 5 |
| Classification | ||||
| Job-Skill Classification | ESCOJob2SkillClassification |
multi_label | 3039 samples, 13939 classes | 28 |
| Model Name | Description | Adaptive Targets |
|---|---|---|
| Embedding Models | ||
| BiEncoderModel | BiEncoder model using sentence-transformers for ranking and classification tasks. | ✅ |
| JobBERTModel | Job-Normalization BiEncoder from Techwolf: https://huggingface.co/TechWolf/JobBERT-v2 | ✅ |
| ConTeXTMatchModel | ConTeXT-Skill-Extraction-base from Techwolf: https://huggingface.co/TechWolf/ConTeXT-Skill-Extraction-base | ✅ |
| CurriculumMatchModel | CurriculumMatch bi-encoder from Aleksandruz: https://huggingface.co/Aleksandruz/skillmatch-mpnet-curriculum-retriever | ✅ |
| Lexical Baselines | ||
| BM25Model | BM25 Okapi probabilistic ranking baseline. | ✅ |
| TfIdfModel | TF-IDF baseline with configurable word-level or character n-gram tokenization. | ✅ |
| EditDistanceModel | Edit distance (Levenshtein ratio) baseline for near-exact matching. | ✅ |
| RandomRankingModel | Random scoring baseline for sanity checking evaluation pipelines. | ✅ |
| Classification Baselines | ||
| RndESCOClassificationModel | Random baseline for multi-label classification with random prediction head for ESCO. | ❌ |
This section covers common usage patterns. Table of Contents:
Add your custom task or model by (1) inheriting from a predefined base class and implementing the abstract methods, and (2) adding it to the registry:
- Custom Tasks: Inherit from
RankingTask,MultilabelClassificationTask,... Implement the abstract methods. Register via@register_task(). - Custom models: Inherit from
ModelInterface. Implement the abstract methods. Register via@register_model().
from workrb.tasks.abstract.ranking_base import RankingTask
from workrb.models.base import ModelInterface
from workrb.registry import register_task, register_model
@register_task()
class MyCustomTask(RankingTask):
name: str = "MyCustomTask"
...
@register_model()
class MyCustomModel(ModelInterface):
name: str = "MyCustomModel"
...
# Use your custom model and task:
model_results = workrb.evaluate(MyCustomModel(),[MyCustomTask()])For detailed examples, see:
- examples/custom_task_example.py for a complete custom task implementation
- examples/custom_model_example.py for a complete custom model implementation
Feel free to make a PR to add your models & tasks to the official package! See CONTRIBUTING guidelines for details.
WorkRB automatically saves result checkpoints after each dataset evaluation within a task.
Automatic Resuming - Simply rerun with the same output_folder:
# Run 1: Gets interrupted after 2 tasks
tasks = [
workrb.tasks.ESCOJob2SkillRanking(
split="val",
languages=["en"],
)
]
results = workrb.evaluate(model, tasks, output_folder="results/my_model")
# Run 2: Automatically resumes from checkpoint
results = workrb.evaluate(model, tasks, output_folder="results/my_model")
# ✓ Skips completed tasks, continues from where it left offExtending Benchmarks - Want to extend your results with additional tasks or languages? Add the new tasks or languages when resuming:
# Resume from previous & extend with new task and languages
tasks_extended = [
workrb.tasks.ESCOJob2SkillRanking( # Add de, fr
split="val",
languages=["en", "de", "fr"],
),
workrb.tasks.ESCOSkillNormRanking( # Add new task
split="val",
languages=["en"],
),
]
results = workrb.evaluate(model, tasks_extended, output_folder="results/my_model")
# ✓ Reuses English results, only evaluates new languages/tasks❌You cannot reduce scope when resuming. This is by design to avoid ambiguity. Finished tasks in the checkpoint should also be included in your WorkRB initialization. If you want to start fresh in the same output folder, use force_restart=True:
results = workrb.evaluate(model, tasks, output_folder="results/my_model", force_restart=True)Results are automatically saved to your output_folder:
results/my_model/
├── checkpoint.json # Incremental checkpoint (for resuming)
├── results.json # Final results dump
└── config.yaml # Final benchmark configuration dump
To load & parse results from a run:
results = workrb.load_results("results/my_model/results.json")
print(results)The final benchmark score mean_benchmark/<metric>/mean is computed via the following chain:
dataset → language → task → task_group → task_type.
This enables sequential macro-averaging in each of the stages:
dataset: Is the individual unit to start aggregation from. Each task contains a set of datasets, each with a uniquedataset_id. Example: The MELO task language/region subsetsusa_q_en_c_enandswe_q_sv_c_en.language: Aggregate over languages within the task's datasets. Example: Group all monolingual French datasets in ESCOSkill2JobRankingtask: Aggregate over tasks in the same task group. Example: HouseSkillExtractRanking and TechSkillExtractRanking tasks in the Skill Extraction task group.task_group: Aggregate over task groups under a specific task type. Example: Skill Extraction, Skill Normalization, and Job Normalization task groups, are all part of the ranking task typetask_type: Aggregate over different task types for final benchmark performance, e.g. the Ranking and Classification task types.
Per-language performance is available under language-grouped modes: mean_per_language/<lang>/<metric>/mean.
Each aggregation provides 95% confidence intervals (replace mean with ci_margin).
Ranking metrics (used in RankingTask):
| Metric | Description |
|---|---|
map |
Mean Average Precision |
mrr |
Mean Reciprocal Rank |
recall@k |
Recall at k (e.g. recall@5, recall@10) |
hit@k |
Hit rate at k — binary: is any relevant item in the top-k? |
rp@k |
R-Precision at k — precision relative to total relevant items |
Classification metrics (used in ClassificationTask):
| Metric | Description |
|---|---|
f1_macro, f1_micro, f1_weighted |
F1 score variants |
f1_samples |
Per-sample F1 (multilabel only) |
precision_macro, precision_micro, precision_weighted |
Precision variants |
recall_macro, recall_micro, recall_weighted |
Recall variants |
accuracy |
Overall accuracy (subset accuracy for multilabel) |
roc_auc, roc_auc_micro |
Area under ROC curve (threshold-independent) |
You can override the default metrics per task via the metrics parameter of evaluate():
results = workrb.evaluate(
model, tasks, output_folder="results/my_model",
metrics={"ESCOJob2SkillRanking": ["map", "mrr", "recall@5", "recall@10"]},
)The language_aggregation_mode parameter controls how dataset results are grouped during metric aggregation. There are 4 modes (LanguageAggregationMode):
| Mode | Behavior |
|---|---|
MONOLINGUAL_ONLY (default) |
Group by language; only include monolingual datasets (input lang == output lang). Cross-lingual datasets are filtered out. |
CROSSLINGUAL_GROUP_INPUT_LANGUAGES |
Group by the input/query language. Requires a single input language per dataset (skip multilingual inputs). |
CROSSLINGUAL_GROUP_OUTPUT_LANGUAGES |
Group by the output/target language. Requires a single output language per dataset (skip multilingual outputs). |
SKIP_LANGUAGE_AGGREGATION |
No language grouping or filtering. All datasets are directly macro-averaged per task. No per-language metrics are produced. |
As datasets may be filtered out by the aggregation mode, you may want to skip evaluations that are not used in the final metrics. The execution_mode parameter controls whether incompatible datasets are evaluated:
| Mode | Behavior |
|---|---|
ExecutionMode.LAZY (default) |
Skip datasets that are incompatible with the chosen language_aggregation_mode, saving compute. |
ExecutionMode.ALL |
Evaluate all datasets regardless. Useful when you want to store all results and re-aggregate later with a different mode. |
Note: Under
SKIP_LANGUAGE_AGGREGATION, no datasets are ever incompatible, soexecution_modecan be ignored.
An example of why you could choose for ExecutionMode.ALL:
from workrb.types import LanguageAggregationMode, ExecutionMode
# Benchmark returns a detailed Pydantic model
results: BenchmarkResults = workrb.evaluate(
model,
tasks,
language_aggregation_mode=LanguageAggregationMode.MONOLINGUAL_ONLY,
execution_mode=ExecutionMode.ALL, # Execute all so we can later switch language_aggregation_mode
)
# No lazy mode was used; We can override the aggregation mode at summary time
summary = results.get_summary_metrics(
language_aggregation_mode=LanguageAggregationMode.CROSSLINGUAL_GROUP_INPUT_LANGUAGES,
)For a complete runnable example of different aggregation strategies, see examples/run_benchmark_aggregation.py.
Evaluate multiple models in a single call with evaluate_multiple_models(). Each model runs the evaluate() and automatically gets its own output folder (with checkpointing) based on the model name:
results = workrb.evaluate_multiple_models(
models=[model_a, model_b],
tasks=tasks,
output_folder_template="results/{model_name}", # Use explicit '{model_name}', used for templating
)
# results["model_a_name"] -> BenchmarkResults
# results["model_b_name"] -> BenchmarkResultsAll keyword arguments from evaluate() (e.g. language_aggregation_mode, execution_mode) are passed through. See examples/run_multiple_models.py for a complete example.
Want to contribute new tasks, models, or metrics? Read our CONTRIBUTING.md guide for all details.
# Clone repository
git clone https://github.com/techwolf-ai/workrb.git && cd workrb
# Create and install a virtual environment
uv sync --all-extras
# Activate the virtual environment
source .venv/bin/activate
# Install the pre-commit hooks
pre-commit install --install-hooks
# Run tests (excludes model benchmarking by default)
uv run poe test
# Run model benchmark tests only, checks reproducibility of original results
uv run poe test-benchmarkDeveloping details
- This project follows the Conventional Commits standard to automate Semantic Versioning and Keep A Changelog with Commitizen.
- Run
poefrom within the development environment to print a list of Poe the Poet tasks available to run on this project. - Run
uv add {package}from within the development environment to install a run time dependency and add it topyproject.tomlanduv.lock. Add--devto install a development dependency. - Run
uv sync --upgradefrom within the development environment to upgrade all dependencies to the latest versions allowed bypyproject.toml. Add--only-devto upgrade the development dependencies only. - Run
cz bumpto bump the package's version, update theCHANGELOG.md, and create a git tag. Then push the changes and the git tag withgit push origin main --tags.
WorkRB builds upon the unifying WorkBench benchmark, consider citing:
@misc{delange2025unifiedworkembeddings,
title={Unified Work Embeddings: Contrastive Learning of a Bidirectional Multi-task Ranker},
author={Matthias De Lange and Jens-Joris Decorte and Jeroen Van Hautte},
year={2025},
eprint={2511.07969},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2511.07969},
}WorkRB has a community paper coming up!
WIPApache 2.0 License - see LICENSE for details.
