Skip to content

Commit de4b6a0

Browse files
committed
regression 2 renamed regression + documentation updated
1 parent 4ae3959 commit de4b6a0

32 files changed

Lines changed: 148 additions & 130 deletions

docs/build/html/_sources/evaluation.rst.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ GRN evaluation
33
=================
44
The evaluation metrics used in geneRNIB are summarized below. For a detailed description of each metric, refer to the geneRNIB paper.
55

6-
We originally defined **eight evaluation metrics**, grouped into three categories: **Regression 1, Regression 2, and Wasserstein Distance**.
6+
We originally defined **eight evaluation metrics**, grouped into three categories: **Regression 1, Regression, and Wasserstein Distance**.
77
However, we recently removed **Regression 1** as it did not prove to be effective for perturbational settings.
88

99
- The **regression-based metrics** assess the predictive power of an inferred GRN by using regression models to predict perturbation data (evaluation data) based on the feature space constructed from the inferred network.

docs/build/html/evaluation.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -76,7 +76,7 @@
7676
<section id="grn-evaluation">
7777
<h1>GRN evaluation<a class="headerlink" href="#grn-evaluation" title="Link to this heading"></a></h1>
7878
<p>The evaluation metrics used in geneRNIB are summarized below. For a detailed description of each metric, refer to the geneRNIB paper.</p>
79-
<p>We originally defined <strong>eight evaluation metrics</strong>, grouped into three categories: <strong>Regression 1, Regression 2, and Wasserstein Distance</strong>.
79+
<p>We originally defined <strong>eight evaluation metrics</strong>, grouped into three categories: <strong>Regression 1, Regression, and Wasserstein Distance</strong>.
8080
However, we recently removed <strong>Regression 1</strong> as it did not prove to be effective for perturbational settings.</p>
8181
<ul class="simple">
8282
<li><p>The <strong>regression-based metrics</strong> assess the predictive power of an inferred GRN by using regression models to predict perturbation data (evaluation data) based on the feature space constructed from the inferred network.</p></li>

docs/source/evaluation.rst

Lines changed: 23 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,15 +1,22 @@
11

22
GRN evaluation
33
=================
4-
The evaluation metrics used in geneRNIB are summarized below. For a detailed description of each metric, refer to the geneRNIB paper.
5-
4+
The evaluation metrics used in geneRNIB are summarized below.
65

76

87
.. image:: images/metrics.png
98
:width: 90%
109
:align: center
1110
----
1211

12+
.. image:: images/datasets_metrics.png
13+
:width: 90%
14+
:align: center
15+
----
16+
17+
18+
For a detailed description of each metric, refer to the geneRNIB paper.
19+
1320
The evaluation metrics expect the inferred network to be in the form of an AnnData object with specific format as explained here.
1421
It should be noted that the metric currently evaluate only the **top TF-gene pairs**, currently limited to **50,000 edges**, ranked by their assigned weight.
1522

@@ -21,6 +28,7 @@ The inferred network should have a tabular format with the following columns:
2128

2229
See `resources/grn_benchmark/prior/collectri.h5ad` for an example of the expected format.
2330

31+
## Running GRN evaluation using standard pipeline
2432

2533
To run the evalution for a given GRN and dataset, use the following command:
2634
```bash
@@ -31,3 +39,16 @@ example command:
3139
```bash
3240
bash scripts/run_grn_evaluation.sh --prediction=resources/grn_models/op/collectri.h5ad --save_dir=output/ --dataset=op --build_images=true
3341
```
42+
43+
44+
## Running GRN evaluation without docker
45+
Considering that Docker is not supported by certtain systems, you can run the evaluation without Docker by following these steps:
46+
47+
```bash
48+
bash src/metrics/all_metrics/run_local.sh --dataset <dataset_name> --prediction=<inferred GRN (e.g.collectri.h5ad)> --score <output_score_file.h5ad> --num_workers <number_of_workers>
49+
```
50+
51+
example command:
52+
```bash
53+
bash src/metrics/all_metrics/run_local.sh --dataset op --prediction=resources/grn_models/op/collectri.h5ad --score=output_score_file.h5ad --num_workers=20
54+
```
295 KB
Loading

docs/source/index.rst

Lines changed: 21 additions & 22 deletions
Original file line numberDiff line numberDiff line change
@@ -2,40 +2,39 @@ Documentation for Gene Regulatory Network Inference Benchmark (geneRNIB)
22
========================================================================
33

44

5-
geneRNIB is a living benchmark platform for GRN inference. This platform provides curated datasets for GRN inference and evaluation, standardized evaluation protocols and metrics, computational infrastructure, and a dynamically updated leaderboard to track state-of-the-art methods. It runs novel GRNs in the cloud, offers competition scores, and stores them for future comparisons, reflecting new developments over time.
5+
geneRNIB is a living benchmark platform for GRN inference. This platform provides curated datasets for GRN inference and evaluation, standardized evaluation protocols and metrics, computational infrastructure, and a dynamically updated leaderboard to track state-of-the-art methods.
6+
It runs novel GRNs in the cloud, offers competition scores, and stores them for future comparisons, reflecting new developments over time.
67

7-
The platform supports the integration of new inference methods, datasets, and protocols. When a new feature is added, previously evaluated GRNs are re-assessed, and the leaderboard is updated accordingly. The aim is to evaluate both the accuracy and completeness of inferred GRNs. It is designed for both single-modality and multi-omics GRN inference.
8+
The platform supports the integration of new inference methods, datasets, and protocols. When a new feature is added, previously evaluated GRNs are re-assessed, and the leaderboard is updated accordingly.
9+
It is designed for both single-modality and multi-omics GRN inference.
810

911
.. image:: images/overview.png
1012
:width: 70%
1113
:align: center
1214
----
1315

14-
This documentation is supplementary to the paper `geneRNIB: a living benchmark for gene regulatory network inference <add a link here>`_ and the `GitHub page <https://github.com/openproblems-bio/task_grn_inference>`_ on the OpenProblems platform.
16+
This documentation is supplementary to the paper `geneRNIB: a living benchmark for gene regulatory network inference <https://www.biorxiv.org/content/10.1101/2025.02.25.640181v1.full.pdf>`_ and the `GitHub page <https://github.com/openproblems-bio/task_grn_inference>`_ on the OpenProblems platform.
1517

16-
To install geneRNIB, see the `GitHub page <https://github.com/openproblems-bio/task_grn_inference>`_.
17-
18-
For instructions on how to download and access datasets, refer to the :doc:`dataset` section.
18+
- To install geneRNIB, see the `GitHub page <https://github.com/openproblems-bio/task_grn_inference>`_
19+
- To download, see :doc:`dataset` page
20+
- To perform GRN inference using our integrated methods, see :doc:`inference` page
21+
- To run evaluation metrics, see :doc:`evaluation` page
22+
- To extend geneRNIB with new methods, metrics, or datasets, see :doc:`extending` page
23+
- To view the leaderboard of integrated methods, see :doc:`leaderboard` page
1924

20-
For information on evaluation metrics, refer to the :doc:`evaluation` section.
25+
.. .. image:: images/grn_models.png
26+
.. :width: 70%
27+
.. :align: center
28+
.. ----
2129
22-
To integrate your GRN inference method, metric, or dataset, follow the instructions in the :doc:`extending` section.
2330
24-
To see the comparitive performance of the integrated GRN inference methods, refer to the :doc:`leaderboard` section.
25-
26-
.. image:: images/grn_models.png
27-
:width: 70%
28-
:align: center
29-
----
31+
.. Pls see the GitHub page for the list of currently integrated methods. The methods are implemented in Python and R, and they can be used to infer GRNs from the datasets provided by geneRNIB.
3032
33+
.. In addition, three baseline methods are integrated into geneRNIB. These methods are used to evaluate the performance of new methods. The baseline methods are:
3134
32-
Pls see the GitHub page for the list of currently integrated methods. The methods are implemented in Python and R, and they can be used to infer GRNs from the datasets provided by geneRNIB.
33-
34-
In addition, three baseline methods are integrated into geneRNIB. These methods are used to evaluate the performance of new methods. The baseline methods are:
35-
36-
- **Negative control**: Randomly assigns weights to edges. GRN inference methods should outperform this method.
37-
- **Pearson correlation**: Assigns weights based on the Pearson correlation between genes.
38-
- **Positive control**: Similar to Pearson correlation with the exception that it uses both inference and evaluation dataset to infer the GRN. This method is expected to outperform most methods.
35+
.. - **Negative control**: Randomly assigns weights to edges. GRN inference methods should outperform this method.
36+
.. - **Pearson correlation**: Assigns weights based on the Pearson correlation between genes.
37+
.. - **Positive control**: Similar to Pearson correlation with the exception that it uses both inference and evaluation dataset to infer the GRN. This method is expected to outperform most methods.
3938
4039
4140
.. .. list-table:: Authors & contributors
@@ -59,8 +58,8 @@ Contents
5958
--------
6059

6160
.. toctree::
62-
6361
dataset
62+
inference
6463
evaluation
6564
extending
6665
leaderboard

scripts/prior/run_consensus.sh

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -30,12 +30,12 @@ for model in "${models[@]}"; do
3030
done
3131
printf '%s\n' "${predictions[@]}"
3232

33-
echo "Running consensus for regression 2"
33+
echo "Running consensus for Regression"
3434
datasets=(${DATASET})
3535
for dataset in "${datasets[@]}"; do
3636
echo "Running reg2 consensus for dataset: $dataset"
3737

38-
python src/metrics/regression_2/consensus/script.py \
38+
python src/metrics/regression/consensus/script.py \
3939
--dataset "$dataset" \
4040
--regulators_consensus "resources/grn_benchmark/prior/regulators_consensus_${dataset}.json" \
4141
--evaluation_data "resources/grn_benchmark/evaluation_data/${dataset}_bulk.h5ad" \

scripts/repo/run_benchmark_all.sh

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -97,19 +97,19 @@ HERE
9797

9898
# --------- COMBINATIONS TO ADD ----------
9999

100-
# append_entry "op" "[regression_1,regression_2, ws_distance]" "[pearson_corr, negative_control, positive_control,
100+
# append_entry "op" "[regression_1,regression, ws_distance]" "[pearson_corr, negative_control, positive_control,
101101
# portia, ppcor, scenic, scprint, grnboost,
102102
# scenicplus, scglue, granie, figr, celloracle]"
103-
# append_entry "norman" "[regression_1,regression_2, ws_distance]" "[pearson_corr, negative_control, positive_control,
103+
# append_entry "norman" "[regression_1,regression, ws_distance]" "[pearson_corr, negative_control, positive_control,
104104
# portia, ppcor, scenic, scprint, grnboost]"
105-
# append_entry "adamson" "[regression_1,regression_2, ws_distance]" "[pearson_corr, negative_control, positive_control,
105+
# append_entry "adamson" "[regression_1,regression, ws_distance]" "[pearson_corr, negative_control, positive_control,
106106
# portia, ppcor, scenic, grnboost]"
107-
# append_entry "nakatake" "[regression_1,regression_2]" "[pearson_corr, negative_control, positive_control,
107+
# append_entry "nakatake" "[regression_1,regression]" "[pearson_corr, negative_control, positive_control,
108108
# portia, scenic, grnboost]"
109-
# append_entry "replogle" "[regression_1, regression_2, ws_distance]" "[pearson_corr, negative_control, positive_control, portia, ppcor, scenic, grnboost]"
110-
# append_entry "replogle" "[regression_1, regression_2, ws_distance]" "[scprint]" "special_case"
109+
# append_entry "replogle" "[regression_1, regression, ws_distance]" "[pearson_corr, negative_control, positive_control, portia, ppcor, scenic, grnboost]"
110+
# append_entry "replogle" "[regression_1, regression, ws_distance]" "[scprint]" "special_case"
111111

112-
append_entry "xaira_HCT116" "[regression_1, regression_2]" "[pearson_corr, negative_control, positive_control]"
112+
append_entry "xaira_HCT116" "[regression_1, regression]" "[pearson_corr, negative_control, positive_control]"
113113

114114
# --- Final configuration ---
115115
if [ "$run_local" = true ]; then

scripts/repo/run_benchmark_all_repo.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ apply_tf_methods=true
1313
apply_skeleton=false
1414
# - specify inputs
1515
dataset_ids=" op "
16-
metric_ids="[regression_1, regression_2, ws_distance]"
16+
metric_ids="[regression_1, regression, ws_distance]"
1717
method_ids="[pearson_corr,
1818
negative_control,
1919
positive_control,

scripts/repo/run_grn_evaluation copy.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ grn_models_folder="${resources_dir}/grn_models"
1414
subsample=-2
1515
max_workers=10
1616
layer=scgen_pearson
17-
metric_ids="[regression_1, regression_2]"
17+
metric_ids="[regression_1, regression]"
1818

1919
param_file="./params/${RUN_ID}.yaml"
2020

0 commit comments

Comments
 (0)