Skip to content

Commit 585b92a

Browse files
committed
Remove unnessasary step from README.md
1 parent b2a0fac commit 585b92a

File tree

1 file changed

+3
-15
lines changed

1 file changed

+3
-15
lines changed

README.md

Lines changed: 3 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -126,19 +126,7 @@ less ${KOMP_PATH}/impc_statistical_pipeline/IMPC_DRs/stats_pipeline_logs/stats_p
126126
```
127127

128128
## Outdated Steps
129-
### Step 2. Run the Report Generating Pipeline
130-
This process generates statistical reports typically utilized by the IMPC working groups.
131-
1. Navigate to `${KOMP_PATH}/impc_statistical_pipeline/IMPC_DRs/stats_pipeline_input_drXX.y/SP/jobs/Results_IMPC_SP_Windowed`
132-
2. Allocate a high memory machine on cluster and initialise an interactive shell:
133-
`bsub –M 300000 –e errReportGeneratingPipeline –o outReportGeneratingPipeline –Is /bin/bash`
134-
3. The commands below will generate two CSV files in the `${KOMP_PATH}/impc_statistical_pipeline/IMPC_DRs/stats_pipeline_input_drXX.y/SP/jobs/Results_IMPC_SP_Windowed` directory for the unidimentional and categorical results. The files can be gzip and moved to the FTP directory. You can decorate and format the files by using one of the formatted files in the previous data releases.
135-
```console
136-
R
137-
DRrequiredAgeing:::IMPC_statspipelinePostProcess(mp_chooser_file=${MP_CHOOSER_FILE})
138-
DRrequiredAgeing:::ClearReportsAfterCreation()
139-
```
140-
141-
### Step 3. Run the Extraction of Risky Genes Pipeline
129+
### Step 2. Run the Extraction of Risky Genes Pipeline
142130
This process generates a list of risky genes to check manually.
143131
1. Allocate a machine on codon cluster: `bsub –M 8000 –Is /bin/bash`
144132
2. Open an R session: `R`
@@ -151,7 +139,7 @@ This process generates a list of risky genes to check manually.
151139
When there are no jobs running under the mi_stats user and all jobs are successfully completed.<br><br>
152140
- ***Do you expect any errors from the stats pipeline?***<br>
153141
Yes, having a few errors is normal. If you observe more than a few errors, you may want to run the GapFilling pipeline. Refer to the Step_1.2_RunGapFillingPipeline.mp4 [video](https://www.ebi.ac.uk/seqdb/confluence/display/MouseInformatics/How+to+run+the+IMPC+statistical+pipeline). Make sure to log in to Confluence first.<br><br>
154-
## Step 1 FAQ
142+
## Statistical Pipeline FAQ
155143
- ***How can you determine on step 1 if the pipeline is still running?***<br>
156144
The simplest method is to execute the `squeue` command. During the first 4 days of running the pipeline, there should be less than 20 jobs running. Otherwise, there should be 5000+ jobs running on the codon cluster.<br><br>
157145
- ***How to determine if step 1 is finished?***<br>
@@ -167,7 +155,7 @@ find ./*/*_RawData/ClusterErr/ -name *ClusterErr -type f | xargs cp --backup=num
167155
- ***When should you run the gap filling pipeline after completing step 1?***<br>
168156
In very rare cases, the stats pipeline may fail for unknown reasons.To resume the pipeline from the point of failure, you can use the GapFilling Pipeline. This is equivalent to running the pipeline by navigating to `<stats pipeline directory>/SP/jobs` and executing `AllJobs.bch`. Before doing so, make sure to edit function.R and set the parameter `onlyFillNonExistingResults` to TRUE. After making this change, run the pipeline by executing `./AllJobs.bch` and wait for the pipeline to fill in the missing analyses. Please note that this process may take up to 2 days.<br><br>
169157

170-
## Step 2 Annotation Pipeline FAQ
158+
## Annotation Pipeline FAQ
171159
The `IMPC_HadoopLoad` command uses the power of cluster to assign the annotations to the StatPackets and transfers the files to the Hadoop cluster (transfer=TRUE). The files will be transferred to `Hadoop:/hadoop/user/mi_stats/impc/statpackets/DRXX`.
172160
**Note**: we run annotation pipeline with transfer=FALSE, so we don't transfer it now.
173161

0 commit comments

Comments
 (0)