Skip to content
This repository was archived by the owner on Feb 9, 2025. It is now read-only.

Commit 5857fc5

Browse files
authored
clarify it is "your" root
1 parent 370f839 commit 5857fc5

1 file changed

Lines changed: 1 addition & 1 deletion

File tree

Lab 2 - RDD, DataFrame, ML pipeline, and parallelization.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ Firstly, we follow the standard steps as in Task 2 of Lab 1 but with some variat
3939

4040
```sh
4141
qrshx -P rse-com6012 -pe smp 4 # request 4 CPU cores using our reserved queue
42-
source myspark.sh # assuming HPC/myspark.sh is under the root directory, otherwise, see Lab 1 Task 2
42+
source myspark.sh # assuming HPC/myspark.sh is under your root directory, otherwise, see Lab 1 Task 2
4343
conda install -y numpy # install numpy, to be used in Task 3. This ONLY needs to be done ONCE. NOT every time.
4444
cd com6012/ScalableML # our main working directory
4545
pyspark --master local[4] # start pyspark with 4 cores requested above.

0 commit comments

Comments
 (0)