You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+6-6Lines changed: 6 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -25,7 +25,7 @@ This repository is composed of the following elements:
25
25
*`PANG.py`: script that implements the Pang method.
26
26
*`ProcessingPattern.py`: script that computes the number of occurences and the set of induced patterns.
27
27
*`Pattern.sh`: script that computes the patterns of a dataset.
28
-
*`CORKcpp.zip`: archive containing the CORK source code (used in `EMCL.py`) cf. Section [Installation](#installation).
28
+
*`CORKcpp.zip`: archive containing the CORK source code (used in `EMCL.py`), cf. Section [Installation](#installation).
29
29
*`data`: folder containing the input data. Each subfolder corresponds to a distinct dataset, cf. Section [Datasets](#datasets).
30
30
*`results`: files produced by the processing.
31
31
@@ -47,7 +47,7 @@ Second, one of the dependencies, SPMF, is not a Python package, but rather a Jav
47
47
48
48
Note that SPMF is available both as a JAR and as source code archive. However, the former does not contain all the features required by Pang, so one should use only the latter.
49
49
50
-
In order to run the script that reproduces our ECML PKDD experiments, you also need to install CORK. This is done by unzipping the archive `CORKcpp.zip` in the `src` folder. The documentation for installing CORK is available on the Readme file in the `CORKcpp` folder.
50
+
In order to run the script that reproduces our ECML PKDD experiments, you also need to install CORK. This is done by unzipping the archive `CORKcpp.zip` in the `src` folder. File `Readme` in this archive contains the instruction for compiling the C++ source code.
51
51
52
52
53
53
## Data
@@ -115,9 +115,9 @@ Once the data are ready, you need to run a script to identify the patterns, and
115
115
1. Open the `Python` console.
116
116
2. Run the script `Patterns.sh` in order to create the files `XXX_patterns.txt`.
117
117
3. Run `ProcessingPattern.py`with the option `-d XXX` in order to create the files `XXX_mono.txt` and `XXX_iso.txt`.
118
-
4. Run `PANG.py`. 2 parameters are required:
119
-
*`-d XXX` : the name of the dataset
120
-
*`-k k` : the number of patterns to consider in the PANG representations. User can provide a single value, or a list of values separated by commas.
118
+
4. Run `PANG.py`, specifying both following parameters:
119
+
*`-d XXX`: name of the dataset
120
+
*`-k k`: number of patterns to consider in the PANG representations. User can provide a single value, or a list of values separated by commas.
121
121
122
122
For each value of the parameter `k`, Pang will create a file `KResults.txt` containing the results of the classification and a file `KPatterns.txt` containing the patterns.
123
123
@@ -142,7 +142,7 @@ For the ECML PKDD assessment, we use the following algorithms for the sake of co
142
142
* The `WL`[[S'11](#references)] and `WL_OA`[[K'16](#references)] algorithms are included in the `Grakel` library, documentation available [here](https://ysig.github.io/GraKeL/0.1a8/benchmarks.html)
143
143
*`Graph2Vec`[[N'17](#references)] is included in the `karateclub` library, documentation available [here](https://karateclub.readthedocs.io/en/latest/)
144
144
*`DGCNN`[[Z'18](#references)] is included in the `stellargraph` library, documentation available [here](https://stellargraph.readthedocs.io/en/stable/).
145
-
* We use the implementation of `CORK`[[T'09](#references)]from Marisa Thoma. This implementation is available in the `CORKcpp.zip` archive.
145
+
* We use the implementation of `CORK`[[T'09](#references)]by Marisa Thoma, which is available [here](http://www.dbs.ifi.lmu.de/~thoma/pub/sam2010/sam2010.zip).
0 commit comments