@@ -23,7 +23,7 @@ def testPlaintextTokenizer(self):
2323 output_file = Path (temp_dir , "output.json" )
2424 expected_output_file = Path (test_data_dir , "tokenizer" , "plaintext" , "expected_output" , "output.json" )
2525
26- subprocess .check_call (f"{ str (Path (lichen_installation_dir , 'bin' , 'plaintext_tokenizer.out ' ))} < { str (input_file )} > { str (output_file )} " , shell = True )
26+ subprocess .check_call (f"python3 { str (Path (lichen_installation_dir , 'bin' , 'plaintext_tokenizer.py ' ))} { str (input_file )} > { str (output_file )} " , shell = True )
2727
2828 with open (output_file ) as file :
2929 actual_output = json .load (file )
@@ -41,7 +41,7 @@ def testPlaintextTokenizerIgnorePunctuation(self):
4141 output_file = Path (temp_dir , "output.json" )
4242 expected_output_file = Path (test_data_dir , "tokenizer" , "plaintext" , "expected_output" , "output_ignore_punctuation.json" )
4343
44- subprocess .check_call (f"{ str (Path (lichen_installation_dir , 'bin' , 'plaintext_tokenizer.out ' ))} --ignore_punctuation < { str (input_file )} > { str (output_file )} " , shell = True )
44+ subprocess .check_call (f"python3 { str (Path (lichen_installation_dir , 'bin' , 'plaintext_tokenizer.py ' ))} --ignore_punctuation { str (input_file )} > { str (output_file )} " , shell = True )
4545
4646 with open (output_file ) as file :
4747 actual_output = json .load (file )
@@ -59,7 +59,7 @@ def testPlaintextTokenizerToLower(self):
5959 output_file = Path (temp_dir , "output.json" )
6060 expected_output_file = Path (test_data_dir , "tokenizer" , "plaintext" , "expected_output" , "output_to_lower.json" )
6161
62- subprocess .check_call (f"{ str (Path (lichen_installation_dir , 'bin' , 'plaintext_tokenizer.out ' ))} --to_lower < { str (input_file )} > { str (output_file )} " , shell = True )
62+ subprocess .check_call (f"python3 { str (Path (lichen_installation_dir , 'bin' , 'plaintext_tokenizer.py ' ))} --to_lower { str (input_file )} > { str (output_file )} " , shell = True )
6363
6464 with open (output_file ) as file :
6565 actual_output = json .load (file )
@@ -77,7 +77,7 @@ def testPlaintextTokenizerIgnoreNewlines(self):
7777 output_file = Path (temp_dir , "output.json" )
7878 expected_output_file = Path (test_data_dir , "tokenizer" , "plaintext" , "expected_output" , "output_ignore_newlines.json" )
7979
80- subprocess .check_call (f"{ str (Path (lichen_installation_dir , 'bin' , 'plaintext_tokenizer.out ' ))} --ignore_newlines < { str (input_file )} > { str (output_file )} " , shell = True )
80+ subprocess .check_call (f"python3 { str (Path (lichen_installation_dir , 'bin' , 'plaintext_tokenizer.py ' ))} --ignore_newlines { str (input_file )} > { str (output_file )} " , shell = True )
8181
8282 with open (output_file ) as file :
8383 actual_output = json .load (file )
@@ -95,7 +95,7 @@ def testPlaintextTokenizerIgnoreEverything(self):
9595 output_file = Path (temp_dir , "output.json" )
9696 expected_output_file = Path (test_data_dir , "tokenizer" , "plaintext" , "expected_output" , "output_ignore_everything.json" )
9797
98- subprocess .check_call (f"{ str (Path (lichen_installation_dir , 'bin' , 'plaintext_tokenizer.out ' ))} --ignore_punctuation --to_lower --ignore_numbers --ignore_newlines < { str (input_file )} > { str (output_file )} " , shell = True )
98+ subprocess .check_call (f"python3 { str (Path (lichen_installation_dir , 'bin' , 'plaintext_tokenizer.py ' ))} --ignore_punctuation --to_lower --ignore_numbers --ignore_newlines { str (input_file )} > { str (output_file )} " , shell = True )
9999
100100 with open (output_file ) as file :
101101 actual_output = json .load (file )
0 commit comments