Skip to content

Commit 7c4575e

Browse files
committed
update stub for typing
1 parent c47328c commit 7c4575e

File tree

9 files changed

+1284
-46
lines changed

9 files changed

+1284
-46
lines changed

bindings/python/py_src/tokenizers/__init__.pyi

Lines changed: 318 additions & 1 deletion
Large diffs are not rendered by default.

bindings/python/py_src/tokenizers/decoders/__init__.pyi

Lines changed: 120 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -52,6 +52,16 @@ class BPEDecoder(Decoder):
5252
"""
5353
pass
5454

55+
@property
56+
def suffix(self):
57+
""" """
58+
pass
59+
60+
@suffix.setter
61+
def suffix(self, value):
62+
""" """
63+
pass
64+
5565
class ByteFallback(Decoder):
5666
"""
5767
ByteFallback Decoder
@@ -115,6 +125,16 @@ class CTC(Decoder):
115125
def __init__(self, pad_token="<pad>", word_delimiter_token="|", cleanup=True):
116126
pass
117127

128+
@property
129+
def cleanup(self):
130+
""" """
131+
pass
132+
133+
@cleanup.setter
134+
def cleanup(self, value):
135+
""" """
136+
pass
137+
118138
def decode(self, tokens):
119139
"""
120140
Decode the given list of tokens to a final string
@@ -128,6 +148,26 @@ class CTC(Decoder):
128148
"""
129149
pass
130150

151+
@property
152+
def pad_token(self):
153+
""" """
154+
pass
155+
156+
@pad_token.setter
157+
def pad_token(self, value):
158+
""" """
159+
pass
160+
161+
@property
162+
def word_delimiter_token(self):
163+
""" """
164+
pass
165+
166+
@word_delimiter_token.setter
167+
def word_delimiter_token(self, value):
168+
""" """
169+
pass
170+
131171
class Fuse(Decoder):
132172
"""
133173
Fuse Decoder
@@ -182,6 +222,36 @@ class Metaspace(Decoder):
182222
"""
183223
pass
184224

225+
@property
226+
def prepend_scheme(self):
227+
""" """
228+
pass
229+
230+
@prepend_scheme.setter
231+
def prepend_scheme(self, value):
232+
""" """
233+
pass
234+
235+
@property
236+
def replacement(self):
237+
""" """
238+
pass
239+
240+
@replacement.setter
241+
def replacement(self, value):
242+
""" """
243+
pass
244+
245+
@property
246+
def split(self):
247+
""" """
248+
pass
249+
250+
@split.setter
251+
def split(self, value):
252+
""" """
253+
pass
254+
185255
class Replace(Decoder):
186256
"""
187257
Replace Decoder
@@ -237,6 +307,16 @@ class Strip(Decoder):
237307
def __init__(self, content, left=0, right=0):
238308
pass
239309

310+
@property
311+
def content(self):
312+
""" """
313+
pass
314+
315+
@content.setter
316+
def content(self, value):
317+
""" """
318+
pass
319+
240320
def decode(self, tokens):
241321
"""
242322
Decode the given list of tokens to a final string
@@ -250,6 +330,26 @@ class Strip(Decoder):
250330
"""
251331
pass
252332

333+
@property
334+
def start(self):
335+
""" """
336+
pass
337+
338+
@start.setter
339+
def start(self, value):
340+
""" """
341+
pass
342+
343+
@property
344+
def stop(self):
345+
""" """
346+
pass
347+
348+
@stop.setter
349+
def stop(self, value):
350+
""" """
351+
pass
352+
253353
class WordPiece(Decoder):
254354
"""
255355
WordPiece Decoder
@@ -265,6 +365,16 @@ class WordPiece(Decoder):
265365
def __init__(self, prefix="##", cleanup=True):
266366
pass
267367

368+
@property
369+
def cleanup(self):
370+
""" """
371+
pass
372+
373+
@cleanup.setter
374+
def cleanup(self, value):
375+
""" """
376+
pass
377+
268378
def decode(self, tokens):
269379
"""
270380
Decode the given list of tokens to a final string
@@ -277,3 +387,13 @@ class WordPiece(Decoder):
277387
:obj:`str`: The decoded string
278388
"""
279389
pass
390+
391+
@property
392+
def prefix(self):
393+
""" """
394+
pass
395+
396+
@prefix.setter
397+
def prefix(self, value):
398+
""" """
399+
pass

bindings/python/py_src/tokenizers/models/__init__.pyi

Lines changed: 110 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -131,6 +131,46 @@ class BPE(Model):
131131
):
132132
pass
133133

134+
@property
135+
def byte_fallback(self):
136+
""" """
137+
pass
138+
139+
@byte_fallback.setter
140+
def byte_fallback(self, value):
141+
""" """
142+
pass
143+
144+
@property
145+
def continuing_subword_prefix(self):
146+
""" """
147+
pass
148+
149+
@continuing_subword_prefix.setter
150+
def continuing_subword_prefix(self, value):
151+
""" """
152+
pass
153+
154+
@property
155+
def dropout(self):
156+
""" """
157+
pass
158+
159+
@dropout.setter
160+
def dropout(self, value):
161+
""" """
162+
pass
163+
164+
@property
165+
def end_of_word_suffix(self):
166+
""" """
167+
pass
168+
169+
@end_of_word_suffix.setter
170+
def end_of_word_suffix(self, value):
171+
""" """
172+
pass
173+
134174
@staticmethod
135175
def from_file(cls, vocab, merge, **kwargs):
136176
"""
@@ -157,6 +197,16 @@ class BPE(Model):
157197
"""
158198
pass
159199

200+
@property
201+
def fuse_unk(self):
202+
""" """
203+
pass
204+
205+
@fuse_unk.setter
206+
def fuse_unk(self, value):
207+
""" """
208+
pass
209+
160210
def get_trainer(self):
161211
"""
162212
Get the associated :class:`~tokenizers.trainers.Trainer`
@@ -182,6 +232,16 @@ class BPE(Model):
182232
"""
183233
pass
184234

235+
@property
236+
def ignore_merges(self):
237+
""" """
238+
pass
239+
240+
@ignore_merges.setter
241+
def ignore_merges(self, value):
242+
""" """
243+
pass
244+
185245
@staticmethod
186246
def read_file(self, vocab, merges):
187247
"""
@@ -250,6 +310,16 @@ class BPE(Model):
250310
"""
251311
pass
252312

313+
@property
314+
def unk_token(self):
315+
""" """
316+
pass
317+
318+
@unk_token.setter
319+
def unk_token(self, value):
320+
""" """
321+
pass
322+
253323
class Unigram(Model):
254324
"""
255325
An implementation of the Unigram algorithm
@@ -460,6 +530,16 @@ class WordLevel(Model):
460530
"""
461531
pass
462532

533+
@property
534+
def unk_token(self):
535+
""" """
536+
pass
537+
538+
@unk_token.setter
539+
def unk_token(self, value):
540+
""" """
541+
pass
542+
463543
class WordPiece(Model):
464544
"""
465545
An implementation of the WordPiece algorithm
@@ -477,6 +557,16 @@ class WordPiece(Model):
477557
def __init__(self, vocab, unk_token, max_input_chars_per_word):
478558
pass
479559

560+
@property
561+
def continuing_subword_prefix(self):
562+
""" """
563+
pass
564+
565+
@continuing_subword_prefix.setter
566+
def continuing_subword_prefix(self, value):
567+
""" """
568+
pass
569+
480570
@staticmethod
481571
def from_file(vocab, **kwargs):
482572
"""
@@ -525,6 +615,16 @@ class WordPiece(Model):
525615
"""
526616
pass
527617

618+
@property
619+
def max_input_chars_per_word(self):
620+
""" """
621+
pass
622+
623+
@max_input_chars_per_word.setter
624+
def max_input_chars_per_word(self, value):
625+
""" """
626+
pass
627+
528628
@staticmethod
529629
def read_file(vocab):
530630
"""
@@ -589,3 +689,13 @@ class WordPiece(Model):
589689
A :obj:`List` of :class:`~tokenizers.Token`: The generated tokens
590690
"""
591691
pass
692+
693+
@property
694+
def unk_token(self):
695+
""" """
696+
pass
697+
698+
@unk_token.setter
699+
def unk_token(self, value):
700+
""" """
701+
pass

0 commit comments

Comments
 (0)