chrome_models / 19 /libtextclassifier3 /GrammarTokenizerOptions.py
dejanseo's picture
Upload 51 files
4c8c729 verified
# automatically generated by the FlatBuffers compiler, do not modify
# namespace: libtextclassifier3
import flatbuffers
from flatbuffers.compat import import_numpy
np = import_numpy()
class GrammarTokenizerOptions(object):
__slots__ = ['_tab']
@classmethod
def GetRootAsGrammarTokenizerOptions(cls, buf, offset):
n = flatbuffers.encode.Get(flatbuffers.packer.uoffset, buf, offset)
x = GrammarTokenizerOptions()
x.Init(buf, n + offset)
return x
@classmethod
def GrammarTokenizerOptionsBufferHasIdentifier(cls, buf, offset, size_prefixed=False):
return flatbuffers.util.BufferHasIdentifier(buf, offset, b"\x54\x43\x32\x20", size_prefixed=size_prefixed)
# GrammarTokenizerOptions
def Init(self, buf, pos):
self._tab = flatbuffers.table.Table(buf, pos)
# GrammarTokenizerOptions
def TokenizationType(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(4))
if o != 0:
return self._tab.Get(flatbuffers.number_types.Int32Flags, o + self._tab.Pos)
return 2
# GrammarTokenizerOptions
def IcuPreserveWhitespaceTokens(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(6))
if o != 0:
return bool(self._tab.Get(flatbuffers.number_types.BoolFlags, o + self._tab.Pos))
return False
# GrammarTokenizerOptions
def TokenizationCodepointConfig(self, j):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(8))
if o != 0:
x = self._tab.Vector(o)
x += flatbuffers.number_types.UOffsetTFlags.py_type(j) * 4
x = self._tab.Indirect(x)
from libtextclassifier3.TokenizationCodepointRange import TokenizationCodepointRange
obj = TokenizationCodepointRange()
obj.Init(self._tab.Bytes, x)
return obj
return None
# GrammarTokenizerOptions
def TokenizationCodepointConfigLength(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(8))
if o != 0:
return self._tab.VectorLen(o)
return 0
# GrammarTokenizerOptions
def TokenizationCodepointConfigIsNone(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(8))
return o == 0
# GrammarTokenizerOptions
def InternalTokenizerCodepointRanges(self, j):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(10))
if o != 0:
x = self._tab.Vector(o)
x += flatbuffers.number_types.UOffsetTFlags.py_type(j) * 4
x = self._tab.Indirect(x)
from libtextclassifier3.CodepointRange import CodepointRange
obj = CodepointRange()
obj.Init(self._tab.Bytes, x)
return obj
return None
# GrammarTokenizerOptions
def InternalTokenizerCodepointRangesLength(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(10))
if o != 0:
return self._tab.VectorLen(o)
return 0
# GrammarTokenizerOptions
def InternalTokenizerCodepointRangesIsNone(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(10))
return o == 0
# GrammarTokenizerOptions
def TokenizeOnScriptChange(self):
o = flatbuffers.number_types.UOffsetTFlags.py_type(self._tab.Offset(12))
if o != 0:
return bool(self._tab.Get(flatbuffers.number_types.BoolFlags, o + self._tab.Pos))
return False
def GrammarTokenizerOptionsStart(builder): builder.StartObject(5)
def GrammarTokenizerOptionsAddTokenizationType(builder, tokenizationType): builder.PrependInt32Slot(0, tokenizationType, 2)
def GrammarTokenizerOptionsAddIcuPreserveWhitespaceTokens(builder, icuPreserveWhitespaceTokens): builder.PrependBoolSlot(1, icuPreserveWhitespaceTokens, 0)
def GrammarTokenizerOptionsAddTokenizationCodepointConfig(builder, tokenizationCodepointConfig): builder.PrependUOffsetTRelativeSlot(2, flatbuffers.number_types.UOffsetTFlags.py_type(tokenizationCodepointConfig), 0)
def GrammarTokenizerOptionsStartTokenizationCodepointConfigVector(builder, numElems): return builder.StartVector(4, numElems, 4)
def GrammarTokenizerOptionsAddInternalTokenizerCodepointRanges(builder, internalTokenizerCodepointRanges): builder.PrependUOffsetTRelativeSlot(3, flatbuffers.number_types.UOffsetTFlags.py_type(internalTokenizerCodepointRanges), 0)
def GrammarTokenizerOptionsStartInternalTokenizerCodepointRangesVector(builder, numElems): return builder.StartVector(4, numElems, 4)
def GrammarTokenizerOptionsAddTokenizeOnScriptChange(builder, tokenizeOnScriptChange): builder.PrependBoolSlot(4, tokenizeOnScriptChange, 0)
def GrammarTokenizerOptionsEnd(builder): return builder.EndObject()