from maskrcnn_benchmark.modeling.language_backbone import build_tokenizer if __name__ == '__main__': tokenizer2 = build_tokenizer("clip") tokenized2 = tokenizer2( ["Detectest : fishid. jellyfishioasod. penguinasd. puffin.asd shark. starfish. round stingray"]) print(tokenized2)