chromium/third_party/tflite_support/src/tensorflow_lite_support/metadata/python/tests/testdata/sentence_piece_tokenizer_meta.json

{
  "subgraph_metadata": [
    {
      "input_process_units": [
        {
          "options_type": "SentencePieceTokenizerOptions",
          "options": {
            "sentencePiece_model": [
              {
                "name": "sp.model",
                "description": "The sentence piece model file."
              }
            ],
            "vocab_file": [
              {
                "name": "vocab.txt",
                "description": "Vocabulary file to convert natural language words to embedding vectors. This file is optional during tokenization, while the sentence piece model is mandatory.",
                "type": "VOCABULARY"
              }
            ]
          }
        }
      ]
    }
  ]
}