//- 💫 DOCS > USAGE > MODELS > LANGUAGE SUPPORT

p spaCy currently provides models for the following languages:

+table(["Language", "Code", "Language data", "Models"])
    for models, code in MODELS
        - var count = Object.keys(models).length
        +row
            +cell=LANGUAGES[code]
            +cell #[code=code]
            +cell
                +src(gh("spaCy", "spacy/lang/" + code)) #[code lang/#{code}]
            +cell
                +a("/models/" + code) #{count} #{(count == 1) ? "model" : "models"}

+h(3, "alpha-support") Alpha tokenization support

p
    |  Work has started on the following languages. You can help by
    |  #[+a("/usage/adding-languages#language-data") improving the existing language data]
    |  and extending the tokenization patterns.

+aside("Usage note")
    |  Note that the alpha languages don't yet come with a language model. In
    |  order to use them, you have to import them directly, or use
    |  #[+api("spacy#blank") #[code spacy.blank]]:

    +code.o-no-block.
        from spacy.lang.fi import Finnish
        nlp = Finnish()  # use directly
        nlp = spacy.blank('fi')  # blank instance

+table(["Language", "Code", "Language data"])
    for lang, code in LANGUAGES
        if !Object.keys(MODELS).includes(code)
            +row
                +cell #{LANGUAGES[code]}
                +cell #[code=code]
                +cell
                    +src(gh("spaCy", "spacy/lang/" + code)) #[code lang/#{code}]

+infobox("Dependencies")
    .o-block-small Some language tokenizers require external dependencies.

    +list.o-no-block
        +item #[strong Chinese]: #[+a("https://github.com/fxsjy/jieba") Jieba]
        +item #[strong Japanese]: #[+a("https://github.com/mocobeta/janome") Janome]
        +item #[strong Thai]: #[+a("https://github.com/wannaphongcom/pythainlp") pythainlp]
        +item #[strong Russian]: #[+a("https://github.com/kmike/pymorphy2") pymorphy2]

+h(3, "multi-language") Multi-language support
    +tag-new(2)

p
    |  As of v2.0, spaCy supports models trained on more than one language. This
    |  is especially useful for named entity recognition. The language ID used
    |  for multi-language or language-neutral models is #[code xx]. The
    |  language class, a generic subclass containing only the base language data,
    |  can be found in #[+src(gh("spaCy", "spacy/lang/xx")) #[code lang/xx]].

p
    |  To load your model with the neutral, multi-language class, simply set
    |  #[code "language": "xx"] in your
    |  #[+a("/usage/training#models-generating") model package]'s
    |  meta.json. You can also import the class directly, or call
    |  #[+api("util#get_lang_class") #[code util.get_lang_class()]] for
    |  lazy-loading.

+code("Standard import").
    from spacy.lang.xx import MultiLanguage
    nlp = MultiLanguage()

+code("With lazy-loading").
    from spacy.util import get_lang_class
    nlp = get_lang_class('xx')
