Pre training

#1
by vhh23 - opened

Is it possible to pre train this model on a new data ?

Databricks org

Yes. See the Github repo associated with this model: https://github.com/databrickslabs/dolly You can simply supply a different data set in the same format, to fine-tune the base model differently.

Should we use type as "classification" if we just want to give data as a pre training data ?

Databricks org

No, this is not a classification model. It is a text-generation model. See the repo for the full training script and more information.

vhh23 changed discussion status to closed

Is databricks/dolly-v2-12b multilingua?

Databricks org

@ammontenegrod no it is based on the Pythia model, which is pretrained on English text from The Pile.

https://huggingface.co/EleutherAI/pythia-12b

Databricks org

You might find it has some non-English tokens and works a little bit, from snippets of non-English language in the training data, but generally no.

Can we use this model for 'question-answering' task?

For ex:
task="question-answering"

model, tokenizer = load_model_tokenizer_for_generate(input_model='databricks/dolly-v2-3b')
llm = HuggingFacePipeline(
    pipeline=InstructionTextGenerationPipeline(
        # Return the full text, because this is what the HuggingFacePipeline expects.
        model=model, tokenizer=tokenizer, return_full_text=True, task="question-answering",
        torch_dtype=torch.bfloat16, max_new_tokens=512, top_p=0.95, top_k=50),
    )

This model is using AutoModelForCausalLM. Can we use AutoModelForQuestionAnswering when retraining the model?

Databricks org

(Please use new threads for new questions)
question-answering means extractive QA, and no it is not that type of model. You can answer questions, but not in the sense of that task.

Sign up or log in to comment