# from transformers import (
#     BertTokenizer,
#     BertLMHeadModel,
#     GPT2Tokenizer,
#     GPT2LMHeadModel,
#     GPTNeoForCausalLM,
#     AutoTokenizer,
#     GPTJForCausalLM,
#     LlamaTokenizer,
#     LlamaForCausalLM,
#     LlamaTokenizerFast
# )
# import torch
from .knowledge_neurons import KnowledgeNeurons, initialize_model_and_tokenizer, model_type
# from .data import pararel, pararel_expanded, PARAREL_RELATION_NAMES

# BERT_MODELS = ["bert-base-uncased", "bert-base-multilingual-uncased"]
# GPT2_MODELS = ["gpt2"]
# GPT_NEO_MODELS = [
#     "EleutherAI/gpt-neo-125M",
#     "EleutherAI/gpt-neo-1.3B",
#     "EleutherAI/gpt-neo-2.7B",
# ]
# ALL_MODELS = BERT_MODELS + GPT2_MODELS + GPT_NEO_MODELS

