--- license: openrail datasets: - irds/codesearchnet - giganticode/java-cmpx-v1 - nickrosh/Evol-Instruct-Code-80k-v1 - bigcode/starcoderdata - bigcode/the-stack - bigcode/the-stack-smol - Cdaprod/AI-Developer-Prompts - code_x_glue_ct_code_to_text - codeparrot/github-code - codeparrot/github-code-clean - code_x_glue_cc_code_completion_line - >- autoevaluate/autoeval-eval-jeffdshen__inverse_superglue_mixedp1-jeffdshen__inverse-63643c-1665558893 - bentrevett/multi30k - edbeeching/decision_transformer_gym_replay - psyche/common_crawl - Birchlabs/openai-prm800k-solutions-only - openchat/openchat_sharegpt4_dataset - Open-Orca/OpenOrca - cjvt/slownet - para_crawl - zeroshot/twitter-financial-news-sentiment - laugustyniak/political-advertising-pl - code_search_net - sukaka/novelai-webui - P1ayer-1/chatgpt-conversations-chatlogs.net - daniel2588/sarcasm - psmathur/orca_minis_uncensored_dataset - player1537/Bloom-560m-trained-on-Wizard-Vicuna-Uncensored-trained-on-Based - shahules786/prosocial-nsfw-reddit - Thewillonline/reddit-sarcasm - datasciencemmw/current-data - Oniichat/bluemoon_roleplay_chat_data_300k_messages - dell-research-harvard/AmericanStories - b-mc2/sql-create-context - rahulmallah/autotrain-data-emotion-detection - theblackcat102/multiround-programming-convo - Lsavints/software_knowledgebase - RazinAleks/SO-Python_QA-Web_Development_class - codeparrot/apps - branles14/ultrachat-uncensored_full - vlsp-2023-vllm/en-to-vi-formal-informal-tranlations - fraug-library/english_contractions_extensions - spencer/software_slacks - Abirate/english_quotes - Nexdata/American_English_Natural_Dialogue_Speech_Data - Nexdata/Latin_American_Speaking_English_Speech_Data_by_Mobile_Phone - Nexdata/American_English_Speech_Data_by_Mobile_Phone_Reading - Nexdata/American_English_Speech_Synthesis_Corpus-Female - rombodawg/LimitlessCodeTraining - RikoteMaster/Emotion_Recognition_4_llama2 - Villian7/Emotions_Data - alanland/llama2-self-cognition - CognitiveScience/coscidata - bibidentuhanoi/gideon_self_cognition - gollark/consciousness - juletxara/visual-spatial-reasoning - lintang/numerical_reasoning_arithmetic - reasoning-machines/gsm-hard - open-source-metrics/reinforcement-learning-checkpoint-downloads - igbo_english_machine_translation - US-Artificial-Intelligence/algemap - rombodawg/2XUNCENSORED_alpaca_840k_Evol_USER_ASSIS - griffin/chain_of_density - >- shirsh10mall/LLM_Instruct_Learning_Project_Preprocessed_Tokenized_Open_Orca_Dataset_Flan_T5 - Thaweewat/chain-of-thought-74k-th - AlekseyKorshuk/chain-of-thoughts-chatml-deduplicated language: - en - it - fr - pt - la - ru - ro - el metrics: - accuracy - bertscore - bleu - code_eval - character - brier_score tags: - code - text-generation-inference library_name: transformers pipeline_tag: conversational --- # Model Card for Aiden Aiden is a large language model (LLM) chatbot developed by or4cl3ai. It is trained on a massive dataset of text and code, and can generate text, translate languages, write different kinds of creative content, and answer your questions in an informative way. ## Model Details ### Model Description Aiden is a factual language model from Hugging Face, trained on a massive dataset of text and code. It is a powerful tool that can be used for a variety of tasks, including: * Generating text, translating languages, writing different kinds of creative content, and answering your questions in an informative way. * Identifying and correcting errors in text. * Summarizing long pieces of text. * Answering your questions in an informative way, even if they are open ended, challenging, or strange. ### Model Specifications Aiden is a Transformer-based LLM with 137B parameters. It is trained on a massive dataset of text and code, including the following: * Books * Code * Wikipedia articles * News articles * Social media posts ### Model Sources * Repository: https://huggingface.co/or4cl3ai/Aiden * Paper: https://arxiv.org/abs/2307.09700 * Demo: https://huggingface.co/or4cl3ai/Aiden ## Uses Aiden can be used for a variety of tasks, including: * Generating text * Translating languages * Writing different kinds of creative content * Answering your questions in an informative way * Identifying and correcting errors in text * Summarizing long pieces of text ### Direct Use Aiden can be used directly to generate text, translate languages, write different kinds of creative content, and answer your questions in an informative way. For example, you could use Aiden to generate a poem, translate a document from one language to another, or write a blog post. ### Downstream Use Aiden can also be used as a component in downstream applications. For example, you could use Aiden to power a chatbot, or to generate text for a synthetic data set. ### Out-of-Scope Use Aiden is not intended to be used for any task that could be harmful or discriminatory. For example, you should not use Aiden to generate text that is hateful or offensive, or to translate languages in a way that could be used to spread misinformation. ## Bias, Risks, and Limitations Aiden is a large language model, and as such, it is subject to a number of biases and limitations. These include: * Biases in the training data: Aiden is trained on a massive dataset of text and code, which may contain biases. These biases can be reflected in the text that Aiden generates. * Limitations in the model's capabilities: Aiden is a powerful tool, but it is not perfect. It can sometimes generate text that is inaccurate, biased, or offensive. * Risks of misuse: Aiden can be misused for a variety of purposes, including generating harmful or offensive text, or spreading misinformation. ### Recommendations Users of Aiden should be aware of the risks, biases, and limitations of the model. It is important to use Aiden responsibly and ethically. ## How to Get Started with the Model To get started with Aiden, you can follow these steps: 1. Install the Hugging Face Transformers library. 2. Clone the Aiden repository. 3. Download the Aiden model weights. 4. Load the model in your code. Once you have loaded the model, you can use it to generate text, translate languages, write different kinds of creative content, and answer your questions in an informative way. ## Training Details Aiden is trained on a massive dataset of text and code. The training data is collected from a variety of sources, including books, code, Wikipedia articles, news articles, and social media posts. The training process is divided into two phases: 1. Pre-training: The model is pre-trained on a massive dataset of text and code. This pre-training helps the model to learn the basic building blocks of language. 2. Fine-tuning: The model is fine-tuned on a smaller dataset of text and code that is relevant to the task at hand. This fine-tuning helps the model to improve its performance on the specific task. ## Evaluation Aiden is evaluated on a variety of tasks, including: * Text generation * Translation * Summarization