V41

WeMake πŸ’™ Llama-3 8B V41 Instruct 1048k

Welcome to the official repository for WEMAKE-CX/Llama-3-8B-Instruct-V41-1048k, WeMake's pioneering 1 Million Token Large Language Model (LLM). This model represents a significant milestone in the evolution of natural language understanding and generation, combining the robust foundation of Meta's Llama-3 architecture with the nuanced alignment and emotional intelligence of WeMake's V41.

Overview

WEMAKE-CX/Llama-3-8B-Instruct-V41-1048k is a state-of-the-art language model designed to understand and generate human-like text with an unprecedented level of emotional intelligence and alignment. This model is a fork of both "gradientai/Llama-3-8B-Instruct-Gradient-1048k" and "meta-llama/Meta-Llama-3-8B", enhanced with the unique capabilities of WeMake's V41 and trained using the proprietary WeMake ICU method.

Our model is engineered to serve a wide array of applications, from advanced conversational agents and content creation tools to sophisticated data analysis and insight generation platforms. It embodies WeMake's commitment to pushing the boundaries of AI to create more empathetic, understanding, and useful technologies.

Key Features

  • Emotional Intelligence: Integrates WeMake's V41 emotional intelligence, enabling the model to understand and generate responses that consider emotional context and nuances.
  • Alignment with Human Values: Trained using the WeMake ICU method, ensuring the model's outputs are aligned with ethical standards and human values.
  • Extensive Knowledge Base: Leverages a vast dataset, encompassing a wide range of topics, to provide accurate and contextually relevant responses.
  • Highly Configurable: Offers extensive customization options to cater to specific application requirements, including adjustable generation settings and fine-tuning capabilities.
  • Multilingual Support: Capable of understanding and generating text in multiple languages, making it a versatile tool for global applications.

Model Specifications

  • Model Path: WEMAKE-CX/Llama-3-8B-Instruct-V41-1048k
  • Architecture: LlamaForCausalLM
  • Hidden Size: 4096
  • Number of Attention Heads: 32
  • Number of Hidden Layers: 32
  • Max Position Embeddings: 1048576
  • Vocabulary Size: 128256
  • Torch Data Type: bfloat16

License

WEMAKE-CX/Llama-3-8B-Instruct-V41-1048k is distributed under the "llama3" license. For more details, please refer to the LICENSE file in this repository.

Contributing

We welcome contributions from the community, including bug reports, feature requests, and code contributions. Please refer to the CONTRIBUTING.md file for more information on how to get involved.

Acknowledgments

This model is built upon the foundational work of Meta's Llama-3 and the enhancements made by Gradient's Llama-3-8B-Instruct-Gradient-1048k. We extend our gratitude to the researchers and developers behind these projects for their contributions to the field of AI.

Contact

For any inquiries, please contact us at hey@wemake.cx.

Join us in exploring the possibilities of emotionally intelligent and ethically aligned AI with WEMAKE-CX/Llama-3-8B-Instruct-V41-1048k. Together, let's shape the future of human-AI interaction.

Downloads last month
11
Safetensors
Model size
8.03B params
Tensor type
BF16
Β·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for WeMake/Llama-3-8B-Instruct-V41-1048k

Quantizations
2 models

Dataset used to train WeMake/Llama-3-8B-Instruct-V41-1048k