Pragna-1b

pragna-1b on huggingface

Architecture Overview

Pragna-1B is a decoder-only transformer model inspired by TinyLlama, featuring the following specifications:

  • Layers: 22
  • Attention Heads: 32
  • Context Length: 2048
  • Hidden Dimension: 2048
  • Expansion Dimension: 5632
  • Vocabulary Size: 69632

This model incorporates Rotary Positional Encoding to infuse positional information into the embeddings, utilising a base of 10,000. It employs RSNorm with an epsilon value of 1e-5 and the Sigmoid Activation Unit (SiLU) as the activation function. Additionally, Pragna-1B adopts Grouped Query Attention, an alternative to Multi-Head Attention, which enhances training and inference speed while reducing memory bandwidth. This also supports the use of lower-compute devices for inference tasks.

Pragna-1B is trained on our proprietary platform, GenAI Studio, a modular AI Developer Platform designed to support any GenAI model architecture. It is capable of scaling across thousands of GPUs or accelerators and is built to be fault-tolerant. The development of this model leveraged Triton, an open-source language from OpenAI, for crafting high-performance custom fused CUDA Kernels for various operations. Furthermore, the model uses Fully Sharded Data Parallel (FSDP) for distributed and parallel training and incorporates the state-of-the-art FlashAttention2 to accelerate training and inference.

Model Description

  • Developed by: Soket AI Labs
  • Language(s) (NLP): Hindi, Bangla, Gujarati and English
  • License: Apache 2.0

Bias, Risks, and Limitations

[More Information Needed]

How to Get Started with the Model

Use the code below to get started with the model.

from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("soketlabs/pragna-1b")
model = AutoModelForCausalLM.from_pretrained(
    "soketlabs/pragna-1b", torch_dtype=torch.bfloat16
)

Training Details

Training Data

  1. Bhasha-wiki
  2. SlimPajama
  3. Sangraha-Verified

Training Procedure

[To be added]

Training Hyperparameters

  • Precision: BFloat16
  • Batch Size: 2k - 2.5k
  • Context Length: 2,048
  • Learning Rate: 3e-5
  • Optimizer: AdamW
  • LR Scheduler: Cosine
  • Mixed Precision Training

Evaluation

Hindi

Arc-Easy Arc-Challenge Hellaswag Average
pragna-1b 0.33 0.22 0.35 0.30
sarvamai/OpenHathi-7B-Hi-v0.1-Base 0.3582 0.2645 0.4315 0.35
meta-llama/Llama-2-7b-hf 0.295 0.2406 0.3789 0.30
google/gemma-7b 0.5926 0.4258 0.6341 0.55
meta-llama/Meta-Llama-3-8B 0.5354 0.3541 0.6072 0.50

Gujarati

Arc-Easy Arc-Challenge Hellaswag Average
pragna-1b 0.32 0.22 0.37 0.30
google/gemma-7b 0.4954 0.3208 0.5673 0.46

English

Arc Hellaswag MMLU TruthfulQA Winogrande GSM8K Average
pragna-1b 0.3 0.51 0.27 0.38 0.56 0 0.34
TinyLlama/TinyLlama-1.1B-Chat-v1.0 0.36 0.61 0.25 0.37 0.61 0.02 0.37
meta-llama/Meta-Llama-3-8B-Instruct 0.6 0.82 0.67 0.51 0.77 0.68 0.66
meta-llama/Llama-2-7b-hf 0.53 0.78 0.46 0.39 0.74 0.14 0.51

Results

Context Tokens Predicted
India is one of the most India is one of the most exciting places to travel to in Asia. The country hosts a wide variety of cultures and offers a plethora of attractions across the country. There's a lot you can do here and the tourists are always on the lookout for the new things to do and see. If you're looking to travel to India for your next adventure, you'll be pleased to know that you can find a ton of things to do in the country. From the temples to the beaches, there's a lot you can see and do here. One of the best things for visitors to do in India is to visit Jaipur. The capital city of Rajputana, Jaipur is a beautiful city that's known as the "City of Lives".
नरेंद्र मोदी का जन्म नरेंद्र मोदी का जन्मदिन आज मनाया जा रहा है। इस मौके पर प्रधानमंत्री नरेंद्र मोदी ने ट्वीट कर लिखा कि 'आज देश को एक नया संसद भवन मिला है, जिसे आप लोगों ने बधाई दी है। भारत और भारतवासियों को यह संसद बहुत ही महत्त्वपूर्ण है और इसका उपयोग देश के विकास में बहुत अधिक योगदान देगा। आज हमारे देश में एक संसद बनी हुई है जो हमारे देश की लोकतांत्रिक प्रक्रियाओं को मजबूत बनाती है।
Prime minister Modi Prime minister Modi's visit to China: PM Modi is scheduled to arrive in Beijing on Thursday. PM Narendra Modi will be the first Indian prime minister to visit China since former President Mao Zedong, who visited India in 1959 and was assassinated by Chinese soldiers a year later. The two leaders are expected to discuss regional security issues such as terrorism, trade and energy cooperation during their talks.
महात्मा गांधी महात्मा गांधी एक स्वतंत्र व्यक्ति, एक आत्मनिर्भर और एक अत्यंत लोकप्रिय नेता थे। उन्होंने अपने जीवनकाल में देश को एक नई दिशा देने के लिए बहुत कुछ किया। समाज के विभिन्न क्षेत्रों में उनके योगदान के लिए उन्हें सम्मानित किया गया था।
ब्रिटिश राजशाही ब्रिटिश राजशाही के विघटन के बाद, इंग्लैंड और वेल्स में ब्रिटेन की संसद का गठन हुआ। संसद ने 1807 में अपने पहले अधिनियम, "द क्राउन की विधेयक" (एक्ट ऑफ द डेमोक्रैटिक) को मंजूरी दी, जिसमें प्रत्येक राज्य से एक निर्वाचित सदस्य चुना गया।
মহাত্মা গান্ধী মহাত্মা গান্ধী ছিলেন আন্তর্জাতিক সমাজতান্ত্রিক আন্দোলনের অন্যতম প্রধান নেতা। তিনি ভারতের স্বাধীনতা সংগ্রামে তাঁর ভূমিকা অপরিসীম। ১৯৪৭ সালের ১৫ আগস্ট ভারতে বিপ্লব ঘটে। ১৯৪৮ সালে মহাত্মা গান্ধীর মৃত্যু হয়। এরপর থেকে মহাত্মা আর দেখা যায় না।
নরেন্দ্র মোদীর জন্ম নরেন্দ্র মোদীর জন্মদিনের শুভেচ্ছা জানিয়েছেন অমিতাভ বচ্চন। প্রসঙ্গত, গোটা দেশ জুড়ে চলছে নববর্ষ উৎসব। আর এই উৎসবে নিজেকে সামিল করতে চান না কেউই। এমনটাই জানালেন বলিউড অভিনেত্রী সারা আলি খান।
એક વખત ક વખતે, તેમણે તેમના મિત્ર અને સહાયક વિલિયમ હેનરી મેકડોનાલ્ડ સાથે મળીને "ધ ઓરિજિનલ મેન ઓફ ધ ડેડ" નામના પુસ્તકનું નિર્માણ કર્યું હતું; તે 1961માં પ્રકાશિત થયું હતું. 1958માં, તેમણે ધ ન્યૂ યોર્કર માટે લેખક અને સંપાદક પિતા તરીકે કામ કર્યું હતું.
એક સમયે એક રાજા જંગલમાં રહેતો હતો. તેમણે એક સમયે એક રાજા જંગલમાં રહેતો હતો. તેમણે, એક દિવસ રાજાને કહ્યું, "હું તમને મારી જાતને એક સ્ત્રી તરીકે આપું છું." રાજા ખુશ થયો. તેમણે તેને તેના ઘરમાં લઈ ગયા. તે દિવસે રાણી પોતાના પતિ સાથે ઘરે આવી અને કહ્યું કે તે તેની પાસે છે. રાજાએ તેને પૂછ્યું કે શું તેણે તેની સાથે લગ્ન કર્યા છે ? રાજા અને રાણીએ પૂછ્યું - હા. ત્યારે તેમણે કહ્યું કે, "તમે મને તમારી સાથે લગ્ન કરવાનું વચન આપ્યું છે. હવે મારે તમારા પર વિશ્વાસ રાખવો જોઈએ નહીં".

Citation

@ONLINE{bhasha-wiki,
    author = "Soket Labs Technology and Research Private Limited",
    title  = "pragna-1b",
    url    = "https://soket.ai"
}

Model Card Contact

connect@soket.ai

Downloads last month
87
Safetensors
Model size
1.25B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for soketlabs/pragna-1b

Finetunes
7 models
Quantizations
2 models

Datasets used to train soketlabs/pragna-1b

Spaces using soketlabs/pragna-1b 3