Kasper Groes Albin Ludvigsen

ThatsGroes

AI & ML interests

Open source; LLM; time series forecasting; green AI

Recent Activity

liked a model about 16 hours ago
Finnish-NLP/Ahma-7B-Instruct
updated a dataset 5 days ago
ThatsGroes/syntetisk-dialog-opsummering-raw
updated a dataset 5 days ago
ThatsGroes/wiki_views
View all activity

Organizations

Flax Community's profile picture Dansk Data Science Community's profile picture Merge Crew's profile picture

ThatsGroes's activity

reacted to tomaarsen's post with ā¤ļø 7 days ago
view post
Post
2529
That didn't take long! Nomic AI has finetuned the new ModernBERT-base encoder model into a strong embedding model for search, classification, clustering and more!

Details:
šŸ¤– Based on ModernBERT-base with 149M parameters.
šŸ“Š Outperforms both nomic-embed-text-v1 and nomic-embed-text-v1.5 on MTEB!
šŸŽļø Immediate FA2 and unpacking support for super efficient inference.
šŸŖ† Trained with Matryoshka support, i.e. 2 valid output dimensionalities: 768 and 256.
āž”ļø Maximum sequence length of 8192 tokens!
2ļøāƒ£ Trained in 2 stages: unsupervised contrastive data -> high quality labeled datasets.
āž• Integrated in Sentence Transformers, Transformers, LangChain, LlamaIndex, Haystack, etc.
šŸ›ļø Apache 2.0 licensed: fully commercially permissible

Try it out here: nomic-ai/modernbert-embed-base

Very nice work by Zach Nussbaum and colleagues at Nomic AI.
New activity in indsigt-ai/MeDa-BERT about 1 month ago

Github?

2
#1 opened about 1 month ago by
ThatsGroes