Evgeniy Hristoforu's picture

Evgeniy Hristoforu

ehristoforu

AI & ML interests

Diffusers, LLM and others ML.

Recent Activity

updated a collection about 22 hours ago
FluentlyLM Prinum
liked a model about 22 hours ago
mradermacher/FluentlyLM-Prinum-abliterated-GGUF
updated a collection about 22 hours ago
FluentlyLM Prinum
View all activity

Organizations

Samsung Electronics's profile picture video-p2p-library's profile picture lora concepts library's profile picture Open-Source AI Meetup's profile picture Tune a video concepts library's profile picture Stable Diffusion Dreambooth Concepts Library's profile picture Blog-explorers's profile picture OpenSky's profile picture OpenLLM France's profile picture DreamDrop AI's profile picture ZeroGPU Explorers's profile picture Project Fluently's profile picture test-org's profile picture LocalLLaMA's profile picture Project Dreamly's profile picture ReLP's profile picture TensorLlama's profile picture TensorMistral's profile picture MLX Community's profile picture Theme repo for Gradio's profile picture GuardAI (GAI)'s profile picture Public League AI's profile picture Social Post Explorers's profile picture MiniSearch's profile picture Project Fluently LM's profile picture C4AI Community's profile picture Dev Mode Explorers's profile picture Different LLM Sizes's profile picture Different LLM's profile picture ChatGPT Community (Unoffical, Non-profit)'s profile picture Stable Diffusion Community (Unofficial)'s profile picture Stable Diffusion Community (Unofficial, Non-profit)'s profile picture Synergetic AI's profile picture Diffusers Testing's profile picture LoRA Factory's profile picture Hugging Face Discord Community's profile picture Enhanced By Ehristoforu (EBE)'s profile picture LM Layers (Research)'s profile picture Fluently Lab.'s profile picture Puregen AI's profile picture Experiments with LLMs's profile picture Gramota.AI - Russian language's profile picture Need For LLM (NFL Project)'s profile picture Fluently Datasets's profile picture

ehristoforu's activity

replied to their post 6 days ago
view reply

Perhaps the seed was unsuccessful or perhaps the temperature you set or the system prompt had an effect. We are happy to answer your other questions if they appear)

replied to their post 6 days ago
view reply

  • I have 3 apples. Last week I ate 2 apples. How many apples do I have?

  • You currently have 3 apples. The fact that you ate 2 apples last week does not affect the number of apples you have now. So, you still have 3 apples.

replied to their post 10 days ago
view reply

Thank you for your interest in our model! I hope you like it)

reacted to their post with ๐Ÿ”ฅ 10 days ago
view post
Post
2648
Introducing our first standalone model โ€“ FluentlyLM Prinum

Introducing the first standalone model from Project Fluently LM! We worked on it for several months, used different approaches and eventually found the optimal one.

General characteristics:
- Model type: Causal language models (QwenForCausalLM, LM Transformer)
- Number of parameters: 32.5B
- Number of parameters (not embedded): 31.0B
- Number of layers: 64
- Context: 131,072 tokens
- Language(s) (NLP): English, French, Spanish, Russian, Chinese, Japanese, Persian (officially supported)
- License: MIT

Creation strategy:
The basis of the strategy is shown in Pic. 2.
We used Axolotl & Unsloth for SFT-finetuning with PEFT LoRA (rank=64, alpha=64) and Mergekit for SLERP and TIES mergers.

Evolution:
๐Ÿ† 12th place in the Open LLM Leaderboard ( open-llm-leaderboard/open_llm_leaderboard) (21.02.2025)

Detailed results and comparisons are presented in Pic. 3.

Links:
- Model: fluently-lm/FluentlyLM-Prinum
- GGUF version: mradermacher/FluentlyLM-Prinum-GGUF
- Demo on ZeroGPU: ehristoforu/FluentlyLM-Prinum-demo
  • 7 replies
ยท
posted an update 10 days ago
view post
Post
2648
Introducing our first standalone model โ€“ FluentlyLM Prinum

Introducing the first standalone model from Project Fluently LM! We worked on it for several months, used different approaches and eventually found the optimal one.

General characteristics:
- Model type: Causal language models (QwenForCausalLM, LM Transformer)
- Number of parameters: 32.5B
- Number of parameters (not embedded): 31.0B
- Number of layers: 64
- Context: 131,072 tokens
- Language(s) (NLP): English, French, Spanish, Russian, Chinese, Japanese, Persian (officially supported)
- License: MIT

Creation strategy:
The basis of the strategy is shown in Pic. 2.
We used Axolotl & Unsloth for SFT-finetuning with PEFT LoRA (rank=64, alpha=64) and Mergekit for SLERP and TIES mergers.

Evolution:
๐Ÿ† 12th place in the Open LLM Leaderboard ( open-llm-leaderboard/open_llm_leaderboard) (21.02.2025)

Detailed results and comparisons are presented in Pic. 3.

Links:
- Model: fluently-lm/FluentlyLM-Prinum
- GGUF version: mradermacher/FluentlyLM-Prinum-GGUF
- Demo on ZeroGPU: ehristoforu/FluentlyLM-Prinum-demo
  • 7 replies
ยท
reacted to their post with ๐Ÿค— 2 months ago
view post
Post
3532
โœ’๏ธ Ultraset - all-in-one dataset for SFT training in Alpaca format.
fluently-sets/ultraset

โ“ Ultraset is a comprehensive dataset for training Large Language Models (LLMs) using the SFT (instruction-based Fine-Tuning) method. This dataset consists of over 785 thousand entries in eight languages, including English, Russian, French, Italian, Spanish, German, Chinese, and Korean.

๐Ÿคฏ Ultraset solves the problem faced by users when selecting an appropriate dataset for LLM training. It combines various types of data required to enhance the model's skills in areas such as text writing and editing, mathematics, coding, biology, medicine, finance, and multilingualism.

๐Ÿค— For effective use of the dataset, it is recommended to utilize only the "instruction," "input," and "output" columns and train the model for 1-3 epochs. The dataset does not include DPO or Instruct data, making it suitable for training various types of LLM models.

โ‡๏ธ Ultraset is an excellent tool to improve your language model's skills in diverse knowledge areas.
posted an update 2 months ago
view post
Post
3532
โœ’๏ธ Ultraset - all-in-one dataset for SFT training in Alpaca format.
fluently-sets/ultraset

โ“ Ultraset is a comprehensive dataset for training Large Language Models (LLMs) using the SFT (instruction-based Fine-Tuning) method. This dataset consists of over 785 thousand entries in eight languages, including English, Russian, French, Italian, Spanish, German, Chinese, and Korean.

๐Ÿคฏ Ultraset solves the problem faced by users when selecting an appropriate dataset for LLM training. It combines various types of data required to enhance the model's skills in areas such as text writing and editing, mathematics, coding, biology, medicine, finance, and multilingualism.

๐Ÿค— For effective use of the dataset, it is recommended to utilize only the "instruction," "input," and "output" columns and train the model for 1-3 epochs. The dataset does not include DPO or Instruct data, making it suitable for training various types of LLM models.

โ‡๏ธ Ultraset is an excellent tool to improve your language model's skills in diverse knowledge areas.
reacted to their post with โค๏ธ๐Ÿ‘€โž•๐Ÿง ๐Ÿค๐Ÿคฏ๐Ÿค—๐Ÿš€๐Ÿ‘๐Ÿ˜Ž๐Ÿ”ฅ 7 months ago
view post
Post
4545
๐Ÿ˜ Hello from Project Fluently Team!

โœจ Finally we can give you some details about Supple Diffusion. We worked on it for a long time and we have little left, we apologize that we had to increase the work time.

๐Ÿ› ๏ธ Some technical information. The first version will be the Small version (there will also be Medium, Large, Huge, possibly Tiny), it will be based on the SD1 architecture, that is, one text encoder, U-net, VAE. Now about each component, the first is a text encoder, it will be a CLIP model (perhaps not CLIP-L-path14), CLIP was specially retrained by us in order to achieve the universality of the model in understanding completely different styles and to simplify the prompt as much as possible. Next, we did U-net, U-net in a rather complicated way, first we trained different parts (types) of data with different U-nets, then we carried out merging using different methods, then we trained DPO and SPO using methods, and then we looked at the remaining shortcomings and further trained model, details will come later. We left VAE the same as in SD1 architecture.

๐Ÿ™Œ Compatibility. Another goal of the Supple model series is full compatibility with Auto1111 and ComfyUI already at the release stage, the model is fully supported by these interfaces and the diffusers library and does not require adaptation, your usual Sampling methods are also compatible, such as DPM++ 2M Karras, DPM++ SDE and others.

๐Ÿง Today, without demo images (there wasnโ€™t much time), final work is underway on the model and we are already preparing to develop the Medium version, the release of the Small version will most likely be in mid-August or earlier.

๐Ÿ˜ป Feel free to ask your questions in the comments below the post, we will be happy to answer them, have a nice day!
  • 1 reply
ยท
posted an update 7 months ago
view post
Post
4545
๐Ÿ˜ Hello from Project Fluently Team!

โœจ Finally we can give you some details about Supple Diffusion. We worked on it for a long time and we have little left, we apologize that we had to increase the work time.

๐Ÿ› ๏ธ Some technical information. The first version will be the Small version (there will also be Medium, Large, Huge, possibly Tiny), it will be based on the SD1 architecture, that is, one text encoder, U-net, VAE. Now about each component, the first is a text encoder, it will be a CLIP model (perhaps not CLIP-L-path14), CLIP was specially retrained by us in order to achieve the universality of the model in understanding completely different styles and to simplify the prompt as much as possible. Next, we did U-net, U-net in a rather complicated way, first we trained different parts (types) of data with different U-nets, then we carried out merging using different methods, then we trained DPO and SPO using methods, and then we looked at the remaining shortcomings and further trained model, details will come later. We left VAE the same as in SD1 architecture.

๐Ÿ™Œ Compatibility. Another goal of the Supple model series is full compatibility with Auto1111 and ComfyUI already at the release stage, the model is fully supported by these interfaces and the diffusers library and does not require adaptation, your usual Sampling methods are also compatible, such as DPM++ 2M Karras, DPM++ SDE and others.

๐Ÿง Today, without demo images (there wasnโ€™t much time), final work is underway on the model and we are already preparing to develop the Medium version, the release of the Small version will most likely be in mid-August or earlier.

๐Ÿ˜ป Feel free to ask your questions in the comments below the post, we will be happy to answer them, have a nice day!
  • 1 reply
ยท
replied to their post 8 months ago
view reply

Detailed technical information will come later.