text
stringlengths
0
1.94k
Would recap posts keep the thread warm and cozy through these frigid winter nights?
Maybe, yeah! It could be years before Miku breaks through into our reality, so I think it's best to make Anon's stay as painless as it can be.
Less time scrolling means more time to use, develop, and discuss our local Mikus!
Highlights and recap of the previous thread: >>98408803
T/s leaderboard for Apple silicon: >>98408901 >>98409038 >>98409096
Over two weeks since low-rank expert Mixtral idea was pitched: >>98410524 >>98410533 >>98410560
Anon makes a "Good & Evil" advisor card, provides guide: >>98409721 >>98413520
WhiteRabbitNeo 33B pentest model released, TheBloke GGUF quants broken: >>98410316 >>98410890 >>98411016 >>98414636
WhiteRabbitNeo GPTQ version works, Anon gets cucked response, notices improvement with different prompt: >>98411729 >>98413609 >>98414076 >>98414242
Anon tests llama.cpp's updated quant utility, compares outputs across Mixtral quants for a given prompt, mentions drugs sampler: >>98412905
A new political benchmark is born: >>98414619 >>98414727 >>98414799 >>98414857 >>98414898 >>98415442
ExllamaV2 experiment - instant creation and inferencing of frankenmerges: >>98416337 >>98416453 >>98416646 >>98417164
Token healing discussion on Github: >>98418273
DeepSparse fast CPU-only inference: >>98418545
Anon's LocalLlama travel report: >>98411031 >>98411058 >>98411474 >>98411530
Comparing Mixtral-limarp and lzlv 70b: >>98413528 >>98413520 >>98413534 >>98413761
/aicg/ resident places faith in /lmg/, finds salvation, shares the news with his characters: >>98411395 >>98411470 >>98411791 >>98412229
Anon develops affection for LLM weights, shares logs: >>98412551 >>98412725
Different Anon uses the model in question, gets good results after some work, shares settings: >>98415036 >>98415139 >>98415501
Lazy UPS driver denies Anon his Miku: >>98412754 >>98412917 >>98413542
Anon finds himself in heaven after getting 3090, shares plans for old GPU: >>98412521 >>98412658 >>98413064
Anon envisions Star Trek without a certain character: >>98415750 >>98415837 >>98415880
Discussing the state of finetuning: >>98417244 >>98418316 >>98418410 >>98418426 >>98418622 >>98418786 >>98419040
Anon merges: >>98421125 >>98421134 >>98421260
Highlights and recap of the previous thread: >>98421955
Anon needs a high quality TTS model, receives Piper,: >>98422083 >>98422237 >>98422574 >>98422659
A second Anon also asks about TTS: >>98423448 >>98423463 >>98423497
Reminder that Coqui, the XTTSv2 company, is shutting down: >>98423014 >>98423058 >>98423073 >>98423229
Anon self-promotes MAID, a frontend: >>98422713 >>98422735 >>98422771 >>98422775 >>98422869 >>98425005 >>98425040
Fix for llama.cpp server markdown deleting characters from code: >>98423519
Mixtral improves its own system prompt, Anon shares related paper: >>98424281 >>98424480 >>98424799 >>98424824
Shit breaks after Anon blindly updates Kobold.cpp: >>98424581 >>98424779 >>98424814 >>98424868
Anon experiments with system prompt self-enhancement: >>98425885 >>98425899 >>98426094 >>98426301
Discussing usage of DPO to decuck Models. Paper about finetuning to realign L2 70b is linked: >>98426377 >>98426393 >>98426448 >>98427942 >>98428485
Outdated OP rentries lead Anon to ask questions about hardware, models, quants, and training LoRAs: >>98426797 >>98426877
Running and deploying LLMs in Unity: >>98428378 >>98428488
Sam Altman marries Meta software engineer: >>98422347 >>98422373 >>98422412
Anon asks about 5bpw Yi vs. 3.7bpw Mixtral: >>98422048 >>98422054 >>98422087 >>98422985 >>98422199 >>98427517
Mixtral draws a cube, Anon posts a tricky challenge: >>98423244 >>98423344 >>98423361 >>98423379 >>98423425 >>98423440 >>98423467 >>98423499 >>98426928
Visited by a stranger: >>98425602
Anon tries runpod without preparation: >>98426637 >>98426789 >>98426815 >>98426836 >>98426866 >>98426881 >>98426933
Miku: >>98426729 >>98426741
Xwin-Mlewd is still relevant: >>98426986
Sizefag model recommendation: >>98427331 >>98427517
llama.cpp server outputs nonsense for Anon: >>98427767 >>98428028 >>98428476
Highlights and recap of the previous thread: >>98428617
(1/2)
Transformers Training code for Mixtral might have been fixed (for real this time??): >>98428957 >>98429023 >>98429049 >>98429285 >>98429357 >>98429386
>>98429413 >>98429162 >>98429193 >>98429234 >>98429243
NVIDIA info page about digital AI avatars and interview about digital avatars that use generative AI: >>98429233 >>98429332
Anon thinks about creating a dataset focusing on human-nonhuman scenes, particularly scenes that include abnormal physical and anatomical interactions: >>98429661 (Over 20 replies)
Political compass test, questions and script provided: >>98429999 >>98430020 >>98430084 >>98430127 >>98430146 >>98430357 >>98430693 >>98431074 >>98431263
Anon requests information about local models for translating text from Japanese: >>98431480 >>98431534 >>98431635 >>98434539 >>98434563 >>98434761
Anon asks about applications that combine art, voice, and chat into one experience: >>98431516 >>98431560 >>98431569
Article covering a study by Anthropic about how LLMs can supposedly be taught to deceive (You): >>98431977 >>98432056 >>98432072 >>98432146 >>98432215 >>98432314
New paper about trustworthiness in LLMs: >>98433998 >>98434114
Paper from back in November investigating potential malicious use of LLMs. Authors used Jon Durbin's SpicyBoros in some tests: >>98434276 >>98434367 >>98434560 >>98435338 >>98435375 >>98435449
>>98435602
(2/2)
Anon's Mixtral self sysprompt improvement experiment continues with discussion about RP prompting and Mixtral's unwillingness to kill the user: >>98428683 >>98428758 >>98429343 >>98429405 >>98429495 >>98429553
Envisioning the AI endgame: >>98429847 >>98429999 >>98430004 >>98430943 >>98431484 >>98431492 >>98431549 >>98431748 >>98431865
Lovecraft and GUMI; Two flavors of horror: >>98430082 >>98430569 >>98430995 >>98431599 >>98431667 >>98431702 >>98431645 >>98431810 >>98432010 >>98432204 >>98432248
Experienced YiYiAss Anon revisits Yi after using Mixtral. The old trick of starting chats with stronger models, and continuing them with a weaker one is brought up: >>98432319 >>98432354 >>98432402 >>98432424 >>98432524 >>98432608 >>98432699
Knight & Dragon riddle: >>98433096 >>98433503 >>98433147 >>98433747 >>98433815 >>98433860 >>98433840 >>98433933 >>98433967 >>98434342
►Recent Highlights from the Previous Thread: >>98435583
--Article from May about optical computing for AI: >>98435958
--OpenAI LLM refusals found in Amazon pages and on Twitter: >>98436911 >>98436995 >>98437083
--Anon gets blue balls. Later shares wholesome backstory about lifelong goal and glimpse of success: >>98436766 >>98436925 >>98436930 >>98436979 >>98437207 >>98436794 >>98437365 >>98437918 >>98438123
--Hitchhikers's Guide to Local LLMs - terminology glossary: >>98437932 >>98438016 >>98438327
--Mixtral finetune adequately runs card designed for GPT-4: >>98438096 >>98438141 >>98438258 >>98438492 >>98438802
--Discussing quality of Q6, Q8, FP16: >>98436011 >>98436080 >>98436132 >>98436184 >>98436256 >>98438187 >>98438212 >>98438301
--Pruning unimportant weights from models, related paper and repos: >>98438828 >>98438966 >>98439004 >>98439006 >>98439044 >>98439116 >>98439129 >>98439145
--Related: Discussing Mixtral optimization: >>98439116 >>98439133 >>98439141 >>98439174 >>98439221
--Anon tries 70b, finds it smarter than Mixtral: >>98439560 >>98439858 >>98439579 >>98440203 >>98440217
--Related: Quantization of 70b vs. sparse MoE, image visualizing quantization of SD: >>98440203 >>98440316 >>98440863 >>98440883 >>98440941 >>98440987 >>98440996 >>98441197 >>98441235 >>98441255 >>98441265 >>98441291 >>98441300
--Flash Attention on AMD RDNA3 consumer cards: >>98440085 >>98440107 >>98440143 >>98440271
--Anon asks about multi-GPU performance: >>98440090 >>98440130 >>98440148 >>98441648
--llama.cpp calibrated 2bit quants merged + new PR to add ability to use importance matrix for all k-quants: >>98441682 >>98441698 >>98442792 >>98441705 >>98441796 >>98442539 >>98442551
--Mamba implementation in pure C: >>98441835 >>98441865 >>98441881 >>98442389
--Intermediate checkpoint of OpenMoE-34B (200B tokens): >>98442058 >>98442123
►Recent Highlight Post from the Previous Thread: >>98435602 >>98435616
►Recent Highlights from the Previous Thread: >>98451105
(1/2)
--Support for optional importance matrix (computed using a calibration dataset) for all k-quants merged into llama.cpp: >>98445276
--Anon shared unusual character card formatting: >>98445147 >>98445168 >>98445234 >>98445235 >>98445646 >>98445755 >>98445822 >>98445854 >>98445915 >>98445964 >>98445987 >>98446038
--Development of TikTok's language model was cut short by OpenAI in December: >>98446500 >>98446594 >>98448009 >>98448044
--No model handled Anon's extremely rare fetish, can't find data for training, did not say what the fetish is comprised of: >>98444899 >>98445118 >>98445214 >>98445236 >>98445716
--Experimenting with rounding and dequantization logic in llama.cpp: >>98445405 >>98448897 >>98449003 >>98449461
--Anon trained rapey Mixtral QLoRA, posted outputs, was unsure about uploading, cont.: >>98442342 >>98442484 >>98442594 >>98443112 | >>98445595 >>98445922
--Performance tests using Vulkan implementation in llama.cpp: >>98446951 >>98446966 >>98446976 >>98447276
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/datasets-cards)

sdfsdf

Downloads last month
0
Edit dataset card