Distilabel and synthetic data community interviews - the outcomes
We've been doing some interview with community members to understand the needs surrounding synthetic data. Many thanks to the participants. Note that, given they interviewees were sourced from our community, so the results will likely represent that.
Things distilabel does well - security and reliability by caching generations and having serializable pipelines. - scaling up generation by parallelising inference and Anyscale Ray - solid implementations of state of the art research papers
Things to improve - communication about the fact we support structured generation - customization of existing prompt implementations are difficult - creation of new tasks prove difficult - arguments and parameters for tasks aren't available at first glance - the learning curve can be steep - more tutorials that represent real-life usage
Things to note - create small scale and large scale dataset to Millions of records - people use synthetic data to move away from frontier model providers - people mostly use 7B or 70B models for generating
I can solve the Traveling Salesman Problem using the same methods the scientists used to solve it with 1 qubit, except I do not need quantum computers to do it. I am kind of tired of screaming this from the rooftops at this point. I can create an imaginary probability space, then I can put a bunch of imaginary agents in the imaginary box, and solve real problems in seconds. Problems that would take minutes, hours, or years to solve via other algorithms. Here is a demo of me solving the Traveling Salesman problem using 50 agents to probabilistically sample at once: https://colab.research.google.com/drive/1XplG72nQDO_-2h4DUllERLp0Dr2pI2J2?usp=sharing