.gitattributes CHANGED
@@ -53,3 +53,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
53
  *.jpg filter=lfs diff=lfs merge=lfs -text
54
  *.jpeg filter=lfs diff=lfs merge=lfs -text
55
  *.webp filter=lfs diff=lfs merge=lfs -text
 
 
 
53
  *.jpg filter=lfs diff=lfs merge=lfs -text
54
  *.jpeg filter=lfs diff=lfs merge=lfs -text
55
  *.webp filter=lfs diff=lfs merge=lfs -text
56
+ all_cluster_centroids_10000.json filter=lfs diff=lfs merge=lfs -text
57
+ all_cluster_centroids_20000.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,14 @@
1
  ---
2
  license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+
5
+ # Subjects-for-curricular
6
+
7
+ ---
8
+
9
+ Subjects-for-curricular is a clustered set of book titles and concepts from https://huggingface.co/datasets/benxh/opensyllabus-tagged-libgen and https://huggingface.co/datasets/benxh/us-library-of-congress-subjects.
10
+
11
+
12
+ These two datasets are combined, taking the 'text' field from the US library of congress subjects (subjects_fixed.jsonl), and the 'name' field from the opensyllabus data. This text data is then converted into embeddings using the [bge-large-en-v1.5](https://huggingface.co/BAAI/bge-large-en-v1.5) model.
13
+
14
+ The embeddings are clustered into topics using K-means clustering. We include data for k=10,000 (all_topic_clusters_10000.parquet) and k=20,000 (all_topic_clusters_20000.parquet). We also include the centroid embeddings for each cluster. Both runs used 100 iterations of the k-means algorithm.
all_cluster_centroids_10000.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8fabf6293f29e7fce73fd56657119aae35065cf4e0509717ddf1ae79c8d4de9
3
+ size 351345061
all_cluster_centroids_20000.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:512163b16d8a848f2517c5494599833a34426b228cc9a7218c55ac3a50d3f2f1
3
+ size 702335914