File size: 1,923 Bytes
fc81a95
 
 
 
 
 
 
baf29fc
fc81a95
 
 
 
 
 
 
 
 
 
 
 
 
 
 
baf29fc
 
 
 
 
fc81a95
 
 
 
 
 
 
baf29fc
 
 
 
fc81a95
 
 
 
 
 
baf29fc
 
 
 
 
 
 
 
 
fc81a95
baf29fc
 
fc81a95
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
#!/bin/env python

""" Work in progress
Plan:
   Modded version of graph-embeddings.py
   Just to see if using different CLIP module changes values significantly
   (It does not)
   This code requires
       pip install git+https://github.com/openai/CLIP.git
"""


import sys
import json
import torch
import clip

import PyQt5
import matplotlib
matplotlib.use('QT5Agg')  # Set the backend to TkAgg

import matplotlib.pyplot as plt

# Available models:
#  'RN50', 'RN101', 'RN50x4', 'RN50x16', 'RN50x64', 'ViT-B/32', 'ViT-B/16', 'ViT-L/14', 'ViT-L/14@336px'
#CLIPname= "ViT-L/14"
CLIPname= "ViT-B/16"
#CLIPname= "ViT-L/14@336px"

device=torch.device("cuda")
print("loading CLIP model")
model, processor = clip.load(CLIPname,device=device)
model.cuda().eval()
print("done")

def embed_from_tokenid(num):
    # A bit sleazy, but, eh.
    tokens = clip.tokenize("dummy").to(device)
    tokens[0][1]=num

    with torch.no_grad():
        embed = model.encode_text(tokens)
        return embed



def embed_from_text(text):
    if text[0]=="#":
        print("Converting string to number")
        return embed_from_tokenid(int(text[1:]))

    tokens = clip.tokenize(text).to(device)
    print("Tokens for",text,"=",tokens)

    with torch.no_grad():
        embed = model.encode_text(tokens)
        return embed


fig, ax = plt.subplots()


text1 = input("First word or prompt: ")
text2 = input("Second prompt(or leave blank): ")


print("generating embeddings for each now")
emb1 = embed_from_text(text1)
print("shape of emb1:",emb1.shape)

graph1=emb1[0].tolist()
ax.plot(graph1, label=text1[:20])

if len(text2) >0:
    emb2 = embed_from_text(text2)
    graph2=emb2[0].tolist()
    ax.plot(graph2, label=text2[:20])

# Add labels, title, and legend
#ax.set_xlabel('Index')
ax.set_ylabel('Values')
ax.set_title('Comparative Graph of Two Embeddings')
ax.legend()

# Display the graph
print("Pulling up the graph")
plt.show()