yuxi-liu-wired commited on
Commit
cd2dac3
1 Parent(s): 386dad3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -46,7 +46,9 @@ def _transform(n_px):
46
 
47
  See the documentation for [`CLIPImageProcessor` for details](https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPImageProcessor).
48
 
49
- Also, despite the names `style vector` and `content vector`, I have noticed by visual inspection that both are basically equally good for style embedding. I don't know why, but I guess that's life?
 
 
50
 
51
  ## How to use it
52
 
 
46
 
47
  See the documentation for [`CLIPImageProcessor` for details](https://huggingface.co/docs/transformers/model_doc/clip#transformers.CLIPImageProcessor).
48
 
49
+ Also, despite the names `style vector` and `content vector`, I have noticed by visual inspection that both are basically equally good for style embedding. I don't know why, but I guess that's life? (No, it's actually not supposed to happen. I don't know why it didn't really disentangle style and content. Maybe that's a question for a small research paper.)
50
+
51
+ You can see for yourself by changing the line `style_output = output["style_output"].squeeze(0)` to `style_output = output["content_output"].squeeze(0)` in the demo. The resulting t-SNE is still clustering by style, to my eyes equally well.
52
 
53
  ## How to use it
54