--- title: Which Frame? emoji: πŸ” colorFrom: purple colorTo: purple sdk: streamlit app_file: whichframe.py --- # Which Frame? Search a video **semantically** with AI. For example, try a natural language search query like "a person with sunglasses". You can also search with images like Google's reverse image search and also a combined text + image. The underlying querying is powered by OpenAI’s CLIP neural network for "zero-shot" image classification. --- ## Try it out! http://whichframe.chuanenlin.com --- ## Setting up 1. Clone the repository. ```python git clone https://github.com/chuanenlin/whichframe.git cd whichframe ``` 2. Install package dependencies. ```python pip install -r requirements.txt ``` 3. Run the app. ```python streamlit run whichframe.py ``` --- ## Examples ### πŸ”€ Text Search #### Query "three red cars side by side" #### Result ![three-red-cars-side-by-side](examples/three-red-cars-side-by-side.jpeg) ### πŸŒ… Image Search #### Query ![police-car-query](examples/helicopter-query.jpeg) #### Result ![police-car-result](examples/helicopter-result.jpeg) ### πŸ”€ Text + πŸŒ… Image Search #### Query "a red subaru" + ![police-car-query](examples/police-car-query.jpeg) #### Result ![subaru-and-police-car-result](examples/subaru-and-police-car-result.jpeg)