nyap commited on
Commit
eab8e58
·
verified ·
1 Parent(s): 9a5ee34

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -3
README.md CHANGED
@@ -1,8 +1,42 @@
1
  ---
2
  license: mit
3
  ---
4
- This model was created for Cosmocook 2, a project at the 2024 Berkeley AI Hackathon.
 
5
 
6
- Cosmo-demo is a fine-tuned version of the Moondream vision model. This model was trained on a set of images of avocadoes, bananas, doritos, and onions. With special attention to counting the number of each food item in an image.
7
 
8
- This model was trained in Intel's Jupyter notebooks using IPEX.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
  ---
4
+ ## Development Process
5
+ This model was created with the intention of supplmenting our mobile application as an ingredient counter. Such that a user could take an image and have AI determine whether they have the necessary ingredients to create a recipe. As a result of the straightforward usecase, we were motivated to use a lightweight model for this classification task. We were recommended Moondream by a mentor from Intel, which we fine-tuned in this project.
6
 
7
+ ## Model Details
8
 
9
+ - **Model Name**: Moondream Fine-tuned Variant
10
+ - **Task**: Ingredient Classification and Counting
11
+ - **Repository Name**: nyap/cosmo-demo
12
+ - **Original Model**: vikhyatk/moondream2
13
+
14
+ ## Intended Use
15
+
16
+ - **Primary Use Case**: This model counts and classifies ingredients (exclusively avocados, bananas, doritos, and onions during CalHacks) in images.
17
+ - **Associated Application**: This model is applied through a cooking assistant mobile application built in React-native.
18
+
19
+ ## Data
20
+
21
+ - **Dataset**: Used a very small dataset of a little less than 100 images collected during the Hackathon. Including images of avocados, bananas, doritos, and onions from various angles and in different combinations.
22
+ - **Data Collection**: Images were taken of various tables with the ingredients arranged, before being manually labelled with the count/type of ingredients in each image.
23
+
24
+ ## Evaluation Metrics
25
+
26
+ - **Performance Metrics**: Performance metrics did not provide any valuable feedback due to the size of the dataset (overfit)
27
+
28
+ ## Training
29
+
30
+ - **Training Procedure**: Created a basic dataloader class to read image data and expected output. PIL images were then passed through the VIT provided with Moondream. Moondream was then fine-tuned using the default Adam optimizer with a batch size of 8 for 30 epochs.
31
+ - **Hardware**: The model was trained with XPU (Intel(R) Data Center GPU Max 1100) provided by Intel mentors at the Berkeley AI Hackathon.
32
+
33
+ ## Limitations
34
+
35
+ - **Current Limitations**: The model may struggle with images containing overlapping ingredients or items not seen during training. It also only has a limited dataset, so different backgrounds or types of ingredients would not likely not be counted.
36
+ - **Potential Improvements**: Incorporate a full-scale dataset with labelled ingredient counts to make the model more knowledgable and applicable within our application. Tune hyperparameters beyond what was done during the Hackathon to optimize performance.
37
+
38
+ ## Additional Information
39
+
40
+ - **License**: Model released under MIT License.
41
+ - **Contact Information**: For inquiries, contact [nyap@umich.edu](mailto:nyap@umich.edu).
42
+ - **Acknowledgments**: Acknowledge Intel team for starter code and assistance in fine-tuning.