dennisjooo commited on
Commit
170a155
1 Parent(s): 2409a1c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -9
README.md CHANGED
@@ -4,7 +4,7 @@ base_model: google/vit-base-patch16-224-in21k
4
  tags:
5
  - generated_from_trainer
6
  datasets:
7
- - imagefolder
8
  metrics:
9
  - accuracy
10
  - precision
@@ -16,8 +16,8 @@ model-index:
16
  name: Image Classification
17
  type: image-classification
18
  dataset:
19
- name: imagefolder
20
- type: imagefolder
21
  config: default
22
  split: train
23
  args: default
@@ -38,7 +38,8 @@ should probably proofread and complete it, then remove this comment. -->
38
 
39
  # emotion_classification
40
 
41
- This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the imagefolder dataset.
 
42
  It achieves the following results on the evaluation set:
43
  - Loss: 1.1031
44
  - Accuracy: 0.6312
@@ -47,15 +48,25 @@ It achieves the following results on the evaluation set:
47
 
48
  ## Model description
49
 
50
- More information needed
 
51
 
52
- ## Intended uses & limitations
53
 
54
- More information needed
55
 
56
- ## Training and evaluation data
 
 
 
 
 
 
57
 
58
- More information needed
 
 
 
59
 
60
  ## Training procedure
61
 
 
4
  tags:
5
  - generated_from_trainer
6
  datasets:
7
+ - FastJobs/Visual_Emotional_Analysis
8
  metrics:
9
  - accuracy
10
  - precision
 
16
  name: Image Classification
17
  type: image-classification
18
  dataset:
19
+ name: FastJobs/Visual_Emotional_Analysis
20
+ type: FastJobs/Visual_Emotional_Analysis
21
  config: default
22
  split: train
23
  args: default
 
38
 
39
  # emotion_classification
40
 
41
+ This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k)
42
+ on the [FastJobs/Visual_Emotional_Analysis](https://huggingface.co/datasets/FastJobs/Visual_Emotional_Analysis) dataset.
43
  It achieves the following results on the evaluation set:
44
  - Loss: 1.1031
45
  - Accuracy: 0.6312
 
48
 
49
  ## Model description
50
 
51
+ The Vision Transformer base version trained on ImageNet-21K released by Google.
52
+ Further details can be found on their [repo]((https://huggingface.co/google/vit-base-patch16-224-in21k))
53
 
54
+ ## Training and evaluation data
55
 
56
+ ### Data Split
57
 
58
+ Used a 4:1 ratio for training and development sets and a seed of 42.
59
+
60
+ ### Pre-processing Augmentation
61
+
62
+ The main pre-processing phase for both training and evaluation includes:
63
+ - Resizing to (224, 224, 3) because it uses ImageNet images to train the original model
64
+ - Normalizing images using a mean and standard deviation of [0.5, 0.5, 0.5]
65
 
66
+ Other than the aforementioned pre-processing, the training set was augmented using:
67
+ - Random horizontal & vertical flip
68
+ - Color jitter
69
+ - Random resized crop
70
 
71
  ## Training procedure
72