adnanaman commited on
Commit
784f6ba
·
verified ·
1 Parent(s): 0f59632

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -10
README.md CHANGED
@@ -86,38 +86,38 @@ result = classifier(product_description)
86
  print(result)
87
  ```
88
 
89
- #Training Details
90
- ##Training Data
91
  The model was trained on an e-commerce dataset that includes various product categories such as CPUs, Digital Cameras, Dishwashers, Fridge Freezers, Microwaves, Mobile Phones, TVs, and Washing Machines. The data was preprocessed by removing duplicates, lowercasing, and tokenizing text.
92
 
93
- ##Training Procedure
94
  1. Preprocessing: Text data was cleaned, lowercased, and tokenized. Product descriptions were truncated to 128 tokens for uniformity.
95
  2. Hyperparameters: Fine-tuning was conducted with a learning rate of 2e-5 and batch size of 16 for 3 epochs.
96
  3. Training Hardware: The model was trained on a single NVIDIA Tesla V100 GPU for approximately 3 hours.
97
 
98
- ##Evaluation
99
  Testing Data, Factors & Metrics
100
  The model was evaluated on a separate test set of product descriptions, using precision, recall, and F1-score as the evaluation metrics.
101
 
102
- ##Summary
103
  The model achieved an overall accuracy of 96.16%, with strong performance across multiple product categories. The F1-scores indicate that the model performs particularly well in the "CPUs" and "Digital Cameras" categories.
104
 
105
- ##Technical Specifications
106
  Model Architecture and Objective
107
  The DistilBERT-ProductClassifier model utilizes the DistilBERT architecture, fine-tuned with a text classification head for e-commerce product categorization tasks.
108
 
109
- ##Compute Infrastructure
110
  The model is optimized to run efficiently on CPUs and small GPUs, making it suitable for real-time applications.
111
 
112
- ##Hardware
113
  This model requires a minimum of 4GB of RAM for efficient inference, and a modern CPU or basic GPU is recommended.
114
 
115
- ##Software
116
  Transformers library: v4.26.0
117
  Python version: 3.8 or higher
118
  CUDA [optional]: 10.2 or higher (if running on GPU)
119
 
120
- ##Citation
121
  If you use this model, please cite as follows:
122
 
123
 
 
86
  print(result)
87
  ```
88
 
89
+ # Training Details
90
+ ## Training Data
91
  The model was trained on an e-commerce dataset that includes various product categories such as CPUs, Digital Cameras, Dishwashers, Fridge Freezers, Microwaves, Mobile Phones, TVs, and Washing Machines. The data was preprocessed by removing duplicates, lowercasing, and tokenizing text.
92
 
93
+ ## Training Procedure
94
  1. Preprocessing: Text data was cleaned, lowercased, and tokenized. Product descriptions were truncated to 128 tokens for uniformity.
95
  2. Hyperparameters: Fine-tuning was conducted with a learning rate of 2e-5 and batch size of 16 for 3 epochs.
96
  3. Training Hardware: The model was trained on a single NVIDIA Tesla V100 GPU for approximately 3 hours.
97
 
98
+ ## Evaluation
99
  Testing Data, Factors & Metrics
100
  The model was evaluated on a separate test set of product descriptions, using precision, recall, and F1-score as the evaluation metrics.
101
 
102
+ ## Summary
103
  The model achieved an overall accuracy of 96.16%, with strong performance across multiple product categories. The F1-scores indicate that the model performs particularly well in the "CPUs" and "Digital Cameras" categories.
104
 
105
+ ## Technical Specifications
106
  Model Architecture and Objective
107
  The DistilBERT-ProductClassifier model utilizes the DistilBERT architecture, fine-tuned with a text classification head for e-commerce product categorization tasks.
108
 
109
+ ## Compute Infrastructure
110
  The model is optimized to run efficiently on CPUs and small GPUs, making it suitable for real-time applications.
111
 
112
+ ## Hardware
113
  This model requires a minimum of 4GB of RAM for efficient inference, and a modern CPU or basic GPU is recommended.
114
 
115
+ ## Software
116
  Transformers library: v4.26.0
117
  Python version: 3.8 or higher
118
  CUDA [optional]: 10.2 or higher (if running on GPU)
119
 
120
+ ## Citation
121
  If you use this model, please cite as follows:
122
 
123