geninhu commited on
Commit
66f942c
1 Parent(s): faabbba

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -6
README.md CHANGED
@@ -2,26 +2,45 @@
2
  library_name: keras
3
  tags:
4
  - structured-data-classification
 
5
  ---
6
 
7
  ## Model description
8
 
9
  More information needed
 
 
 
10
 
11
- ## Intended uses & limitations
12
 
13
- More information needed
 
 
 
 
 
14
 
15
- ## Training and evaluation data
 
16
 
17
- More information needed
 
 
18
 
19
  ## Training procedure
20
-
21
  ### Training hyperparameters
22
 
23
  The following hyperparameters were used during training:
24
- - optimizer: {'name': 'AdamW', 'learning_rate': 0.001, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay': 1e-04, 'exclude_from_weight_decay': None}
 
 
 
 
 
 
 
 
25
  - training_precision: float32
26
 
27
  ## Training Metrics
 
2
  library_name: keras
3
  tags:
4
  - structured-data-classification
5
+ - transformer
6
  ---
7
 
8
  ## Model description
9
 
10
  More information needed
11
+ ### Keras Implementation of Structured data learning with TabTransformer
12
+ This repo contains the trained model of [Structured data learning with TabTransformer](https://keras.io/examples/structured_data/tabtransformer/#define-dataset-metadata).
13
+ The full credit goes to: [Khalid Salama](https://www.linkedin.com/in/khalid-salama-24403144/)
14
 
15
+ Spaces Link:
16
 
17
+ ### Model summary:
18
+ - The trained model uses self-attention based Transformers structure following by multiple feed forward layers in order to serve supervised and semi-supervised learning.
19
+ - The model's inputs can contain both numerical and categorical features.
20
+ - All the categorical features will be encoded into embedding vector with the same number of embedding dimensions, before adding (point-wise) with each other and feeding into a stack of Transformer blocks.
21
+ - The contextual embeddings of the categorical features after the final Transformer layer, are concatenated with the input numerical features, and fed into a final MLP block.
22
+ - A SoftMax function is applied at the end of the model.
23
 
24
+ ## Intended uses & limitations:
25
+ - This model can be used for both supervised and semi-supervised tasks on tabular data.
26
 
27
+ ## Training and evaluation data:
28
+ - This model was trained using the [United States Census Income Dataset](https://archive.ics.uci.edu/ml/datasets/census+income) provided by the UC Irvine Machine Learning Repository. The task of the dataset is to predict whether a person is likely to be making over USD 50,000 a year (binary classification).
29
+ - The dataset consists of 14 input features: 5 numerical features and 9 categorical features.
30
 
31
  ## Training procedure
 
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
+ - optimizer: 'AdamW'
36
+ - learning_rate: 0.001
37
+ - weight decay: 1e-04
38
+ - loss: 'sparse_categorical_crossentropy'
39
+ - beta_1: 0.9
40
+ - beta_2: 0.999
41
+ - epsilon: 1e-07
42
+ - epochs: 50
43
+ - batch_size: 16
44
  - training_precision: float32
45
 
46
  ## Training Metrics