SivilTaram
commited on
Commit
•
8338aff
1
Parent(s):
a755f6f
Update README.md
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ TAPEX was proposed in [TAPEX: Table Pre-training via Learning a Neural SQL Execu
|
|
16 |
|
17 |
TAPEX (**Ta**ble **P**re-training via **Ex**ecution) is a conceptually simple and empirically powerful pre-training approach to empower existing models with *table reasoning* skills. TAPEX realizes table pre-training by learning a neural SQL executor over a synthetic corpus, which is obtained by automatically synthesizing executable SQL queries.
|
18 |
|
19 |
-
TAPEX is based on the BART architecture, the transformer encoder-
|
20 |
|
21 |
This model is the `tapex-base` model fine-tuned on the [WikiTableQuestions](https://huggingface.co/datasets/wikitablequestions) dataset.
|
22 |
|
|
|
16 |
|
17 |
TAPEX (**Ta**ble **P**re-training via **Ex**ecution) is a conceptually simple and empirically powerful pre-training approach to empower existing models with *table reasoning* skills. TAPEX realizes table pre-training by learning a neural SQL executor over a synthetic corpus, which is obtained by automatically synthesizing executable SQL queries.
|
18 |
|
19 |
+
TAPEX is based on the BART architecture, the transformer encoder-decoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder.
|
20 |
|
21 |
This model is the `tapex-base` model fine-tuned on the [WikiTableQuestions](https://huggingface.co/datasets/wikitablequestions) dataset.
|
22 |
|