HCZhang commited on
Commit
a626d77
·
1 Parent(s): 249253a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -22
README.md CHANGED
@@ -15,20 +15,28 @@ We fine-tuned the [Open-Orca/OpenOrca-Platypus2-13B](https://huggingface.co/Open
15
  Its performance is competitive, rivaling previous state-of-the-art algorithms and LLMs such as OpenAI's GPT 3.5 and GPT 4 ([as demonstrated in our earlier studies](https://arxiv.org/abs/2308.16361)).
16
  It is notable that, as a 13B model, Jellyfish allows for cost-effective local execution without compromising data security.
17
 
18
- | Task | Dataset | Non-LLM SoTA<sup>1</sup> | GPT-3.5<sup>2</sup> | GPT-4<sup>2</sup> | Jellyfish-13B| Jellyfish-13B-Resoning |
19
- | ---- | ---- | ---- | ---- | ---- | ---- | ---- |
20
- | Entity Matching | Fodors-Zagats | 100 | 100 | 100 | 100 | 100 |
21
- | Entity Matching | Beer | 94.37| 96.30 | 100 | 93.33 | 100 |
22
- | Entity Matching | iTunes-Amazon | 97.06| 96.43 | 100 | 96.30 | 96.15 |
23
- | Entity Matching | Walmart-Amazon | 86.76| 86.17 | 90.27 | 80.71 | 85.16 |
24
- | Entity Matching | DBLP-ACM | 98.99| 96.99 | 97.44 | 97.35 | 95.74 |
25
- | Entity Matching | DBLP-GoogleScholar | 95.60| 76.12 | 91.87 | 92.83 | 89.45 |
26
- | Entity Matching | Amazon-Google | 75.58| 66.53 | 74.21 | 72.69 | 56.64 |
27
- | Data Imputation | Restaurant | 77.20| 94.19 | 97.67 | 94.19 | 93.02 |
28
- | Data Imputation | Buy | 96.50| 98.46 | 100 | 100 | 100 |
29
- | Error Detection | Hosptial | 99.10| 90.74 | 90.74 | 92.21 | 65.66 |
30
- | Error Detection | Adult | 94.40| 92.01 | 92.01 | 96.62 | 90.13 |
31
- | Schema Matching | Sythea | 38.50| 57.14 | 66.67 | 36.36 | 30.77 |
 
 
 
 
 
 
 
 
32
 
33
  _Accuracy as the metric for data imputation and the F1 score for other tasks._
34
  _For GPT-3.5, GPT-4 we used the few-shot approach, while for Jellyfish and Jellyfish-Reasoning, the zero-shot approach was employed._
@@ -39,15 +47,8 @@ _For GPT-3.5, GPT-4 we used the few-shot approach, while for Jellyfish and Jelly
39
  [HoloClean](https://arxiv.org/abs/1702.00820) for Data Imputation
40
  2.
41
  [Large Language Models as Data Preprocessors](https://arxiv.org/abs/2308.16361)
42
-
43
 
44
- We release two distinct versions of Jellyfish: Jellyfish-13B (the main branch) and Jellyfish-13B-Reasoning.
45
- As the names suggest, Jellyfish-13B is tailored to deliver precise, straightforward answers.
46
- In contrast, Jellyfish-13B-Reasoning, is fine-tuned with data that includes reasoning and sequential thought processes for handling data preprocessing tasks, distilling knowledge from GPT-4.
47
-
48
- The two versions are designed for different application scenarios.
49
- Jellyfish-13B is suitable for integration into larger data management systems due to its simple and clear responses that can be easily transformed into code.
50
- On the other hand, Jellyfish-13B-Reasoning is more user-oriented, with responses that provide them with in-depth data insights without the necessity for advanced coding skills or an intricate grasp of statistics.
51
 
52
  **Jellyfish paper will be coming soon!**
53
 
 
15
  Its performance is competitive, rivaling previous state-of-the-art algorithms and LLMs such as OpenAI's GPT 3.5 and GPT 4 ([as demonstrated in our earlier studies](https://arxiv.org/abs/2308.16361)).
16
  It is notable that, as a 13B model, Jellyfish allows for cost-effective local execution without compromising data security.
17
 
18
+ We release two distinct versions of Jellyfish: Jellyfish-13B (the main branch) and Jellyfish-13B-Reasoning.
19
+ As the names suggest, Jellyfish-13B is tailored to deliver precise, straightforward answers.
20
+ In contrast, Jellyfish-13B-Reasoning, is fine-tuned with data that includes reasoning and sequential thought processes for handling data preprocessing tasks, distilling knowledge from GPT-4.
21
+
22
+ The two versions are designed for different application scenarios.
23
+ Jellyfish-13B is suitable for integration into larger data management systems due to its simple and clear responses that can be easily transformed into code.
24
+ On the other hand, Jellyfish-13B-Reasoning is more user-oriented, with responses that provide them with in-depth data insights without the necessity for advanced coding skills or an intricate grasp of statistics.
25
+
26
+ | Task | Dataset | Non-LLM SoTA<sup>1</sup> | GPT-3.5<sup>2</sup> | GPT-4<sup>2</sup> | Jellyfish-13B| Jellyfish-13B-Resoning | Jellyfish-13B-1.1<sup>3</sup> |
27
+ | ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- |
28
+ | Entity Matching | Fodors-Zagats | 100 | 100 | 100 | 100 | 100 | 100 |
29
+ | Entity Matching | Beer | 94.37| 96.30 | 100 | 93.33 | 100 | 96.55 |
30
+ | Entity Matching | iTunes-Amazon | 97.06| 96.43 | 100 | 96.30 | 96.15 | 100 |
31
+ | Entity Matching | Walmart-Amazon | 86.76| 86.17 | 90.27 | 80.71 | 85.16 | 89.18 |
32
+ | Entity Matching | DBLP-ACM | 98.99| 96.99 | 97.44 | 97.35 | 95.74 | 99.32 |
33
+ | Entity Matching | DBLP-GoogleScholar | 95.60| 76.12 | 91.87 | 92.83 | 89.45 | 95.16 |
34
+ | Entity Matching | Amazon-Google | 75.58| 66.53 | 74.21 | 72.69 | 56.64 | 80.25 |
35
+ | Data Imputation | Restaurant | 77.20| 94.19 | 97.67 | 94.19 | 93.02 | 93.02 |
36
+ | Data Imputation | Buy | 96.50| 98.46 | 100 | 100 | 100 | 100 |
37
+ | Error Detection | Hosptial | 99.10| 90.74 | 90.74 | 92.21 | 65.66 | 86.59 |
38
+ | Error Detection | Adult | 94.40| 92.01 | 92.01 | 96.62 | 90.13 | 99.20 |
39
+ | Schema Matching | Sythea | 38.50| 57.14 | 66.67 | 36.36 | 30.77 | NA |
40
 
41
  _Accuracy as the metric for data imputation and the F1 score for other tasks._
42
  _For GPT-3.5, GPT-4 we used the few-shot approach, while for Jellyfish and Jellyfish-Reasoning, the zero-shot approach was employed._
 
47
  [HoloClean](https://arxiv.org/abs/1702.00820) for Data Imputation
48
  2.
49
  [Large Language Models as Data Preprocessors](https://arxiv.org/abs/2308.16361)
50
+ 3. Jellyfish-13B-1.1 is set to be the next iteration of Jellyfish-13B and is presently under development.We're showcasing its performance at this stage to highlight its impressive potential.As demonstrated in the table, it has already outperformed Non-LLM methods on the majority of benchmark datasets. We've optimized the training data for this 1.1 version, and its release is on the horizon.
51
 
 
 
 
 
 
 
 
52
 
53
  **Jellyfish paper will be coming soon!**
54