gokceuludogan commited on
Commit
fe59bc3
1 Parent(s): d5d5255

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +131 -0
README.md ADDED
@@ -0,0 +1,131 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ language:
4
+ - tr
5
+ library_name: transformers
6
+ pipeline_tag: text2text-generation
7
+ ---
8
+
9
+
10
+ # Model Card for TURNA
11
+
12
+ <!-- Provide a quick summary of what the model is/does. -->
13
+
14
+ TURNA is a Turkish language model based on the UL2 framework which is suitable for both understanding and generation tasks.
15
+
16
+ Evaluations across three generation and six understanding tasks in Turkish show that TURNA outperforms several multilingual models and competes with monolingual Turkish models in understanding tasks.
17
+
18
+ The model is shared with the public to be used solely for non-commercial academic research purposes.
19
+
20
+ ## Model Details
21
+
22
+ ### Model Description
23
+
24
+ <!-- Provide a longer summary of what this model is. -->
25
+
26
+ - **Developed by:** Bogazici University Computer Engineering Department TABILAB (special thanks to VNGRS-AI for sharing their tokenizer)
27
+ - **Funded by:** We thank the Google TPU Research Cloud program for providing us with credits to pretrain our model on TPU v3-8 machines.
28
+ <!-- - **Shared by [optional]:** [More Information Needed] -->
29
+ - **Model type:** Transformer-based encoder-decoder
30
+ - **Language(s) (NLP):** Turkish
31
+ - **License:** The model is shared with the public to be used solely for non-commercial academic research purposes.
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** Coming soon! TBA
38
+ - **Paper [optional]:** Coming soon! TBA
39
+
40
+ ## Uses
41
+
42
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
43
+
44
+ ### Direct Use
45
+
46
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
47
+
48
+ This model can be used for research purposes. You give some text and this model tries to predict the next words.
49
+
50
+ ### Downstream Use
51
+
52
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
53
+
54
+ This model can be finetuned using [our library](https://example.com) to solve your own task involving Turkish language.
55
+
56
+ This model can be further trained for behaving more helpful, less harmful and better for dialog use cases.
57
+
58
+ ### Out-of-Scope Use
59
+
60
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
61
+
62
+ Any commercial or malicious activity.
63
+
64
+ ## Bias, Risks, and Limitations
65
+
66
+ We refer to the Flan-T5's [official model card](https://arxiv.org/pdf/2210.11416.pdf):
67
+
68
+ > Language models, including Flan-T5, can potentially be used for language generation in a harmful way, according to Rae et al. (2021). Flan-T5 should not be used directly in any application, without a prior assessment of safety and fairness concerns specific to the application.
69
+
70
+ ### Ethical considerations and risks
71
+
72
+ > ... (ed. The model) is fine-tuned on a large corpus of text data that was not filtered for explicit content or assessed for existing biases. As a result the model itself is potentially vulnerable to generating equivalently inappropriate content or replicating inherent biases in the underlying data.
73
+
74
+ ### Known Limitations
75
+
76
+ > ... (ed. The model) has not been tested in real world applications.
77
+
78
+ ### Sensitive Use:
79
+
80
+ > ... (ed. The model) should not be applied for any unacceptable use cases, e.g., generation of abusive speech.
81
+
82
+
83
+ ## How to Get Started with the Model
84
+
85
+ You can find the technical usage guidance at our library's Github [page](https://example.com).
86
+
87
+ ## Training Details
88
+
89
+ Refer to the paper for more information.
90
+
91
+ ## Evaluation
92
+
93
+ Refer to the paper for more information.
94
+
95
+ ## Environmental Impact
96
+
97
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
98
+
99
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
100
+
101
+ - **Hardware Type:** TPU v3-8
102
+ - **Hours used:** About 400 hours
103
+ - **Cloud Provider:** Google Cloud
104
+ - **Compute Region:** europe-west4-a
105
+ - **Carbon Emitted:** 64.52 kg CO2_2
106
+
107
+ ## Technical Specifications
108
+
109
+ Refer to the paper for more information.
110
+
111
+ ## Citation
112
+
113
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
114
+
115
+ **BibTeX:**
116
+
117
+ Coming soon!
118
+
119
+ **APA:**
120
+
121
+ Coming soon!
122
+
123
+ ## Model Card Authors
124
+
125
+ Paper authors.
126
+
127
+ ## Model Card Contact
128
+
129
+ Onur Güngör
130
+
131
+