psych-search / README.md
datawrestler's picture
initial sycn of psych-search model and documentation
cdb0643
|
raw
history blame
2.89 kB
metadata
language:
  - en
tags:
  - mental-health
license: Apache 2.0
datasets:
  - PubMed

Psych-Search

Model description

This model is an extension of allenai/scibert_scivocab_uncased. Continued pretraining was done using SciBERT as the base model using abstract text only from Pyschology and Psychiatry PubMed research. Training was done on approximately 3.5 million papers for 10 epochs and evaluated on a task similar to BioASQ Task A.

Intended uses & limitations

How to use

from transformers import AutoTokenizer, AutoModel

mname = "datawrestler/psych-search"
tokenizer = AutoTokenizer.from_pretrained(mname)
model = AutoModel.from_pretrained(mname)

Limitations and bias

This model was trained on all PubMed abstracts categorized under Psychology and Psychiatry. As of March 1, this corresponded to approximately 3.2 million papers that contained abstract text. Of these 3.2 million papers, relevant sparse categories were back translated to increase the representation of sparser mental health categories. This included backtranslating the following:

Training data

This model was trained on all PubMed abstracts categorized under Psychology and Psychiatry. As of March 1, this corresponded to approximately 3.2 million papers that contained abstract text. Of these 3.2 million papers, relevant sparse categories were back translated from english to french and from french to english to increase the representation of sparser mental health categories. This included backtranslating the following papers with the following categories:

  • Female
  • Adult
  • Middle Aged
  • Depressive Disorder
  • Risk Factors
  • Mental Disorders
  • Child, Preschool
  • Mental Health

In aggregate, this process added 557,980 additional papers to our training data.

Training procedure

Continued pretraining was on Psychology and Psychiatry PubMed papers for 10 epochs. Default parameters were used with the exception of gradient accumulation steps which was set at 4, with a per device train batch size of 32. 2 x Nvidia 3090's were used in the development of this model.

Eval results

To evaluate the utility of psych-search within the mental health domain, an evaluation task was constructed by finetuning psych-search for a task similar to BioASQ Task A. Here we perform large scale biomedical indexing using the MESH taxonomy associated with each paper underneath Psychology and Psychiatry. The evaluation metric is the micro F1 score across all second level descriptors under Psychology and Psychiatry. This corresponds to 38 different MESH categories used during evaluation.

bert-base-uncased SciBERT Scivocab Uncased Psych-Search
0.7348 0.7394 0.7415