File size: 1,450 Bytes
5300c16
 
6e64108
 
5300c16
 
 
 
 
 
65e9f20
5300c16
16660e8
afd2a60
7a26d32
afd2a60
 
 
2c1d107
cbab2c9
2cc6a74
 
 
 
 
 
cbab2c9
424fd94
2cc6a74
424fd94
6d85436
afd2a60
18847ff
4d53d1c
16660e8
707e4f7
 
5300c16
7cc6ee0
5300c16
707e4f7
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
FROM python:3.11

#ARG GBIF_DOWNLOAD_ID="0032228-231002084531237"
ARG GBIF_DOWNLOAD_ID="0237899-210914110416597"
WORKDIR /code

COPY ./requirements.txt /code/requirements.txt

RUN pip install --no-cache-dir --upgrade -r /code/requirements.txt

ADD https://docs.google.com/uc?export=download&id=17mI5W0qiGiBp_RV1jy3QR3KtN7Ah-1Ha /code/ihinst.db

# Download GBIF occurrences and prepare for use with datasette
RUN mkdir /data
ADD https://api.gbif.org/v1/occurrence/download/request/${GBIF_DOWNLOAD_ID}.zip /data/gbif-occs.zip
RUN ls -l /data
RUN unzip /data/gbif-occs.zip -d /data
RUN ls -l /data
COPY ./tab2csv.py /code/tab2csv.py

## Setup to parse collector names using Bionomia utils (reqs Ruby)
## Install ruby
#RUN \
#  apt-get update && \
#  apt-get install -y ruby
#RUN gem install dwc_agent

#COPY ./extractcollectorname.py /code/extractcollectorname.py
RUN python tab2csv.py --createcols /data/${GBIF_DOWNLOAD_ID}.csv /data/gbifocc.csv 
#RUN python extractcollectorname.py /data/gbifocc-temp.csv /data/gbifocc.csv 
RUN csvs-to-sqlite /data/gbifocc.csv /code/gbifocc.db
RUN ls -l /code
RUN sqlite-utils tables /code/gbifocc.db --counts
RUN sqlite-utils enable-fts /code/gbifocc.db gbifocc collectorNameAndNumber

#RUN sqlite-utils tables /code/ihinst.db --counts
RUN chmod 755 /code/gbifocc.db

COPY ./metadata.json /code/metadata.json

CMD ["datasette", "/code/gbifocc.db", "-m", "/code/metadata.json", "--host", "0.0.0.0", "--port", "7860"]