Unverified Commit 877da82b authored by takatost's avatar takatost Committed by GitHub

feat: cache huggingface gpt2 tokenizer files (#1138)

parent 66376290
...@@ -26,7 +26,7 @@ EXPOSE 5001 ...@@ -26,7 +26,7 @@ EXPOSE 5001
WORKDIR /app/api WORKDIR /app/api
RUN apt-get update \ RUN apt-get update \
&& apt-get install -y --no-install-recommends bash curl wget vim nodejs \ && apt-get install -y --no-install-recommends bash curl wget vim nodejs \
&& apt-get autoremove \ && apt-get autoremove \
&& rm -rf /var/lib/apt/lists/* && rm -rf /var/lib/apt/lists/*
...@@ -34,6 +34,8 @@ RUN apt-get update \ ...@@ -34,6 +34,8 @@ RUN apt-get update \
COPY --from=base /pkg /usr/local COPY --from=base /pkg /usr/local
COPY . /app/api/ COPY . /app/api/
RUN python -c "from transformers import GPT2TokenizerFast; GPT2TokenizerFast.from_pretrained('gpt2')"
COPY docker/entrypoint.sh /entrypoint.sh COPY docker/entrypoint.sh /entrypoint.sh
RUN chmod +x /entrypoint.sh RUN chmod +x /entrypoint.sh
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment