| FROM nvidia/cuda:11.8.0-cudnn8-devel-ubuntu20.04 | |
| LABEL maintainer="Hugging Face" | |
| ENV PYTHONUNBUFFERED 1 | |
| EXPOSE 7860 | |
| ARG DEBIAN_FRONTEND=noninteractive | |
| # Use login shell to read variables from `~/.profile` (to pass dynamic created variables between RUN commands) | |
| SHELL ["sh", "-lc"] | |
| RUN apt update | |
| RUN apt --yes install build-essential | |
| RUN apt --yes install curl | |
| RUN curl -fsSL https://deb.nodesource.com/setup_18.x | bash - | |
| RUN apt --yes install nodejs | |
| # we need Rust | |
| RUN curl https://sh.rustup.rs -sSf | sh -s -- --default-toolchain nightly -y | |
| # configure PNPM | |
| RUN corepack enable | |
| RUN corepack prepare [email protected] --activate | |
| RUN apt --yes install git git-lfs libsndfile1-dev tesseract-ocr espeak-ng python3 python3-pip ffmpeg | |
| RUN git lfs install | |
| RUN python3 -m pip install --no-cache-dir --upgrade pip | |
| WORKDIR /tmp | |
| # ok! let's try to compile llama-node | |
| RUN git clone https://github.com/Atome-FE/llama-node.git | |
| WORKDIR /tmp/llama-node | |
| RUN git submodule update --init --recursive | |
| RUN pnpm install --ignore-scripts | |
| # let's try to build backends, to see | |
| RUN pnpm build:llama-cpp | |
| RUN pnpm:build | |
| WORKDIR /tmp/llama-node/packages/llama-cpp | |
| RUN pnpm build:cuda | |
| ENV LD_LIBRARY_PATH $LD_LIBRARY_PATH:$HOME/.llama-node | |
| RUN useradd -m -u 1000 user | |
| USER user | |
| ENV HOME=/home/user \ | |
| PATH=/home/user/.local/bin:$PATH | |
| WORKDIR $HOME/app | |
| # prepare to install the Node app | |
| COPY --chown=user package*.json . | |
| RUN pnpm --version | |
| RUN npm install | |
| # ok.. should be good? | |
| COPY --chown=user . . | |
| ADD --chown=user https://huggingface.co/TheBloke/airoboros-13b-gpt4-GGML/resolve/main/airoboros-13b-gpt4.ggmlv3.q4_0.bin models/airoboros-13b-gpt4.ggmlv3.q4_0.bin | |
| RUN python3 test.py | |
| CMD [ "pnpm", "run", "start" ] |