Commit
·
acd252b
1
Parent(s):
6b01a9a
move order of commands
Browse files- Dockerfile +7 -12
Dockerfile
CHANGED
@@ -12,12 +12,19 @@ SHELL ["sh", "-lc"]
|
|
12 |
|
13 |
RUN apt update
|
14 |
|
|
|
|
|
15 |
RUN apt --yes install curl
|
16 |
|
17 |
RUN curl -fsSL https://deb.nodesource.com/setup_18.x | bash -
|
18 |
|
19 |
RUN apt --yes install nodejs
|
20 |
|
|
|
|
|
|
|
|
|
|
|
21 |
RUN corepack enable
|
22 |
RUN corepack prepare [email protected] --activate
|
23 |
|
@@ -39,18 +46,6 @@ COPY --chown=user package*.json .
|
|
39 |
|
40 |
RUN pnpm install
|
41 |
|
42 |
-
|
43 |
-
# OK, now the hell begins.. we need to build llama-node with CUDA support
|
44 |
-
|
45 |
-
# we need Rust
|
46 |
-
RUN curl https://sh.rustup.rs -sSf | sh -s -- --default-toolchain nightly -y
|
47 |
-
|
48 |
-
# we need PNP
|
49 |
-
# RUN curl -fsSL https://get.pnpm.io/install.sh | bash -
|
50 |
-
|
51 |
-
# we also need this (not sure we need musl-tools as it is for cross-compilation)
|
52 |
-
RUN apt --yes install build-essential musl-tools
|
53 |
-
|
54 |
# ok! let's try to compile llama-node
|
55 |
RUN git clone https://github.com/Atome-FE/llama-node.git
|
56 |
|
|
|
12 |
|
13 |
RUN apt update
|
14 |
|
15 |
+
RUN apt --yes install build-essential
|
16 |
+
|
17 |
RUN apt --yes install curl
|
18 |
|
19 |
RUN curl -fsSL https://deb.nodesource.com/setup_18.x | bash -
|
20 |
|
21 |
RUN apt --yes install nodejs
|
22 |
|
23 |
+
# we need Rust
|
24 |
+
RUN curl https://sh.rustup.rs -sSf | sh -s -- --default-toolchain nightly -y
|
25 |
+
|
26 |
+
# configure PNPM
|
27 |
+
|
28 |
RUN corepack enable
|
29 |
RUN corepack prepare [email protected] --activate
|
30 |
|
|
|
46 |
|
47 |
RUN pnpm install
|
48 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
49 |
# ok! let's try to compile llama-node
|
50 |
RUN git clone https://github.com/Atome-FE/llama-node.git
|
51 |
|