# ---------------------------------------------------- # 2. Inference function # ---------------------------------------------------- function predict(json_msg::String) # Expect: "x": [0.1, 0.2, …, 0.128] payload = JSON.parse(json_msg) x = Float32.(payload["x"]) |> gpu y = model(x) # GPU forward pass prob = softmax(y) |> cpu # back to CPU for serialization return JSON.json(Dict("prob" => vec(prob))) end
using Flux, CUDA, JSON, WebSockets, HTTP, Revise
# ---------------------------------------------------- # 4. Start the server (in a background task) # ---------------------------------------------------- @async run_ws(8081)
# ---------------------------------------------------- # 2. Inference function # ---------------------------------------------------- function predict(json_msg::String) # Expect: "x": [0.1, 0.2, …, 0.128] payload = JSON.parse(json_msg) x = Float32.(payload["x"]) |> gpu y = model(x) # GPU forward pass prob = softmax(y) |> cpu # back to CPU for serialization return JSON.json(Dict("prob" => vec(prob))) end
using Flux, CUDA, JSON, WebSockets, HTTP, Revise
# ---------------------------------------------------- # 4. Start the server (in a background task) # ---------------------------------------------------- @async run_ws(8081)
⬤ | We are offline | c |
| E-mail: |