When I examin openops-localai docker container stdout, with docker logs
, I see this error:
12:34PM DBG Request received: {"model":"gpt4all-j","file":"","language":"","response_format":"","size":"","prompt":null,"instruction":"","input":null,"stop":null,"messages":[{"role":"system","content":"You are a helpful assistant."},{"role":"user","content":"Hello."}],"stream":true,"echo":false,"top_p":1,"top_k":0,"temperature":1,"max_tokens":0,"n":0,"batch":0,"f16":false,"ignore_eos":false,"repeat_penalty":0,"n_keep":0,"mirostat_eta":0,"mirostat_tau":0,"mirostat":0,"frequency_penalty":0,"tfz":0,"seed":0,"mode":0,"step":0,"typical_p":0}
12:34PM DBG Parameter Config: &{OpenAIRequest:{Model:gpt4all-j File: Language: ResponseFormat: Size: Prompt:<nil> Instruction: Input:<nil> Stop:<nil> Messages:[]
Stream:false Echo:false TopP:1 TopK:80 Temperature:1 Maxtokens:512 N:0 Batch:0 F16:false IgnoreEOS:false RepeatPenalty:0 Keep:0 MirostatETA:0 MirostatTAU:0 Mirostat:0 FrequencyPenalty:0 TFZ:0 Seed:0 Mode:0 Step:0 TypicalP:0} Name: StopWords:[] Cutstrings:[] TrimSpace:[] ContextSize:512 F16:false Threads:4 Debug:true Roles:map[] Embeddings:false Backend: TemplateConfig:{Completion: Chat: Edit:} MirostatETA:0 MirostatTAU:0 Mirostat:0 NGPULayers:0 MMap:false MMlock:false TensorSplit:
MainGPU: ImageGenerationAssets: PromptCachePath: PromptCacheAll:false PromptCacheRO:false PromptStrings:[] InputStrings:[] InputToken:[]}
12:34PM DBG Stream request received
12:34PM DBG Template found, input modified to: The prompt below is a question to answer, a task to complete, or a conversation to respond to; decide which and write an appropriate response
### Prompt:
You are a helpful assistant.
Hello.
### Response:
12:34PM DBG Loading model 'gpt4all-j' greedly
12:34PM DBG [llama] Attempting to load
12:34PM DBG Loading model llama from gpt4all-j
12:34PM DBG Sending chunk: {"object":"chat.completion.chunk","model":"gpt4all-j","choices":[{"delta":{"role":"assistant"}}],"usage":{"prompt_tokens":0,"completion_tokens":0,"total_tokens":0}}
12:34PM DBG Loading model in memory from file: /models/gpt4all-j
llama.cpp: loading model from /models/gpt4all-j
error loading model: unexpectedly reached end of file
llama_init_from_file: failed to load model
12:34PM DBG [llama] Fails: failed loading model
12:34PM DBG [gpt4all] Attempting to load
12:34PM DBG Loading model gpt4all from gpt4all-j
12:34PM DBG Loading model in memory from file: /models/gpt4all-j
12:34PM DBG [gpt4all] Loads OK
which indicates the model was not correctly loaded.
Would you know why?