Added a cleanup failed button for Processing Queue in the Knowledge Base
since documents that fail to process tend to get stuck and then can't be
cleared.
Fixed the ingestion of documents for OpenAI servers.
Updated some text in the chat and chat settings since user will need to
manually download models when using a non-ollama remote gpu server.
Exisiting Ollama API support still functions as before. OpenAI vs
Ollama API mostly have the same features, however model file size is not
support with OpenAI's API so when a user chooses one of those then the
models will just show up as the model name without the size.
`npm install openai` triggered some updates in admin/package-lock.json
such as adding many instances of "dev: true".
This further enhances the user's ability to run the LLM on a different
host.