Skip to content

Need assistance - installed via bash...chatbox empty response #3595

Closed Answered by helf-charles
gl2007 asked this question in Q&A
Discussion options

You must be logged in to vote

So, with LLMs, the models are effectively just the probabilistic data bank. They don't do anything, in of themselves. They require an inference engine to load portions of the model into memory, and then use those portions of the model to transform user input into response output. The inference engine is a series of low-level processes meant to execute these transformations. Generally, efficiency is extremely important, as language models require truly massive numbers of calculations.

Thus, the faster and more lightweight the inference engine's architecture, the more rapidly it will be able to use the loaded portions of a language model to generate responses. Llama.cpp is basically the gol…

Replies: 2 comments 3 replies

Comment options

You must be logged in to vote
1 reply
@helf-charles
Comment options

Comment options

You must be logged in to vote
2 replies
@helf-charles
Comment options

Answer selected by gl2007
@gl2007
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants