- list_loras() : llama_ros::Llama
- list_loras_service_ : llama_ros::LlamaNode
- list_loras_service_callback() : llama_ros::LlamaNode
-- Llama() : llama_ros::Llama
+- Llama() : llama_ros::Llama
- llama : llama_ros::LlamaNode
- llama_client : llama_ros.langchain.llama_ros_common.LlamaROSCommon, llama_ros.langchain.llama_ros_embeddings.LlamaROSEmbeddings, llama_ros.langchain.llama_ros_reranker.LlamaROSReranker
- llama_init : llama_ros::Llama
diff --git a/functions_r.html b/functions_r.html
index 8b6e0f7d..3a3d425e 100644
--- a/functions_r.html
+++ b/functions_r.html
@@ -93,7 +93,7 @@ - r -
diff --git a/functions_s.html b/functions_s.html
index cd26605c..57e7815d 100644
--- a/functions_s.html
+++ b/functions_s.html
@@ -103,6 +103,7 @@ - s -
diff --git a/functions_vars.html b/functions_vars.html
index 2b2110db..832d49e0 100644
--- a/functions_vars.html
+++ b/functions_vars.html
@@ -302,6 +302,7 @@ - s -
diff --git a/llama_8hpp_source.html b/llama_8hpp_source.html
index e5665bda..b5d9b764 100644
--- a/llama_8hpp_source.html
+++ b/llama_8hpp_source.html
@@ -285,174 +285,175 @@
-
+
-
-
-
-
175 std::vector<llama_token>
tokenize(
const std::string &text,
bool add_bos,
-
176 bool special =
false);
-
177 std::string
detokenize(
const std::vector<llama_token> &tokens);
-
-
-
-
-
-
-
-
-
-
187 std::vector<llama_token>
-
188 truncate_tokens(
const std::vector<llama_token> &tokens,
int limit_size,
-
189 bool add_eos =
true);
-
-
191 int normalization = 2);
-
-
-
194 int normalization = 2);
-
195 float
rank_document(const std::string &query, const std::string &document);
-
-
197 const std::vector<std::string> &documents);
-
-
-
-
201 struct common_params_sampling sparams,
-
202 GenerateResponseCallback callbakc = nullptr,
-
203 std::vector<std::string> stop = {});
-
-
-
206 GenerateResponseCallback callbakc = nullptr,
-
207 std::vector<std::string> stop = {});
-
-
-
-
-
-
-
-
-
216 std::string
get_metadata(
const std::string &key,
size_t size);
-
-
218 const std::string &key,
size_t size);
-
-
-
-
-
-
224 const std::string &key,
size_t size);
-
-
-
-
-
-
-
-
-
-
-
-
-
237 struct llama_context *
ctx;
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
255 virtual void load_prompt(
const std::string &input_prompt,
bool add_pfx,
-
-
-
-
259 find_stop(std::vector<struct CompletionOutput> completion_result_list,
-
260 std::vector<std::string> stopping_words);
-
-
262 find_stop_word(std::vector<struct CompletionOutput> completion_result_list,
-
263 std::string stopping_word);
-
-
-
-
-
-
269 bool eval(std::vector<llama_token> tokens);
-
270 bool eval(
struct llama_batch batch);
-
-
272 std::vector<struct TokenProb>
get_probs();
-
-
-
-
-
-
+
+
173 bool initial_reset =
true);
+
+
+
176 std::vector<llama_token>
tokenize(
const std::string &text,
bool add_bos,
+
177 bool special =
false);
+
178 std::string
detokenize(
const std::vector<llama_token> &tokens);
+
+
180 virtual void reset();
+
+
+
+
+
+
+
+
188 std::vector<llama_token>
+
189 truncate_tokens(
const std::vector<llama_token> &tokens,
int limit_size,
+
190 bool add_eos =
true);
+
+
192 int normalization = 2);
+
+
+
195 int normalization = 2);
+
196 float
rank_document(const std::string &query, const std::string &document);
+
+
198 const std::vector<std::string> &documents);
+
+
+
+
202 struct common_params_sampling sparams,
+
203 GenerateResponseCallback callbakc = nullptr,
+
204 std::vector<std::string> stop = {});
+
+
+
207 GenerateResponseCallback callbakc = nullptr,
+
208 std::vector<std::string> stop = {});
+
+
+
+
+
+
+
+
+
217 std::string
get_metadata(
const std::string &key,
size_t size);
+
+
219 const std::string &key,
size_t size);
+
+
+
+
+
+
225 const std::string &key,
size_t size);
+
+
+
+
+
+
+
+
+
+
+
+
+
238 struct llama_context *
ctx;
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
+
256 virtual void load_prompt(
const std::string &input_prompt,
bool add_pfx,
+
+
+
+
260 find_stop(std::vector<struct CompletionOutput> completion_result_list,
+
261 std::vector<std::string> stopping_words);
+
+
263 find_stop_word(std::vector<struct CompletionOutput> completion_result_list,
+
264 std::string stopping_word);
+
+
+
+
+
+
270 virtual bool eval(std::vector<llama_token> tokens);
+
271 bool eval(
struct llama_batch batch);
+
+
273 std::vector<struct TokenProb>
get_probs();
+
+
+
+
+
+
-
-
+
+
-
-
-StopType find_stop(std::vector< struct CompletionOutput > completion_result_list, std::vector< std::string > stopping_words)
Definition llama.cpp:787
-std::string detokenize(const std::vector< llama_token > &tokens)
Definition llama.cpp:386
-int get_n_embd()
Definition llama.hpp:213
-int get_n_ctx()
Definition llama.hpp:211
-int32_t n_consumed
Definition llama.hpp:252
-std::string system_prompt
Definition llama.hpp:245
-int32_t ga_i
Definition llama.hpp:253
-struct Metadata get_metadata()
Definition llama.cpp:200
-llama_token get_token_eos()
Definition llama.hpp:230
-struct EmbeddingsOuput generate_embeddings(const std::string &input_prompt, int normalization=2)
Definition llama.cpp:473
-int get_int_metadata(const std::string &key, size_t size)
Definition llama.cpp:178
-StopType find_stop_word(std::vector< struct CompletionOutput > completion_result_list, std::string stopping_word)
Definition llama.cpp:860
-struct common_sampler * sampler
Definition llama.hpp:240
-std::vector< llama_token > tokenize(const std::string &text, bool add_bos, bool special=false)
Definition llama.cpp:380
-int32_t n_past
Definition llama.hpp:251
-std::recursive_mutex mutex
Definition llama.hpp:277
-void cancel()
Definition llama.cpp:398
-Llama(const struct common_params ¶ms, std::string system_prompt="")
Definition llama.cpp:37
-bool is_embedding()
Definition llama.hpp:227
-const struct llama_model * get_model()
Definition llama.hpp:210
-struct ggml_threadpool * threadpool_batch
Definition llama.hpp:242
-std::vector< float > rank_documents(const std::string &query, const std::vector< std::string > &documents)
Definition llama.cpp:529
-int get_n_vocab()
Definition llama.hpp:214
-void reset()
Definition llama.cpp:348
-float get_float_metadata(const std::string &key, size_t size)
Definition llama.cpp:189
-bool canceled
Definition llama.hpp:246
-std::vector< llama_token > truncate_tokens(const std::vector< llama_token > &tokens, int limit_size, bool add_eos=true)
Definition llama.cpp:482
-struct llama_model * model
Definition llama.hpp:238
-virtual void load_prompt(const std::string &input_prompt, bool add_pfx, bool add_sfx)
Definition llama.cpp:735
-std::vector< llama_token > prompt_tokens
Definition llama.hpp:248
-struct common_init_result llama_init
Definition llama.hpp:236
-bool eval_system_prompt()
Definition llama.cpp:894
-bool is_reranking()
Definition llama.hpp:228
-struct common_params params
Definition llama.hpp:233
-virtual bool eval_prompt()
Definition llama.cpp:909
-std::vector< struct LoRA > list_loras()
Definition llama.cpp:563
-bool add_bos_token()
Definition llama.hpp:229
-bool eval(std::vector< llama_token > tokens)
Definition llama.cpp:940
-bool eval_token(llama_token token)
Definition llama.cpp:936
-virtual ~Llama()
Definition llama.cpp:131
-llama_utils::Spinner spinner
Definition llama.hpp:247
-void update_loras(std::vector< struct LoRA > loras)
Definition llama.cpp:583
-struct ResponseOutput generate_response(const std::string &input_prompt, struct common_params_sampling sparams, GenerateResponseCallback callbakc=nullptr, std::vector< std::string > stop={})
Definition llama.cpp:629
-const struct llama_context * get_ctx()
Definition llama.hpp:209
-int get_n_ctx_train()
Definition llama.hpp:212
-std::vector< struct TokenProb > get_probs()
Definition llama.cpp:1031
-struct llama_context * ctx
Definition llama.hpp:237
-struct CompletionOutput sample()
Definition llama.cpp:1048
-std::vector< common_lora_adapter_info > lora_adapters
Definition llama.hpp:239
-std::string format_chat_prompt(std::vector< struct common_chat_msg > chat_msgs, bool add_ass)
Definition llama.cpp:553
-struct ggml_threadpool * threadpool
Definition llama.hpp:241
-float rank_document(const std::string &query, const std::string &document)
Definition llama.cpp:506
+
+
+StopType find_stop(std::vector< struct CompletionOutput > completion_result_list, std::vector< std::string > stopping_words)
Definition llama.cpp:790
+std::string detokenize(const std::vector< llama_token > &tokens)
Definition llama.cpp:389
+int get_n_embd()
Definition llama.hpp:214
+int get_n_ctx()
Definition llama.hpp:212
+int32_t n_consumed
Definition llama.hpp:253
+std::string system_prompt
Definition llama.hpp:246
+int32_t ga_i
Definition llama.hpp:254
+struct Metadata get_metadata()
Definition llama.cpp:235
+llama_token get_token_eos()
Definition llama.hpp:231
+Llama(const struct common_params ¶ms, std::string system_prompt="", bool initial_reset=true)
Definition llama.cpp:37
+struct EmbeddingsOuput generate_embeddings(const std::string &input_prompt, int normalization=2)
Definition llama.cpp:476
+int get_int_metadata(const std::string &key, size_t size)
Definition llama.cpp:213
+StopType find_stop_word(std::vector< struct CompletionOutput > completion_result_list, std::string stopping_word)
Definition llama.cpp:863
+struct common_sampler * sampler
Definition llama.hpp:241
+std::vector< llama_token > tokenize(const std::string &text, bool add_bos, bool special=false)
Definition llama.cpp:383
+int32_t n_past
Definition llama.hpp:252
+std::recursive_mutex mutex
Definition llama.hpp:278
+void cancel()
Definition llama.cpp:401
+bool is_embedding()
Definition llama.hpp:228
+const struct llama_model * get_model()
Definition llama.hpp:211
+struct ggml_threadpool * threadpool_batch
Definition llama.hpp:243
+std::vector< float > rank_documents(const std::string &query, const std::vector< std::string > &documents)
Definition llama.cpp:532
+int get_n_vocab()
Definition llama.hpp:215
+virtual void reset()
Definition llama.cpp:160
+float get_float_metadata(const std::string &key, size_t size)
Definition llama.cpp:224
+bool canceled
Definition llama.hpp:247
+std::vector< llama_token > truncate_tokens(const std::vector< llama_token > &tokens, int limit_size, bool add_eos=true)
Definition llama.cpp:485
+struct llama_model * model
Definition llama.hpp:239
+virtual void load_prompt(const std::string &input_prompt, bool add_pfx, bool add_sfx)
Definition llama.cpp:738
+std::vector< llama_token > prompt_tokens
Definition llama.hpp:249
+struct common_init_result llama_init
Definition llama.hpp:237
+bool eval_system_prompt()
Definition llama.cpp:897
+bool is_reranking()
Definition llama.hpp:229
+struct common_params params
Definition llama.hpp:234
+virtual bool eval_prompt()
Definition llama.cpp:912
+std::vector< struct LoRA > list_loras()
Definition llama.cpp:566
+bool add_bos_token()
Definition llama.hpp:230
+virtual bool eval(std::vector< llama_token > tokens)
Definition llama.cpp:944
+bool eval_token(llama_token token)
Definition llama.cpp:940
+virtual ~Llama()
Definition llama.cpp:134
+llama_utils::Spinner spinner
Definition llama.hpp:248
+void update_loras(std::vector< struct LoRA > loras)
Definition llama.cpp:586
+struct ResponseOutput generate_response(const std::string &input_prompt, struct common_params_sampling sparams, GenerateResponseCallback callbakc=nullptr, std::vector< std::string > stop={})
Definition llama.cpp:632
+const struct llama_context * get_ctx()
Definition llama.hpp:210
+int get_n_ctx_train()
Definition llama.hpp:213
+std::vector< struct TokenProb > get_probs()
Definition llama.cpp:1035
+struct llama_context * ctx
Definition llama.hpp:238
+struct CompletionOutput sample()
Definition llama.cpp:1052
+std::vector< common_lora_adapter_info > lora_adapters
Definition llama.hpp:240
+std::string format_chat_prompt(std::vector< struct common_chat_msg > chat_msgs, bool add_ass)
Definition llama.cpp:556
+struct ggml_threadpool * threadpool
Definition llama.hpp:242
+float rank_document(const std::string &query, const std::string &document)
Definition llama.cpp:509
Definition spinner.hpp:31
std::function< void(struct CompletionOutput)> GenerateResponseCallback
Definition llama.hpp:167
diff --git a/llava_8hpp_source.html b/llava_8hpp_source.html
index 4760de49..2bede218 100644
--- a/llava_8hpp_source.html
+++ b/llava_8hpp_source.html
@@ -147,45 +147,51 @@
-
- 54 struct llava_image_embed *
-
-
-
- 58 void load_prompt(
const std::string &input_prompt,
bool add_pfx,
- 59 bool add_sfx)
override;
-
-
-
-
-
-
-
-
-
-
-
+ 53 void reset()
override;
+
+ 55 struct llava_image_embed *
+
+
+
+ 59 void load_prompt(
const std::string &input_prompt,
bool add_pfx,
+ 60 bool add_sfx)
override;
+
+
+ 63 bool eval(std::vector<llama_token> tokens)
override;
+
+
+
+
+
+
+
+
+
+
-
-
+
+
-
-
+
+
-std::string system_prompt
Definition llama.hpp:245
-struct common_params params
Definition llama.hpp:233
-struct LlavaParams llava_params
Definition llava.hpp:65
-struct llava_image_embed * image_embed
Definition llava.hpp:63
-bool load_image(std::string base64_str)
Definition llava.cpp:102
-bool eval_image(struct llava_image_embed *image_embed)
Definition llava.cpp:148
-void free_image()
Definition llava.cpp:117
-struct clip_ctx * ctx_clip
Definition llava.hpp:64
-~Llava()
Definition llava.cpp:45
-bool eval_prompt()
Definition llava.cpp:182
-void load_prompt(const std::string &input_prompt, bool add_pfx, bool add_sfx) override
Definition llava.cpp:55
+std::string system_prompt
Definition llama.hpp:246
+struct common_params params
Definition llama.hpp:234
+struct LlavaParams llava_params
Definition llava.hpp:67
+bool eval(std::vector< llama_token > tokens) override
Definition llava.cpp:268
+struct llava_image_embed * image_embed
Definition llava.hpp:65
+bool load_image(std::string base64_str)
Definition llava.cpp:115
+bool eval_image(struct llava_image_embed *image_embed)
Definition llava.cpp:161
+void reset() override
Definition llava.cpp:56
+void free_image()
Definition llava.cpp:130
+struct clip_ctx * ctx_clip
Definition llava.hpp:66
+~Llava()
Definition llava.cpp:49
+int st_pos_id
Definition llava.hpp:72
+bool eval_prompt()
Definition llava.cpp:234
+void load_prompt(const std::string &input_prompt, bool add_pfx, bool add_sfx) override
Definition llava.cpp:68
Llava(const struct common_params ¶ms, const struct LlavaParams &llava_params, std::string system_prompt="")
Definition llava.cpp:35
-struct llava_image_embed * base64_image_to_embed(const std::string &base64_str)
Definition llava.cpp:125
-int image_pose
Definition llava.hpp:69
+struct llava_image_embed * base64_image_to_embed(const std::string &base64_str)
Definition llava.cpp:138
+int image_pose
Definition llava.hpp:71
diff --git a/search/all_12.js b/search/all_12.js
index 5635eddc..2489b613 100644
--- a/search/all_12.js
+++ b/search/all_12.js
@@ -18,7 +18,7 @@ var searchData=
['reranker_15',['llama_ros (LLM + RAG + Reranker)',['../index.html#autotoc_md28',1,'']]],
['reranking_20demo_16',['Reranking Demo',['../index.html#autotoc_md33',1,'']]],
['reranking_20stream_17',['Full Demo (LLM + chat template + RAG + Reranking + Stream)',['../index.html#autotoc_md36',1,'']]],
- ['reset_18',['reset',['../classllama__ros_1_1Llama.html#a7c2cd8a9a1361837e6f7661a4f63910c',1,'llama_ros::Llama']]],
+ ['reset_18',['reset',['../classllama__ros_1_1Llama.html#a7c2cd8a9a1361837e6f7661a4f63910c',1,'llama_ros::Llama::reset()'],['../classllava__ros_1_1Llava.html#a5708e11372b049375bd981443d2d8ac9',1,'llava_ros::Llava::reset()']]],
['response_19',['Generate Response',['../index.html#autotoc_md20',1,'']]],
['response_20llava_20',['Generate Response (llava)',['../index.html#autotoc_md21',1,'']]],
['responseoutput_21',['ResponseOutput',['../structllama__ros_1_1ResponseOutput.html',1,'llama_ros']]],
diff --git a/search/all_13.js b/search/all_13.js
index fede986a..3156e42f 100644
--- a/search/all_13.js
+++ b/search/all_13.js
@@ -26,8 +26,9 @@ var searchData=
['spinner_23',['spinner',['../classllama__ros_1_1Llama.html#ac5758af17e629830eb7dad343115f93b',1,'llama_ros::Llama::spinner'],['../classllama__utils_1_1Spinner.html#ab51213c449e3926314d35d8c7183863e',1,'llama_utils::Spinner::spinner']]],
['spinner_2ehpp_24',['spinner.hpp',['../spinner_8hpp.html',1,'']]],
['splits_25',['splits',['../namespacellama__demos_1_1llama__rag__demo__node.html#a34fb77601e84c7dec99c9fb16734d55a',1,'llama_demos::llama_rag_demo_node']]],
- ['stop_26',['stop',['../structllama__ros_1_1ResponseOutput.html#a2b4cf7d6cb0d2d5a02c3b605fe5e7729',1,'llama_ros::ResponseOutput']]],
- ['stoptype_27',['StopType',['../namespacellama__ros.html#aa0ac7656a188e47404db8ced957790d0',1,'llama_ros']]],
- ['stream_28',['Stream',['../index.html#autotoc_md36',1,'Full Demo (LLM + chat template + RAG + Reranking + Stream)'],['../index.html#autotoc_md24',1,'llama_ros (Stream)']]],
- ['system_5fprompt_29',['system_prompt',['../classllama__ros_1_1Llama.html#a2159f2355b8f059933f59d1138e1d3c2',1,'llama_ros::Llama::system_prompt'],['../structllama__utils_1_1LlamaParams.html#a89a676517fb802fced98ede2c3796a01',1,'llama_utils::LlamaParams::system_prompt']]]
+ ['st_5fpos_5fid_26',['st_pos_id',['../classllava__ros_1_1Llava.html#a818a8f591e619f89aa77241742578aa9',1,'llava_ros::Llava']]],
+ ['stop_27',['stop',['../structllama__ros_1_1ResponseOutput.html#a2b4cf7d6cb0d2d5a02c3b605fe5e7729',1,'llama_ros::ResponseOutput']]],
+ ['stoptype_28',['StopType',['../namespacellama__ros.html#aa0ac7656a188e47404db8ced957790d0',1,'llama_ros']]],
+ ['stream_29',['Stream',['../index.html#autotoc_md36',1,'Full Demo (LLM + chat template + RAG + Reranking + Stream)'],['../index.html#autotoc_md24',1,'llama_ros (Stream)']]],
+ ['system_5fprompt_30',['system_prompt',['../classllama__ros_1_1Llama.html#a2159f2355b8f059933f59d1138e1d3c2',1,'llama_ros::Llama::system_prompt'],['../structllama__utils_1_1LlamaParams.html#a89a676517fb802fced98ede2c3796a01',1,'llama_utils::LlamaParams::system_prompt']]]
];
diff --git a/search/all_6.js b/search/all_6.js
index 6a273265..be21f77f 100644
--- a/search/all_6.js
+++ b/search/all_6.js
@@ -11,7 +11,7 @@ var searchData=
['entry_5fpoints_8',['entry_points',['../namespacesetup.html#ada7058afc98897f073d3f3b8b9157059',1,'setup']]],
['eos_5ftoken_5fid_9',['eos_token_id',['../structllama__ros_1_1Metadata_1_1TokenizerInfo.html#a0893a43ad19ca68221d909a35be54ca1',1,'llama_ros::Metadata::TokenizerInfo']]],
['error_10',['ERROR',['../namespacellama__utils.html#a386024dced31ce4ae6a2bb5e0b97fefda1029b57592d6b0f89d84cafcba4079d9',1,'llama_utils']]],
- ['eval_11',['eval',['../classllama__ros_1_1Llama.html#ab41df18b60dc8bac07cc1ae306082f10',1,'llama_ros::Llama::eval(std::vector< llama_token > tokens)'],['../classllama__ros_1_1Llama.html#a8a16e099f81537e166018abd08df63e3',1,'llama_ros::Llama::eval(struct llama_batch batch)']]],
+ ['eval_11',['eval',['../classllama__ros_1_1Llama.html#ab41df18b60dc8bac07cc1ae306082f10',1,'llama_ros::Llama::eval(std::vector< llama_token > tokens)'],['../classllama__ros_1_1Llama.html#a8a16e099f81537e166018abd08df63e3',1,'llama_ros::Llama::eval(struct llama_batch batch)'],['../classllava__ros_1_1Llava.html#a1629b8fcd0726b58467d7aee7687b849',1,'llava_ros::Llava::eval(std::vector< llama_token > tokens) override']]],
['eval_5fimage_12',['eval_image',['../classllava__ros_1_1Llava.html#a5116423210df12eeaa0c07d9cda72f76',1,'llava_ros::Llava']]],
['eval_5fprompt_13',['eval_prompt',['../classllama__ros_1_1Llama.html#a9de8440b62c2e93c03a0453e6c8c2f3d',1,'llama_ros::Llama::eval_prompt()'],['../classllama__ros_1_1Llama.html#a04b5be2cefb14b0e824d47ae432ea720',1,'llama_ros::Llama::eval_prompt(std::vector< llama_token > prompt_tokens)'],['../classllava__ros_1_1Llava.html#a9f815d7b2bc9312644f18ce859fe27fc',1,'llava_ros::Llava::eval_prompt()']]],
['eval_5fsystem_5fprompt_14',['eval_system_prompt',['../classllama__ros_1_1Llama.html#a9791d8dcd3c72b3159fb3e968d739c61',1,'llama_ros::Llama']]],
diff --git a/search/all_c.js b/search/all_c.js
index 402efb9e..969bfc90 100644
--- a/search/all_c.js
+++ b/search/all_c.js
@@ -16,7 +16,7 @@ var searchData=
['list_5floras_13',['list_loras',['../classllama__ros_1_1Llama.html#aa17aaf9615a03f65ead934cd8a3c33e0',1,'llama_ros::Llama']]],
['list_5floras_5fservice_5f_14',['list_loras_service_',['../classllama__ros_1_1LlamaNode.html#ae48613d70306d3100ea7d332b0a5040c',1,'llama_ros::LlamaNode']]],
['list_5floras_5fservice_5fcallback_15',['list_loras_service_callback',['../classllama__ros_1_1LlamaNode.html#a5a79bd20864384f74c8d6faa9a8d2386',1,'llama_ros::LlamaNode']]],
- ['llama_16',['Llama',['../classllama__ros_1_1Llama.html',1,'llama_ros::Llama'],['../classllama__ros_1_1Llama.html#a67b1d1838dc98f67c03051cc2067b290',1,'llama_ros::Llama::Llama()']]],
+ ['llama_16',['Llama',['../classllama__ros_1_1Llama.html',1,'llama_ros::Llama'],['../classllama__ros_1_1Llama.html#a36707047a228c3a9278337b1426938c3',1,'llama_ros::Llama::Llama()']]],
['llama_17',['llama',['../classllama__ros_1_1LlamaNode.html#add09929e4bb019eac008da2bdd4b8427',1,'llama_ros::LlamaNode']]],
['llama_2ecpp_18',['llama.cpp',['../llama_8cpp.html',1,'']]],
['llama_2ehpp_19',['llama.hpp',['../llama_8hpp.html',1,'']]],
diff --git a/search/functions_5.js b/search/functions_5.js
index 4715811d..0ce0bcbf 100644
--- a/search/functions_5.js
+++ b/search/functions_5.js
@@ -2,7 +2,7 @@ var searchData=
[
['embed_5fdocuments_0',['embed_documents',['../classllama__ros_1_1langchain_1_1llama__ros__embeddings_1_1LlamaROSEmbeddings.html#a2f2f921bef86561a6f52e83edc4fc50c',1,'llama_ros::langchain::llama_ros_embeddings::LlamaROSEmbeddings']]],
['embed_5fquery_1',['embed_query',['../classllama__ros_1_1langchain_1_1llama__ros__embeddings_1_1LlamaROSEmbeddings.html#a6af644d86edae69a1e59f72ef84726e9',1,'llama_ros::langchain::llama_ros_embeddings::LlamaROSEmbeddings']]],
- ['eval_2',['eval',['../classllama__ros_1_1Llama.html#ab41df18b60dc8bac07cc1ae306082f10',1,'llama_ros::Llama::eval(std::vector< llama_token > tokens)'],['../classllama__ros_1_1Llama.html#a8a16e099f81537e166018abd08df63e3',1,'llama_ros::Llama::eval(struct llama_batch batch)']]],
+ ['eval_2',['eval',['../classllama__ros_1_1Llama.html#ab41df18b60dc8bac07cc1ae306082f10',1,'llama_ros::Llama::eval(std::vector< llama_token > tokens)'],['../classllama__ros_1_1Llama.html#a8a16e099f81537e166018abd08df63e3',1,'llama_ros::Llama::eval(struct llama_batch batch)'],['../classllava__ros_1_1Llava.html#a1629b8fcd0726b58467d7aee7687b849',1,'llava_ros::Llava::eval(std::vector< llama_token > tokens) override']]],
['eval_5fimage_3',['eval_image',['../classllava__ros_1_1Llava.html#a5116423210df12eeaa0c07d9cda72f76',1,'llava_ros::Llava']]],
['eval_5fprompt_4',['eval_prompt',['../classllama__ros_1_1Llama.html#a9de8440b62c2e93c03a0453e6c8c2f3d',1,'llama_ros::Llama::eval_prompt()'],['../classllama__ros_1_1Llama.html#a04b5be2cefb14b0e824d47ae432ea720',1,'llama_ros::Llama::eval_prompt(std::vector< llama_token > prompt_tokens)'],['../classllava__ros_1_1Llava.html#a9f815d7b2bc9312644f18ce859fe27fc',1,'llava_ros::Llava::eval_prompt()']]],
['eval_5fsystem_5fprompt_5',['eval_system_prompt',['../classllama__ros_1_1Llama.html#a9791d8dcd3c72b3159fb3e968d739c61',1,'llama_ros::Llama']]],
diff --git a/search/functions_a.js b/search/functions_a.js
index ea52cdaa..eef1910d 100644
--- a/search/functions_a.js
+++ b/search/functions_a.js
@@ -3,7 +3,7 @@ var searchData=
['launch_5fllm_0',['launch_llm',['../namespacellama__cli_1_1api.html#ad208ac7099874aa35e0670e6237fe312',1,'llama_cli::api']]],
['list_5floras_1',['list_loras',['../classllama__ros_1_1Llama.html#aa17aaf9615a03f65ead934cd8a3c33e0',1,'llama_ros::Llama']]],
['list_5floras_5fservice_5fcallback_2',['list_loras_service_callback',['../classllama__ros_1_1LlamaNode.html#a5a79bd20864384f74c8d6faa9a8d2386',1,'llama_ros::LlamaNode']]],
- ['llama_3',['Llama',['../classllama__ros_1_1Llama.html#a67b1d1838dc98f67c03051cc2067b290',1,'llama_ros::Llama']]],
+ ['llama_3',['Llama',['../classllama__ros_1_1Llama.html#a36707047a228c3a9278337b1426938c3',1,'llama_ros::Llama']]],
['llamanode_4',['LlamaNode',['../classllama__ros_1_1LlamaNode.html#adc90c82214422e4d5c11bfc480e5ff93',1,'llama_ros::LlamaNode']]],
['llava_5',['Llava',['../classllava__ros_1_1Llava.html#ad19043ea9fed380a9e9b8e7bbbe8393c',1,'llava_ros::Llava']]],
['llavanode_6',['LlavaNode',['../classllava__ros_1_1LlavaNode.html#aeeaefad9b9adeebe6d58e668693f21e4',1,'llava_ros::LlavaNode']]],
diff --git a/search/functions_e.js b/search/functions_e.js
index 58529819..4f7bf3c1 100644
--- a/search/functions_e.js
+++ b/search/functions_e.js
@@ -5,5 +5,5 @@ var searchData=
['replace_5fall_2',['replace_all',['../llama__params_8cpp.html#a6022dd052437b9f97f6c1ffc13f5e518',1,'llama_params.cpp']]],
['rerank_5fdocuments_3',['rerank_documents',['../classllama__ros_1_1llama__client__node_1_1LlamaClientNode.html#a14d2cbeae980c6811349e5c946f5a452',1,'llama_ros::llama_client_node::LlamaClientNode']]],
['rerank_5fdocuments_5fservice_5fcallback_4',['rerank_documents_service_callback',['../classllama__ros_1_1LlamaNode.html#a1cf41258ed6f008022420773eb208c61',1,'llama_ros::LlamaNode']]],
- ['reset_5',['reset',['../classllama__ros_1_1Llama.html#a7c2cd8a9a1361837e6f7661a4f63910c',1,'llama_ros::Llama']]]
+ ['reset_5',['reset',['../classllama__ros_1_1Llama.html#a7c2cd8a9a1361837e6f7661a4f63910c',1,'llama_ros::Llama::reset()'],['../classllava__ros_1_1Llava.html#a5708e11372b049375bd981443d2d8ac9',1,'llava_ros::Llava::reset()']]]
];
diff --git a/search/variables_12.js b/search/variables_12.js
index 489989aa..54239e6c 100644
--- a/search/variables_12.js
+++ b/search/variables_12.js
@@ -13,6 +13,7 @@ var searchData=
['size_5flabel_10',['size_label',['../structllama__ros_1_1Metadata_1_1GeneralInfo.html#a06e4db77275f710862e0976e88cdb3bc',1,'llama_ros::Metadata::GeneralInfo']]],
['spinner_11',['spinner',['../classllama__ros_1_1Llama.html#ac5758af17e629830eb7dad343115f93b',1,'llama_ros::Llama::spinner'],['../classllama__utils_1_1Spinner.html#ab51213c449e3926314d35d8c7183863e',1,'llama_utils::Spinner::spinner']]],
['splits_12',['splits',['../namespacellama__demos_1_1llama__rag__demo__node.html#a34fb77601e84c7dec99c9fb16734d55a',1,'llama_demos::llama_rag_demo_node']]],
- ['stop_13',['stop',['../structllama__ros_1_1ResponseOutput.html#a2b4cf7d6cb0d2d5a02c3b605fe5e7729',1,'llama_ros::ResponseOutput']]],
- ['system_5fprompt_14',['system_prompt',['../classllama__ros_1_1Llama.html#a2159f2355b8f059933f59d1138e1d3c2',1,'llama_ros::Llama::system_prompt'],['../structllama__utils_1_1LlamaParams.html#a89a676517fb802fced98ede2c3796a01',1,'llama_utils::LlamaParams::system_prompt']]]
+ ['st_5fpos_5fid_13',['st_pos_id',['../classllava__ros_1_1Llava.html#a818a8f591e619f89aa77241742578aa9',1,'llava_ros::Llava']]],
+ ['stop_14',['stop',['../structllama__ros_1_1ResponseOutput.html#a2b4cf7d6cb0d2d5a02c3b605fe5e7729',1,'llama_ros::ResponseOutput']]],
+ ['system_5fprompt_15',['system_prompt',['../classllama__ros_1_1Llama.html#a2159f2355b8f059933f59d1138e1d3c2',1,'llama_ros::Llama::system_prompt'],['../structllama__utils_1_1LlamaParams.html#a89a676517fb802fced98ede2c3796a01',1,'llama_utils::LlamaParams::system_prompt']]]
];