Class: Ollama
This class both implements the LLM and Embedding interfaces.
Hierarchy
-
↳
Ollama
Implements
Constructors
constructor
• new Ollama(params
): Ollama
Parameters
Name | Type |
---|---|
params | OllamaParams |
Returns
Overrides
Defined in
packages/core/src/llm/ollama.ts:57
Properties
embedBatchSize
• embedBatchSize: number
= DEFAULT_EMBED_BATCH_SIZE
Inherited from
Defined in
packages/core/src/embeddings/types.ts:11
hasStreaming
• Readonly
hasStreaming: true
Defined in
packages/core/src/llm/ollama.ts:45
model
• model: string
Defined in
packages/core/src/llm/ollama.ts:48
options
• options: Partial
<Omit
<Options
, "temperature"
| "top_p"
| "num_ctx"
>> & Pick
<Options
, "temperature"
| "top_p"
| "num_ctx"
>
Defined in
packages/core/src/llm/ollama.ts:50
Accessors
metadata
• get
metadata(): LLMMetadata
Returns
Implementation of
Defined in
packages/core/src/llm/ollama.ts:68
Methods
chat
▸ chat(params
): Promise
<AsyncIterable
<ChatResponseChunk
>>
Get a chat response from the LLM
Parameters
Name | Type |
---|---|
params | LLMChatParamsStreaming <object , object > |
Returns
Promise
<AsyncIterable
<ChatResponseChunk
>>