Skip to main content

Class: Gemini

ToolCallLLM for Gemini

Extends

Constructors

new Gemini()

new Gemini(init?): Gemini

Parameters

init?: GeminiConfig

Returns

Gemini

Overrides

ToolCallLLM.constructor

Defined in

packages/llamaindex/src/llm/gemini/base.ts:194

Properties

maxTokens?

optional maxTokens: number

Defined in

packages/llamaindex/src/llm/gemini/base.ts:191


model

model: GEMINI_MODEL

Defined in

packages/llamaindex/src/llm/gemini/base.ts:188


session

session: IGeminiSession

Defined in

packages/llamaindex/src/llm/gemini/base.ts:192


temperature

temperature: number

Defined in

packages/llamaindex/src/llm/gemini/base.ts:189


topP

topP: number

Defined in

packages/llamaindex/src/llm/gemini/base.ts:190

Accessors

metadata

get metadata(): LLMMetadata

Returns

LLMMetadata

Overrides

ToolCallLLM.metadata

Defined in

packages/llamaindex/src/llm/gemini/base.ts:207


supportToolCall

get supportToolCall(): boolean

Returns

boolean

Overrides

ToolCallLLM.supportToolCall

Defined in

packages/llamaindex/src/llm/gemini/base.ts:203

Methods

chat()

chat(params)

chat(params): Promise<GeminiChatStreamResponse>

Get a chat response from the LLM

Parameters

params: GeminiChatParamsStreaming

Returns

Promise<GeminiChatStreamResponse>

Overrides

ToolCallLLM.chat

Defined in

packages/llamaindex/src/llm/gemini/base.ts:288

chat(params)

chat(params): Promise<GeminiChatNonStreamResponse>

Get a chat response from the LLM

Parameters

params: GeminiChatParamsNonStreaming

Returns

Promise<GeminiChatNonStreamResponse>

Overrides

ToolCallLLM.chat

Defined in

packages/llamaindex/src/llm/gemini/base.ts:289


complete()

complete(params)

complete(params): Promise<AsyncIterable<CompletionResponse, any, any>>

Get a prompt completion from the LLM

Parameters

params: LLMCompletionParamsStreaming

Returns

Promise<AsyncIterable<CompletionResponse, any, any>>

Overrides

ToolCallLLM.complete

Defined in

packages/llamaindex/src/llm/gemini/base.ts:300

complete(params)

complete(params): Promise<CompletionResponse>

Get a prompt completion from the LLM

Parameters

params: LLMCompletionParamsNonStreaming

Returns

Promise<CompletionResponse>

Overrides

ToolCallLLM.complete

Defined in

packages/llamaindex/src/llm/gemini/base.ts:303


nonStreamChat()

protected nonStreamChat(params): Promise<GeminiChatNonStreamResponse>

Parameters

params: GeminiChatParamsNonStreaming

Returns

Promise<GeminiChatNonStreamResponse>

Defined in

packages/llamaindex/src/llm/gemini/base.ts:218


streamChat()

protected streamChat(params): GeminiChatStreamResponse

Parameters

params: GeminiChatParamsStreaming

Returns

GeminiChatStreamResponse

Defined in

packages/llamaindex/src/llm/gemini/base.ts:261