Firebase. AI. GenerativeModel
A type that represents a remote multimodal model (like Gemini), with the ability to generate content based on various input types.
Summary
Public functions |
|
---|---|
CountTokensAsync(ModelContent content, CancellationToken cancellationToken)
|
Task< CountTokensResponse >
Counts the number of tokens in a prompt using the model's tokenizer.
|
CountTokensAsync(string text, CancellationToken cancellationToken)
|
Task< CountTokensResponse >
Counts the number of tokens in a prompt using the model's tokenizer.
|
CountTokensAsync(IEnumerable< ModelContent > content, CancellationToken cancellationToken)
|
Task< CountTokensResponse >
Counts the number of tokens in a prompt using the model's tokenizer.
|
GenerateContentAsync(ModelContent content, CancellationToken cancellationToken)
|
Task< GenerateContentResponse >
Generates new content from input
ModelContent given to the model as a prompt. |
GenerateContentAsync(string text, CancellationToken cancellationToken)
|
Task< GenerateContentResponse >
Generates new content from input text given to the model as a prompt.
|
GenerateContentAsync(IEnumerable< ModelContent > content, CancellationToken cancellationToken)
|
Task< GenerateContentResponse >
Generates new content from input
ModelContent given to the model as a prompt. |
GenerateContentStreamAsync(ModelContent content, CancellationToken cancellationToken)
|
IAsyncEnumerable< GenerateContentResponse >
Generates new content as a stream from input
ModelContent given to the model as a prompt. |
GenerateContentStreamAsync(string text, CancellationToken cancellationToken)
|
IAsyncEnumerable< GenerateContentResponse >
Generates new content as a stream from input text given to the model as a prompt.
|
GenerateContentStreamAsync(IEnumerable< ModelContent > content, CancellationToken cancellationToken)
|
IAsyncEnumerable< GenerateContentResponse >
Generates new content as a stream from input
ModelContent given to the model as a prompt. |
StartChat(params ModelContent[] history)
|
Creates a new chat conversation using this model with the provided history.
|
StartChat(IEnumerable< ModelContent > history)
|
Creates a new chat conversation using this model with the provided history.
|
Public functions
CountTokensAsync
Task< CountTokensResponse > CountTokensAsync( ModelContent content, CancellationToken cancellationToken )
Counts the number of tokens in a prompt using the model's tokenizer.
Details | |||
---|---|---|---|
Parameters |
|
||
Exceptions |
|
||
Returns |
The
CountTokensResponse of running the model's tokenizer on the input. |
CountTokensAsync
Task< CountTokensResponse > CountTokensAsync( string text, CancellationToken cancellationToken )
Counts the number of tokens in a prompt using the model's tokenizer.
Details | |||||
---|---|---|---|---|---|
Parameters |
|
||||
Exceptions |
|
||||
Returns |
The
CountTokensResponse of running the model's tokenizer on the input. |
CountTokensAsync
Task< CountTokensResponse > CountTokensAsync( IEnumerable< ModelContent > content, CancellationToken cancellationToken )
Counts the number of tokens in a prompt using the model's tokenizer.
Details | |||||
---|---|---|---|---|---|
Parameters |
|
||||
Exceptions |
|
||||
Returns |
The
CountTokensResponse of running the model's tokenizer on the input. |
GenerateContentAsync
Task< GenerateContentResponse > GenerateContentAsync( ModelContent content, CancellationToken cancellationToken )
Generates new content from input ModelContent
given to the model as a prompt.
Details | |||||
---|---|---|---|---|---|
Parameters |
|
||||
Exceptions |
|
||||
Returns |
The generated content response from the model.
|
GenerateContentAsync
Task< GenerateContentResponse > GenerateContentAsync( string text, CancellationToken cancellationToken )
Generates new content from input text given to the model as a prompt.
Details | |||||
---|---|---|---|---|---|
Parameters |
|
||||
Exceptions |
|
||||
Returns |
The generated content response from the model.
|
GenerateContentAsync
Task< GenerateContentResponse > GenerateContentAsync( IEnumerable< ModelContent > content, CancellationToken cancellationToken )
Generates new content from input ModelContent
given to the model as a prompt.
Details | |||||
---|---|---|---|---|---|
Parameters |
|
||||
Exceptions |
|
||||
Returns |
The generated content response from the model.
|
GenerateContentStreamAsync
IAsyncEnumerable< GenerateContentResponse > GenerateContentStreamAsync( ModelContent content, CancellationToken cancellationToken )
Generates new content as a stream from input ModelContent
given to the model as a prompt.
Details | |||||
---|---|---|---|---|---|
Parameters |
|
||||
Exceptions |
|
||||
Returns |
A stream of generated content responses from the model.
|
GenerateContentStreamAsync
IAsyncEnumerable< GenerateContentResponse > GenerateContentStreamAsync( string text, CancellationToken cancellationToken )
Generates new content as a stream from input text given to the model as a prompt.
Details | |||||
---|---|---|---|---|---|
Parameters |
|
||||
Exceptions |
|
||||
Returns |
A stream of generated content responses from the model.
|
GenerateContentStreamAsync
IAsyncEnumerable< GenerateContentResponse > GenerateContentStreamAsync( IEnumerable< ModelContent > content, CancellationToken cancellationToken )
Generates new content as a stream from input ModelContent
given to the model as a prompt.
Details | |||||
---|---|---|---|---|---|
Parameters |
|
||||
Exceptions |
|
||||
Returns |
A stream of generated content responses from the model.
|
StartChat
Chat StartChat( params ModelContent[] history )
Creates a new chat conversation using this model with the provided history.
Details | |||
---|---|---|---|
Parameters |
|
StartChat
Chat StartChat( IEnumerable< ModelContent > history )
Creates a new chat conversation using this model with the provided history.
Details | |||
---|---|---|---|
Parameters |
|