diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.BatchesUpdateEmbedContentBatch.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.BatchesUpdateEmbedContentBatch.g.cs
index 2fb4a69..3e9fb9b 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.BatchesUpdateEmbedContentBatch.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.BatchesUpdateEmbedContentBatch.g.cs
@@ -397,14 +397,11 @@ partial void ProcessBatchesUpdateEmbedContentBatchResponseContent(
///
///
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
- ///
///
/// Stats about the batch.
///
- ///
- /// The output of a batch request. This is returned in the `AsyncBatchEmbedContentResponse` or the `EmbedContentBatch.output` field.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
///
/// Optional. The priority of the batch. Batches with a higher priority value will be processed before batches with a lower priority value. Negative values are allowed. Default is 0.
@@ -412,6 +409,9 @@ partial void ProcessBatchesUpdateEmbedContentBatchResponseContent(
///
/// Required. The user-defined name of this batch.
///
+ ///
+ /// The output of a batch request. This is returned in the `AsyncBatchEmbedContentResponse` or the `EmbedContentBatch.output` field.
+ ///
///
/// Configures the input to the batch request.
///
@@ -421,22 +421,22 @@ partial void ProcessBatchesUpdateEmbedContentBatchResponseContent(
public async global::System.Threading.Tasks.Task BatchesUpdateEmbedContentBatchAsync(
string batchesId,
string? updateMask = default,
- string? model = default,
global::Google.Gemini.EmbedContentBatchStats? batchStats = default,
- global::Google.Gemini.EmbedContentBatchOutput? output = default,
+ string? model = default,
string? priority = default,
string? displayName = default,
+ global::Google.Gemini.EmbedContentBatchOutput? output = default,
global::Google.Gemini.InputEmbedContentConfig? inputConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.EmbedContentBatch
{
- Model = model,
BatchStats = batchStats,
- Output = output,
+ Model = model,
Priority = priority,
DisplayName = displayName,
+ Output = output,
InputConfig = inputConfig,
};
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.BatchesUpdateGenerateContentBatch.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.BatchesUpdateGenerateContentBatch.g.cs
index 5a074aa..a44ff8d 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.BatchesUpdateGenerateContentBatch.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.BatchesUpdateGenerateContentBatch.g.cs
@@ -397,21 +397,21 @@ partial void ProcessBatchesUpdateGenerateContentBatchResponseContent(
///
///
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
- ///
///
/// Stats about the batch.
///
- ///
- /// The output of a batch request. This is returned in the `BatchGenerateContentResponse` or the `GenerateContentBatch.output` field.
- ///
///
/// Optional. The priority of the batch. Batches with a higher priority value will be processed before batches with a lower priority value. Negative values are allowed. Default is 0.
///
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
///
/// Required. The user-defined name of this batch.
///
+ ///
+ /// The output of a batch request. This is returned in the `BatchGenerateContentResponse` or the `GenerateContentBatch.output` field.
+ ///
///
/// Configures the input to the batch request.
///
@@ -421,22 +421,22 @@ partial void ProcessBatchesUpdateGenerateContentBatchResponseContent(
public async global::System.Threading.Tasks.Task BatchesUpdateGenerateContentBatchAsync(
string batchesId,
string? updateMask = default,
- string? model = default,
global::Google.Gemini.BatchStats? batchStats = default,
- global::Google.Gemini.GenerateContentBatchOutput? output = default,
string? priority = default,
+ string? model = default,
string? displayName = default,
+ global::Google.Gemini.GenerateContentBatchOutput? output = default,
global::Google.Gemini.InputConfig? inputConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.GenerateContentBatch
{
- Model = model,
BatchStats = batchStats,
- Output = output,
Priority = priority,
+ Model = model,
DisplayName = displayName,
+ Output = output,
InputConfig = inputConfig,
};
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CachedContentsCreate.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CachedContentsCreate.g.cs
index 34464cc..b96f45e 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CachedContentsCreate.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CachedContentsCreate.g.cs
@@ -380,60 +380,60 @@ partial void ProcessCachedContentsCreateResponseContent(
///
/// Creates CachedContent resource.
///
- ///
- /// Metadata on the usage of the cached content.
+ ///
+ /// Input only. New TTL for this resource, input only.
+ ///
+ ///
+ /// Optional. Input only. Immutable. The content to cache.
+ ///
+ ///
+ /// Timestamp in UTC of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input.
///
///
/// Optional. Immutable. The user-generated meaningful display name of the cached content. Maximum 128 Unicode characters.
///
- ///
- /// Required. Immutable. The name of the `Model` to use for cached content Format: `models/{model}`
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
- ///
- /// Timestamp in UTC of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input.
+ ///
+ /// Metadata on the usage of the cached content.
///
- ///
- /// Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response
+ ///
+ /// Required. Immutable. The name of the `Model` to use for cached content Format: `models/{model}`
///
///
/// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Input only. New TTL for this resource, input only.
- ///
- ///
- /// Optional. Input only. Immutable. The content to cache.
- ///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Threading.Tasks.Task CachedContentsCreateAsync(
- global::Google.Gemini.CachedContentUsageMetadata? usageMetadata = default,
- string? displayName = default,
- string? model = default,
- string? expireTime = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
string? ttl = default,
global::System.Collections.Generic.IList? contents = default,
+ string? expireTime = default,
+ string? displayName = default,
global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::Google.Gemini.CachedContentUsageMetadata? usageMetadata = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.CachedContent
{
- UsageMetadata = usageMetadata,
- DisplayName = displayName,
- Model = model,
- ExpireTime = expireTime,
- Tools = tools,
- SystemInstruction = systemInstruction,
Ttl = ttl,
Contents = contents,
+ ExpireTime = expireTime,
+ DisplayName = displayName,
ToolConfig = toolConfig,
+ UsageMetadata = usageMetadata,
+ Model = model,
+ SystemInstruction = systemInstruction,
+ Tools = tools,
};
return await CachedContentsCreateAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CachedContentsPatch.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CachedContentsPatch.g.cs
index f527692..3ce38da 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CachedContentsPatch.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CachedContentsPatch.g.cs
@@ -397,32 +397,32 @@ partial void ProcessCachedContentsPatchResponseContent(
///
///
///
- ///
- /// Metadata on the usage of the cached content.
+ ///
+ /// Input only. New TTL for this resource, input only.
+ ///
+ ///
+ /// Optional. Input only. Immutable. The content to cache.
+ ///
+ ///
+ /// Timestamp in UTC of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input.
///
///
/// Optional. Immutable. The user-generated meaningful display name of the cached content. Maximum 128 Unicode characters.
///
- ///
- /// Required. Immutable. The name of the `Model` to use for cached content Format: `models/{model}`
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
- ///
- /// Timestamp in UTC of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input.
+ ///
+ /// Metadata on the usage of the cached content.
///
- ///
- /// Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response
+ ///
+ /// Required. Immutable. The name of the `Model` to use for cached content Format: `models/{model}`
///
///
/// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Input only. New TTL for this resource, input only.
- ///
- ///
- /// Optional. Input only. Immutable. The content to cache.
- ///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
@@ -430,29 +430,29 @@ partial void ProcessCachedContentsPatchResponseContent(
public async global::System.Threading.Tasks.Task CachedContentsPatchAsync(
string cachedContentsId,
string? updateMask = default,
- global::Google.Gemini.CachedContentUsageMetadata? usageMetadata = default,
- string? displayName = default,
- string? model = default,
- string? expireTime = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
string? ttl = default,
global::System.Collections.Generic.IList? contents = default,
+ string? expireTime = default,
+ string? displayName = default,
global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::Google.Gemini.CachedContentUsageMetadata? usageMetadata = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.CachedContent
{
- UsageMetadata = usageMetadata,
- DisplayName = displayName,
- Model = model,
- ExpireTime = expireTime,
- Tools = tools,
- SystemInstruction = systemInstruction,
Ttl = ttl,
Contents = contents,
+ ExpireTime = expireTime,
+ DisplayName = displayName,
ToolConfig = toolConfig,
+ UsageMetadata = usageMetadata,
+ Model = model,
+ SystemInstruction = systemInstruction,
+ Tools = tools,
};
return await CachedContentsPatchAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CorporaPermissionsCreate.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CorporaPermissionsCreate.g.cs
index c2e9902..61a7c63 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CorporaPermissionsCreate.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CorporaPermissionsCreate.g.cs
@@ -387,31 +387,31 @@ partial void ProcessCorporaPermissionsCreateResponseContent(
/// Create a permission to a specific resource.
///
///
- ///
- /// Required. The role granted by this permission.
- ///
///
/// Optional. Immutable. The type of the grantee.
///
///
/// Optional. Immutable. The email address of the user of group which this permission refers. Field is not set when permission's grantee type is EVERYONE.
///
+ ///
+ /// Required. The role granted by this permission.
+ ///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Threading.Tasks.Task CorporaPermissionsCreateAsync(
string corporaId,
- global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.PermissionGranteeType? granteeType = default,
string? emailAddress = default,
+ global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.Permission
{
- Role = role,
GranteeType = granteeType,
EmailAddress = emailAddress,
+ Role = role,
};
return await CorporaPermissionsCreateAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CorporaPermissionsPatch.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CorporaPermissionsPatch.g.cs
index 6a5e165..d104c54 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CorporaPermissionsPatch.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.CorporaPermissionsPatch.g.cs
@@ -404,15 +404,15 @@ partial void ProcessCorporaPermissionsPatchResponseContent(
///
///
///
- ///
- /// Required. The role granted by this permission.
- ///
///
/// Optional. Immutable. The type of the grantee.
///
///
/// Optional. Immutable. The email address of the user of group which this permission refers. Field is not set when permission's grantee type is EVERYONE.
///
+ ///
+ /// Required. The role granted by this permission.
+ ///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
@@ -420,17 +420,17 @@ partial void ProcessCorporaPermissionsPatchResponseContent(
string permissionsId,
string corporaId,
string? updateMask = default,
- global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.PermissionGranteeType? granteeType = default,
string? emailAddress = default,
+ global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.Permission
{
- Role = role,
GranteeType = granteeType,
EmailAddress = emailAddress,
+ Role = role,
};
return await CorporaPermissionsPatchAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.DynamicGenerateContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.DynamicGenerateContent.g.cs
index e589842..79d6732 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.DynamicGenerateContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.DynamicGenerateContent.g.cs
@@ -387,66 +387,66 @@ partial void ProcessDynamicGenerateContentResponseContent(
/// Generates a model response given an input `GenerateContentRequest`. Refer to the [text generation guide](https://ai.google.dev/gemini-api/docs/text-generation) for detailed usage information. Input capabilities differ between models, including tuned models. Refer to the [model guide](https://ai.google.dev/gemini-api/docs/models/gemini) and [tuning guide](https://ai.google.dev/gemini-api/docs/model-tuning) for details.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Threading.Tasks.Task DynamicGenerateContentAsync(
string dynamicId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.GenerateContentRequest
{
- CachedContent = cachedContent,
- GenerationConfig = generationConfig,
ServiceTier = serviceTier,
- Model = model,
- Tools = tools,
- SystemInstruction = systemInstruction,
- SafetySettings = safetySettings,
Contents = contents,
- ToolConfig = toolConfig,
Store = store,
+ ToolConfig = toolConfig,
+ SafetySettings = safetySettings,
+ Model = model,
+ SystemInstruction = systemInstruction,
+ Tools = tools,
+ CachedContent = cachedContent,
+ GenerationConfig = generationConfig,
};
return await DynamicGenerateContentAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.DynamicStreamGenerateContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.DynamicStreamGenerateContent.g.cs
index 73a22d0..6dde78c 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.DynamicStreamGenerateContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.DynamicStreamGenerateContent.g.cs
@@ -387,66 +387,66 @@ partial void ProcessDynamicStreamGenerateContentResponseContent(
/// Generates a [streamed response](https://ai.google.dev/gemini-api/docs/text-generation?lang=python#generate-a-text-stream) from the model given an input `GenerateContentRequest`.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Threading.Tasks.Task DynamicStreamGenerateContentAsync(
string dynamicId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.GenerateContentRequest
{
- CachedContent = cachedContent,
- GenerationConfig = generationConfig,
ServiceTier = serviceTier,
- Model = model,
- Tools = tools,
- SystemInstruction = systemInstruction,
- SafetySettings = safetySettings,
Contents = contents,
- ToolConfig = toolConfig,
Store = store,
+ ToolConfig = toolConfig,
+ SafetySettings = safetySettings,
+ Model = model,
+ SystemInstruction = systemInstruction,
+ Tools = tools,
+ CachedContent = cachedContent,
+ GenerationConfig = generationConfig,
};
return await DynamicStreamGenerateContentAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.DynamicStreamGenerateContentAsStream.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.DynamicStreamGenerateContentAsStream.g.cs
index a4603e9..601cbe5 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.DynamicStreamGenerateContentAsStream.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.DynamicStreamGenerateContentAsStream.g.cs
@@ -360,66 +360,66 @@ partial void ProcessDynamicStreamGenerateContentAsStreamResponse(
/// Generates a [streamed response](https://ai.google.dev/gemini-api/docs/text-generation?lang=python#generate-a-text-stream) from the model given an input `GenerateContentRequest`.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Collections.Generic.IAsyncEnumerable DynamicStreamGenerateContentAsStreamAsync(
string dynamicId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
[global::System.Runtime.CompilerServices.EnumeratorCancellation] global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.GenerateContentRequest
{
- CachedContent = cachedContent,
- GenerationConfig = generationConfig,
ServiceTier = serviceTier,
- Model = model,
- Tools = tools,
- SystemInstruction = systemInstruction,
- SafetySettings = safetySettings,
Contents = contents,
- ToolConfig = toolConfig,
Store = store,
+ ToolConfig = toolConfig,
+ SafetySettings = safetySettings,
+ Model = model,
+ SystemInstruction = systemInstruction,
+ Tools = tools,
+ CachedContent = cachedContent,
+ GenerationConfig = generationConfig,
};
var __enumerable = DynamicStreamGenerateContentAsStreamAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsCountTokens.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsCountTokens.g.cs
index 5e4fce8..63289cd 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsCountTokens.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsCountTokens.g.cs
@@ -387,26 +387,26 @@ partial void ProcessModelsCountTokensResponseContent(
/// Runs a model's tokenizer on input `Content` and returns the token count. Refer to the [tokens guide](https://ai.google.dev/gemini-api/docs/tokens) to learn more about tokens.
///
///
- ///
- /// Optional. The input given to the model as a prompt. This field is ignored when `generate_content_request` is set.
- ///
///
/// Request to generate a completion from the model.
///
+ ///
+ /// Optional. The input given to the model as a prompt. This field is ignored when `generate_content_request` is set.
+ ///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Threading.Tasks.Task ModelsCountTokensAsync(
string modelsId,
- global::System.Collections.Generic.IList? contents = default,
global::Google.Gemini.GenerateContentRequest? generateContentRequest = default,
+ global::System.Collections.Generic.IList? contents = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.CountTokensRequest
{
- Contents = contents,
GenerateContentRequest = generateContentRequest,
+ Contents = contents,
};
return await ModelsCountTokensAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsEmbedContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsEmbedContent.g.cs
index 64f9dfe..1e9694b 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsEmbedContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsEmbedContent.g.cs
@@ -387,8 +387,8 @@ partial void ProcessModelsEmbedContentResponseContent(
/// Generates a text embedding vector from the input `Content` using the specified [Gemini Embedding model](https://ai.google.dev/gemini-api/docs/models/gemini#text-embedding).
///
///
- ///
- /// Required. The model's resource name. This serves as an ID for the Model to use. This name should match a model name returned by the `ListModels` method. Format: `models/{model}`
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
///
/// Optional. An optional title for the text. Only applicable when TaskType is `RETRIEVAL_DOCUMENT`. Note: Specifying a `title` for `RETRIEVAL_DOCUMENT` provides better quality embeddings for retrieval.
@@ -399,29 +399,29 @@ partial void ProcessModelsEmbedContentResponseContent(
///
/// Optional. Optional reduced dimension for the output embedding. If set, excessive values in the output embedding are truncated from the end. Supported by newer models since 2024 only. You cannot set this value if using the earlier model (`models/embedding-001`).
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// Required. The model's resource name. This serves as an ID for the Model to use. This name should match a model name returned by the `ListModels` method. Format: `models/{model}`
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Threading.Tasks.Task ModelsEmbedContentAsync(
string modelsId,
- string? model = default,
+ global::Google.Gemini.Content? content = default,
string? title = default,
global::Google.Gemini.EmbedContentRequestTaskType? taskType = default,
int? outputDimensionality = default,
- global::Google.Gemini.Content? content = default,
+ string? model = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.EmbedContentRequest
{
- Model = model,
+ Content = content,
Title = title,
TaskType = taskType,
OutputDimensionality = outputDimensionality,
- Content = content,
+ Model = model,
};
return await ModelsEmbedContentAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsGenerateContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsGenerateContent.g.cs
index 441ed51..cb14a1c 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsGenerateContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsGenerateContent.g.cs
@@ -387,66 +387,66 @@ partial void ProcessModelsGenerateContentResponseContent(
/// Generates a model response given an input `GenerateContentRequest`. Refer to the [text generation guide](https://ai.google.dev/gemini-api/docs/text-generation) for detailed usage information. Input capabilities differ between models, including tuned models. Refer to the [model guide](https://ai.google.dev/gemini-api/docs/models/gemini) and [tuning guide](https://ai.google.dev/gemini-api/docs/model-tuning) for details.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Threading.Tasks.Task ModelsGenerateContentAsync(
string modelsId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.GenerateContentRequest
{
- CachedContent = cachedContent,
- GenerationConfig = generationConfig,
ServiceTier = serviceTier,
- Model = model,
- Tools = tools,
- SystemInstruction = systemInstruction,
- SafetySettings = safetySettings,
Contents = contents,
- ToolConfig = toolConfig,
Store = store,
+ ToolConfig = toolConfig,
+ SafetySettings = safetySettings,
+ Model = model,
+ SystemInstruction = systemInstruction,
+ Tools = tools,
+ CachedContent = cachedContent,
+ GenerationConfig = generationConfig,
};
return await ModelsGenerateContentAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsStreamGenerateContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsStreamGenerateContent.g.cs
index 5232c60..6fa1ad6 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsStreamGenerateContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsStreamGenerateContent.g.cs
@@ -387,66 +387,66 @@ partial void ProcessModelsStreamGenerateContentResponseContent(
/// Generates a [streamed response](https://ai.google.dev/gemini-api/docs/text-generation?lang=python#generate-a-text-stream) from the model given an input `GenerateContentRequest`.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Threading.Tasks.Task ModelsStreamGenerateContentAsync(
string modelsId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.GenerateContentRequest
{
- CachedContent = cachedContent,
- GenerationConfig = generationConfig,
ServiceTier = serviceTier,
- Model = model,
- Tools = tools,
- SystemInstruction = systemInstruction,
- SafetySettings = safetySettings,
Contents = contents,
- ToolConfig = toolConfig,
Store = store,
+ ToolConfig = toolConfig,
+ SafetySettings = safetySettings,
+ Model = model,
+ SystemInstruction = systemInstruction,
+ Tools = tools,
+ CachedContent = cachedContent,
+ GenerationConfig = generationConfig,
};
return await ModelsStreamGenerateContentAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsStreamGenerateContentAsStream.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsStreamGenerateContentAsStream.g.cs
index c6840ba..581de3d 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsStreamGenerateContentAsStream.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.ModelsStreamGenerateContentAsStream.g.cs
@@ -360,66 +360,66 @@ partial void ProcessModelsStreamGenerateContentAsStreamResponse(
/// Generates a [streamed response](https://ai.google.dev/gemini-api/docs/text-generation?lang=python#generate-a-text-stream) from the model given an input `GenerateContentRequest`.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Collections.Generic.IAsyncEnumerable ModelsStreamGenerateContentAsStreamAsync(
string modelsId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
[global::System.Runtime.CompilerServices.EnumeratorCancellation] global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.GenerateContentRequest
{
- CachedContent = cachedContent,
- GenerationConfig = generationConfig,
ServiceTier = serviceTier,
- Model = model,
- Tools = tools,
- SystemInstruction = systemInstruction,
- SafetySettings = safetySettings,
Contents = contents,
- ToolConfig = toolConfig,
Store = store,
+ ToolConfig = toolConfig,
+ SafetySettings = safetySettings,
+ Model = model,
+ SystemInstruction = systemInstruction,
+ Tools = tools,
+ CachedContent = cachedContent,
+ GenerationConfig = generationConfig,
};
var __enumerable = ModelsStreamGenerateContentAsStreamAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsCreate.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsCreate.g.cs
index 5ebe449..d972176 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsCreate.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsCreate.g.cs
@@ -390,17 +390,14 @@ partial void ProcessTunedModelsCreateResponseContent(
/// Creates a tuned model. Check intermediate tuning progress (if any) through the [google.longrunning.Operations] service. Access status and results through the Operations service. Example: GET /v1/tunedModels/az2mb0bpw6i/operations/000-111-222
///
///
- ///
- /// Optional. For Nucleus sampling. Nucleus sampling considers the smallest set of tokens whose probability sum is at least `top_p`. This value specifies default to be the one used by the base model while creating the model.
- ///
- ///
- /// Optional. List of project numbers that have read access to the tuned model.
- ///
///
/// Tuned model as a source for training a new model.
///
- ///
- /// Optional. For Top-k sampling. Top-k sampling considers the set of `top_k` most probable tokens. This value specifies default to be used by the backend while making the call to the model. This value specifies default to be the one used by the base model while creating the model.
+ ///
+ /// Optional. For Nucleus sampling. Nucleus sampling considers the smallest set of tokens whose probability sum is at least `top_p`. This value specifies default to be the one used by the base model while creating the model.
+ ///
+ ///
+ /// Tuning tasks that create tuned models.
///
///
/// Immutable. The name of the `Model` to tune. Example: `models/gemini-1.5-flash-001`
@@ -408,8 +405,11 @@ partial void ProcessTunedModelsCreateResponseContent(
///
/// Optional. The name to display for this model in user interfaces. The display name must be up to 40 characters including spaces.
///
- ///
- /// Tuning tasks that create tuned models.
+ ///
+ /// Optional. For Top-k sampling. Top-k sampling considers the set of `top_k` most probable tokens. This value specifies default to be used by the backend while making the call to the model. This value specifies default to be the one used by the base model while creating the model.
+ ///
+ ///
+ /// Optional. List of project numbers that have read access to the tuned model.
///
///
/// Optional. A short description of this model.
@@ -422,13 +422,13 @@ partial void ProcessTunedModelsCreateResponseContent(
///
public async global::System.Threading.Tasks.Task TunedModelsCreateAsync(
string? tunedModelId = default,
- float? topP = default,
- global::System.Collections.Generic.IList? readerProjectNumbers = default,
global::Google.Gemini.TunedModelSource? tunedModelSource = default,
- int? topK = default,
+ float? topP = default,
+ global::Google.Gemini.TuningTask? tuningTask = default,
string? baseModel = default,
string? displayName = default,
- global::Google.Gemini.TuningTask? tuningTask = default,
+ int? topK = default,
+ global::System.Collections.Generic.IList? readerProjectNumbers = default,
string? description = default,
float? temperature = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
@@ -436,13 +436,13 @@ partial void ProcessTunedModelsCreateResponseContent(
{
var __request = new global::Google.Gemini.TunedModel
{
- TopP = topP,
- ReaderProjectNumbers = readerProjectNumbers,
TunedModelSource = tunedModelSource,
- TopK = topK,
+ TopP = topP,
+ TuningTask = tuningTask,
BaseModel = baseModel,
DisplayName = displayName,
- TuningTask = tuningTask,
+ TopK = topK,
+ ReaderProjectNumbers = readerProjectNumbers,
Description = description,
Temperature = temperature,
};
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsGenerateContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsGenerateContent.g.cs
index 98dff96..c236b33 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsGenerateContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsGenerateContent.g.cs
@@ -387,66 +387,66 @@ partial void ProcessTunedModelsGenerateContentResponseContent(
/// Generates a model response given an input `GenerateContentRequest`. Refer to the [text generation guide](https://ai.google.dev/gemini-api/docs/text-generation) for detailed usage information. Input capabilities differ between models, including tuned models. Refer to the [model guide](https://ai.google.dev/gemini-api/docs/models/gemini) and [tuning guide](https://ai.google.dev/gemini-api/docs/model-tuning) for details.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Threading.Tasks.Task TunedModelsGenerateContentAsync(
string tunedModelsId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.GenerateContentRequest
{
- CachedContent = cachedContent,
- GenerationConfig = generationConfig,
ServiceTier = serviceTier,
- Model = model,
- Tools = tools,
- SystemInstruction = systemInstruction,
- SafetySettings = safetySettings,
Contents = contents,
- ToolConfig = toolConfig,
Store = store,
+ ToolConfig = toolConfig,
+ SafetySettings = safetySettings,
+ Model = model,
+ SystemInstruction = systemInstruction,
+ Tools = tools,
+ CachedContent = cachedContent,
+ GenerationConfig = generationConfig,
};
return await TunedModelsGenerateContentAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsPatch.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsPatch.g.cs
index 07ccf8f..5989e72 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsPatch.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsPatch.g.cs
@@ -397,17 +397,14 @@ partial void ProcessTunedModelsPatchResponseContent(
///
///
///
- ///
- /// Optional. For Nucleus sampling. Nucleus sampling considers the smallest set of tokens whose probability sum is at least `top_p`. This value specifies default to be the one used by the base model while creating the model.
- ///
- ///
- /// Optional. List of project numbers that have read access to the tuned model.
- ///
///
/// Tuned model as a source for training a new model.
///
- ///
- /// Optional. For Top-k sampling. Top-k sampling considers the set of `top_k` most probable tokens. This value specifies default to be used by the backend while making the call to the model. This value specifies default to be the one used by the base model while creating the model.
+ ///
+ /// Optional. For Nucleus sampling. Nucleus sampling considers the smallest set of tokens whose probability sum is at least `top_p`. This value specifies default to be the one used by the base model while creating the model.
+ ///
+ ///
+ /// Tuning tasks that create tuned models.
///
///
/// Immutable. The name of the `Model` to tune. Example: `models/gemini-1.5-flash-001`
@@ -415,8 +412,11 @@ partial void ProcessTunedModelsPatchResponseContent(
///
/// Optional. The name to display for this model in user interfaces. The display name must be up to 40 characters including spaces.
///
- ///
- /// Tuning tasks that create tuned models.
+ ///
+ /// Optional. For Top-k sampling. Top-k sampling considers the set of `top_k` most probable tokens. This value specifies default to be used by the backend while making the call to the model. This value specifies default to be the one used by the base model while creating the model.
+ ///
+ ///
+ /// Optional. List of project numbers that have read access to the tuned model.
///
///
/// Optional. A short description of this model.
@@ -430,13 +430,13 @@ partial void ProcessTunedModelsPatchResponseContent(
public async global::System.Threading.Tasks.Task TunedModelsPatchAsync(
string tunedModelsId,
string? updateMask = default,
- float? topP = default,
- global::System.Collections.Generic.IList? readerProjectNumbers = default,
global::Google.Gemini.TunedModelSource? tunedModelSource = default,
- int? topK = default,
+ float? topP = default,
+ global::Google.Gemini.TuningTask? tuningTask = default,
string? baseModel = default,
string? displayName = default,
- global::Google.Gemini.TuningTask? tuningTask = default,
+ int? topK = default,
+ global::System.Collections.Generic.IList? readerProjectNumbers = default,
string? description = default,
float? temperature = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
@@ -444,13 +444,13 @@ partial void ProcessTunedModelsPatchResponseContent(
{
var __request = new global::Google.Gemini.TunedModel
{
- TopP = topP,
- ReaderProjectNumbers = readerProjectNumbers,
TunedModelSource = tunedModelSource,
- TopK = topK,
+ TopP = topP,
+ TuningTask = tuningTask,
BaseModel = baseModel,
DisplayName = displayName,
- TuningTask = tuningTask,
+ TopK = topK,
+ ReaderProjectNumbers = readerProjectNumbers,
Description = description,
Temperature = temperature,
};
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsPermissionsCreate.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsPermissionsCreate.g.cs
index badbf15..00e6a37 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsPermissionsCreate.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsPermissionsCreate.g.cs
@@ -387,31 +387,31 @@ partial void ProcessTunedModelsPermissionsCreateResponseContent(
/// Create a permission to a specific resource.
///
///
- ///
- /// Required. The role granted by this permission.
- ///
///
/// Optional. Immutable. The type of the grantee.
///
///
/// Optional. Immutable. The email address of the user of group which this permission refers. Field is not set when permission's grantee type is EVERYONE.
///
+ ///
+ /// Required. The role granted by this permission.
+ ///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Threading.Tasks.Task TunedModelsPermissionsCreateAsync(
string tunedModelsId,
- global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.PermissionGranteeType? granteeType = default,
string? emailAddress = default,
+ global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.Permission
{
- Role = role,
GranteeType = granteeType,
EmailAddress = emailAddress,
+ Role = role,
};
return await TunedModelsPermissionsCreateAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsPermissionsPatch.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsPermissionsPatch.g.cs
index d336db3..03fa5f1 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsPermissionsPatch.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsPermissionsPatch.g.cs
@@ -404,15 +404,15 @@ partial void ProcessTunedModelsPermissionsPatchResponseContent(
///
///
///
- ///
- /// Required. The role granted by this permission.
- ///
///
/// Optional. Immutable. The type of the grantee.
///
///
/// Optional. Immutable. The email address of the user of group which this permission refers. Field is not set when permission's grantee type is EVERYONE.
///
+ ///
+ /// Required. The role granted by this permission.
+ ///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
@@ -420,17 +420,17 @@ partial void ProcessTunedModelsPermissionsPatchResponseContent(
string tunedModelsId,
string permissionsId,
string? updateMask = default,
- global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.PermissionGranteeType? granteeType = default,
string? emailAddress = default,
+ global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.Permission
{
- Role = role,
GranteeType = granteeType,
EmailAddress = emailAddress,
+ Role = role,
};
return await TunedModelsPermissionsPatchAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsStreamGenerateContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsStreamGenerateContent.g.cs
index d3e77d2..51697a9 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsStreamGenerateContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsStreamGenerateContent.g.cs
@@ -387,66 +387,66 @@ partial void ProcessTunedModelsStreamGenerateContentResponseContent(
/// Generates a [streamed response](https://ai.google.dev/gemini-api/docs/text-generation?lang=python#generate-a-text-stream) from the model given an input `GenerateContentRequest`.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Threading.Tasks.Task TunedModelsStreamGenerateContentAsync(
string tunedModelsId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.GenerateContentRequest
{
- CachedContent = cachedContent,
- GenerationConfig = generationConfig,
ServiceTier = serviceTier,
- Model = model,
- Tools = tools,
- SystemInstruction = systemInstruction,
- SafetySettings = safetySettings,
Contents = contents,
- ToolConfig = toolConfig,
Store = store,
+ ToolConfig = toolConfig,
+ SafetySettings = safetySettings,
+ Model = model,
+ SystemInstruction = systemInstruction,
+ Tools = tools,
+ CachedContent = cachedContent,
+ GenerationConfig = generationConfig,
};
return await TunedModelsStreamGenerateContentAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsStreamGenerateContentAsStream.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsStreamGenerateContentAsStream.g.cs
index 3e822d9..9d5edec 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsStreamGenerateContentAsStream.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.GeminiClient.TunedModelsStreamGenerateContentAsStream.g.cs
@@ -360,66 +360,66 @@ partial void ProcessTunedModelsStreamGenerateContentAsStreamResponse(
/// Generates a [streamed response](https://ai.google.dev/gemini-api/docs/text-generation?lang=python#generate-a-text-stream) from the model given an input `GenerateContentRequest`.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
public async global::System.Collections.Generic.IAsyncEnumerable TunedModelsStreamGenerateContentAsStreamAsync(
string tunedModelsId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
[global::System.Runtime.CompilerServices.EnumeratorCancellation] global::System.Threading.CancellationToken cancellationToken = default)
{
var __request = new global::Google.Gemini.GenerateContentRequest
{
- CachedContent = cachedContent,
- GenerationConfig = generationConfig,
ServiceTier = serviceTier,
- Model = model,
- Tools = tools,
- SystemInstruction = systemInstruction,
- SafetySettings = safetySettings,
Contents = contents,
- ToolConfig = toolConfig,
Store = store,
+ ToolConfig = toolConfig,
+ SafetySettings = safetySettings,
+ Model = model,
+ SystemInstruction = systemInstruction,
+ Tools = tools,
+ CachedContent = cachedContent,
+ GenerationConfig = generationConfig,
};
var __enumerable = TunedModelsStreamGenerateContentAsStreamAsync(
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.BatchesUpdateEmbedContentBatch.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.BatchesUpdateEmbedContentBatch.g.cs
index c82ad97..fbd3ba5 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.BatchesUpdateEmbedContentBatch.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.BatchesUpdateEmbedContentBatch.g.cs
@@ -25,14 +25,11 @@ public partial interface IGeminiClient
///
///
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
- ///
///
/// Stats about the batch.
///
- ///
- /// The output of a batch request. This is returned in the `AsyncBatchEmbedContentResponse` or the `EmbedContentBatch.output` field.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
///
/// Optional. The priority of the batch. Batches with a higher priority value will be processed before batches with a lower priority value. Negative values are allowed. Default is 0.
@@ -40,6 +37,9 @@ public partial interface IGeminiClient
///
/// Required. The user-defined name of this batch.
///
+ ///
+ /// The output of a batch request. This is returned in the `AsyncBatchEmbedContentResponse` or the `EmbedContentBatch.output` field.
+ ///
///
/// Configures the input to the batch request.
///
@@ -49,11 +49,11 @@ public partial interface IGeminiClient
global::System.Threading.Tasks.Task BatchesUpdateEmbedContentBatchAsync(
string batchesId,
string? updateMask = default,
- string? model = default,
global::Google.Gemini.EmbedContentBatchStats? batchStats = default,
- global::Google.Gemini.EmbedContentBatchOutput? output = default,
+ string? model = default,
string? priority = default,
string? displayName = default,
+ global::Google.Gemini.EmbedContentBatchOutput? output = default,
global::Google.Gemini.InputEmbedContentConfig? inputConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.BatchesUpdateGenerateContentBatch.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.BatchesUpdateGenerateContentBatch.g.cs
index d439852..379d789 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.BatchesUpdateGenerateContentBatch.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.BatchesUpdateGenerateContentBatch.g.cs
@@ -25,21 +25,21 @@ public partial interface IGeminiClient
///
///
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
- ///
///
/// Stats about the batch.
///
- ///
- /// The output of a batch request. This is returned in the `BatchGenerateContentResponse` or the `GenerateContentBatch.output` field.
- ///
///
/// Optional. The priority of the batch. Batches with a higher priority value will be processed before batches with a lower priority value. Negative values are allowed. Default is 0.
///
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
///
/// Required. The user-defined name of this batch.
///
+ ///
+ /// The output of a batch request. This is returned in the `BatchGenerateContentResponse` or the `GenerateContentBatch.output` field.
+ ///
///
/// Configures the input to the batch request.
///
@@ -49,11 +49,11 @@ public partial interface IGeminiClient
global::System.Threading.Tasks.Task BatchesUpdateGenerateContentBatchAsync(
string batchesId,
string? updateMask = default,
- string? model = default,
global::Google.Gemini.BatchStats? batchStats = default,
- global::Google.Gemini.GenerateContentBatchOutput? output = default,
string? priority = default,
+ string? model = default,
string? displayName = default,
+ global::Google.Gemini.GenerateContentBatchOutput? output = default,
global::Google.Gemini.InputConfig? inputConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CachedContentsCreate.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CachedContentsCreate.g.cs
index 5460339..7f1f954 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CachedContentsCreate.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CachedContentsCreate.g.cs
@@ -19,46 +19,46 @@ public partial interface IGeminiClient
///
/// Creates CachedContent resource.
///
- ///
- /// Metadata on the usage of the cached content.
+ ///
+ /// Input only. New TTL for this resource, input only.
+ ///
+ ///
+ /// Optional. Input only. Immutable. The content to cache.
+ ///
+ ///
+ /// Timestamp in UTC of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input.
///
///
/// Optional. Immutable. The user-generated meaningful display name of the cached content. Maximum 128 Unicode characters.
///
- ///
- /// Required. Immutable. The name of the `Model` to use for cached content Format: `models/{model}`
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
- ///
- /// Timestamp in UTC of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input.
+ ///
+ /// Metadata on the usage of the cached content.
///
- ///
- /// Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response
+ ///
+ /// Required. Immutable. The name of the `Model` to use for cached content Format: `models/{model}`
///
///
/// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Input only. New TTL for this resource, input only.
- ///
- ///
- /// Optional. Input only. Immutable. The content to cache.
- ///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Threading.Tasks.Task CachedContentsCreateAsync(
- global::Google.Gemini.CachedContentUsageMetadata? usageMetadata = default,
- string? displayName = default,
- string? model = default,
- string? expireTime = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
string? ttl = default,
global::System.Collections.Generic.IList? contents = default,
+ string? expireTime = default,
+ string? displayName = default,
global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::Google.Gemini.CachedContentUsageMetadata? usageMetadata = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CachedContentsPatch.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CachedContentsPatch.g.cs
index 330851a..eff7e09 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CachedContentsPatch.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CachedContentsPatch.g.cs
@@ -25,32 +25,32 @@ public partial interface IGeminiClient
///
///
///
- ///
- /// Metadata on the usage of the cached content.
+ ///
+ /// Input only. New TTL for this resource, input only.
+ ///
+ ///
+ /// Optional. Input only. Immutable. The content to cache.
+ ///
+ ///
+ /// Timestamp in UTC of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input.
///
///
/// Optional. Immutable. The user-generated meaningful display name of the cached content. Maximum 128 Unicode characters.
///
- ///
- /// Required. Immutable. The name of the `Model` to use for cached content Format: `models/{model}`
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
- ///
- /// Timestamp in UTC of when this resource is considered expired. This is *always* provided on output, regardless of what was sent on input.
+ ///
+ /// Metadata on the usage of the cached content.
///
- ///
- /// Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response
+ ///
+ /// Required. Immutable. The name of the `Model` to use for cached content Format: `models/{model}`
///
///
/// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Input only. New TTL for this resource, input only.
- ///
- ///
- /// Optional. Input only. Immutable. The content to cache.
- ///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// Optional. Input only. Immutable. A list of `Tools` the model may use to generate the next response
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
@@ -58,15 +58,15 @@ public partial interface IGeminiClient
global::System.Threading.Tasks.Task CachedContentsPatchAsync(
string cachedContentsId,
string? updateMask = default,
- global::Google.Gemini.CachedContentUsageMetadata? usageMetadata = default,
- string? displayName = default,
- string? model = default,
- string? expireTime = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
string? ttl = default,
global::System.Collections.Generic.IList? contents = default,
+ string? expireTime = default,
+ string? displayName = default,
global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::Google.Gemini.CachedContentUsageMetadata? usageMetadata = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CorporaPermissionsCreate.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CorporaPermissionsCreate.g.cs
index e6352a7..92284c1 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CorporaPermissionsCreate.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CorporaPermissionsCreate.g.cs
@@ -22,23 +22,23 @@ public partial interface IGeminiClient
/// Create a permission to a specific resource.
///
///
- ///
- /// Required. The role granted by this permission.
- ///
///
/// Optional. Immutable. The type of the grantee.
///
///
/// Optional. Immutable. The email address of the user of group which this permission refers. Field is not set when permission's grantee type is EVERYONE.
///
+ ///
+ /// Required. The role granted by this permission.
+ ///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Threading.Tasks.Task CorporaPermissionsCreateAsync(
string corporaId,
- global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.PermissionGranteeType? granteeType = default,
string? emailAddress = default,
+ global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CorporaPermissionsPatch.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CorporaPermissionsPatch.g.cs
index f49f85d..07ba825 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CorporaPermissionsPatch.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.CorporaPermissionsPatch.g.cs
@@ -28,15 +28,15 @@ public partial interface IGeminiClient
///
///
///
- ///
- /// Required. The role granted by this permission.
- ///
///
/// Optional. Immutable. The type of the grantee.
///
///
/// Optional. Immutable. The email address of the user of group which this permission refers. Field is not set when permission's grantee type is EVERYONE.
///
+ ///
+ /// Required. The role granted by this permission.
+ ///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
@@ -44,9 +44,9 @@ public partial interface IGeminiClient
string permissionsId,
string corporaId,
string? updateMask = default,
- global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.PermissionGranteeType? granteeType = default,
string? emailAddress = default,
+ global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.DynamicGenerateContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.DynamicGenerateContent.g.cs
index 32e0e9a..614c29d 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.DynamicGenerateContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.DynamicGenerateContent.g.cs
@@ -22,51 +22,51 @@ public partial interface IGeminiClient
/// Generates a model response given an input `GenerateContentRequest`. Refer to the [text generation guide](https://ai.google.dev/gemini-api/docs/text-generation) for detailed usage information. Input capabilities differ between models, including tuned models. Refer to the [model guide](https://ai.google.dev/gemini-api/docs/models/gemini) and [tuning guide](https://ai.google.dev/gemini-api/docs/model-tuning) for details.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Threading.Tasks.Task DynamicGenerateContentAsync(
string dynamicId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.DynamicStreamGenerateContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.DynamicStreamGenerateContent.g.cs
index 94a2c89..61bab0e 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.DynamicStreamGenerateContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.DynamicStreamGenerateContent.g.cs
@@ -22,51 +22,51 @@ public partial interface IGeminiClient
/// Generates a [streamed response](https://ai.google.dev/gemini-api/docs/text-generation?lang=python#generate-a-text-stream) from the model given an input `GenerateContentRequest`.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Threading.Tasks.Task DynamicStreamGenerateContentAsync(
string dynamicId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.DynamicStreamGenerateContentAsStream.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.DynamicStreamGenerateContentAsStream.g.cs
index cab906f..c9dc80a 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.DynamicStreamGenerateContentAsStream.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.DynamicStreamGenerateContentAsStream.g.cs
@@ -22,51 +22,51 @@ public partial interface IGeminiClient
/// Generates a [streamed response](https://ai.google.dev/gemini-api/docs/text-generation?lang=python#generate-a-text-stream) from the model given an input `GenerateContentRequest`.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Collections.Generic.IAsyncEnumerable DynamicStreamGenerateContentAsStreamAsync(
string dynamicId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsCountTokens.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsCountTokens.g.cs
index 3de073d..5fd7ed1 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsCountTokens.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsCountTokens.g.cs
@@ -22,19 +22,19 @@ public partial interface IGeminiClient
/// Runs a model's tokenizer on input `Content` and returns the token count. Refer to the [tokens guide](https://ai.google.dev/gemini-api/docs/tokens) to learn more about tokens.
///
///
- ///
- /// Optional. The input given to the model as a prompt. This field is ignored when `generate_content_request` is set.
- ///
///
/// Request to generate a completion from the model.
///
+ ///
+ /// Optional. The input given to the model as a prompt. This field is ignored when `generate_content_request` is set.
+ ///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Threading.Tasks.Task ModelsCountTokensAsync(
string modelsId,
- global::System.Collections.Generic.IList? contents = default,
global::Google.Gemini.GenerateContentRequest? generateContentRequest = default,
+ global::System.Collections.Generic.IList? contents = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsEmbedContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsEmbedContent.g.cs
index 481513d..915a5b9 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsEmbedContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsEmbedContent.g.cs
@@ -22,8 +22,8 @@ public partial interface IGeminiClient
/// Generates a text embedding vector from the input `Content` using the specified [Gemini Embedding model](https://ai.google.dev/gemini-api/docs/models/gemini#text-embedding).
///
///
- ///
- /// Required. The model's resource name. This serves as an ID for the Model to use. This name should match a model name returned by the `ListModels` method. Format: `models/{model}`
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
///
/// Optional. An optional title for the text. Only applicable when TaskType is `RETRIEVAL_DOCUMENT`. Note: Specifying a `title` for `RETRIEVAL_DOCUMENT` provides better quality embeddings for retrieval.
@@ -34,19 +34,19 @@ public partial interface IGeminiClient
///
/// Optional. Optional reduced dimension for the output embedding. If set, excessive values in the output embedding are truncated from the end. Supported by newer models since 2024 only. You cannot set this value if using the earlier model (`models/embedding-001`).
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// Required. The model's resource name. This serves as an ID for the Model to use. This name should match a model name returned by the `ListModels` method. Format: `models/{model}`
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Threading.Tasks.Task ModelsEmbedContentAsync(
string modelsId,
- string? model = default,
+ global::Google.Gemini.Content? content = default,
string? title = default,
global::Google.Gemini.EmbedContentRequestTaskType? taskType = default,
int? outputDimensionality = default,
- global::Google.Gemini.Content? content = default,
+ string? model = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsGenerateContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsGenerateContent.g.cs
index 34979d4..66a769b 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsGenerateContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsGenerateContent.g.cs
@@ -22,51 +22,51 @@ public partial interface IGeminiClient
/// Generates a model response given an input `GenerateContentRequest`. Refer to the [text generation guide](https://ai.google.dev/gemini-api/docs/text-generation) for detailed usage information. Input capabilities differ between models, including tuned models. Refer to the [model guide](https://ai.google.dev/gemini-api/docs/models/gemini) and [tuning guide](https://ai.google.dev/gemini-api/docs/model-tuning) for details.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Threading.Tasks.Task ModelsGenerateContentAsync(
string modelsId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsStreamGenerateContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsStreamGenerateContent.g.cs
index 8096b3a..7b21ea5 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsStreamGenerateContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsStreamGenerateContent.g.cs
@@ -22,51 +22,51 @@ public partial interface IGeminiClient
/// Generates a [streamed response](https://ai.google.dev/gemini-api/docs/text-generation?lang=python#generate-a-text-stream) from the model given an input `GenerateContentRequest`.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Threading.Tasks.Task ModelsStreamGenerateContentAsync(
string modelsId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsStreamGenerateContentAsStream.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsStreamGenerateContentAsStream.g.cs
index dbdfba0..c495761 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsStreamGenerateContentAsStream.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.ModelsStreamGenerateContentAsStream.g.cs
@@ -22,51 +22,51 @@ public partial interface IGeminiClient
/// Generates a [streamed response](https://ai.google.dev/gemini-api/docs/text-generation?lang=python#generate-a-text-stream) from the model given an input `GenerateContentRequest`.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Collections.Generic.IAsyncEnumerable ModelsStreamGenerateContentAsStreamAsync(
string modelsId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsCreate.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsCreate.g.cs
index 582d46e..d8a97fc 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsCreate.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsCreate.g.cs
@@ -22,17 +22,14 @@ public partial interface IGeminiClient
/// Creates a tuned model. Check intermediate tuning progress (if any) through the [google.longrunning.Operations] service. Access status and results through the Operations service. Example: GET /v1/tunedModels/az2mb0bpw6i/operations/000-111-222
///
///
- ///
- /// Optional. For Nucleus sampling. Nucleus sampling considers the smallest set of tokens whose probability sum is at least `top_p`. This value specifies default to be the one used by the base model while creating the model.
- ///
- ///
- /// Optional. List of project numbers that have read access to the tuned model.
- ///
///
/// Tuned model as a source for training a new model.
///
- ///
- /// Optional. For Top-k sampling. Top-k sampling considers the set of `top_k` most probable tokens. This value specifies default to be used by the backend while making the call to the model. This value specifies default to be the one used by the base model while creating the model.
+ ///
+ /// Optional. For Nucleus sampling. Nucleus sampling considers the smallest set of tokens whose probability sum is at least `top_p`. This value specifies default to be the one used by the base model while creating the model.
+ ///
+ ///
+ /// Tuning tasks that create tuned models.
///
///
/// Immutable. The name of the `Model` to tune. Example: `models/gemini-1.5-flash-001`
@@ -40,8 +37,11 @@ public partial interface IGeminiClient
///
/// Optional. The name to display for this model in user interfaces. The display name must be up to 40 characters including spaces.
///
- ///
- /// Tuning tasks that create tuned models.
+ ///
+ /// Optional. For Top-k sampling. Top-k sampling considers the set of `top_k` most probable tokens. This value specifies default to be used by the backend while making the call to the model. This value specifies default to be the one used by the base model while creating the model.
+ ///
+ ///
+ /// Optional. List of project numbers that have read access to the tuned model.
///
///
/// Optional. A short description of this model.
@@ -54,13 +54,13 @@ public partial interface IGeminiClient
///
global::System.Threading.Tasks.Task TunedModelsCreateAsync(
string? tunedModelId = default,
- float? topP = default,
- global::System.Collections.Generic.IList? readerProjectNumbers = default,
global::Google.Gemini.TunedModelSource? tunedModelSource = default,
- int? topK = default,
+ float? topP = default,
+ global::Google.Gemini.TuningTask? tuningTask = default,
string? baseModel = default,
string? displayName = default,
- global::Google.Gemini.TuningTask? tuningTask = default,
+ int? topK = default,
+ global::System.Collections.Generic.IList? readerProjectNumbers = default,
string? description = default,
float? temperature = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsGenerateContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsGenerateContent.g.cs
index ae70bfa..1f08912 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsGenerateContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsGenerateContent.g.cs
@@ -22,51 +22,51 @@ public partial interface IGeminiClient
/// Generates a model response given an input `GenerateContentRequest`. Refer to the [text generation guide](https://ai.google.dev/gemini-api/docs/text-generation) for detailed usage information. Input capabilities differ between models, including tuned models. Refer to the [model guide](https://ai.google.dev/gemini-api/docs/models/gemini) and [tuning guide](https://ai.google.dev/gemini-api/docs/model-tuning) for details.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Threading.Tasks.Task TunedModelsGenerateContentAsync(
string tunedModelsId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsPatch.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsPatch.g.cs
index a3ade84..f796626 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsPatch.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsPatch.g.cs
@@ -25,17 +25,14 @@ public partial interface IGeminiClient
///
///
///
- ///
- /// Optional. For Nucleus sampling. Nucleus sampling considers the smallest set of tokens whose probability sum is at least `top_p`. This value specifies default to be the one used by the base model while creating the model.
- ///
- ///
- /// Optional. List of project numbers that have read access to the tuned model.
- ///
///
/// Tuned model as a source for training a new model.
///
- ///
- /// Optional. For Top-k sampling. Top-k sampling considers the set of `top_k` most probable tokens. This value specifies default to be used by the backend while making the call to the model. This value specifies default to be the one used by the base model while creating the model.
+ ///
+ /// Optional. For Nucleus sampling. Nucleus sampling considers the smallest set of tokens whose probability sum is at least `top_p`. This value specifies default to be the one used by the base model while creating the model.
+ ///
+ ///
+ /// Tuning tasks that create tuned models.
///
///
/// Immutable. The name of the `Model` to tune. Example: `models/gemini-1.5-flash-001`
@@ -43,8 +40,11 @@ public partial interface IGeminiClient
///
/// Optional. The name to display for this model in user interfaces. The display name must be up to 40 characters including spaces.
///
- ///
- /// Tuning tasks that create tuned models.
+ ///
+ /// Optional. For Top-k sampling. Top-k sampling considers the set of `top_k` most probable tokens. This value specifies default to be used by the backend while making the call to the model. This value specifies default to be the one used by the base model while creating the model.
+ ///
+ ///
+ /// Optional. List of project numbers that have read access to the tuned model.
///
///
/// Optional. A short description of this model.
@@ -58,13 +58,13 @@ public partial interface IGeminiClient
global::System.Threading.Tasks.Task TunedModelsPatchAsync(
string tunedModelsId,
string? updateMask = default,
- float? topP = default,
- global::System.Collections.Generic.IList? readerProjectNumbers = default,
global::Google.Gemini.TunedModelSource? tunedModelSource = default,
- int? topK = default,
+ float? topP = default,
+ global::Google.Gemini.TuningTask? tuningTask = default,
string? baseModel = default,
string? displayName = default,
- global::Google.Gemini.TuningTask? tuningTask = default,
+ int? topK = default,
+ global::System.Collections.Generic.IList? readerProjectNumbers = default,
string? description = default,
float? temperature = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsPermissionsCreate.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsPermissionsCreate.g.cs
index 0b0f721..10d0529 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsPermissionsCreate.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsPermissionsCreate.g.cs
@@ -22,23 +22,23 @@ public partial interface IGeminiClient
/// Create a permission to a specific resource.
///
///
- ///
- /// Required. The role granted by this permission.
- ///
///
/// Optional. Immutable. The type of the grantee.
///
///
/// Optional. Immutable. The email address of the user of group which this permission refers. Field is not set when permission's grantee type is EVERYONE.
///
+ ///
+ /// Required. The role granted by this permission.
+ ///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Threading.Tasks.Task TunedModelsPermissionsCreateAsync(
string tunedModelsId,
- global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.PermissionGranteeType? granteeType = default,
string? emailAddress = default,
+ global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsPermissionsPatch.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsPermissionsPatch.g.cs
index 7becdff..8cd2556 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsPermissionsPatch.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsPermissionsPatch.g.cs
@@ -28,15 +28,15 @@ public partial interface IGeminiClient
///
///
///
- ///
- /// Required. The role granted by this permission.
- ///
///
/// Optional. Immutable. The type of the grantee.
///
///
/// Optional. Immutable. The email address of the user of group which this permission refers. Field is not set when permission's grantee type is EVERYONE.
///
+ ///
+ /// Required. The role granted by this permission.
+ ///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
@@ -44,9 +44,9 @@ public partial interface IGeminiClient
string tunedModelsId,
string permissionsId,
string? updateMask = default,
- global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.PermissionGranteeType? granteeType = default,
string? emailAddress = default,
+ global::Google.Gemini.PermissionRole? role = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsStreamGenerateContent.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsStreamGenerateContent.g.cs
index b4bef5b..05857bd 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsStreamGenerateContent.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsStreamGenerateContent.g.cs
@@ -22,51 +22,51 @@ public partial interface IGeminiClient
/// Generates a [streamed response](https://ai.google.dev/gemini-api/docs/text-generation?lang=python#generate-a-text-stream) from the model given an input `GenerateContentRequest`.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Threading.Tasks.Task TunedModelsStreamGenerateContentAsync(
string tunedModelsId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsStreamGenerateContentAsStream.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsStreamGenerateContentAsStream.g.cs
index bf7424c..8d1fa9a 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsStreamGenerateContentAsStream.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.IGeminiClient.TunedModelsStreamGenerateContentAsStream.g.cs
@@ -22,51 +22,51 @@ public partial interface IGeminiClient
/// Generates a [streamed response](https://ai.google.dev/gemini-api/docs/text-generation?lang=python#generate-a-text-stream) from the model given an input `GenerateContentRequest`.
///
///
- ///
- /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
- ///
- ///
- /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
- ///
///
/// Optional. The service tier of the request.
///
- ///
- /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
+ ///
+ /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
///
- ///
- /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
///
- ///
- /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
+ ///
+ /// The Tool configuration containing parameters for specifying `Tool` use in the request.
///
///
/// Optional. A list of unique `SafetySetting` instances for blocking unsafe content. This will be enforced on the `GenerateContentRequest.contents` and `GenerateContentResponse.candidates`. There should not be more than one setting for each `SafetyCategory` type. The API will block any contents and responses that fail to meet the thresholds set by these settings. This list overrides the default settings for each `SafetyCategory` specified in the safety_settings. If there is no `SafetySetting` for a given `SafetyCategory` provided in the list, the API will use the default safety setting for that category. Harm categories HARM_CATEGORY_HATE_SPEECH, HARM_CATEGORY_SEXUALLY_EXPLICIT, HARM_CATEGORY_DANGEROUS_CONTENT, HARM_CATEGORY_HARASSMENT, HARM_CATEGORY_CIVIC_INTEGRITY are supported. Refer to the [guide](https://ai.google.dev/gemini-api/docs/safety-settings) for detailed information on available safety settings. Also refer to the [Safety guidance](https://ai.google.dev/gemini-api/docs/safety-guidance) to learn how to incorporate safety considerations in your AI applications.
///
- ///
- /// Required. The content of the current conversation with the model. For single-turn queries, this is a single instance. For multi-turn queries like [chat](https://ai.google.dev/gemini-api/docs/text-generation#chat), this is a repeated field that contains the conversation history and the latest request.
+ ///
+ /// Required. The name of the `Model` to use for generating the completion. Format: `models/{model}`.
///
- ///
- /// The Tool configuration containing parameters for specifying `Tool` use in the request.
+ ///
+ /// The base structured datatype containing multi-part content of a message. A `Content` includes a `role` field designating the producer of the `Content` and a `parts` field containing multi-part data that contains the content of the message turn.
///
- ///
- /// Optional. Configures the logging behavior for a given request. If set, it takes precedence over the project-level logging config.
+ ///
+ /// Optional. A list of `Tools` the `Model` may use to generate the next response. A `Tool` is a piece of code that enables the system to interact with external systems to perform an action, or set of actions, outside of knowledge and scope of the `Model`. Supported `Tool`s are `Function` and `code_execution`. Refer to the [Function calling](https://ai.google.dev/gemini-api/docs/function-calling) and the [Code execution](https://ai.google.dev/gemini-api/docs/code-execution) guides to learn more.
+ ///
+ ///
+ /// Optional. The name of the content [cached](https://ai.google.dev/gemini-api/docs/caching) to use as context to serve the prediction. Format: `cachedContents/{cachedContent}`
+ ///
+ ///
+ /// Configuration options for model generation and outputs. Not all parameters are configurable for every model.
///
/// Per-request overrides such as headers, query parameters, timeout, retries, and response buffering.
/// The token to cancel the operation with
///
global::System.Collections.Generic.IAsyncEnumerable TunedModelsStreamGenerateContentAsStreamAsync(
string tunedModelsId,
- string? cachedContent = default,
- global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.GenerateContentRequestServiceTier? serviceTier = default,
- string? model = default,
- global::System.Collections.Generic.IList? tools = default,
- global::Google.Gemini.Content? systemInstruction = default,
- global::System.Collections.Generic.IList? safetySettings = default,
global::System.Collections.Generic.IList? contents = default,
- global::Google.Gemini.ToolConfig? toolConfig = default,
bool? store = default,
+ global::Google.Gemini.ToolConfig? toolConfig = default,
+ global::System.Collections.Generic.IList? safetySettings = default,
+ string? model = default,
+ global::Google.Gemini.Content? systemInstruction = default,
+ global::System.Collections.Generic.IList? tools = default,
+ string? cachedContent = default,
+ global::Google.Gemini.GenerationConfig? generationConfig = default,
global::Google.Gemini.AutoSDKRequestOptions? requestOptions = default,
global::System.Threading.CancellationToken cancellationToken = default);
}
diff --git a/src/libs/Google.Gemini/Generated/Google.Gemini.JsonSerializerContext.g.cs b/src/libs/Google.Gemini/Generated/Google.Gemini.JsonSerializerContext.g.cs
index 75b2c20..f14d017 100644
--- a/src/libs/Google.Gemini/Generated/Google.Gemini.JsonSerializerContext.g.cs
+++ b/src/libs/Google.Gemini/Generated/Google.Gemini.JsonSerializerContext.g.cs
@@ -13,10 +13,6 @@ namespace Google.Gemini
DefaultIgnoreCondition = global::System.Text.Json.Serialization.JsonIgnoreCondition.WhenWritingNull,
Converters = new global::System.Type[]
{
- typeof(global::Google.Gemini.JsonConverters.FunctionCallingConfigModeJsonConverter),
-
- typeof(global::Google.Gemini.JsonConverters.FunctionCallingConfigModeNullableJsonConverter),
-
typeof(global::Google.Gemini.JsonConverters.GenerationConfigMediaResolutionJsonConverter),
typeof(global::Google.Gemini.JsonConverters.GenerationConfigMediaResolutionNullableJsonConverter),
@@ -25,50 +21,34 @@ namespace Google.Gemini
typeof(global::Google.Gemini.JsonConverters.GenerationConfigResponseModalitieNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.FileSourceJsonConverter),
-
- typeof(global::Google.Gemini.JsonConverters.FileSourceNullableJsonConverter),
-
- typeof(global::Google.Gemini.JsonConverters.FileStateJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.GenerateContentBatchStateJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.FileStateNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.GenerateContentBatchStateNullableJsonConverter),
typeof(global::Google.Gemini.JsonConverters.CandidateFinishReasonJsonConverter),
typeof(global::Google.Gemini.JsonConverters.CandidateFinishReasonNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.GeneratedFileStateJsonConverter),
-
- typeof(global::Google.Gemini.JsonConverters.GeneratedFileStateNullableJsonConverter),
-
- typeof(global::Google.Gemini.JsonConverters.ComputerUseEnvironmentJsonConverter),
-
- typeof(global::Google.Gemini.JsonConverters.ComputerUseEnvironmentNullableJsonConverter),
-
- typeof(global::Google.Gemini.JsonConverters.DynamicRetrievalConfigModeJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ToolResponseToolTypeJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.DynamicRetrievalConfigModeNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ToolResponseToolTypeNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.ThinkingConfigThinkingLevelJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ModalityTokenCountModalityJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.ThinkingConfigThinkingLevelNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ModalityTokenCountModalityNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.ModelStatusModelStageJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.GenerateContentRequestServiceTierJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.ModelStatusModelStageNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.GenerateContentRequestServiceTierNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.CodeExecutionResultOutcomeJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.UrlMetadataUrlRetrievalStatusJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.CodeExecutionResultOutcomeNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.UrlMetadataUrlRetrievalStatusNullableJsonConverter),
typeof(global::Google.Gemini.JsonConverters.TunedModelStateJsonConverter),
typeof(global::Google.Gemini.JsonConverters.TunedModelStateNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.DocumentStateJsonConverter),
-
- typeof(global::Google.Gemini.JsonConverters.DocumentStateNullableJsonConverter),
-
typeof(global::Google.Gemini.JsonConverters.SafetySettingThresholdJsonConverter),
typeof(global::Google.Gemini.JsonConverters.SafetySettingThresholdNullableJsonConverter),
@@ -77,34 +57,46 @@ namespace Google.Gemini
typeof(global::Google.Gemini.JsonConverters.SafetySettingCategoryNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.PermissionRoleJsonConverter),
-
- typeof(global::Google.Gemini.JsonConverters.PermissionRoleNullableJsonConverter),
-
- typeof(global::Google.Gemini.JsonConverters.PermissionGranteeTypeJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.PromptFeedbackBlockReasonJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.PermissionGranteeTypeNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.PromptFeedbackBlockReasonNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.ToolCallToolTypeJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ComputerUseEnvironmentJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.ToolCallToolTypeNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ComputerUseEnvironmentNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.ToolResponseToolTypeJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.CodeExecutionResultOutcomeJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.ToolResponseToolTypeNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.CodeExecutionResultOutcomeNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.ExecutableCodeLanguageJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.DocumentStateJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.ExecutableCodeLanguageNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.DocumentStateNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.FunctionResponseSchedulingJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.FunctionCallingConfigModeJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.FunctionResponseSchedulingNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.FunctionCallingConfigModeNullableJsonConverter),
typeof(global::Google.Gemini.JsonConverters.EmbedContentBatchStateJsonConverter),
typeof(global::Google.Gemini.JsonConverters.EmbedContentBatchStateNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.GeneratedFileStateJsonConverter),
+
+ typeof(global::Google.Gemini.JsonConverters.GeneratedFileStateNullableJsonConverter),
+
+ typeof(global::Google.Gemini.JsonConverters.DynamicRetrievalConfigModeJsonConverter),
+
+ typeof(global::Google.Gemini.JsonConverters.DynamicRetrievalConfigModeNullableJsonConverter),
+
+ typeof(global::Google.Gemini.JsonConverters.PermissionGranteeTypeJsonConverter),
+
+ typeof(global::Google.Gemini.JsonConverters.PermissionGranteeTypeNullableJsonConverter),
+
+ typeof(global::Google.Gemini.JsonConverters.PermissionRoleJsonConverter),
+
+ typeof(global::Google.Gemini.JsonConverters.PermissionRoleNullableJsonConverter),
+
typeof(global::Google.Gemini.JsonConverters.SafetyRatingCategoryJsonConverter),
typeof(global::Google.Gemini.JsonConverters.SafetyRatingCategoryNullableJsonConverter),
@@ -113,319 +105,327 @@ namespace Google.Gemini
typeof(global::Google.Gemini.JsonConverters.SafetyRatingProbabilityNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.FunctionDeclarationBehaviorJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ToolCallToolTypeJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.FunctionDeclarationBehaviorNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ToolCallToolTypeNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.PromptFeedbackBlockReasonJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ModelStatusModelStageJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.PromptFeedbackBlockReasonNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ModelStatusModelStageNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.UrlMetadataUrlRetrievalStatusJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ExecutableCodeLanguageJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.UrlMetadataUrlRetrievalStatusNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ExecutableCodeLanguageNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.EmbedContentRequestTaskTypeJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.FileStateJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.EmbedContentRequestTaskTypeNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.FileStateNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.SchemaTypeJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.FileSourceJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.SchemaTypeNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.FileSourceNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.GenerateContentRequestServiceTierJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ThinkingConfigThinkingLevelJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.GenerateContentRequestServiceTierNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.ThinkingConfigThinkingLevelNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.GenerateContentBatchStateJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.FunctionDeclarationBehaviorJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.GenerateContentBatchStateNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.FunctionDeclarationBehaviorNullableJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.ModalityTokenCountModalityJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.EmbedContentRequestTaskTypeJsonConverter),
- typeof(global::Google.Gemini.JsonConverters.ModalityTokenCountModalityNullableJsonConverter),
+ typeof(global::Google.Gemini.JsonConverters.EmbedContentRequestTaskTypeNullableJsonConverter),
+
+ typeof(global::Google.Gemini.JsonConverters.FunctionResponseSchedulingJsonConverter),
+
+ typeof(global::Google.Gemini.JsonConverters.FunctionResponseSchedulingNullableJsonConverter),
+
+ typeof(global::Google.Gemini.JsonConverters.SchemaTypeJsonConverter),
+
+ typeof(global::Google.Gemini.JsonConverters.SchemaTypeNullableJsonConverter),
typeof(global::Google.Gemini.JsonConverters.UnixTimestampJsonConverter),
})]
[global::System.Text.Json.Serialization.JsonSerializable(typeof(global::Google.Gemini.JsonSerializerContextTypes))]
- [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::Google.Gemini.FunctionCallingConfig))]
- [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::System.Collections.Generic.IList))]
- [global::System.Text.Json.Serialization.JsonSerializable(typeof(string))]
- [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::Google.Gemini.FunctionCallingConfigMode), TypeInfoPropertyName = "FunctionCallingConfigMode2")]
- [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::Google.Gemini.EmbedContentBatchOutput))]
- [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::Google.Gemini.InlinedEmbedContentResponses))]
- [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::Google.Gemini.Status))]
- [global::System.Text.Json.Serialization.JsonSerializable(typeof(global::System.Collections.Generic.IList