diff --git a/src/libs/Ollama/Generated/Ollama.ChatClient.GenerateChatCompletion.g.cs b/src/libs/Ollama/Generated/Ollama.ChatClient.GenerateChatCompletion.g.cs index 19f428b..adda58f 100644 --- a/src/libs/Ollama/Generated/Ollama.ChatClient.GenerateChatCompletion.g.cs +++ b/src/libs/Ollama/Generated/Ollama.ChatClient.GenerateChatCompletion.g.cs @@ -120,7 +120,7 @@ partial void ProcessGenerateChatCompletionResponse( global::System.Collections.Generic.IList messages, global::Ollama.ResponseFormat? format = default, global::Ollama.RequestOptions? options = default, - bool stream = true, + bool? stream = true, int? keepAlive = default, global::System.Collections.Generic.IList? tools = default, [global::System.Runtime.CompilerServices.EnumeratorCancellation] global::System.Threading.CancellationToken cancellationToken = default) diff --git a/src/libs/Ollama/Generated/Ollama.CompletionsClient.GenerateCompletion.g.cs b/src/libs/Ollama/Generated/Ollama.CompletionsClient.GenerateCompletion.g.cs index 36763b4..8782c33 100644 --- a/src/libs/Ollama/Generated/Ollama.CompletionsClient.GenerateCompletion.g.cs +++ b/src/libs/Ollama/Generated/Ollama.CompletionsClient.GenerateCompletion.g.cs @@ -142,8 +142,8 @@ partial void ProcessGenerateCompletionResponse( global::System.Collections.Generic.IList? context = default, global::Ollama.RequestOptions? options = default, global::Ollama.ResponseFormat? format = default, - bool raw = default, - bool stream = true, + bool? raw = default, + bool? stream = true, int? keepAlive = default, [global::System.Runtime.CompilerServices.EnumeratorCancellation] global::System.Threading.CancellationToken cancellationToken = default) { diff --git a/src/libs/Ollama/Generated/Ollama.IChatClient.GenerateChatCompletion.g.cs b/src/libs/Ollama/Generated/Ollama.IChatClient.GenerateChatCompletion.g.cs index 9b6fe84..fb4ec5e 100644 --- a/src/libs/Ollama/Generated/Ollama.IChatClient.GenerateChatCompletion.g.cs +++ b/src/libs/Ollama/Generated/Ollama.IChatClient.GenerateChatCompletion.g.cs @@ -56,7 +56,7 @@ public partial interface IChatClient global::System.Collections.Generic.IList messages, global::Ollama.ResponseFormat? format = default, global::Ollama.RequestOptions? options = default, - bool stream = true, + bool? stream = true, int? keepAlive = default, global::System.Collections.Generic.IList? tools = default, global::System.Threading.CancellationToken cancellationToken = default); diff --git a/src/libs/Ollama/Generated/Ollama.ICompletionsClient.GenerateCompletion.g.cs b/src/libs/Ollama/Generated/Ollama.ICompletionsClient.GenerateCompletion.g.cs index 8905b6f..f3f24bf 100644 --- a/src/libs/Ollama/Generated/Ollama.ICompletionsClient.GenerateCompletion.g.cs +++ b/src/libs/Ollama/Generated/Ollama.ICompletionsClient.GenerateCompletion.g.cs @@ -78,8 +78,8 @@ public partial interface ICompletionsClient global::System.Collections.Generic.IList? context = default, global::Ollama.RequestOptions? options = default, global::Ollama.ResponseFormat? format = default, - bool raw = default, - bool stream = true, + bool? raw = default, + bool? stream = true, int? keepAlive = default, global::System.Threading.CancellationToken cancellationToken = default); } diff --git a/src/libs/Ollama/Generated/Ollama.IModelsClient.CreateModel.g.cs b/src/libs/Ollama/Generated/Ollama.IModelsClient.CreateModel.g.cs index 3cc63b5..514fdef 100644 --- a/src/libs/Ollama/Generated/Ollama.IModelsClient.CreateModel.g.cs +++ b/src/libs/Ollama/Generated/Ollama.IModelsClient.CreateModel.g.cs @@ -45,7 +45,7 @@ public partial interface IModelsClient string modelfile, string? path = default, string? quantize = default, - bool stream = true, + bool? stream = true, global::System.Threading.CancellationToken cancellationToken = default); } } \ No newline at end of file diff --git a/src/libs/Ollama/Generated/Ollama.IModelsClient.PullModel.g.cs b/src/libs/Ollama/Generated/Ollama.IModelsClient.PullModel.g.cs index 89c765f..175c008 100644 --- a/src/libs/Ollama/Generated/Ollama.IModelsClient.PullModel.g.cs +++ b/src/libs/Ollama/Generated/Ollama.IModelsClient.PullModel.g.cs @@ -43,10 +43,10 @@ public partial interface IModelsClient /// global::System.Collections.Generic.IAsyncEnumerable PullModelAsync( string model, - bool insecure = false, + bool? insecure = false, string? username = default, string? password = default, - bool stream = true, + bool? stream = true, global::System.Threading.CancellationToken cancellationToken = default); } } \ No newline at end of file diff --git a/src/libs/Ollama/Generated/Ollama.IModelsClient.PushModel.g.cs b/src/libs/Ollama/Generated/Ollama.IModelsClient.PushModel.g.cs index e9099a5..3f6feeb 100644 --- a/src/libs/Ollama/Generated/Ollama.IModelsClient.PushModel.g.cs +++ b/src/libs/Ollama/Generated/Ollama.IModelsClient.PushModel.g.cs @@ -42,10 +42,10 @@ public partial interface IModelsClient /// global::System.Collections.Generic.IAsyncEnumerable PushModelAsync( string model, - bool insecure = false, + bool? insecure = false, string? username = default, string? password = default, - bool stream = true, + bool? stream = true, global::System.Threading.CancellationToken cancellationToken = default); } } \ No newline at end of file diff --git a/src/libs/Ollama/Generated/Ollama.Models.CreateModelRequest.g.cs b/src/libs/Ollama/Generated/Ollama.Models.CreateModelRequest.g.cs index 34d533e..51aa1af 100644 --- a/src/libs/Ollama/Generated/Ollama.Models.CreateModelRequest.g.cs +++ b/src/libs/Ollama/Generated/Ollama.Models.CreateModelRequest.g.cs @@ -42,7 +42,7 @@ public sealed partial class CreateModelRequest /// Default Value: true /// [global::System.Text.Json.Serialization.JsonPropertyName("stream")] - public bool Stream { get; set; } = true; + public bool? Stream { get; set; } = true; /// /// Additional properties that are not explicitly defined in the schema diff --git a/src/libs/Ollama/Generated/Ollama.Models.GenerateChatCompletionRequest.g.cs b/src/libs/Ollama/Generated/Ollama.Models.GenerateChatCompletionRequest.g.cs index cf518d3..e61f7ea 100644 --- a/src/libs/Ollama/Generated/Ollama.Models.GenerateChatCompletionRequest.g.cs +++ b/src/libs/Ollama/Generated/Ollama.Models.GenerateChatCompletionRequest.g.cs @@ -44,7 +44,7 @@ public sealed partial class GenerateChatCompletionRequest /// Default Value: true /// [global::System.Text.Json.Serialization.JsonPropertyName("stream")] - public bool Stream { get; set; } = true; + public bool? Stream { get; set; } = true; /// /// How long (in minutes) to keep the model loaded in memory.
diff --git a/src/libs/Ollama/Generated/Ollama.Models.GenerateChatCompletionResponse.g.cs b/src/libs/Ollama/Generated/Ollama.Models.GenerateChatCompletionResponse.g.cs index 54c2524..ec9919c 100644 --- a/src/libs/Ollama/Generated/Ollama.Models.GenerateChatCompletionResponse.g.cs +++ b/src/libs/Ollama/Generated/Ollama.Models.GenerateChatCompletionResponse.g.cs @@ -51,42 +51,42 @@ public sealed partial class GenerateChatCompletionResponse /// Example: 5589157167L ///
[global::System.Text.Json.Serialization.JsonPropertyName("total_duration")] - public long TotalDuration { get; set; } + public long? TotalDuration { get; set; } /// /// Time spent in nanoseconds loading the model.
/// Example: 3013701500L ///
[global::System.Text.Json.Serialization.JsonPropertyName("load_duration")] - public long LoadDuration { get; set; } + public long? LoadDuration { get; set; } /// /// Number of tokens in the prompt.
/// Example: 46 ///
[global::System.Text.Json.Serialization.JsonPropertyName("prompt_eval_count")] - public int PromptEvalCount { get; set; } + public int? PromptEvalCount { get; set; } /// /// Time spent in nanoseconds evaluating the prompt.
/// Example: 1160282000L ///
[global::System.Text.Json.Serialization.JsonPropertyName("prompt_eval_duration")] - public long PromptEvalDuration { get; set; } + public long? PromptEvalDuration { get; set; } /// /// Number of tokens the response.
/// Example: 113 ///
[global::System.Text.Json.Serialization.JsonPropertyName("eval_count")] - public int EvalCount { get; set; } + public int? EvalCount { get; set; } /// /// Time in nanoseconds spent generating the response.
/// Example: 1325948000L ///
[global::System.Text.Json.Serialization.JsonPropertyName("eval_duration")] - public long EvalDuration { get; set; } + public long? EvalDuration { get; set; } /// /// Additional properties that are not explicitly defined in the schema diff --git a/src/libs/Ollama/Generated/Ollama.Models.GenerateCompletionRequest.g.cs b/src/libs/Ollama/Generated/Ollama.Models.GenerateCompletionRequest.g.cs index fbbba4a..5b20c0f 100644 --- a/src/libs/Ollama/Generated/Ollama.Models.GenerateCompletionRequest.g.cs +++ b/src/libs/Ollama/Generated/Ollama.Models.GenerateCompletionRequest.g.cs @@ -75,14 +75,14 @@ public sealed partial class GenerateCompletionRequest /// You may choose to use the `raw` parameter if you are specifying a full templated prompt in your request to the API, and are managing history yourself. /// [global::System.Text.Json.Serialization.JsonPropertyName("raw")] - public bool Raw { get; set; } + public bool? Raw { get; set; } /// /// If `false` the response will be returned as a single response object, otherwise the response will be streamed as a series of objects.
/// Default Value: true ///
[global::System.Text.Json.Serialization.JsonPropertyName("stream")] - public bool Stream { get; set; } = true; + public bool? Stream { get; set; } = true; /// /// How long (in minutes) to keep the model loaded in memory.
diff --git a/src/libs/Ollama/Generated/Ollama.Models.GenerateCompletionResponse.g.cs b/src/libs/Ollama/Generated/Ollama.Models.GenerateCompletionResponse.g.cs index 83eece2..f4d972c 100644 --- a/src/libs/Ollama/Generated/Ollama.Models.GenerateCompletionResponse.g.cs +++ b/src/libs/Ollama/Generated/Ollama.Models.GenerateCompletionResponse.g.cs @@ -20,7 +20,7 @@ public sealed partial class GenerateCompletionResponse /// Date on which a model was created. ///
[global::System.Text.Json.Serialization.JsonPropertyName("created_at")] - public global::System.DateTime CreatedAt { get; set; } + public global::System.DateTime? CreatedAt { get; set; } /// /// The response for a given prompt with a provided model.
@@ -34,7 +34,7 @@ public sealed partial class GenerateCompletionResponse /// Example: true ///
[global::System.Text.Json.Serialization.JsonPropertyName("done")] - public bool Done { get; set; } + public bool? Done { get; set; } /// /// An encoding of the conversation used in this response, this can be sent in the next request to keep a conversational memory.
@@ -48,42 +48,42 @@ public sealed partial class GenerateCompletionResponse /// Example: 5589157167L ///
[global::System.Text.Json.Serialization.JsonPropertyName("total_duration")] - public long TotalDuration { get; set; } + public long? TotalDuration { get; set; } /// /// Time spent in nanoseconds loading the model.
/// Example: 3013701500L ///
[global::System.Text.Json.Serialization.JsonPropertyName("load_duration")] - public long LoadDuration { get; set; } + public long? LoadDuration { get; set; } /// /// Number of tokens in the prompt.
/// Example: 46 ///
[global::System.Text.Json.Serialization.JsonPropertyName("prompt_eval_count")] - public int PromptEvalCount { get; set; } + public int? PromptEvalCount { get; set; } /// /// Time spent in nanoseconds evaluating the prompt.
/// Example: 1160282000L ///
[global::System.Text.Json.Serialization.JsonPropertyName("prompt_eval_duration")] - public long PromptEvalDuration { get; set; } + public long? PromptEvalDuration { get; set; } /// /// Number of tokens the response.
/// Example: 113 ///
[global::System.Text.Json.Serialization.JsonPropertyName("eval_count")] - public int EvalCount { get; set; } + public int? EvalCount { get; set; } /// /// Time in nanoseconds spent generating the response.
/// Example: 1325948000L ///
[global::System.Text.Json.Serialization.JsonPropertyName("eval_duration")] - public long EvalDuration { get; set; } + public long? EvalDuration { get; set; } /// /// Additional properties that are not explicitly defined in the schema diff --git a/src/libs/Ollama/Generated/Ollama.Models.Model.g.cs b/src/libs/Ollama/Generated/Ollama.Models.Model.g.cs index 22d5e67..d39eeaf 100644 --- a/src/libs/Ollama/Generated/Ollama.Models.Model.g.cs +++ b/src/libs/Ollama/Generated/Ollama.Models.Model.g.cs @@ -20,14 +20,14 @@ public sealed partial class Model /// Model modification date. /// [global::System.Text.Json.Serialization.JsonPropertyName("modified_at")] - public global::System.DateTime ModifiedAt { get; set; } + public global::System.DateTime? ModifiedAt { get; set; } /// /// Size of the model on disk.
/// Example: 7323310500L ///
[global::System.Text.Json.Serialization.JsonPropertyName("size")] - public long Size { get; set; } + public long? Size { get; set; } /// /// The model's digest.
diff --git a/src/libs/Ollama/Generated/Ollama.Models.ProcessModel.g.cs b/src/libs/Ollama/Generated/Ollama.Models.ProcessModel.g.cs index 1f6385a..bb397e4 100644 --- a/src/libs/Ollama/Generated/Ollama.Models.ProcessModel.g.cs +++ b/src/libs/Ollama/Generated/Ollama.Models.ProcessModel.g.cs @@ -21,7 +21,7 @@ public sealed partial class ProcessModel /// Example: 7323310500L ///
[global::System.Text.Json.Serialization.JsonPropertyName("size")] - public long Size { get; set; } + public long? Size { get; set; } /// /// The model's digest.
@@ -40,14 +40,14 @@ public sealed partial class ProcessModel /// ///
[global::System.Text.Json.Serialization.JsonPropertyName("expires_at")] - public global::System.DateTime ExpiresAt { get; set; } + public global::System.DateTime? ExpiresAt { get; set; } /// /// Size of the model on disk.
/// Example: 7323310500L ///
[global::System.Text.Json.Serialization.JsonPropertyName("size_vram")] - public long SizeVram { get; set; } + public long? SizeVram { get; set; } /// /// Additional properties that are not explicitly defined in the schema diff --git a/src/libs/Ollama/Generated/Ollama.Models.PullModelRequest.g.cs b/src/libs/Ollama/Generated/Ollama.Models.PullModelRequest.g.cs index bfc708f..0f9ac6b 100644 --- a/src/libs/Ollama/Generated/Ollama.Models.PullModelRequest.g.cs +++ b/src/libs/Ollama/Generated/Ollama.Models.PullModelRequest.g.cs @@ -23,7 +23,7 @@ public sealed partial class PullModelRequest /// Default Value: false /// [global::System.Text.Json.Serialization.JsonPropertyName("insecure")] - public bool Insecure { get; set; } = false; + public bool? Insecure { get; set; } = false; /// /// Ollama username. @@ -42,7 +42,7 @@ public sealed partial class PullModelRequest /// Default Value: true /// [global::System.Text.Json.Serialization.JsonPropertyName("stream")] - public bool Stream { get; set; } = true; + public bool? Stream { get; set; } = true; /// /// Additional properties that are not explicitly defined in the schema diff --git a/src/libs/Ollama/Generated/Ollama.Models.PullModelResponse.g.cs b/src/libs/Ollama/Generated/Ollama.Models.PullModelResponse.g.cs index 0a7846a..97df866 100644 --- a/src/libs/Ollama/Generated/Ollama.Models.PullModelResponse.g.cs +++ b/src/libs/Ollama/Generated/Ollama.Models.PullModelResponse.g.cs @@ -30,14 +30,14 @@ public sealed partial class PullModelResponse /// Example: 2142590208L /// [global::System.Text.Json.Serialization.JsonPropertyName("total")] - public long Total { get; set; } + public long? Total { get; set; } /// /// Total bytes transferred.
/// Example: 2142590208L ///
[global::System.Text.Json.Serialization.JsonPropertyName("completed")] - public long Completed { get; set; } + public long? Completed { get; set; } /// /// Additional properties that are not explicitly defined in the schema diff --git a/src/libs/Ollama/Generated/Ollama.Models.PushModelRequest.g.cs b/src/libs/Ollama/Generated/Ollama.Models.PushModelRequest.g.cs index a457da3..0a111d6 100644 --- a/src/libs/Ollama/Generated/Ollama.Models.PushModelRequest.g.cs +++ b/src/libs/Ollama/Generated/Ollama.Models.PushModelRequest.g.cs @@ -22,7 +22,7 @@ public sealed partial class PushModelRequest /// Default Value: false /// [global::System.Text.Json.Serialization.JsonPropertyName("insecure")] - public bool Insecure { get; set; } = false; + public bool? Insecure { get; set; } = false; /// /// Ollama username. @@ -41,7 +41,7 @@ public sealed partial class PushModelRequest /// Default Value: true /// [global::System.Text.Json.Serialization.JsonPropertyName("stream")] - public bool Stream { get; set; } = true; + public bool? Stream { get; set; } = true; /// /// Additional properties that are not explicitly defined in the schema diff --git a/src/libs/Ollama/Generated/Ollama.Models.PushModelResponse.g.cs b/src/libs/Ollama/Generated/Ollama.Models.PushModelResponse.g.cs index 9dd0560..665ea3e 100644 --- a/src/libs/Ollama/Generated/Ollama.Models.PushModelResponse.g.cs +++ b/src/libs/Ollama/Generated/Ollama.Models.PushModelResponse.g.cs @@ -29,14 +29,14 @@ public sealed partial class PushModelResponse /// Example: 2142590208L /// [global::System.Text.Json.Serialization.JsonPropertyName("total")] - public long Total { get; set; } + public long? Total { get; set; } /// /// Total bytes transferred.
/// Example: 2142590208L ///
[global::System.Text.Json.Serialization.JsonPropertyName("completed")] - public long Completed { get; set; } + public long? Completed { get; set; } /// /// Additional properties that are not explicitly defined in the schema diff --git a/src/libs/Ollama/Generated/Ollama.ModelsClient.CreateModel.g.cs b/src/libs/Ollama/Generated/Ollama.ModelsClient.CreateModel.g.cs index 1f1dba6..b51f5a1 100644 --- a/src/libs/Ollama/Generated/Ollama.ModelsClient.CreateModel.g.cs +++ b/src/libs/Ollama/Generated/Ollama.ModelsClient.CreateModel.g.cs @@ -109,7 +109,7 @@ partial void ProcessCreateModelResponse( string modelfile, string? path = default, string? quantize = default, - bool stream = true, + bool? stream = true, [global::System.Runtime.CompilerServices.EnumeratorCancellation] global::System.Threading.CancellationToken cancellationToken = default) { var request = new global::Ollama.CreateModelRequest diff --git a/src/libs/Ollama/Generated/Ollama.ModelsClient.PullModel.g.cs b/src/libs/Ollama/Generated/Ollama.ModelsClient.PullModel.g.cs index ccecc39..091de0f 100644 --- a/src/libs/Ollama/Generated/Ollama.ModelsClient.PullModel.g.cs +++ b/src/libs/Ollama/Generated/Ollama.ModelsClient.PullModel.g.cs @@ -107,10 +107,10 @@ partial void ProcessPullModelResponse( /// public async global::System.Collections.Generic.IAsyncEnumerable PullModelAsync( string model, - bool insecure = false, + bool? insecure = false, string? username = default, string? password = default, - bool stream = true, + bool? stream = true, [global::System.Runtime.CompilerServices.EnumeratorCancellation] global::System.Threading.CancellationToken cancellationToken = default) { var request = new global::Ollama.PullModelRequest diff --git a/src/libs/Ollama/Generated/Ollama.ModelsClient.PushModel.g.cs b/src/libs/Ollama/Generated/Ollama.ModelsClient.PushModel.g.cs index d8d20e6..89c416e 100644 --- a/src/libs/Ollama/Generated/Ollama.ModelsClient.PushModel.g.cs +++ b/src/libs/Ollama/Generated/Ollama.ModelsClient.PushModel.g.cs @@ -106,10 +106,10 @@ partial void ProcessPushModelResponse( /// public async global::System.Collections.Generic.IAsyncEnumerable PushModelAsync( string model, - bool insecure = false, + bool? insecure = false, string? username = default, string? password = default, - bool stream = true, + bool? stream = true, [global::System.Runtime.CompilerServices.EnumeratorCancellation] global::System.Threading.CancellationToken cancellationToken = default) { var request = new global::Ollama.PushModelRequest