Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -5,14 +5,14 @@ namespace tryAGI.OpenAI
{
public partial class AudioClient
{
partial void PrepareCreateSpeechAsStreamArguments(
partial void PrepareCreateSpeechArguments(
global::System.Net.Http.HttpClient httpClient,
global::tryAGI.OpenAI.CreateSpeechRequest request);
partial void PrepareCreateSpeechAsStreamRequest(
partial void PrepareCreateSpeechRequest(
global::System.Net.Http.HttpClient httpClient,
global::System.Net.Http.HttpRequestMessage httpRequestMessage,
global::tryAGI.OpenAI.CreateSpeechRequest request);
partial void ProcessCreateSpeechAsStreamResponse(
partial void ProcessCreateSpeechResponse(
global::System.Net.Http.HttpClient httpClient,
global::System.Net.Http.HttpResponseMessage httpResponseMessage);

Expand All @@ -23,7 +23,7 @@ partial void ProcessCreateSpeechAsStreamResponse(
/// <param name="request"></param>
/// <param name="cancellationToken">The token to cancel the operation with</param>
/// <exception cref="global::tryAGI.OpenAI.ApiException"></exception>
public async global::System.Collections.Generic.IAsyncEnumerable<global::tryAGI.OpenAI.CreateSpeechResponseStreamEvent> CreateSpeechAsStreamAsync(
public async global::System.Collections.Generic.IAsyncEnumerable<global::tryAGI.OpenAI.CreateSpeechResponseStreamEvent> CreateSpeechAsync(

global::tryAGI.OpenAI.CreateSpeechRequest request,
[global::System.Runtime.CompilerServices.EnumeratorCancellation] global::System.Threading.CancellationToken cancellationToken = default)
Expand All @@ -32,7 +32,7 @@ partial void ProcessCreateSpeechAsStreamResponse(

PrepareArguments(
client: HttpClient);
PrepareCreateSpeechAsStreamArguments(
PrepareCreateSpeechArguments(
httpClient: HttpClient,
request: request);

Expand Down Expand Up @@ -73,7 +73,7 @@ partial void ProcessCreateSpeechAsStreamResponse(
PrepareRequest(
client: HttpClient,
request: __httpRequest);
PrepareCreateSpeechAsStreamRequest(
PrepareCreateSpeechRequest(
httpClient: HttpClient,
httpRequestMessage: __httpRequest,
request: request);
Expand All @@ -86,7 +86,7 @@ partial void ProcessCreateSpeechAsStreamResponse(
ProcessResponse(
client: HttpClient,
response: __response);
ProcessCreateSpeechAsStreamResponse(
ProcessCreateSpeechResponse(
httpClient: HttpClient,
httpResponseMessage: __response);

Expand Down Expand Up @@ -160,7 +160,7 @@ partial void ProcessCreateSpeechAsStreamResponse(
/// </param>
/// <param name="cancellationToken">The token to cancel the operation with</param>
/// <exception cref="global::System.InvalidOperationException"></exception>
public async global::System.Collections.Generic.IAsyncEnumerable<global::tryAGI.OpenAI.CreateSpeechResponseStreamEvent> CreateSpeechAsStreamAsync(
public async global::System.Collections.Generic.IAsyncEnumerable<global::tryAGI.OpenAI.CreateSpeechResponseStreamEvent> CreateSpeechAsync(
global::tryAGI.OpenAI.AnyOf<string, global::tryAGI.OpenAI.CreateSpeechRequestModel?> model,
string input,
global::tryAGI.OpenAI.VoiceIdsOrCustomVoice voice,
Expand All @@ -181,7 +181,7 @@ partial void ProcessCreateSpeechAsStreamResponse(
StreamFormat = streamFormat,
};

var __enumerable = CreateSpeechAsStreamAsync(
var __enumerable = CreateSpeechAsync(
request: __request,
cancellationToken: cancellationToken);

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -36,6 +36,23 @@ partial void ProcessCreateTranscriptionResponseContent(
{
request = request ?? throw new global::System.ArgumentNullException(nameof(request));


request = new global::tryAGI.OpenAI.CreateTranscriptionRequest
{
File = request.File,
Filename = request.Filename,
Model = request.Model,
Language = request.Language,
Prompt = request.Prompt,
ResponseFormat = request.ResponseFormat,
Temperature = request.Temperature,
Include = request.Include,
TimestampGranularities = request.TimestampGranularities,
Stream = false,
ChunkingStrategy = request.ChunkingStrategy,
KnownSpeakerNames = request.KnownSpeakerNames,
KnownSpeakerReferences = request.KnownSpeakerReferences,
};
PrepareArguments(
client: HttpClient);
PrepareCreateTranscriptionArguments(
Expand Down Expand Up @@ -287,7 +304,6 @@ partial void ProcessCreateTranscriptionResponseContent(
/// This option is not available for `gpt-4o-transcribe-diarize`.<br/>
/// Default Value: [segment]
/// </param>
/// <param name="stream"></param>
/// <param name="chunkingStrategy"></param>
/// <param name="knownSpeakerNames">
/// Optional list of speaker names that correspond to the audio samples provided in `known_speaker_references[]`. Each entry should be a short identifier (for example `customer` or `agent`). Up to 4 speakers are supported.
Expand All @@ -307,7 +323,6 @@ partial void ProcessCreateTranscriptionResponseContent(
double? temperature = default,
global::System.Collections.Generic.IList<global::tryAGI.OpenAI.TranscriptionInclude>? include = default,
global::System.Collections.Generic.IList<global::tryAGI.OpenAI.CreateTranscriptionRequestTimestampGranularitie>? timestampGranularities = default,
bool? stream = default,
global::tryAGI.OpenAI.AnyOf<global::tryAGI.OpenAI.CreateTranscriptionRequestChunkingStrategyVariant1?, global::tryAGI.OpenAI.VadConfig>? chunkingStrategy = default,
global::System.Collections.Generic.IList<string>? knownSpeakerNames = default,
global::System.Collections.Generic.IList<string>? knownSpeakerReferences = default,
Expand All @@ -324,7 +339,7 @@ partial void ProcessCreateTranscriptionResponseContent(
Temperature = temperature,
Include = include,
TimestampGranularities = timestampGranularities,
Stream = stream,
Stream = false,
ChunkingStrategy = chunkingStrategy,
KnownSpeakerNames = knownSpeakerNames,
KnownSpeakerReferences = knownSpeakerReferences,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -31,6 +31,23 @@ partial void ProcessCreateTranscriptionAsStreamResponse(
{
request = request ?? throw new global::System.ArgumentNullException(nameof(request));


request = new global::tryAGI.OpenAI.CreateTranscriptionRequest
{
File = request.File,
Filename = request.Filename,
Model = request.Model,
Language = request.Language,
Prompt = request.Prompt,
ResponseFormat = request.ResponseFormat,
Temperature = request.Temperature,
Include = request.Include,
TimestampGranularities = request.TimestampGranularities,
Stream = true,
ChunkingStrategy = request.ChunkingStrategy,
KnownSpeakerNames = request.KnownSpeakerNames,
KnownSpeakerReferences = request.KnownSpeakerReferences,
};
PrepareArguments(
client: HttpClient);
PrepareCreateTranscriptionAsStreamArguments(
Expand Down Expand Up @@ -250,7 +267,6 @@ partial void ProcessCreateTranscriptionAsStreamResponse(
/// This option is not available for `gpt-4o-transcribe-diarize`.<br/>
/// Default Value: [segment]
/// </param>
/// <param name="stream"></param>
/// <param name="chunkingStrategy"></param>
/// <param name="knownSpeakerNames">
/// Optional list of speaker names that correspond to the audio samples provided in `known_speaker_references[]`. Each entry should be a short identifier (for example `customer` or `agent`). Up to 4 speakers are supported.
Expand All @@ -270,7 +286,6 @@ partial void ProcessCreateTranscriptionAsStreamResponse(
double? temperature = default,
global::System.Collections.Generic.IList<global::tryAGI.OpenAI.TranscriptionInclude>? include = default,
global::System.Collections.Generic.IList<global::tryAGI.OpenAI.CreateTranscriptionRequestTimestampGranularitie>? timestampGranularities = default,
bool? stream = default,
global::tryAGI.OpenAI.AnyOf<global::tryAGI.OpenAI.CreateTranscriptionRequestChunkingStrategyVariant1?, global::tryAGI.OpenAI.VadConfig>? chunkingStrategy = default,
global::System.Collections.Generic.IList<string>? knownSpeakerNames = default,
global::System.Collections.Generic.IList<string>? knownSpeakerReferences = default,
Expand All @@ -287,7 +302,7 @@ partial void ProcessCreateTranscriptionAsStreamResponse(
Temperature = temperature,
Include = include,
TimestampGranularities = timestampGranularities,
Stream = stream,
Stream = true,
ChunkingStrategy = chunkingStrategy,
KnownSpeakerNames = knownSpeakerNames,
KnownSpeakerReferences = knownSpeakerReferences,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ public partial interface IAudioClient
/// <param name="request"></param>
/// <param name="cancellationToken">The token to cancel the operation with</param>
/// <exception cref="global::tryAGI.OpenAI.ApiException"></exception>
global::System.Collections.Generic.IAsyncEnumerable<global::tryAGI.OpenAI.CreateSpeechResponseStreamEvent> CreateSpeechAsStreamAsync(
global::System.Collections.Generic.IAsyncEnumerable<global::tryAGI.OpenAI.CreateSpeechResponseStreamEvent> CreateSpeechAsync(

global::tryAGI.OpenAI.CreateSpeechRequest request,
global::System.Threading.CancellationToken cancellationToken = default);
Expand Down Expand Up @@ -46,7 +46,7 @@ public partial interface IAudioClient
/// </param>
/// <param name="cancellationToken">The token to cancel the operation with</param>
/// <exception cref="global::System.InvalidOperationException"></exception>
global::System.Collections.Generic.IAsyncEnumerable<global::tryAGI.OpenAI.CreateSpeechResponseStreamEvent> CreateSpeechAsStreamAsync(
global::System.Collections.Generic.IAsyncEnumerable<global::tryAGI.OpenAI.CreateSpeechResponseStreamEvent> CreateSpeechAsync(
global::tryAGI.OpenAI.AnyOf<string, global::tryAGI.OpenAI.CreateSpeechRequestModel?> model,
string input,
global::tryAGI.OpenAI.VoiceIdsOrCustomVoice voice,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,6 @@ public partial interface IAudioClient
/// This option is not available for `gpt-4o-transcribe-diarize`.<br/>
/// Default Value: [segment]
/// </param>
/// <param name="stream"></param>
/// <param name="chunkingStrategy"></param>
/// <param name="knownSpeakerNames">
/// Optional list of speaker names that correspond to the audio samples provided in `known_speaker_references[]`. Each entry should be a short identifier (for example `customer` or `agent`). Up to 4 speakers are supported.
Expand All @@ -78,7 +77,6 @@ public partial interface IAudioClient
double? temperature = default,
global::System.Collections.Generic.IList<global::tryAGI.OpenAI.TranscriptionInclude>? include = default,
global::System.Collections.Generic.IList<global::tryAGI.OpenAI.CreateTranscriptionRequestTimestampGranularitie>? timestampGranularities = default,
bool? stream = default,
global::tryAGI.OpenAI.AnyOf<global::tryAGI.OpenAI.CreateTranscriptionRequestChunkingStrategyVariant1?, global::tryAGI.OpenAI.VadConfig>? chunkingStrategy = default,
global::System.Collections.Generic.IList<string>? knownSpeakerNames = default,
global::System.Collections.Generic.IList<string>? knownSpeakerReferences = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,6 @@ public partial interface IAudioClient
/// This option is not available for `gpt-4o-transcribe-diarize`.<br/>
/// Default Value: [segment]
/// </param>
/// <param name="stream"></param>
/// <param name="chunkingStrategy"></param>
/// <param name="knownSpeakerNames">
/// Optional list of speaker names that correspond to the audio samples provided in `known_speaker_references[]`. Each entry should be a short identifier (for example `customer` or `agent`). Up to 4 speakers are supported.
Expand All @@ -78,7 +77,6 @@ public partial interface IAudioClient
double? temperature = default,
global::System.Collections.Generic.IList<global::tryAGI.OpenAI.TranscriptionInclude>? include = default,
global::System.Collections.Generic.IList<global::tryAGI.OpenAI.CreateTranscriptionRequestTimestampGranularitie>? timestampGranularities = default,
bool? stream = default,
global::tryAGI.OpenAI.AnyOf<global::tryAGI.OpenAI.CreateTranscriptionRequestChunkingStrategyVariant1?, global::tryAGI.OpenAI.VadConfig>? chunkingStrategy = default,
global::System.Collections.Generic.IList<string>? knownSpeakerNames = default,
global::System.Collections.Generic.IList<string>? knownSpeakerReferences = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -56,13 +56,6 @@ public partial interface IImagesClient
/// Default Value: 100<br/>
/// Example: 100
/// </param>
/// <param name="stream">
/// Generate the image in streaming mode. Defaults to `false`. See the<br/>
/// [Image generation guide](/docs/guides/image-generation) for more information.<br/>
/// This parameter is only supported for the GPT image models.<br/>
/// Default Value: false<br/>
/// Example: false
/// </param>
/// <param name="partialImages"></param>
/// <param name="size">
/// The size of the generated images. Must be one of `1024x1024`, `1536x1024` (landscape), `1024x1536` (portrait), or `auto` (default value) for the GPT image models, one of `256x256`, `512x512`, or `1024x1024` for `dall-e-2`, and one of `1024x1024`, `1792x1024`, or `1024x1792` for `dall-e-3`.<br/>
Expand Down Expand Up @@ -103,7 +96,6 @@ public partial interface IImagesClient
global::tryAGI.OpenAI.CreateImageRequestResponseFormat? responseFormat = default,
global::tryAGI.OpenAI.CreateImageRequestOutputFormat? outputFormat = default,
int? outputCompression = default,
bool? stream = default,
int? partialImages = default,
global::tryAGI.OpenAI.CreateImageRequestSize? size = default,
global::tryAGI.OpenAI.CreateImageRequestModeration? moderation = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -56,13 +56,6 @@ public partial interface IImagesClient
/// Default Value: 100<br/>
/// Example: 100
/// </param>
/// <param name="stream">
/// Generate the image in streaming mode. Defaults to `false`. See the<br/>
/// [Image generation guide](/docs/guides/image-generation) for more information.<br/>
/// This parameter is only supported for the GPT image models.<br/>
/// Default Value: false<br/>
/// Example: false
/// </param>
/// <param name="partialImages"></param>
/// <param name="size">
/// The size of the generated images. Must be one of `1024x1024`, `1536x1024` (landscape), `1024x1536` (portrait), or `auto` (default value) for the GPT image models, one of `256x256`, `512x512`, or `1024x1024` for `dall-e-2`, and one of `1024x1024`, `1792x1024`, or `1024x1792` for `dall-e-3`.<br/>
Expand Down Expand Up @@ -103,7 +96,6 @@ public partial interface IImagesClient
global::tryAGI.OpenAI.CreateImageRequestResponseFormat? responseFormat = default,
global::tryAGI.OpenAI.CreateImageRequestOutputFormat? outputFormat = default,
int? outputCompression = default,
bool? stream = default,
int? partialImages = default,
global::tryAGI.OpenAI.CreateImageRequestSize? size = default,
global::tryAGI.OpenAI.CreateImageRequestModeration? moderation = default,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -92,12 +92,6 @@ public partial interface IImagesClient
/// Example: user-1234
/// </param>
/// <param name="inputFidelity"></param>
/// <param name="stream">
/// Edit the image in streaming mode. Defaults to `false`. See the<br/>
/// [Image generation guide](/docs/guides/image-generation) for more information.<br/>
/// Default Value: false<br/>
/// Example: false
/// </param>
/// <param name="partialImages"></param>
/// <param name="quality">
/// The quality of the image that will be generated for GPT image models. Defaults to `auto`.<br/>
Expand All @@ -120,7 +114,6 @@ public partial interface IImagesClient
int? outputCompression = default,
string? user = default,
global::tryAGI.OpenAI.InputFidelity? inputFidelity = default,
bool? stream = default,
int? partialImages = default,
global::tryAGI.OpenAI.CreateImageEditRequestQuality? quality = default,
global::System.Threading.CancellationToken cancellationToken = default);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -92,12 +92,6 @@ public partial interface IImagesClient
/// Example: user-1234
/// </param>
/// <param name="inputFidelity"></param>
/// <param name="stream">
/// Edit the image in streaming mode. Defaults to `false`. See the<br/>
/// [Image generation guide](/docs/guides/image-generation) for more information.<br/>
/// Default Value: false<br/>
/// Example: false
/// </param>
/// <param name="partialImages"></param>
/// <param name="quality">
/// The quality of the image that will be generated for GPT image models. Defaults to `auto`.<br/>
Expand All @@ -120,7 +114,6 @@ public partial interface IImagesClient
int? outputCompression = default,
string? user = default,
global::tryAGI.OpenAI.InputFidelity? inputFidelity = default,
bool? stream = default,
int? partialImages = default,
global::tryAGI.OpenAI.CreateImageEditRequestQuality? quality = default,
global::System.Threading.CancellationToken cancellationToken = default);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ public partial interface ISkillsClient
/// </param>
/// <param name="cancellationToken">The token to cancel the operation with</param>
/// <exception cref="global::tryAGI.OpenAI.ApiException"></exception>
global::System.Threading.Tasks.Task<byte[]> GetSkillContentAsync(
global::System.Threading.Tasks.Task<string> GetSkillContentAsync(
string skillId,
global::System.Threading.CancellationToken cancellationToken = default);
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ public partial interface ISkillsClient
/// </param>
/// <param name="cancellationToken">The token to cancel the operation with</param>
/// <exception cref="global::tryAGI.OpenAI.ApiException"></exception>
global::System.Threading.Tasks.Task<byte[]> GetSkillVersionContentAsync(
global::System.Threading.Tasks.Task<string> GetSkillVersionContentAsync(
string skillId,
string version,
global::System.Threading.CancellationToken cancellationToken = default);
Expand Down
Loading
Loading