From 30aaaee00b73b7af5999fdcab8eb82ef09f9af57 Mon Sep 17 00:00:00 2001 From: Caio Saldanha Date: Fri, 27 Sep 2024 15:14:37 -0700 Subject: [PATCH] [Examples] Updating orange picture links --- README.md | 4 ++-- examples/Assistants/Example05_AssistantsWithVision.cs | 2 +- examples/Assistants/Example05_AssistantsWithVisionAsync.cs | 2 +- 3 files changed, 4 insertions(+), 4 deletions(-) diff --git a/README.md b/README.md index c61697db..688f72c8 100644 --- a/README.md +++ b/README.md @@ -670,7 +670,7 @@ For this example, we will use both image data from a local file as well as an im OpenAIFile pictureOfAppleFile = fileClient.UploadFile( Path.Combine("Assets", "picture-of-apple.png"), FileUploadPurpose.Vision); -Uri linkToPictureOfOrange = new("https://platform.openai.com/fictitious-files/picture-of-orange.png"); +Uri linkToPictureOfOrange = new("https://raw.githubusercontent.com/openai/openai-dotnet/refs/heads/main/examples/Assets/picture-of-orange.png"); ``` Next, create a new assistant with a vision-capable model like `gpt-4o` and a thread with the image information referenced: @@ -731,7 +731,7 @@ This will yield streamed output from the run like the following: ```text --- Run started! --- -The first image shows a red apple with a smooth skin and a single leaf, while the second image depicts an orange with a rough, textured skin and a leaf with droplets of water. Comparing them might seem impossible - it's like apples and oranges! +The first image depicts a multicolored apple with a blend of red and green hues, while the second image shows an orange with a bright, textured orange peel; one might say it’s comparing apples to oranges! ``` ## How to work with Azure OpenAI diff --git a/examples/Assistants/Example05_AssistantsWithVision.cs b/examples/Assistants/Example05_AssistantsWithVision.cs index 9114405c..b7822319 100644 --- a/examples/Assistants/Example05_AssistantsWithVision.cs +++ b/examples/Assistants/Example05_AssistantsWithVision.cs @@ -21,7 +21,7 @@ public void Example05_AssistantsWithVision() OpenAIFile pictureOfAppleFile = fileClient.UploadFile( Path.Combine("Assets", "picture-of-apple.png"), FileUploadPurpose.Vision); - Uri linkToPictureOfOrange = new("https://platform.openai.com/fictitious-files/picture-of-orange.png"); + Uri linkToPictureOfOrange = new("https://raw.githubusercontent.com/openai/openai-dotnet/refs/heads/main/examples/Assets/picture-of-orange.png"); Assistant assistant = assistantClient.CreateAssistant( "gpt-4o", diff --git a/examples/Assistants/Example05_AssistantsWithVisionAsync.cs b/examples/Assistants/Example05_AssistantsWithVisionAsync.cs index 88ed4237..7aeead4e 100644 --- a/examples/Assistants/Example05_AssistantsWithVisionAsync.cs +++ b/examples/Assistants/Example05_AssistantsWithVisionAsync.cs @@ -22,7 +22,7 @@ public async Task Example05_AssistantsWithVisionAsync() OpenAIFile pictureOfAppleFile = await fileClient.UploadFileAsync( Path.Combine("Assets", "picture-of-apple.png"), FileUploadPurpose.Vision); - Uri linkToPictureOfOrange = new("https://platform.openai.com/fictitious-files/picture-of-orange.png"); + Uri linkToPictureOfOrange = new("https://raw.githubusercontent.com/openai/openai-dotnet/refs/heads/main/examples/Assets/picture-of-orange.png"); Assistant assistant = await assistantClient.CreateAssistantAsync( "gpt-4o",