/r/OpenAI

Photograph via //r/OpenAI

OpenAI is an AI research and deployment company. OpenAI's mission is to ensure that artificial general intelligence benefits all of humanity. We are an unofficial community. OpenAI makes ChatGPT, GPT-4, and DALL·E 3.

Welcome to /r/OpenAI!

OpenAI is an AI research and deployment company. OpenAI's mission is to ensure that artificial general intelligence benefits all of humanity. We are an unofficial community. OpenAI makes ChatGPT, GPT-4, and DALL·E 3.

Please view the subreddit rules before posting.


Official OpenAI Links

Sora

ChatGPT

DALL·E 3

Blog

Discord

YouTube

GitHub

Careers

Help Center

Docs


Related Subreddits

r/artificial

r/ChatGPT

r/Singularity

r/MachineLearning

r/GPTStore

r/dalle2

/r/OpenAI

1,401,413 Subscribers

1

Can i run theese twe program simultanely if process data in different Moment ?can i run both the AI voice changer and the speech-to-text program on a single 1650 GPU without any issues?

I want run an AI voice changer and a speech-to-text program on a 1650 GPU with 4GB of VRAM,

Both use only 3gb Vram and i can allocate 3GB for the voice changer and 3GB for the speech-to-text program.

The AI voice changer is configured to modify only the voice input from the microphone,

while the speech-to-text program is designed to transcribe voice from the speaker.

Since these programs do not run simultaneously, I can effectively utilize the GPU for each task individually.

can i run both the AI voice changer and the speech-to-text program on a single 1650 GPU without any issues?

0 Comments
2024/05/16
20:06 UTC

2

I ran the MMLU-Pro benchmark on GPT-4o, some notes

(Yes, the creators already did this, but I wanted to try reproducing their results.)

First of all, there's definitely still issues with the dataset.

  1. Questions mostly have 10 possible answers, but some have fewer "to remove unreasonable choices". This is supposed to be done by actually having <10 answers (I assume), but in some cases questions are padded with N/A at the end multiple times. Removing N/A from the end of the options list until only one existed resulted in removing 3314 options from the test. AFAIK N/A is never the correct answer to any question (I checked).
  2. Some questions are duplicates. I kept only the first instance encountered. This removed 86 questions from the test. I didn't check if their answer options were also duplicates.
  3. Some questions have duplicate answer options (i.e. A and C or whichever being the same answer). As this is obviously incorrect I removed these questions from the test. This removed ~400 questions. It should be pretty easy to fix these but I didn't want to spend more time.
  4. Some questions are impossible to answer as they're missing info.
    • For example, question ID 2668 is "Which of the following scans can image brain function?" but it does not give further details, and the answer options are things like "I, II and III only".
    • I only noticed one of these but I'm pretty sure there are more.
  5. Some questions have very bad formatting, like 11797:
    • \textquotedblleftNew business firms can under price older firms, because thenewer firms can buy the latest equipment and are not bur-denedwith the older, less economical equipment which olderfirms must retire before they can economically adopt thenew. This is why continued technological progress contributesto maintaining a competitive economic system." Explain the er-rors in both sentences.
    • Smarter models will probably do better on things like this, but I don't think the formatting is intentional.

For my first run, I used temperature 0.2 (they used 0.1) and this system prompt:

You are a knowledgeable expert. When given a multiple choice question, think it through then give your single final answer as The answer is ....

Here is their system prompt:

You are an knowledge expert, you are supposed to answer the multi-choice question to derive your final answer as The answer is ....

I'm not a fan of the phrasing, but more importantly in my preliminary testing I noticed that GPT-4o interpreted it to mean that it could give multiple answers. My "single final answer" was intended to prevent that. This mostly worked, but when answering questions that involved a lot of LaTeX in the CoT it would sometimes format the answer using LaTeX, making regex matching unreasonably hard. (I didn't fix those, scoring them as failures.)

For my second run, I used temperature 0.5 (which OpenAI uses in their simple-evals repo) and this system prompt:

Answer the following multiple choice question. The last line of your response should be of the following format: 'Answer: ($LETTER)' (without quotes) where LETTER is a single one of the options. Think step by step before answering.

This is based on this prompt from OpenAI's evals. OpenAI's says to pick one of "ABCD" but I was too lazy to generate that dynamically so I changed it a bit. It seems to work well. I used the regex Answer\s*:\s*\(\$?([A-Ja-j])\) to capture the answer.

If the model's output didn't seem to give an answer at all I manually checked it. This was:

  • Mostly GPT-4o saying that none of the answers are correct
  • Sometimes the model entering an infinite loop until it ran out of output space
  • A few times the model giving a numeric answer, often one that had no correspondence to possible answers. Scored as a failure.
  • A few times very slightly bad formatting that my regex didn't match. If the answer it gave was correct I manually fixed it, on the rationale that a slightly better regex could have caught it, otherwise I left it alone (since it wouldn't affect the score either way).

Here are the results I got, plus the original results from TIGER Lab @ UWaterloo.

ModelsPromptingOverallBiologyBusinessChemistryComputerScienceEconomicsEngineeringHealthHistoryLawMathPhilosophyPhysicsPsychologyOther
GPT-4o (TIGER Lab)CoT0.71490.85040.78520.74280.75120.7840.54010.71520.67180.53930.7620.69470.73550.77260.7091
GPT-4o (my first run)CoT0.71000.84410.78020.75020.74040.77630.57400.70610.67610.57530.79010.67600.73490.76110.7117
GPT-4o (my second run)CoT0.71130.84410.77250.76280.75060.76620.55840.69640.66410.55970.77720.64740.73780.74450.7041

So I'd say the results themselves are basically "yep, that sure is how it scores". I don't know if my overall scores are actually slightly worse or if it's the effect of removing almost 500 questions from a 12.3k question dataset.

Total cost: maybe $160 based on token counts, messed up cost tracking at first.

The second run had approx 2.17M input tokens and 5.04M output tokens. Naively scaling this to other models (without actually using their tokenizers, and assuming they'll give identical output lengths) gives these approximate prices:

  • Claude 3 Opus: $410
  • GPT-4-0314: $367
  • GPT-4-32k-0314: $608

I don't intend to run those.

I'll post the inputs/outputs of both runs if anyone knows a good place to upload password-protected zips (to avoid accidental scraping, since everything posted to reddit is trained on).

0 Comments
2024/05/16
20:01 UTC

0

ChatGPT limiting the number of messages sent per hour

1 Comment
2024/05/16
19:49 UTC

0

OpenAI aesthetics are better…

0 Comments
2024/05/16
19:45 UTC

1

Which language model has the highest output token limit?

It appears that most GPT models, including:

  • GPT-3.5-turbo
  • GPT-4-turbo
  • the latest GPT-4o
  • and also Claude 3 (Haiku/Sonnet/Opus)

have a limitation of 4096 output tokens.

Are there any popular GPT models that allow for generating more output tokens?

1 Comment
2024/05/16
19:40 UTC

1

Anyone else having issues with the API's text to speech?

It was working for me this morning, but not it just spins and goes back to text mode. I miss my passenger princess :(

0 Comments
2024/05/16
19:06 UTC

1

"Using plugins" with the API

Hi,

I need help "using plugins" like Wolfram Alpha for ChatGPT in the OpenAI API.

What I understand so far:

  1. Create an assistant on the OpenAI platform.
  2. Define a JSON function.
  3. Call the function as part of the API call.

I am stuck on point #2. How can I find out what the function looks like for the Wolfram plugin?

Essentially, all I want is to have chatgpt + wolfram plugin, except available through an API endpoint and as JSON.

Any help would be greatly appreciated!

3 Comments
2024/05/16
19:01 UTC

2

AtheistGPT

0 Comments
2024/05/16
18:49 UTC

1

ChatGPT frequently gets things wrong when asked about film and TV

This tech is seemingly very interesting. However, the more I've tried using it, it gets things wrong. I've tried using it to track down specific episodes of old shows I can't quite recall or which actors what worked together previously.

And, 9/10, it gives me wrong information. Maybe TV and film trivia is too niche but thinking of AI and just a fun thing to ask it, sad that it's wrong.

An example is I came across two actors on Murder she wrote that worked together later on a soap. I hadn't known they worked together previously so I asked Chat GPT. It tired to tell me about them working on the Soap but it gave the wrong Soap and made up character names.

Does this mean the data just doesn't have this information?

3 Comments
2024/05/16
17:26 UTC

2

2 problems in Android app (beta) right now

Hi,

The 1st one is: ChatGPT 4o can't read txt or PDF files anymore. 2nd one is: the memory can't be updated and isn't available in new chats. (That's what the memory is intended for, or am I wrong?)

Yesterday 4o put Info into the memory. Checking it with the memory manager displays that it's all there. This morning I got an App Update and I tried to let 4o add something new to the memory. It said "memory updated" but the new infos are not there. Letting 4o check about this, it confirmed that the update didn't show up.

Does anybody else having this issues?

0 Comments
2024/05/16
17:20 UTC

156

Thank God. Maybe the posts will stop now?

56 Comments
2024/05/16
17:10 UTC

1

Is 4o only available as API?

Super confused, because i didn't see it available in ChatGPT but i see people talking about it on a Mac app...

5 Comments
2024/05/16
17:07 UTC

3

Account sign in issue

My account is saml with 365, but we recently changed domains old domain still exists, but is secondary to the new primary and saml login no longer works.

I can't raise a ticket with OpenAI as I can't get into the account to do so. I am still logged in via my mobile phone but not many options in the mobile phone app.

Anyone got any ideas how to get in contact with OpenAI to resolve. I've changed my primary address back to the old domain so I can authenticate into 365 with that account now, but still no luck being able to log into my OpenAI account.

Error message I am getting when just trying to sign in with my email address is:

You tried signing in as "email@address.here" using a password, which is not the authentication method you used during sign up. Try again using the authentication method you used during sign up. (error=identity_provider_mismatch)

So I defiantly need to sign in with the saml by the looks of it.

0 Comments
2024/05/16
17:05 UTC

0

AI and Elections 2024 - Thoughts?

What do you think the role of AI will place in elections, specifically on voter canvassing and showcasing how data can be used for campaigns? 


https://app.askhumans.io/study/817 

Do you foresee any negatives of AI's impact in elections, if so what?

https://app.askhumans.io/study/817

Click here to give verbal feedback

3 Comments
2024/05/16
17:02 UTC

1

Gpts

Does the GPT that i created using the latest 4o or it is still using an older version?

3 Comments
2024/05/16
16:58 UTC

0

Can your ChatGPT 4o access the Video camera?

Does anyone have a version of ChatGPT 4o that can access Video, or any of the new features just announced? I have the paid version on Android and iPad but so far neither does anything new. I've been asking a couple of times a day, "Can you access the Video Camera?" It keeps saying no, and its replies seem a little more grouchy each time. Maybe that's one of the new features?

4 Comments
2024/05/16
16:13 UTC

4

What Challenges Do AI Developers Face With Big Companies Releasing New AI Apps Almost Daily?

I've been keeping up with AI and LLM trends for some time, and I must admit it can be quite daunting for the average developer trying to create an AI app. Big companies like Google, OpenAI, and Microsoft are developing AI platforms for nearly every use case. How can we compete with these giants when they make parts of their products freely accessible to consumers?

4 Comments
2024/05/16
16:05 UTC

3

Thoughts on GPT-4o?

1 Comment
2024/05/16
16:03 UTC

77

The Mac App is awesome

I just got done drafting an email with the Chat Mac app and yeah, I never wanna do this any other way again. This was not a dictation; this was me fumbling and stuttering through a hazy recollection of a meeting I'd had as the model listened ("no worries," it said at one point, "take your time"); answering a few follow up questions as it sought clarity and further details; and then listening to it read back a drafted email based on my recollection of events. I know the previous voice model could do some of this already but it didn't feel this organic; more like a juiced up Siri with a slightly more natural voic. The fact that this isn't even the dynamic voice model we saw in the showcase is wild to me.

I've also been running generated text from 4o through AI detectors and its output performs way better than model 4 did.

It all leads me to believe that they prioritized mimicking humanness with this 4o model than other kinds of performance. So while I might revert to 4 for more complex tasks (I don't think 4o peforms quite as well), 4o really does feel great when interaction is involved.

Any one else had a different experience?

52 Comments
2024/05/16
15:41 UTC

11

A man and a goat

9 Comments
2024/05/16
15:32 UTC

1

Does Free Tier 4o support code interpreter and document reader?

I tried uploading some documents on free tier 4o to visualise data, and it didn't even read the document.

3 Comments
2024/05/16
14:45 UTC

0

the new desktop app requires m2 chip to install is total bs

2 Comments
2024/05/16
14:34 UTC

4

Vibe: Free Offline Transcription with Whisper AI

Hey everyone, just wanted to let you know about Vibe!

It's a new transcription app I created that's open source and works seamlessly on macOS, Windows, and Linux. The best part? It runs on your device using the Whisper AI model, so you don't even need the internet for top-notch transcriptions! Plus, it's designed to be super user-friendly. Check it out on the Vibe website and see for yourself!

And for those interested in diving into the code or contributing, you can find the project on GitHub at github.com/thewh1teagle/vibe. Happy transcribing!

https://preview.redd.it/xn81bqqkqs0d1.png?width=1200&format=png&auto=webp&s=03389d75d2852b016225c68cec1ffd5115455de1

0 Comments
2024/05/16
14:10 UTC

6

Is anyone working on an open source windows version of the ChatGPT 4o app?

Like many, I was floored by the recent ChatGPT 4O desktop app demo. Specifically how you can share your screen with it and have it help you in real time.

I could really use this on a project that I'm working on but they've decided not to release a Windows app until later.

I thought I saw someone mention somewhere about an open source version of it being worked on?

Anyone know anything about that?

3 Comments
2024/05/16
14:03 UTC

28

Farfalle: Open-source Perplexity

5 Comments
2024/05/16
13:15 UTC

4

Yeah, Dalle is pretty cool.

0 Comments
2024/05/16
13:09 UTC

3

Is the Mac App really slow?

Hey, I just downloaded the Mac app and it seems the answers are being printed REALLY slow in comparison to the website. I am curious if anyone else is experiencing this. It is taking over a minute to get me an answer sometimes, and it's mostly due to the printing on the screen.

Thoughts?

4 Comments
2024/05/16
13:06 UTC

75

What does this mean? — "Frontier modelS coming soon" — Why is the word "models" pluralized. Are there really going to be 2+ frontier models this year?

44 Comments
2024/05/16
12:31 UTC

9

My GPT-4 has suddenly disappeared; I only find GPT-4o now.

3 Comments
2024/05/16
12:16 UTC

Back To Top