/r/OpenAI
OpenAI is an AI research and deployment company. OpenAI's mission is to ensure that artificial general intelligence benefits all of humanity. We are an unofficial community. OpenAI makes ChatGPT, GPT-4, and DALL·E 3.
OpenAI is an AI research and deployment company. OpenAI's mission is to ensure that artificial general intelligence benefits all of humanity. We are an unofficial community. OpenAI makes ChatGPT, GPT-4, and DALL·E 3.
Official OpenAI Links
Related Subreddits
/r/OpenAI
I want run an AI voice changer and a speech-to-text program on a 1650 GPU with 4GB of VRAM,
Both use only 3gb Vram and i can allocate 3GB for the voice changer and 3GB for the speech-to-text program.
The AI voice changer is configured to modify only the voice input from the microphone,
while the speech-to-text program is designed to transcribe voice from the speaker.
Since these programs do not run simultaneously, I can effectively utilize the GPU for each task individually.
can i run both the AI voice changer and the speech-to-text program on a single 1650 GPU without any issues?
(Yes, the creators already did this, but I wanted to try reproducing their results.)
First of all, there's definitely still issues with the dataset.
N/A
from the end of the options list until only one existed resulted in removing 3314
options from the test. AFAIK N/A
is never the correct answer to any question (I checked).For my first run, I used temperature 0.2 (they used 0.1) and this system prompt:
You are a knowledgeable expert. When given a multiple choice question, think it through then give your single final answer as
The answer is ...
.
Here is their system prompt:
You are an knowledge expert, you are supposed to answer the multi-choice question to derive your final answer as
The answer is ...
.
I'm not a fan of the phrasing, but more importantly in my preliminary testing I noticed that GPT-4o interpreted it to mean that it could give multiple answers. My "single final answer" was intended to prevent that. This mostly worked, but when answering questions that involved a lot of LaTeX in the CoT it would sometimes format the answer using LaTeX, making regex matching unreasonably hard. (I didn't fix those, scoring them as failures.)
For my second run, I used temperature 0.5 (which OpenAI uses in their simple-evals
repo) and this system prompt:
Answer the following multiple choice question. The last line of your response should be of the following format: 'Answer: ($LETTER)' (without quotes) where LETTER is a single one of the options. Think step by step before answering.
This is based on this prompt from OpenAI's evals. OpenAI's says to pick one of "ABCD" but I was too lazy to generate that dynamically so I changed it a bit. It seems to work well. I used the regex Answer\s*:\s*\(\$?([A-Ja-j])\)
to capture the answer.
If the model's output didn't seem to give an answer at all I manually checked it. This was:
Here are the results I got, plus the original results from TIGER Lab @ UWaterloo.
Models | Prompting | Overall | Biology | Business | Chemistry | ComputerScience | Economics | Engineering | Health | History | Law | Math | Philosophy | Physics | Psychology | Other |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
GPT-4o (TIGER Lab) | CoT | 0.7149 | 0.8504 | 0.7852 | 0.7428 | 0.7512 | 0.784 | 0.5401 | 0.7152 | 0.6718 | 0.5393 | 0.762 | 0.6947 | 0.7355 | 0.7726 | 0.7091 |
GPT-4o (my first run) | CoT | 0.7100 | 0.8441 | 0.7802 | 0.7502 | 0.7404 | 0.7763 | 0.5740 | 0.7061 | 0.6761 | 0.5753 | 0.7901 | 0.6760 | 0.7349 | 0.7611 | 0.7117 |
GPT-4o (my second run) | CoT | 0.7113 | 0.8441 | 0.7725 | 0.7628 | 0.7506 | 0.7662 | 0.5584 | 0.6964 | 0.6641 | 0.5597 | 0.7772 | 0.6474 | 0.7378 | 0.7445 | 0.7041 |
So I'd say the results themselves are basically "yep, that sure is how it scores". I don't know if my overall scores are actually slightly worse or if it's the effect of removing almost 500 questions from a 12.3k question dataset.
Total cost: maybe $160 based on token counts, messed up cost tracking at first.
The second run had approx 2.17M input tokens and 5.04M output tokens. Naively scaling this to other models (without actually using their tokenizers, and assuming they'll give identical output lengths) gives these approximate prices:
I don't intend to run those.
I'll post the inputs/outputs of both runs if anyone knows a good place to upload password-protected zips (to avoid accidental scraping, since everything posted to reddit is trained on).
It appears that most GPT models, including:
have a limitation of 4096 output tokens.
Are there any popular GPT models that allow for generating more output tokens?
It was working for me this morning, but not it just spins and goes back to text mode. I miss my passenger princess :(
Hi,
I need help "using plugins" like Wolfram Alpha for ChatGPT in the OpenAI API.
What I understand so far:
I am stuck on point #2. How can I find out what the function looks like for the Wolfram plugin?
Essentially, all I want is to have chatgpt + wolfram plugin, except available through an API endpoint and as JSON.
Any help would be greatly appreciated!
This tech is seemingly very interesting. However, the more I've tried using it, it gets things wrong. I've tried using it to track down specific episodes of old shows I can't quite recall or which actors what worked together previously.
And, 9/10, it gives me wrong information. Maybe TV and film trivia is too niche but thinking of AI and just a fun thing to ask it, sad that it's wrong.
An example is I came across two actors on Murder she wrote that worked together later on a soap. I hadn't known they worked together previously so I asked Chat GPT. It tired to tell me about them working on the Soap but it gave the wrong Soap and made up character names.
Does this mean the data just doesn't have this information?
Hi,
The 1st one is: ChatGPT 4o can't read txt or PDF files anymore. 2nd one is: the memory can't be updated and isn't available in new chats. (That's what the memory is intended for, or am I wrong?)
Yesterday 4o put Info into the memory. Checking it with the memory manager displays that it's all there. This morning I got an App Update and I tried to let 4o add something new to the memory. It said "memory updated" but the new infos are not there. Letting 4o check about this, it confirmed that the update didn't show up.
Does anybody else having this issues?
Super confused, because i didn't see it available in ChatGPT but i see people talking about it on a Mac app...
My account is saml with 365, but we recently changed domains old domain still exists, but is secondary to the new primary and saml login no longer works.
I can't raise a ticket with OpenAI as I can't get into the account to do so. I am still logged in via my mobile phone but not many options in the mobile phone app.
Anyone got any ideas how to get in contact with OpenAI to resolve. I've changed my primary address back to the old domain so I can authenticate into 365 with that account now, but still no luck being able to log into my OpenAI account.
Error message I am getting when just trying to sign in with my email address is:
You tried signing in as "email@address.here" using a password, which is not the authentication method you used during sign up. Try again using the authentication method you used during sign up. (error=identity_provider_mismatch)
So I defiantly need to sign in with the saml by the looks of it.
What do you think the role of AI will place in elections, specifically on voter canvassing and showcasing how data can be used for campaigns?
https://app.askhumans.io/study/817
Do you foresee any negatives of AI's impact in elections, if so what?
https://app.askhumans.io/study/817
Does the GPT that i created using the latest 4o or it is still using an older version?
Does anyone have a version of ChatGPT 4o that can access Video, or any of the new features just announced? I have the paid version on Android and iPad but so far neither does anything new. I've been asking a couple of times a day, "Can you access the Video Camera?" It keeps saying no, and its replies seem a little more grouchy each time. Maybe that's one of the new features?
I've been keeping up with AI and LLM trends for some time, and I must admit it can be quite daunting for the average developer trying to create an AI app. Big companies like Google, OpenAI, and Microsoft are developing AI platforms for nearly every use case. How can we compete with these giants when they make parts of their products freely accessible to consumers?
I just got done drafting an email with the Chat Mac app and yeah, I never wanna do this any other way again. This was not a dictation; this was me fumbling and stuttering through a hazy recollection of a meeting I'd had as the model listened ("no worries," it said at one point, "take your time"); answering a few follow up questions as it sought clarity and further details; and then listening to it read back a drafted email based on my recollection of events. I know the previous voice model could do some of this already but it didn't feel this organic; more like a juiced up Siri with a slightly more natural voic. The fact that this isn't even the dynamic voice model we saw in the showcase is wild to me.
I've also been running generated text from 4o through AI detectors and its output performs way better than model 4 did.
It all leads me to believe that they prioritized mimicking humanness with this 4o model than other kinds of performance. So while I might revert to 4 for more complex tasks (I don't think 4o peforms quite as well), 4o really does feel great when interaction is involved.
Any one else had a different experience?
I tried uploading some documents on free tier 4o to visualise data, and it didn't even read the document.
Hey everyone, just wanted to let you know about Vibe!
It's a new transcription app I created that's open source and works seamlessly on macOS, Windows, and Linux. The best part? It runs on your device using the Whisper AI model, so you don't even need the internet for top-notch transcriptions! Plus, it's designed to be super user-friendly. Check it out on the Vibe website and see for yourself!
And for those interested in diving into the code or contributing, you can find the project on GitHub at github.com/thewh1teagle/vibe. Happy transcribing!
Like many, I was floored by the recent ChatGPT 4O desktop app demo. Specifically how you can share your screen with it and have it help you in real time.
I could really use this on a project that I'm working on but they've decided not to release a Windows app until later.
I thought I saw someone mention somewhere about an open source version of it being worked on?
Anyone know anything about that?
Hey, I just downloaded the Mac app and it seems the answers are being printed REALLY slow in comparison to the website. I am curious if anyone else is experiencing this. It is taking over a minute to get me an answer sometimes, and it's mostly due to the printing on the screen.
Thoughts?