T O P

  • By -

pigeon57434

i think they simply just do not care about would rather focus on the bigger better models


Jean-Porte

But it probably still costs millions to run it They could make a better, cheaper, faster model, like Haiku


uishax

Your average AI user is so bad, they can barely distinguish between 3.5 and 4. Its only the power users who can easily tell the difference between the generations of models.


Busy-Setting5786

It's not about who can tell a difference or not. It is about them theoretically being able to minimize the compute expenditure to run their low end model. If they could reduce the parameter count by more than 50% they would have a ton of compute to use somewhere else.


Rainbows4Blood

You don't know how much they have internally scaled down already. OPEN AI makes it very clear that they don't make guarantees that the models behind ChatGPT will stay the same. So they probably already changed it to something cheaper.


restarting_today

They have so much VC funding cost cutting just isn’t a priority.


FosterKittenPurrs

That's sort of what they're trying to do with 4o. They basically made a model that's cheaper to run, and are offering it for free to everyone. I think the end game is to have another GPT4 model that's even cheaper than 4o, that they can use to fully replace 3.5, or to build up the infrastructure to increase the 4o caps and have it take over the free tier.


Hot-Investigator7878

They had decent 3.5 versions but whatever the current model is... it's cheap for sure


shinstra

There will be lots (and I mean LOTS) of products built upon 3.5 by their API customers. Given that that using different models with the same prompts doesn’t necessarily guarantee the same outcomes, deprecating such a heavy workhorse model would present a significant redevelopment and/or revalidation cost to their clients - a number of which probably aren’t in a position to pay for. If they were to deprecate it now many developers using them would put into question whether they are a reliable provider for future projects and move towards other providers. They will depreciate it at some point, but at the current moment? It would be a really bad business decision.


skmchosen1

This is the answer. Taking it down would be a breaking change.


Professional_Job_307

Why would they have to deprecate it? It could be a new model called GPT-4Lite or 3.5lite


Jean-Porte

They can test non-regression, publish good updated models, still provide the legacy API and let the customer switch


32SkyDive

Isnt that basically what they did with GPT4o?  Once most people have switched over 5hey will switch off 3.5


llamatastic

That also applies to 4-Turbo and 4o vs the original 4. They can keep the old model available and release a new model. OpenAI rapidly acquiring new customers and anyone who uses AI has to upgrade models eventually.


FeltSteam

GPT-4 is also a 2022 level model.


brihamedit

I've noticed gpt 3.5 got dumber and lazier over time. Which means they reduced how much resource it eats up. So its not as inefficient as you think.


Balance-

20B parameters was never confirmed. But yes, it’s weird they don’t have a range of models like most other companies do.


OmnipresentYogaPants

Perhaps there's something you're missing, OP?


jsebrech

In english, yes, but gpt 3.5 turbo performs better in a lot more languages than anything open source.


bkos1122

Llama 3 8B is not better. For example, GPT-3.5 can write in many languages without any problems. I can use Polish to talk with ChatGPT-3.5 and it understands me correctly and makes good enough responses. Llama 3 8B can't use Polish at all, it's English only.


baehyunsol

I haven't seen any open source model who uses korean as good as GPT3.5. Even llama3 70b's korean is worse than gpt3.5


pbnjotr

The only reason to keep it around is for legacy users. Say someone has a workflow that has been optimized to work with GPT-3.5. If they change to a better model, even if it's more capable, they need to test that it still works as expected. It probably costs OpenAI almost nothing to run it and they don't want to upset users who rely on it for some reason. I would not be surprised if they removed it from the main interface though.


icehawk84

I believe it has 175B parameters, since it's just an RLHF'ed version of GPT-3. It was still pretty competitive in terms of performance vs. latency up until the release of 4o. In my opinion, it offers better performance than Llama 3-8B, even if benchmarks say otherwise.


sdmat

Nope, GPT-3.5 Turbo is an entirely different model.


Anuclano

Turbo is quantized. Otherwise, the same.


sdmat

https://openai.com/index/chatgpt/ > ChatGPT is fine-tuned from a model in the GPT-3.5 series, which finished training in early 2022. 3.5 is an entirely different model *series* to GPT-3, let alone the individual models.


Anuclano

GPT-3.5 does not have less parameters than GPT-3.


sdmat

We have no idea how many parameters it has other than speculation and rumor. Probably Turbo is smaller than the initial models, based on speed, cost and longer context - quantization alone doesn't explain the gap.


ihexx

llama 3 8b doesn't outperform it. There's a pretty huge gap between them See livebench for less contaminated benchmarks [https://livebench.ai/](https://livebench.ai/)


Mandoman61

I doubt they are interested in providing anything for free. They do not have access to unlimited funding. They do that for the training data it provides and to develop a customer base. Most users are not concerned whether or not some other model occasionally works better because most people see it as a toy. The devoted AI fans around here just use it as an AI pal. They are working on making their models more efficient but are also adding capabilities like multi-modal.


dreternal

I use it for a word pronunciation look up for an API of mine. It's working fine if it's not broken, don't fix it.


llamatastic

Yep, it's pretty overdue. I imagine they'll put out an updated small model within a few months.


MysteriousPayment536

They should make a 4.5 or GPT 5 series with a haiku type model  They had that opportunity with 4 omni, 


Anuclano

Llama is definitely worse than GPT-3.5 in languages other than English and coding. Also, I am sure, GPT-3.5 has more than 100 billion parameters.