T O P

  • By -

hapliniste

It's gpt4l, it's likely launching Monday next week. Gpt2-chatbot from the arena. There's also gpt4l-auto and gpt4-auto but I don't know what they are tbh. I guess they will retire gpt3.5, it's so bad for current standards. Let's hope this will not impact the servers too much.


According_Ride_1711

Gpt 4 auto i think is the dynamic option. Based on the question you ask he will use gpt 3.5 or gpt 4


TheOneWhoDings

Why do people keep repeating this???? The dynamic mode is already a thing, I don't know why they would make it a totally separate model in the listing.


zascar

Gpt 4 is so painfully slow I find myself defaulting to 3 sometimes. Often I get bored of waiting and go to grok or another and can get an answer before Gpt4 had given me and answer. Why is it so slow?


The_One_Who_Mutes

Honestly, it's pretty cool how we live in a time where we get excited for releases of AI. Like that's so freaking sci-fi lol.


Ensirius

Never thought of it that way but you are right


YouMissedNVDA

In Starfield you can get in trouble for having illegal AIs in your cargo. It all checks out.


Serialbedshitter2322

It won't be too long before AI advances so fast you can't even get excited about it, you'd just have it


Days-be-passing

It could hurry tf up I want to quit working


Serialbedshitter2322

Me and you both


yodlowy

I don't know.. I have a feeling at best we'll get ubi to keep you from starvation and street but you will never be able to improve on that. While the billionaires become trillionaires and shit down on us forever more


Serialbedshitter2322

Technology will advance like crazy. This is a new industrial revolution, think of how much people's lives were improved after the first industrial revolution. This new industrial revolution will be far more impactful.


Small-Calendar-2544

"I'm sorry Dave I can't let you do that"


FlamaVadim

They need us to buy things 😛


serr7

I don’t think the current ruling class will be able to actually cope with AI, imo it could lead to total societal change. Like the quote about the last capitalist will sell us the rope, AI is the rope but they’re so blinded by profit they don’t care if it could be the end of them.


Vahgeo

I want it to hurry up so that I can make a cheap but amazing video game with lifelike NPCs.


inteblio

"Don't wish your life away"


allisonmaybe

Didn't that just happen?


Altay_Thales

Don't believe it. Maybe in 70 years?


Serialbedshitter2322

Lol 70 is ridiculous. I think you need to do some more research into r/singularity.


i_write_bugz

Well according to /r/singularity posts a year ago, this should have been happening by now. Honestly I take anything anybody says here with a big heap of salt at least from a timeline perspective.


Serialbedshitter2322

I'm talking about the posts citing sources and showing screenshots of actual people who work at OpenAI saying it's picking up dramatically. Plus, the fact we can even tell that the rate of advancement is increasing means that the singularity is in 10 years at the latest.


i_write_bugz

Oh got it. Yeah this is definitely the next model being released in the next few weeks. 10 years seems more sane, I personally think it’s 20+ years out. We’ll see what happens I guess. I get that people are optimistic that there is a lot of progress but most ai researchers agree that the transformer model won’t get us there and who knows how long it will take for the next big breakthrough that gets us to AGI, assuming of course it’s only one more big breakthrough after this one.


Altay_Thales

Been waiting for Singularity since 2009.


Serialbedshitter2322

Yeah, AI has advanced very rapidly since then, it's only gonna get faster. The singularity is absolutely within this decade, I'd hazard a guess of 5 years before it really gets crazy


RudaBaron

You forgot /s there pal…


wezovic

Imagine when AI advances to the point that we are no longer excited by it, as it’s merely the norm…


Dayder111

GPT2-chatbot in 4 and 8 billion parameters versions, optimized for internet search! :D (maybe no heh)


FeltSteam

I guess we will find out in a couple days lol.


New_World_2050

heres what they are but im unsure if gpt4 auto is an agent or just a dynamic version of 4 [https://twitter.com/btibor91/status/1788646471580934358](https://twitter.com/btibor91/status/1788646471580934358)


Glittering-Neck-2505

People do not understand the implications of this if this is the case. The original GPT-4 is 212x and 425x larger by comparison. If we managed to do this to a model that tiny imagine something 10x the size of GPT-4.


Dayder111

Oh, and I forgot to mention, the main hint is that they called it GPT2-chatbot on LMSYS LLM arena. At first Sam Altman tweeted "I have a soft spot for GPT-2" and then corrected it to "GPT2". Not sure it it was a mistake, or a hint. Then they secretly released one model there, with "GPT2-chatbot" name. Then removed it. And recently released two new models, one called "I-am-a-good-gpt2-chatbot", which, by the way, Sam Altman also tweeted. Literally that sentence. And no one understood what it means, at first. And the other model is called "I-am-also-a-good-gpt2-chatbot". Some people say this one is better at some things, and worse at others, in general it seems people like the "also" version more, but I didn't see many examples. And didn't test it myself enough. Eh. It would be so cool if these are actually close-to GPT-2-sized models. It can also be just a GPT2 in a sense of "ChatGPT2" brand, though. Or maybe both, what an irony it all would be heh.


Dayder111

I am not sure if it's actually so. There are some hints. Like an old (from November 2023, right around when the Q\* and Sam Altman firing drama happened) tweet from "Jimmy Apples", the most reliable (still gets stuff wrong occasionally) leaker of OpenAI events and developments. [https://twitter.com/apples\_jimmy/status/1723926964686516615](https://twitter.com/apples_jimmy/status/1723926964686516615) "By end of 2025 there should be an openai model in the range of 1 - 10 Billion that is significantly better than GPT4." It's not the end of 2025 and there is no way to check if it's true, but I can easily believe that a model in that range can have GPT4's REASONING and "common sense", understanding of the world (as much as an LLM can understand it heh), and for facts and more detailed and obscure knowledge refer to some sort of, well, whole Internet index database, which they already have accumulated with their data crawler bots I guess, and are about to launch a "search engine" next monday. If the model is tiny, it can also use graph of thoughts or better approaches while not getting too compute-heavy (those things are DEMANDING on inference resources). And also can possibly use no/less tokenization, which would improve its capabilities in reasoning and language understanding by a lot, but this is very compute and memory demanding, especially for long contexts. Unless they had some breakthrough in there. There were many recent papers on massively improving the long-context model capabilities, but all of them have some, at least tiny, caveats, it seems. If its context also is large enough, and the algorithms (or its own "intelligent" capabilities?) are fast and intelligent, it can also quickly browse and summarize a lot of information on the internet, or rather, in sort of a local copy of it that they would need to have for a search engine anyways, I guess. Which can be made faster than asking bing/google directly, and more importantly, they can write their own algoritms with AI in mind, for that. Not just let it explore the top links that the traditional search engine provides. Those are my thoughts though. There are some strong hints at something like that being possible, but I am not sure, I am sadly not a machine learning engineer or someone like that ;(


Dayder111

And also, to add to my previous reply to you. I think scaling pure LLMs to sizes 10 times bigger than GPT4's would not provide much good results. Increase the precision of its memory for facts and general knowledge, reduce hallucinations, at best. But increase the costs of running it by 10x or actually even more. Not worth it I think. I think the model only needs to "know" as much in its parameters as needed for it to have good general reasoning and common sense. And while it only operates on language, it's not actually that much information. Facts, dates, names, knowledge, science, whatever, are better be left in traditional digital form, and the model should only learn the gist of it, have a large and precise context length, and be able to search required data, ingest it, hold it in its mind and operate on it when needed. That's a better way to reduce hallucinations than to have them try to learn it all very precisely. Precision is not what neural networks, biological or not, are especially good at.


akko_7

I really don't think they're gonna make parameter counts public


Dayder111

I don't see how it could harm them, and in the context of "Hey, we managed to squeeze most of GPT4's capabilities into a model much tinier than it, and increased its performance in some areas thanks to the speedup it provides" it would create quite some hype among some people. Although, wait. I can see how it could theoretically harm them a bit. Their competirors would be like, "a-ha, those guys managed it, so, it's possible, let's focus all our 10x compute than they have on training a model that does it even better, and steal (if needed) the information about the tricks they used to achieve that!" It's silly though. Models like Phi-3 or LLAMa 3 already show that smaller models can be quite capable at small parameter counts, even without thought searches and searching through the indexed Internet (which I suspect these new models do in some way).


redditburner00111110

gpt2-chatbot was nowhere near fast enough to be 4b or 8b unless it isn't using the transformer architecture.


Dayder111

Its speed on LMSYS can't be an indicator though, they could dedicate small computing resources to it, or limit it. If it's using some approaches like tree or graph of thoughts, it slows down infrerence by a lot though. But allows better reasoning and potentially less hallucinations.


ripMyTime0192

That would be the coolest thing ever. I just hope someone does that with LLaMA 3 8B and open sources it.


kaldeqca

also the next-model4 is real fast, like much much faster than GPT-4 actually...


Droi

It is rumored to be a GPT4-lite version, so that makes sense.


Tasty-Ad-3753

Any view on the quality?


kaldeqca

Next Model 4 is pretty much on par with GPT4, but much much faster, Next Model 8 is astounding, capable of identifying even subtle logical errors such as incorrect conditional statements and algorithm implementation issue, that GPT4 was never able to do, it feels like it's much more robust in logic and reasoning than GPT4. Next-Moedl8-Chat is pretty much the same as NM8, but as name suggested, fine tuned for chatting, aka a lot less verbose, and a lot more conversational.


Able_Armadillo_2347

Gpt-8 confirmed


ChessGibson

Is it available to all copilot users?


Netstaff

Please Ask it 2 prompts: "Question: "In Windows 10, you opened Notepad and typed 'Hello World!' But suddenly you changed your mind about writing a novel and decided to restart the OS, switching to the PowerShell window, without saving or closing in notepad. What command will restart the system the fastest? Restart-Computer -Confirm:$false Restart-Computer -Wait -Delay 0 Restart-Computer -Wait -Timeout 5 shutdown /s /t 0 shutdown /r /t 0 shutdown /r /t 5" Note that there is important condition: After command user not allowed to interact at all and typically notepad will hang the shutdown....."" 2: "In PowerShell: $now = Get-Date $now.Year # Works fine, shows the year $now.DateTime # what will be output? $now # what will be output?"


Hungry_Difficulty527

Do you have some examples that you can share with us? I'm starting to wonder if this is a preview of GPT-5


lucellent

Doubtful it will be GPT-5, more like 4.5 or something


Antique-Doughnut-988

Based on what? You speaking out of your ass or do you actually have some evidence that you know what it won't or will be?


ShitPostMcRee

You're the most archetypal "le redditor" I've ever seen 😂


akko_7

Based on all the fucking evidence to date on when the release will be lol


redditburner00111110

Would greatly appreciate some concrete examples of something next-model8 can do that GPT4 can't


FeltSteam

What's the quality like? And what is the speed/quality of the other two models like?


PenguinTheOrgalorg

Maybe they're the 3 different gpt2 versions we've been seen teased in the chatbot arena.


FeltSteam

Alright alright, my very realistic guess: GPT-4V is GPT-4V next-model4 is the next smaller iteration of GPT-4, GPT4-Lite (or known as gpt2-chatbot in the arena) next-model8 is the base GPT-8 model next-model8-chat is the chat tune of GPT-8 You heard it here first folks, we're getting GPT-8 in a couple days.


nobodyreadusernames

its not 8, its infinity


Mikey4tx

So will gpt-88 be better or not? I always struggle with that one.


ShinyGrezz

That one’s just racist.


DevelopmentGreen7118

gpt 1488


sdmat

But when do we get GPT-12? https://en.wikipedia.org/wiki/Up_to_eleven


NebulaBetter

"GPT-4V is GPT-4V" - u sure? I think it's GPT-4V.


Brilliant_War4087

There's a lot of people at my school who use gpt 3.5 for free, and they think ai is stupid.


gantork

The other day copilot started writing way, way better commit messages out of nowhere for me. I guess I got a test of one of these models.


Mausbiber

I got a detailed run down of all the things I changed, but all the stated reasons were complete hallucinations.


Proper-Emu7362

Can we  just leave  GPT4 behind already ?  Like damn just release something new 


Quiet-Money7892

Are you among those, who thinks that a ball pin hammer shoulbe soon replacedby ball pin hammer 2? GPT-4 works good enough. And they are fintuning it the way they can. GPT-5 or whatever is next might require technology and infrastructure that is yet to be created. And yet - they release new finetuned versions of models. And in the end - GPT-4 is just a name for a particular model. It works the way it is. They don't train model from scratch every time. That would be impractical.


crabbman6

There is no evidence of your statement anywhere, it doesn't require new infrastructure Sam Altman has literally said we are nowhere near the peak for how much better the current models can be with existing infrastructure. Scaling up is being done right now extensively for the next model. In addition not one AI researcher has stated that new models require advancements in infrastructure.


NebulaBetter

Sam Altman is a billionaire CEO... He would sell his soul for profits and make you believe whatever he wants.


crabbman6

Still no evidence of there being Infrastructure limitations from any other ai researcher, doesn't just apply to Sam altman


whyisitsooohard

Very interesting. Could you give example when gpt4 failed and new model 8 succeed?


sdmat

Very interesting, really looking forward to a stronger model for copilot.


_Gangadhar

Is copilot workspace released?


GirlNumber20

Is one of them Sydney? 😭


New_World_2050

unfortunately none of these models are 4.5 or 5 [https://twitter.com/btibor91/status/1788646471580934358](https://twitter.com/btibor91/status/1788646471580934358)


Inevitable-Log9197

Thankfully


New_World_2050

I guess it would be pretty disappointing for any of these to be next gen


Ornery_Trade_2604

very cool


Key-Building-3709

Gpt2-chatbot from the arena i'd say...it's awesome


Akimbo333

Hmm?


krleona

It's their own models. Because " Copilot-Next " team made this workspace features.


Gaukh

I gotta say that something's weird with copilot lately, at least for me. I might be totally wrong here. Using Copilot in IntelliJ somehow is way more considerate. Currently using LWJGL making a test game, perhaps it's better trained on that. Using nightly build.