OpenAI's new model is a "RESEARCH ARTIFACT" | Unlocks "Society of Minds"?

15,439
0
Published 2024-07-21
The latest AI News. Learn about LLMs, Gen AI and get ready for the rollout of AGI. Wes Roth covers the latest happenings in the world of OpenAI, Google, Anthropic, NVIDIA and Open Source AI.

My Links 🔗
➡️ Subscribe:    / @wesroth  
➡️ Twitter: x.com/WesRothMoney
➡️ AI Newsletter: natural20.beehiiv.com/subscribe

#ai #openai #llm

RELATED LINKS:

"mini" announcement by OpenAI
openai.com/index/gpt-4o-mini-advancing-cost-effici…

RouteLLM
lmsys.org/blog/2024-07-01-routellm/

Generative Agents: Interactive Simulacra of Human Behavior
arxiv.org/abs/2304.03442

Generative Agents on GitHub
github.com/joonspk-research/generative_agents

Improving Factuality and Reasoning in Language Models through Multiagent Debate
composable-models.github.io/llm_debate/

All Comments (21)
  • @andrewsilber
    There’s another knock-on benefit to mini: it means that now companies on the fence about integrating Ai into their UX will be less hesitant because the cost is lower and the reliability higher. And more adoption in the space means more momentum and competition in the space, which can only be good for progress :)
  • I'm already doing it. I'm virtualizing all the characters in my game with GTP-4o-mini. And it's really, really good. And it's not bankrupting me. So that is great.
  • @Rhomagus
    No TV and no beer make Homer something something.
  • @danteps3
    Great content as aways! Also, I don't believe that a jump from 95% to 98% in the debate rounds can be considered a plateau. As we approach perfection, improvements become increasingly challenging. Consider the difference between 99% and 99.9%; that's a tenfold improvement. The last few percentiles are significantly more impactful.
  • @kirtjames1353
    Hard to get excited anymore for an openAI drop because they tease and don't ship.
  • @7TheWhiteWolf
    My best guess is OpenAI is trying to optimize as much as possible before they kick off the next model. Perhaps they figure GPT-5 isn’t economically viable enough yet.
  • @0x0404
    They teamed up more with apple right? They were probably tasked to get something running natively on phones. This could be early results for that sort of effort
  • @jaysonp9426
    it blows me away people don't understand how big this was
  • @ekot0419
    Mini is amazing. The cost of using uts api is so much cheaper.
  • @novantha1
    I don't really care how low cost a new model is, personally. If I'm using a large corporate closed source model I'm using it for its quality; if I have something that can be done with mixture of agents, agentic frameworks, agentic workflows, text grad and so on, I can absolutely just run it slowly, locally, for free, with a greater degree of customizability, fine tuning, priority inference, and no need to worry about internet connection. On that note, is there any chance you could do a video on textGrad? I think it's more interesting for people familiar with Pytorch in the sense that it's made in a way that reflects the syntax that an ML engineer would already be used to, but it's a pretty strong formalism of a lot of agentic and reflective workflows that people are looking at lately. As an example, it can do differentiation through text, like you might have a prompt for an LLM "how good is the following text on a scale of 1-10" and it could backpropogate that through the "network" of agents / prompts, and in fact you could even continue propagating it back into a neural network from which the answer originated if you needed to. It seems a bit silly at first because a person can already do a lot of what it does more simply (less formally) with just prompt engineering and plaintext operations, but I honestly think it's really powerful once you get to using it.
  • @KivoMedia
    I have one word for OPENAI - Anthropic!!!! and another word, llmafile... don't forget to install CUDA and use the appropriate command line. Fast 7b or 70b llm running on your laptop. MOE, society loop using python and you're almost at gpt 4 levels.
  • @Yipper64
    5:37 I am currently working on my own project doing this exact thing. Its more of an experiment but im interested to see what would happen if each agent could grow and change based on assessing situations they are put in. Or to put it another way im giving each agent a personality that will be used as context for every prompt, and that personality context can change every time the AI does something. This could be watching TV, taking a nap, having a conversation, playing a game, really anything I can think of. I plan on using open source locally run models for this.
  • @TimothyMusson
    It seems they've switched to 4o-mini as the default "free" chat model, too. GPT-3.5 is no longer listed as an option for new chats.
  • @Duncanate
    I would like for them to release a model that would be indistinguishable from a person for conversations.
  • @2CSST2
    Take good care of yourself, you've stood up as the best AI related stuff covering channel over time for me, second place goes to AI Explained who's fallen a liiiittle bit to the "AI progress is just hype and not so real" mind virus lately. We need you to remain the beast at doing what you for the long term!
  • @dokkey
    If only it was OpenAI so people could research but instead it's ClosedAI.
  • @senju2024
    Society of Minds is a good concept to remember going forward. Thanks for the info.
  • @GNARGNARHEAD
    wow mini is great, paused the video to play around with it, so much faster and seems just as good 👍
  • @jtjames79
    I am and always have been a patient gamer. I don't pay more than $20 for anything. Fundamental economic improvements. Excite me the most. Sometimes I get a stiffy thinking about logistics.