How To Prepare AI For Uses In Science

25,943
0
Published 2024-05-02
Is AI ready for use in the sciences? And if not, how can we get there? Stephen Wolfram, Chairman of Wolfram, spoke at Imagination In Action's 'Forging the Future of Business with AI' Summit and speaks about why AI is better with LLMs and how we can use AI usefully in science.

Subscribe to FORBES: youtube.com/user/Forbes?sub_confirmation=1

Fuel your success with Forbes. Gain unlimited access to premium journalism, including breaking news, groundbreaking in-depth reported stories, daily digests and more. Plus, members get a front-row seat at members-only events with leading thinkers and doers, access to premium video that can help you get ahead, an ad-light experience, early access to select products including NFT drops and more:

account.forbes.com/membership/?utm_source=youtube&…

Stay Connected
Forbes newsletters: newsletters.editorial.forbes.com/
Forbes on Facebook: fb.com/forbes
Forbes Video on Twitter: www.twitter.com/forbes
Forbes Video on Instagram: instagram.com/forbes
More From Forbes: forbes.com/

Forbes covers the intersection of entrepreneurship, wealth, technology, business and lifestyle with a focus on people and success.

All Comments (21)
  • Wow, Joscha Bach and Stephen Wolfram on one stage. Is there more of this discussion?
  • @ReflectionOcean
    By YouSum Live 00:00:00 Science and AI limitations in predicting complex systems. 00:00:30 AI struggles with extrapolation beyond trained data. 00:01:41 Language simplicity aids AI success in text analysis. 00:02:07 AI's limitations in creativity and originality. 00:06:53 Computational exploration of vast possibilities by humans. 00:09:23 Computational language as a tool for formalizing the world. 00:14:52 The importance of computational thinking and automation in work. 00:15:20 Leveraging AI as an interface for computational tasks. 00:16:48 Training AI models for specific computational tasks. 00:17:45 Weak form of computation in llms. 00:17:50 Challenges in guiding proofs using llms. 00:18:00 Limitations of llms in mathematical proofs. 00:18:43 Llms excel in making homework but struggle at edge of human knowledge. 00:19:02 Llms prone to errors in math without guidance. 00:19:37 OpenAI's focus on long-form reasoning surpassing human capabilities. 00:20:20 Building systems to extend human capabilities. 00:20:36 Exploring the fundamental workings of machine learning. 00:21:26 Balancing computational capabilities with human needs. 00:22:11 Challenges in developing effective AI tutoring systems. 00:22:28 Goal for llms to understand and assist human learning. 00:23:00 Conceptualizing beyond human intelligence and AI capabilities. By YouSum Live
  • @mikezooper
    My AI predicted the text “so to speak”. Only joking, I love Stephen’s videos. He’s a true genius.
  • @manit77
    Wolfram is our modern day genius.
  • @eyykendrick
    Anyone know where to find the full talk? Thank you
  • @johnkintree763
    So, there is a plugin for ChatGPT so it can access Wolfram resources. How about an interface to Wolfram resources that can be used by any language model?
  • Also it's pretty discrediting to LLMs to say they are only good because language has (easy) grammar. A lot of tests on LLMs show that they have a (though limited, incomplete) world model. It understands basic mathematics, and some basic things about our world.
  • @jurycould4275
    Thank you Stephen for being a scientist and a man of truth!
  • @XenoZona
    I liked the part where he said "computational"
  • @dr.mikeybee
    Semantic space has a shape. It's a model, so of course it has a similar shape to what is being modeled. I like the idea that only that which is simple or computationally reducible can be modeled sufficiently in current scale foundation models. Rigorous agentic behavior is necessary to deal with computationally difficult activation pathways.
  • @GerardSans
    It’s encoding not compression. The difference is subtle but important for technical rigour and to explain the decoding which holds the generative capacity. Decompression wouldn’t be considered correct either it’s called decoding.
  • @johnkintree763
    Before we can expect an AI to accurately predict meaningful events, it probably needs to be able to accurately describe the present, and prior events. A graph structure is probably a good way to represent the present and the past.
  • There is no specificity regarding the metrics of measuring computational intelligence and representing it.
  • @Morris_MK
    Would be helpful if he could produce a simple example in which LLM plus his calculation engine is better than LLM alone.