ConferenceCall 2024 03 27: Difference between revisions
Ontolog Forum
(→Agenda) |
No edit summary |
||
(4 intermediate revisions by the same user not shown) | |||
Line 35: | Line 35: | ||
== Participants == | == Participants == | ||
* [[BevCorwin|Bev Corwin]] | |||
* [[BruceBray|Bruce Bray]] | |||
* [[DouglasMiles|Douglas Miles]] | |||
* E. W. Schuster | |||
* [[GaryBergCross|Gary Berg-Cross]] | |||
* [[JanetSinger|Janet Singer]] | |||
* [[JohnSowa|John Sowa]] | |||
* Josefina | |||
* [[KenBaclawski|Ken Baclawski]] | |||
* [[MarkUnderwood|Mark Underwood]] | |||
* [[MarkusBuehler|Markus J. Buehler]] | |||
* [[MichaelDeBellis|Michael DeBellis]] | |||
* Phil Jackson | |||
* [[RaviSharma|Ravi Sharma]] | |||
== Discussion == | == Discussion == | ||
=== Precision of Language === | |||
[12:25] John Sowa: There is a huge difference between logic and mush. | |||
* In a mush system, which combines a huge amount of small lumps with probabilities, no single lump can make more than a tiny change to the total. | |||
* But in logic, a single symbol for NOT can reverse the total. | |||
* Language can be precise or mushy. | |||
* It all depends on the problem, the source of information, and the sensitivity to a single word, such as NOT. | |||
* Visual languages, such as diagrams, can be precise or mushy -- in the same way as linear languages. | |||
* The only difference between linear languages and diagrammatic languages is the dimension. | |||
* Mushy or precise depends on the nature of the knowledge source and the way of combining parts. | |||
* Re difference between logic and language: It all depends on the subject matter. | |||
* If the source is precise, logic is precise. Language can be precise, but it's easier to slip into a mushy mode. | |||
=== Logic as a Natural Language === | |||
[12:31] Douglas Miles: Somewhere I read that Bertrand Russell finally confessed that Logic was merely yet another natural language | |||
[12:32] John Sowa: Re Bertie: He had some good ideas, but he wasn't as smart as Whitehead. | |||
[12:34] Michael DeBellis: Replying to "Somewhere I read tha..." Where did Russell say that "logic was merely yet another natural language"? I'm skeptical. If he did say it he was wrong. Natural language and logical languages are fundamentally different. Logic is a context free language. A formula's syntax is independent of any WFFs around it in logic. This is not the case in natural language. In natural language syntax can be context sensitive and the syntax of a sentence can be ambiguous. That's not the case with logic. A WFF is either a WFF or it isn't and if it is the syntax is unambiguous. | |||
[12:36] Douglas Miles: re Berrtie: exactly, I interpreted that to mean logic is designed to be as non-precise as needed to be usable | |||
[12:36] Douglas Miles: non-precise as required to not contradict itself | |||
[12:37] Douglas Miles: or anything else we want to imagine it was supposed to do | |||
[12:39] John Sowa: Replying to "Somewhere I read tha..." I don't know where, whether, or how Russell compared language to logic. But you can express a huge amount of language in a context-free syntax. | |||
[12:40] Michael DeBellis: Replying to "non-precise as requi..." I don't agree that logic needs to be non-precise to not have contradictions. Logic is precise and you can have very complex logical systems that don't have contradictions. That's why ZFC set theory is the foundation for conventional math. Because it has the minimum number of axioms (as far as we currently know) that can support mathematics but doesn't have contradictions the way Frege's system did and it doesn't need the huge complex rules that Russel and Whitehead's logic did. | |||
[12:40] John Sowa: Replying to "Somewhere I read tha..." My favorite comment is by Whitehead: "Bertie thinks that I am muddle-headed. But I think that he is simple-minded." | |||
[12:42] John Sowa: Replying to "Somewhere I read tha..." Whitehead made that comment when he was introducing Russell for a set of lectures at Harvard in the 1930s. | |||
[12:45] Michael DeBellis: Replying to "Somewhere I read tha..." Actually, I would bet if you look at any book there are ambiguous sentences on every page where the syntax is only obvious due to context. We often take these things for granted because we just do it unconsciously. You can also have very simple Natural Language sentences where the syntax is ambiguous: "I saw the man on the hill with the telescope" Who has the telescope? The speaker or the man on the hill? The syntax of that simple sentence has at least two different interpretations. You don't get this kind of ambiguity in logic. | |||
[12:46] John Sowa: Logic cannot be vague, but it is very easy to make mistakes that are horribly bad. | |||
[12:47] John Sowa: It is easy to be vague in ordinary language, and sometimes vagueness is good enough. | |||
[12:48] John Sowa: But in logic, vagueness is not possible. Instead, it easy to be false. | |||
[12:50] Douglas Miles: Replying to "Somewhere I read t..." It was in what is often called his finally (video taped) Interview | |||
[12:51] Douglas Miles: Replying to "Somewhere I read t..." "Final (video taped) Interview" | |||
[13:06] Michael DeBellis: Replying to "Somewhere I read tha..." One question I have is how his work relates to ideas regarding emergent behavior? E.g, Krakauer? | |||
=== Multi-Agent Design Patterns === | |||
[12:38] Ravi Sharma: you [Markus Buehler] are practicing a dynamic visual language | |||
[12:41] Bev Corwin: Reacted to "you are practicing a..." with 👍 | |||
[12:41] Mark Underwood: Very happy to see multi-agent design patterns returning to practice. If only it was better standardized . . . | |||
[12:42] Ravi Sharma: for new designs we not only use Visual dynamics but also in reaching new designs based on generative AI | |||
[12:43] Ravi Sharma: but can you integrate AI and Physics together? | |||
[12:48] Ravi Sharma: Simulate to select the actual material development route. | |||
[12:49] Ravi Sharma: Your work can also help Prosthetics and other bio integrated medical devices. | |||
=== What does an AI system know? === | |||
[12:26] Ravi Sharma: model learns about itself is similar to self healing materials? | |||
[12:27] Douglas Miles: can non "natural language" structures such as logic be enough to teach a model to reason? | |||
[12:48] Phil Jackson: Does the model know what it does not know, and needs to better understand? | |||
[12:51] John Sowa: There are many kinds of ways to use the word 'know' about AI systems. But some systems, such as LLM-based systems, may create an illusion of knowing. | |||
[12:52] John Sowa: It's better to avoid words like 'know' and 'understand' unless you have a very deep understanding of what your AI system is doing. | |||
[12:53] John Sowa: Today, I would not apply the words 'know' or 'understand' to any current AI system. | |||
[12:56] Phil Jackson: Understood. I still wonder whether Markus' system has any knowledge or understanding of its limitations. | |||
=== Standard Tests === | |||
[12:33] Gary Berg-Cross: I wonder if there is a standard test set to see how an AI hybrid system provides useful material models integrating the various levels of granularity. | |||
[12:35] Gary Berg-Cross: A related q might be how to demonstrate that the AI system is learning how to reason about these material issues. | |||
=== Summary Comments === | |||
[12:11] Ravi Sharma: Philosophically Nature is elusive and knowledge is evolutionary | |||
* once we understand local and long range we can understand better how wealth of chemistry and biology -variety can be understood | |||
* graphs glyphs and images are innate to understanding | |||
* language of visual images and graphs needs to be developed but will it be like math or programming language or yet a new paradigm? | |||
* visual language and thought integration will have to be dynamic both in space and time? | |||
* bio development is based on vailability of elements, similarly will other forms evolve differently in exobiology? | |||
* Markus you are really addressing a glaring gap in visual languages. | |||
* therefore bio samples have to be varied as well as subject to determination? How can balance be achieved? | |||
* But you showed only one mode or all modes of protein movement | |||
=== Compliments === | |||
[12:56] Josefina: Amazing course! | |||
[13:05] Bruce Bray: excellent discussion! | |||
[13:06] Bev Corwin: Thank you | |||
[13:06] janet singer: Great material, great presentation. | |||
[13:06] E. W. Schuster: Thanks Markus. | |||
[13:06] Markus Buehler: Thanks everyone! | |||
[13:07] E. W. Schuster: Reacted to "Thanks everyone!" with 👍 | |||
== Resources == | == Resources == | ||
* [https://bit.ly/3IYqiyt Video Recording] | |||
* [https://bit.ly/3TBJcQD Video Recording with Closed Captions] | |||
* [[ConferenceCall_2024_03_27/Transcript|Session Transcript]] | |||
* [https://youtu.be/p_cF35YA0rg YouTube Video] | |||
== Previous Meetings == | == Previous Meetings == |
Latest revision as of 20:21, 27 July 2024
Session | Foundations and Architectures |
---|---|
Duration | 1 hour |
Date/Time | 27 Mar 2024 16:00 GMT |
9:00am PDT/12:00pm EDT | |
4:00pm GMT/5:00pm CST | |
Convener | Ravi Sharma |
Ontology Summit 2024 Foundations and Architectures
Agenda
- Markus J. Buehler Accelerating Scientific Discovery with Generative Knowledge Extraction, Graph-Based Representation, and Multimodal Intelligent Graph Reasoning
- Abstract: For centuries, researchers have sought out ways to connect disparate areas of knowledge. With the advent of Artificial Intelligence (AI), we can now rigorously explore relationships that cut across distinct areas – such as, mechanics and biology, or science and art – to deepen our understanding and to accelerate innovation. To do this we transformed a set of 1,000 scientific papers in the field of biological materials into a detailed ontological knowledge graph. We conduct a detailed analysis of the graph structure and calculate node degrees, communities and community connectivities, as well as clustering coefficients and betweenness centrality of key nodes, and find that the graph has an inherently scale-free nature. Using a large language embedding model we compute deep node representations and use combinatorial node similarity ranking to develop a path sampling strategy that allows us to link dissimilar concepts across the graph that have previously not been related. We apply this method to reveal insights into unprecedented interdisciplinary relationships that can be used to answer queries, identify gaps in knowledge, propose never-before-seen material designs, and predict material behaviors. One comparison revealed detailed structural parallels between biological materials and Beethoven's 9th Symphony, highlighting shared patterns of complexity through isomorphic mapping. In another example, the algorithm proposed an innovative hierarchical mycelium-based composite based on a joint synthesis of path sampling with principles extracted from Kandinsky's `Composition VII' painting, where the resulting composite features balance of chaos and order, adjustable porosity, mechanical strength, and complex patterned chemical functionalization. We uncover other isomorphisms across science, technology and art, revealing a nuanced ontology of immanence that reveal a dynamic, context-dependent heterarchical interplay of entities beyond traditional hierarchical paradigms. Because our method transcends traditional disciplinary boundaries, and because it integrates diverse data modalities (graphs, images, text, numerical data, etc.) we achieve a far higher degree of novelty, explorative capacity, and technical detail, than conventional generative AI. The approach establishes a widely useful framework for innovation, drawing from diverse fields such as materials science, logic, art, and music, by revealing hidden connections that facilitate discovery.
- Bio: Markus J. Buehler is the McAfee Professor of Engineering at MIT. Professor Buehler pursues new modeling, design and manufacturing approaches for advanced biomaterials that offer greater resilience and a wide range of controllable properties from the nano- to the macroscale. He received many distinguished awards, including the Feynman Prize, the ASME Drucker Medal, the J.R. Rice Medal, and many others. Buehler is a member of the National Academy of Engineering.
Conference Call Information
- Date: Wednesday, 27 March 2024
- Start Time: 9:00am PDT / 12:00pm EDT / 5:00pm CET / 4:00pm GMT / 1600 UTC
- ref: World Clock
- Note: The US and Canada are on Daylight Saving Time while Europe has not yet changed.
- Expected Call Duration: 1 hour
- Video Conference URL: https://bit.ly/48lM0Ik
- Conference ID: 876 3045 3240
- Passcode: 464312
The unabbreviated URL is: https://us02web.zoom.us/j/87630453240?pwd=YVYvZHRpelVqSkM5QlJ4aGJrbmZzQT09
Participants
- Bev Corwin
- Bruce Bray
- Douglas Miles
- E. W. Schuster
- Gary Berg-Cross
- Janet Singer
- John Sowa
- Josefina
- Ken Baclawski
- Mark Underwood
- Markus J. Buehler
- Michael DeBellis
- Phil Jackson
- Ravi Sharma
Discussion
Precision of Language
[12:25] John Sowa: There is a huge difference between logic and mush.
- In a mush system, which combines a huge amount of small lumps with probabilities, no single lump can make more than a tiny change to the total.
- But in logic, a single symbol for NOT can reverse the total.
- Language can be precise or mushy.
- It all depends on the problem, the source of information, and the sensitivity to a single word, such as NOT.
- Visual languages, such as diagrams, can be precise or mushy -- in the same way as linear languages.
- The only difference between linear languages and diagrammatic languages is the dimension.
- Mushy or precise depends on the nature of the knowledge source and the way of combining parts.
- Re difference between logic and language: It all depends on the subject matter.
- If the source is precise, logic is precise. Language can be precise, but it's easier to slip into a mushy mode.
Logic as a Natural Language
[12:31] Douglas Miles: Somewhere I read that Bertrand Russell finally confessed that Logic was merely yet another natural language
[12:32] John Sowa: Re Bertie: He had some good ideas, but he wasn't as smart as Whitehead.
[12:34] Michael DeBellis: Replying to "Somewhere I read tha..." Where did Russell say that "logic was merely yet another natural language"? I'm skeptical. If he did say it he was wrong. Natural language and logical languages are fundamentally different. Logic is a context free language. A formula's syntax is independent of any WFFs around it in logic. This is not the case in natural language. In natural language syntax can be context sensitive and the syntax of a sentence can be ambiguous. That's not the case with logic. A WFF is either a WFF or it isn't and if it is the syntax is unambiguous.
[12:36] Douglas Miles: re Berrtie: exactly, I interpreted that to mean logic is designed to be as non-precise as needed to be usable
[12:36] Douglas Miles: non-precise as required to not contradict itself
[12:37] Douglas Miles: or anything else we want to imagine it was supposed to do
[12:39] John Sowa: Replying to "Somewhere I read tha..." I don't know where, whether, or how Russell compared language to logic. But you can express a huge amount of language in a context-free syntax.
[12:40] Michael DeBellis: Replying to "non-precise as requi..." I don't agree that logic needs to be non-precise to not have contradictions. Logic is precise and you can have very complex logical systems that don't have contradictions. That's why ZFC set theory is the foundation for conventional math. Because it has the minimum number of axioms (as far as we currently know) that can support mathematics but doesn't have contradictions the way Frege's system did and it doesn't need the huge complex rules that Russel and Whitehead's logic did.
[12:40] John Sowa: Replying to "Somewhere I read tha..." My favorite comment is by Whitehead: "Bertie thinks that I am muddle-headed. But I think that he is simple-minded."
[12:42] John Sowa: Replying to "Somewhere I read tha..." Whitehead made that comment when he was introducing Russell for a set of lectures at Harvard in the 1930s.
[12:45] Michael DeBellis: Replying to "Somewhere I read tha..." Actually, I would bet if you look at any book there are ambiguous sentences on every page where the syntax is only obvious due to context. We often take these things for granted because we just do it unconsciously. You can also have very simple Natural Language sentences where the syntax is ambiguous: "I saw the man on the hill with the telescope" Who has the telescope? The speaker or the man on the hill? The syntax of that simple sentence has at least two different interpretations. You don't get this kind of ambiguity in logic.
[12:46] John Sowa: Logic cannot be vague, but it is very easy to make mistakes that are horribly bad.
[12:47] John Sowa: It is easy to be vague in ordinary language, and sometimes vagueness is good enough.
[12:48] John Sowa: But in logic, vagueness is not possible. Instead, it easy to be false.
[12:50] Douglas Miles: Replying to "Somewhere I read t..." It was in what is often called his finally (video taped) Interview
[12:51] Douglas Miles: Replying to "Somewhere I read t..." "Final (video taped) Interview"
[13:06] Michael DeBellis: Replying to "Somewhere I read tha..." One question I have is how his work relates to ideas regarding emergent behavior? E.g, Krakauer?
Multi-Agent Design Patterns
[12:38] Ravi Sharma: you [Markus Buehler] are practicing a dynamic visual language
[12:41] Bev Corwin: Reacted to "you are practicing a..." with 👍
[12:41] Mark Underwood: Very happy to see multi-agent design patterns returning to practice. If only it was better standardized . . .
[12:42] Ravi Sharma: for new designs we not only use Visual dynamics but also in reaching new designs based on generative AI
[12:43] Ravi Sharma: but can you integrate AI and Physics together?
[12:48] Ravi Sharma: Simulate to select the actual material development route.
[12:49] Ravi Sharma: Your work can also help Prosthetics and other bio integrated medical devices.
What does an AI system know?
[12:26] Ravi Sharma: model learns about itself is similar to self healing materials?
[12:27] Douglas Miles: can non "natural language" structures such as logic be enough to teach a model to reason?
[12:48] Phil Jackson: Does the model know what it does not know, and needs to better understand?
[12:51] John Sowa: There are many kinds of ways to use the word 'know' about AI systems. But some systems, such as LLM-based systems, may create an illusion of knowing.
[12:52] John Sowa: It's better to avoid words like 'know' and 'understand' unless you have a very deep understanding of what your AI system is doing.
[12:53] John Sowa: Today, I would not apply the words 'know' or 'understand' to any current AI system.
[12:56] Phil Jackson: Understood. I still wonder whether Markus' system has any knowledge or understanding of its limitations.
Standard Tests
[12:33] Gary Berg-Cross: I wonder if there is a standard test set to see how an AI hybrid system provides useful material models integrating the various levels of granularity.
[12:35] Gary Berg-Cross: A related q might be how to demonstrate that the AI system is learning how to reason about these material issues.
Summary Comments
[12:11] Ravi Sharma: Philosophically Nature is elusive and knowledge is evolutionary
- once we understand local and long range we can understand better how wealth of chemistry and biology -variety can be understood
- graphs glyphs and images are innate to understanding
- language of visual images and graphs needs to be developed but will it be like math or programming language or yet a new paradigm?
- visual language and thought integration will have to be dynamic both in space and time?
- bio development is based on vailability of elements, similarly will other forms evolve differently in exobiology?
- Markus you are really addressing a glaring gap in visual languages.
- therefore bio samples have to be varied as well as subject to determination? How can balance be achieved?
- But you showed only one mode or all modes of protein movement
Compliments
[12:56] Josefina: Amazing course!
[13:05] Bruce Bray: excellent discussion!
[13:06] Bev Corwin: Thank you
[13:06] janet singer: Great material, great presentation.
[13:06] E. W. Schuster: Thanks Markus.
[13:06] Markus Buehler: Thanks everyone!
[13:07] E. W. Schuster: Reacted to "Thanks everyone!" with 👍
Resources
Previous Meetings
Session | |
---|---|
ConferenceCall 2024 03 20 | Foundations and Architectures |
ConferenceCall 2024 03 13 | LLMs, Ontologies and KGs |
ConferenceCall 2024 03 06 | LLMs, Ontologies and KGs |
... further results |
Next Meetings
Session | |
---|---|
ConferenceCall 2024 04 03 | Synthesis |
ConferenceCall 2024 04 10 | Synthesis |
ConferenceCall 2024 04 17 | Applications |
... further results |