Will AI End Scientific Thinking?
Artificial Intelligence | Devin Ranasinghe
Artificial intelligence (AI) improvements have made humans more cautious of AI. However, there seems to be one “blind spot.” That is, “What is the effect of AI on scientific thinking?”. We could call the scientific thinking process objective-abstract-imaginative thinking. To generate the above schools of thought, humans need to observe the objective world, utilise imaginative power to develop an idea, and then use abstract concepts to explain a particular phenomenon. Unlike scientific thinking, however, AI absconds core knowledge but still produces robust results. So, will the rise of AI end the era of scientific thinking?
Part 1: The Loss of Gravity
Coming from a different background to New Zealand, I always wondered how different cultures, especially the Western world, perceive the world. One of the key features of Western thought (the Enlightenment) is how to perceive the world through scientific explanations.
One of the most prominent figures of the Enlightenment was Sir Isaac Newton, who discovered the famous laws of motion (LM) and universal gravitation (UG). LM and UG were revolutionary, as we all know, and led to the massive development of technology through classical mechanics and explained planetary motion. But what makes LM and UG scientific? Let’s examine UG further. UG is based on the observation of planetary motion in space (notably the works of Kepler). Using imagination and insights, Sir Isaac derived a hypothesis that actually resembles the UG law. Then, he formulated the laws using abstraction, especially calculus, in his famous book, Philosophiæ Naturalis Principia Mathematica [1]. Additionally, the laws of motion and gravity can be tested and proven by methods such as induction and falsification [2]. This process validates Newton’s laws as scientific and applicable universally (even though the theory of relativity came up with different explanations long after Newton).
Now, let’s examine how we use UG laws generally. We know UG laws by the famous equation:
We know that g (gravity) = GM/R2 ≈ 9.81 ms-2 downwards (since acceleration is a vector). Then, we can derive some simple kinematic equations using the integration of acceleration to obtain the velocity and integrate again to acquire the distance. We can obtain the common kinematic equation, s = ut + 0.5at2. As we know a = g downwards, we can calculate the time the ball took to hit the floor when the height to the floor is a known parameter.
However, there is another simple way of calculating or approximating the time the ball takes to hit the floor. Let’s consider a scenario: someone releases an iron ball from 0 to 49 m heights in 1 m increments and calculates the time taken for the ball to hit the floor at each height. We will have 50 data samples and time against the distance (height). Now, we can use methods like polynomial fitting to find a polynomial expression to fit in the table’s data.
Glossary
Hypothesis: A supposition or proposed explanation made on the basis of limited evidence as a starting point for further experimentation.
Abstraction: Conceptualising a particular phenomenon to make it more understandable. In this context, it is by using scientific concepts and mathematics.
Induction: Generalising a scientific conclusion to a broader vicinity based on limited observation.
Falsification: For a theory to be considered scientific, it must be able to be tested and conceivably proven false. The process of trying to refute a scientific argument is called falsification.
We can utilise MATLAB functions for this. In my case, I used some data on MATLAB and obtained a six-degree polynomial expression (you may add Gaussian noise to the data). As readers, you can repeat the same on MATLAB easily. Then, we can use this polynomial function to estimate the time taken for the ball to reach the floor when it is released from 50 m. We can also use the kinematic equation (s = ut +0.5at2) to calculate the time taken for the ball to reach the floor when released from 50 m. When we compare the result of the polynomial function and the kinematic equation, we can observe some errors between the data, but also a certain pattern that both data from the polynomial fitting and the kinematic equation follow.
Now, let’s recall what we did and did not do. In polynomial fitting, we did not use or input the parameter for gravitational acceleration (g = 9.81). We also did not obtain any kinematic equations. We only observe the height and the time to create a non-linear relation between the two variables. Therefore, we did not utilise UG or LM and their derivation. We only used polynomial fitting to find a relationship between the data, which happens to be non-linear. We are unaware of any relation between the mass of the iron ball and the time taken. We are also unaware of the impact on time when external forces, such as air resistance (speed and direction of wind), are exerted on the ball. We can also dispense the standard metric units like seconds, meters, and even more complex units (ms−2) and use more subjective measurements.
But do these issues matter when using the polynomial expression in real-life free-fall applications? We can add more parameters (mass and air resistance) to fit the polynomial equation and find a relation between the time and the distance (height). This is a fundamental principle of non-scientific thinking: thinking without any core knowledge (no scientific theory) of objects and how they are interrelated in the objective world [3]. It is a method of thinking that is purely based on creating non-theory relations between objects. Some can argue mathematics (abstract concepts) has been used here to create a non-linear relation. However, the mathematical operation used here is simple mathematics to find a relation between numbers. Another argument would be that this experiment belongs to the empirical observation in science. I agree that this is a scientific experiment, but what exactly is the conclusion of this experiment? Still, it absconds from a proper hypothesis fundamental to scientific thinking, which means this experiment is inspired by just day-to-day observation of free-fall objects. Also, it lacks a plausible scientific explanation of why this non-linear relation between time and height exists. This is not explained by a ground-breaking scientific discovery like the universal gravitation.
Part 2: A Box in Black
The OceanGate’s Titan implosion has been a major unfortunate incident reported by the media recently. The implosion has been attributed to different causes, from material failure to communication loss [4]. However, I was gobsmacked to learn that the Titan had made it to the Titanic wrecks in previous explorations. This fact has been overridden by the explanations of how scientific tests have not been adequately conducted during the manufacturing of the Titan. Had the Titan made it to the Titanic wrecks again, this would not have been news!
I am not implying that conducting scientific tests after the implosion is unimportant at all. But there should have been proper scientific conduct on how the Titan made it to the Titanic wrecks before, with possible failures of the submersible. Now, a “Box in Black” or absence of scientific explanation as to how the Titan explored the Titanic wrecks despite its failures that were discovered after the implosion, is present.
The power of science can sometimes be limited by the individuals who conduct it. This is because there would be biases that many would follow unconsciously and are mostly limited to current scientific explanations and experiments. Artificial Intelligence (AI) can be applied to fill this gap where certain biases and ambiguities need to be redeemed.
Let’s take a simple AI, or more precisely, the machine learning (ML) application. An image of a cat needs to be identified correctly from other animals by an ML algorithm. The first step is to acquire a collection of cat images and preprocess them to obtain data regarding rotation, flips, and colour contrast. Next, features such as shape (a cat shape), colour (different colours specific for cats), and unique features (fur, paws) would be extracted by the algorithm. Then, the ML algorithm would train the model to extract cat images from non-cat images. Humans can be involved in any of these processes; they can decide what features should be extracted. After the model is trained, it suddenly, from nowhere, can identify cat images.
Nevertheless, the beauty of ML lies in its architecture. It can occupy different architectures, but a Convolution Neural Network (CNN) would be the most appropriate one to utilise here [5]. CNN is a deep learning architecture where features are selected by feature extraction. It has several layers for functions such as preprocessing and applying non-linearity to emphasise certain features, downsampling, dropping out (retaining only prominent features), generating complex features (dense layer), and calculating the error. Previously, I mentioned humans can be involved in identifying which features need to be extracted, but in a CNN-based ML algorithm (unlike in a shallow learning algorithm, like linear regression, where features might be predefined by humans), the algorithm itself identifies them [6].
Figure 2: An architecture of CNN to identify written numbers [7]. Image by Analytic Vidya
Part 3: A Doomsday for Science?
“Artificial intelligence is growing fast, as are robots whose facial expressions can elicit empathy and make your mirror neurons quiver.” — Diane Ackerman
The main two aspects emphasised in Part 1 and Part 2 are,
1. Artificial intelligence (AI) is just a pattern recognition system.
2. The involvement of scientific theories is absent or extremely limited in AI.
The “Black Box” here mainly lies in the feature extraction and its relationship to the functions used and weights on each connection. In this ML algorithm, no individual can know why each weight was assigned to each connection between neurons to create a relationship between each feature. The features we are discussing could be much more complex, derived from surface features, such as shapes or colour. While it is possible to identify these features after execution, why they were considered more prominent is not evident. Additionally, the functions (such as ReLU, sigmoid, and tanh) employed for extracting features are paramount in deciding which features are prominent [5-6]. It is unclear why the functions remove some features while retaining others. The architecture will use different weights on each connection between neurons to identify and minimise the training error. Therefore, the prime target of the model is to minimise the error in assigning values on each connection. Complex features are just a result of reducing error rather than following specific guidelines to identify a feature. If we revert to the news on the Titan implosion, I always wonder if scientists really need to analyse the Titan based on scientific theories such as theories on fluid dynamics, pressure, material sciences, electromagnetism, and signal processing. Or else, they can just insert the data they have already acquired from the previous explorations of Titan and let an ML algorithm decide what features, say, to reinforce the hull, a submersible should comprise. Suppose the ML is input with data from previous explorations such as pressure, temperature changes, and the design deformations that occurred during exploration. Will it output a hull design based on creating complex connections between data?
There could be arguments, such as the AI model was developed on the basis of science and logic. While it has a regular architecture that follows, it does not mean the output of AI can be routed back to the input through its architecture. Most AI experts are now convinced that AI probably would not be entirely understandable forever. Some precautions for AI ambiguity exist, such as explainable AI (XAI) [8]. These models consist of simple architectures that are more transparent and meaningful in interpretation. I request you to explore the topics of XAI and its recent development to realise the differences between AI and XAI.
What would be the immediate impact of AI on the science? I would say nothing because scientific theories and methods are still popular. Theoretical physics has evolved from classical physics, and AI is not substantially involved in the experimentation of theoretical physics yet, especially in quantum mechanics. AI has the capability to replace many realms of classical physics, but that is in the future’s hands to decide. It should be noted that AI is already more utilised in economic science and business environments, where more uncertainties are present [9]. As an example, AI is being used for stock price predictions where prices are highly unpredictable. However, as I mentioned in Part 1, physics theories make precise predictions of the future based on formulas. Economic theories, on the other hand, are not natural science theories like physics. They involve individual psychology, societal aspects, government interventions, and global alterations like the recent pandemic, which are highly uncertain. AI is rapidly used for economic predictions based on different atmospheres. If AI can be successfully used in such uncertainties, can’t it replace the scientific analysis of different phenomena and explain them better than scientific theories?
Figure 1: Polynomial fitting of height against the time taken to reach the floor generated on MATLAB. Image by Devin Ranasinghe.
[1] Tibees, “Reading Newton’s Principia Mathematica by candlelight,” YouTube. Jun. 17, 2019. [Online]. Available: https://www.youtube.com/watch?v=2DBeFqc6c8Y
[2] S. Mcleod and S. Mcleod, “Karl Popper: Theory of Falsification,” Simply Psychology, Jul. 2023, [Online]. Available: https://www.simplypsychology.org/karl-popper.html#:~:text=The%20Falsification%20Principle%2C%20proposed%20by,by%20observing%20a%20black%20swan.
[3] L. Spinney, “Are we witnessing the dawn of post-theory science?” The Guardian, Jan. 09, 2022. [Online]. Available: https://www.theguardian.com/technology/2022/jan/09/are-we-witnessing-the-dawn-of-post-theory-science
[4] K. Hamilton, “Here’s what we know about OceanGate’s sub that tours Titanic—Using 1 button,” Forbes, Jun. 19, 2023. [Online]. Available: https://www.forbes.com/sites/katherinehamilton/2023/06/19/heres-what-we-know-about-oceangates-sub-that-tours-titanic-using-1-button/?sh=242c33353de9
[5] M. Gurucharan, “Top 12 Commerce Project Topics & Ideas in 2023 [For Freshers],” upGrad blog. https://www.upgrad.com/blog/basic-cnn-architecture/.
[6] “Create simple deep learning neural network for classification - MATLAB & Simulink example - MathWorks Australia.” https://au.mathworks.com/help/deeplearning/ug/create-simple-deep-learning-network-for-classification.html
[7] P. Ratan, “What is the Convolutional Neural Network Architecture?,” Analytics Vidhya, Jan. 2021, [Online]. Available: https://www.analyticsvidhya.com/blog/2020/10/what-is-the-convolutional-neural-network-architecture/
[8] Wikipedia contributors, “Explainable artificial intelligence,” Wikipedia, Sep. 2023, [Online]. Available: https://en.wikipedia.org/wiki/Explainable_artificial_intelligence.
[9] S. Gibbons, “2023 Business Predictions As AI And Automation Rise In Popularity,” Forbes, Feb. 02, 2023. [Online]. Available: https://www.forbes.com/sites/serenitygibbons/2023/02/02/2023-business-predictions-as-ai-and-automation-rise-in-popularity/?sh=358a74cc744b
Devin is an enthusiast of science, technology, and philosophy. He is passionate about discussing ideas related to technological improvements and their effect on human life. As an ardent reader of material related to the above fields, he tries to derive his interpretation of ideas that can sprout new perspectives.