Artificial intelligence (AI) and specifically machine learning (ML), a major subset of AI, are two rapidly advancing technologies dominating today's world. They are still in their early stages, but their practical applications have been significantly transforming various sectors across the globe. People's interest in technology has increased substantially due to major events that took place in these domains. The outstanding performance of AI programs in man versus machine-related esports further piqued public interest, stimulating people's curiosity in the area.
AI and ML were first powered by various theories of computers created in the early 20th century. AI is considered the capacity of a machine or computer to perform tasks that usually require human intelligence. In contrast, ML refers to a field within AI that provides algorithms the ability to study and improve themselves. The recent rise of AI and ML has been made possible primarily due to three breakthroughs: the rise of big data technology, increased computational resources, and advanced algorithms. The sharp progress observed is primarily due to these factors. The growing digital economy is projected to continue driving the demand for digital creation and classification as a result of the substantial expansion of online operations, the increasing complexity and sophistication of many organizations, and the growing mobility of work. AI and ML could readily fill this gap. Digitization and data collection are now a part of everyday life. Many smart, constantly connected devices are present. Routine activities can be performed hands-free at homes, schools, universities, and research institutes because of these interconnected devices. AI and ML are essential for a vibrant and youthful generation. This report delves deeper. In the next part, we'll explore the topics: differences and similarities between AI and ML.
2.Fundamental Concepts
Artificial Intelligence (AI) is the big umbrella under which machine learning (ML) and deep learning (DL) operate. Fundamentally, AI is the broader concept used to describe machines and systems that make decisions and act in a way we consider to be smart, like a human. In general, any algorithm that makes decisions could be considered AI. For instance, a thermostat that can learn what temperature you like your coffee or a car that can parallel park itself. Although they are fundamentally different, the terms AI (Artificial Intelligence) and ML (Machine Learning) are often used interchangeably. To aid the understanding between these closely related concepts, it should be clarified that ML is a type of AI, not the other way around. In other words, AI is the discipline that tries to make computers smart, whereas ML is a current implementation that is observing great success in a number of application areas.
Machine learning (ML) can be viewed as a field of artificial intelligence that has significant emphasis on the construction of algorithmic systems that can learn from and make decisions or predictions on given data. There are several methodologies used in ML, including neural networks. Even when ML itself is not being explicitly used, the underlying data-parallel execution model is similar across the various terms and technologies. AI and ML theories are fundamentally intertwined, but historically AI can exist without ML, while ML cannot exist without AI. Although there are fundamental differences between AI and ML, both fields overlap in many ways. Most fundamentally, both fields attempt to answer the same question: how can we make computers do what humans can do? However, the current approach concurs that computers will not do it the same way humans do.
2.1. Definition and Scope of AI and ML
We begin this sub-section with concise but accurate definitions of artificial intelligence (AI) and machine learning (ML). AI refers to the general endeavor or aspiration to create machines and systems capable of carrying out tasks that usually require human intelligence. AI has a considerable scope, spreading to various other domains and having a considerable impact in areas like natural language processing, robotics, and computer vision, among others. In contrast, ML is actually a part of AI, but it refers more specifically to a certain approach. More specifically, ML is a scientific discipline that attempts to design and create algorithms for a system or machine to learn from data. The central idea is not to program the machine to do certain specific tasks but rather to enable the machine to learn from the input data so as to improve its performance on certain tasks. In more colloquial terms, this refers to the machines or systems having the ability to get smarter or better at performing a particular task.
There are various kinds of ML algorithms that fall under the ML paradigm with varying usages and applications. These include help vector machines, linear regression techniques, decision trees and forests, clustering techniques like k-means clustering, and hierarchical or agglomerative clustering, and neural networks including deep learning networks, and so forth. However, the task of classifying the input in some way or recognizing certain input values is a recurrently occurring theme of ML, regardless of the algorithm type. Researchers and companies are combining AI techniques to address significant real-world challenges. Feature specification networks (FSNs) take a different approach by using world models, where a model of the environment that has been learned solely on observational data is used to facilitate planning. The use of these methods in real-world settings demonstrates the breadth of application of AI. The scope of AI solutions and technologies is attested to by the extensive application domain. AI is therefore one of the most powerful research tools in the world today.
2.2. Key Differences and Similarities
As we have previously stated, AI performs tasks that are characteristic of human intelligence: reasoning, planning, learning, and comprehension. ML algorithms build a mathematical model of sample data, known as training data, to generate predictions or decisions. There are some big differences between these two: Firstly, AI aims to mimic human learning and performing day-to-day activities like thinking and understanding, for example. ML is a subdomain of AI. Fundamentally, AI aims for computer programs to be more general so that they can imitate human intelligence and solve any type of problem with the available information.
ML focuses on developing computer programs to do specific tasks using data without any help. In other words, ML trains models to help make decisions based on data-driven learning. Secondly, the methodologies used in AI are designed to perform like a human, such as reasoning, problem-solving, perception, and understanding languages. The subsets of AI research that are centered around a particular aspect are ML. The methods currently used in ML lean heavily toward the use of algorithms in mathematics, which are supported by computational software and hardware. Machines are fed lots of data, and they are trained to identify patterns in data. Both AI and ML use computational models; for AI, it is more complex. AI requires large amounts of data to be available at all times in order for a machine to be accurate. ML can be more general. The technology is scaling so quickly to meet the continuous need for collecting large amounts of data, having faster computing, and providing a wider range of applications. After all, ML is expected to further advance the development of AI by contributing to basic research and innovative productivity alike, creating a mutually beneficial relationship.
3. Types of AI and ML
One form of classification of AI is based on capabilities. There are two types of AI based on capabilities: Narrow AI: AI that has been trained for a specific scope for a specific task, like AI chatbots, robotic process automation, and intelligent virtual assistants. These AIs can only do specific work that they have been trained for. General AI: AI that is trained to think, learn, and understand any intellectual task that a human can do. These AIs are very close to human-level intelligence. General AI is largely still theoretical, and we don’t have any system built that can be categorized into General AI. Many experts and thinkers in the AI space believe that achieving human-level intelligence is very difficult; they predict that we might never be able to achieve that level, which is called artificial superintelligence.
In machine learning, depending on the types of training signals, there are three types of ML techniques: Supervised Learning: This is based on the input-output pair of example data. In supervised learning, training is done on a labeled dataset. Example: digit recognition, prediction, regression – understanding relationships between inputs and outputs. Unsupervised Learning: In this type, no label is present in the data; the training is done on unlabeled data. Example: clustering – like a recommendation system, pattern recognition. Reinforcement Learning: This technique is based on judgments and decision-making. Example: game AI, robots, self-driving cars. In game AI, it learns to play games or is trained according to winning or losing. In these techniques, allowing a system to learn from itself based on previous actions is possible and effective in solving real-world problems and challenges.
3.1. Narrow AI vs. General AI
One way of understanding AI is to distinguish between generating artificial narrow intelligence and artificial general intelligence. Narrow AI refers to the special version because it is limited and focused on only one subset of human activities and decisions. It can develop technological systems that specifically dig into one aspect of such activities or undertakings. For example, we can see narrow AI at work in many modern business applications. Algorithms running virtual assistants can understand spoken natural language requests, follow the conversation, and execute commands. In addition, the cameras in our mobile phones can recognize our pictures by using image recognition software that has been trained to do so. Narrow AI is already doing a vast amount of work and is present in almost all of our web apps and cloud services or the software and hardware systems we deploy and use to perform the most common activities.
Artificial General Intelligence (AGI), or more often just called General AI, is generally thought of as the type of intelligence found in humans. The development of a General AI system would be able to understand and then effectively and efficiently respond to any intellectual task that a human being can achieve. However, the building of General AI remains a controversial area of AI study, as it is not always clear where one should draw the boundary between General and narrow AI. There are many practical challenges that have yet to be overcome. Machines are currently developing patterns on data, learning the rules to identify and classify attributes or states, and even outperforming humans in several cognitive tasks. Nonetheless, their underlying structures remain fundamentally different from the human process of categorization, getting a purely associative machine that has no general cognitive abilities.
3.2. Supervised Learning vs. Unsupervised Learning vs. Reinforcement Learning
One way to categorize machine learning problems is by separating them into one of three types: supervised learning, unsupervised learning, and reinforcement learning. In supervised learning, the computer is presented with example inputs and their corresponding outputs, given by a teacher. The goal is to learn a general rule that maps inputs to outputs. Given a training set that contains example input-output pairs, a model is created to then make predictions about the output when given new inputs. The ML system also receives feedback in the form of the correct output, which is then used to make necessary corrections in order to improve its performance. This supervised learning method utilizes classification algorithms to identify what category an item belongs to and regression algorithms to generate the outputs based on input.
In unsupervised learning, the training data consists of a set of input vectors without any corresponding target values. The goal here is no longer to learn a function that assigns a label to an input, but rather to model the underlying structure or distribution in the data to learn more about the data. Unsupervised learning methods are therefore used for exploratory data analysis tasks, where the aim is not to predict output, but instead to gain insights from the data and to gain value. Cluster analysis is a primary category of assignments performed using this methodology to identify similar groups within the dataset. While both supervised and unsupervised learning methods learn patterns from input, reinforcement learning relies on the sequential decision-making scenario in an environment. In reinforcement learning, an agent makes observations and takes actions within an environment to receive reward signals as feedback. The goal is to improve the policy of the agent using this feedback in order to minimize the long-term cumulative costs.
4. Applications and Use Cases
Artificial intelligence (AI) and machine learning (ML) technologies have a diverse and wide range of applications. Today, we see many practical examples across different industries. In healthcare, AI can be used to predict medical events like stroke or cardiac arrest. In finance, the equity markets are almost entirely driven by AI-based trading algorithms. In transportation, we increasingly rely on AI to make decisions about ride-sharing. In user-facing applications, AI is also evident when predictions are made about what to watch. These technologies bring millisecond efficiencies and the ability to make repeated high-stakes decisions that are beyond the computational capacity of unaided humans.
Use cases come in several different varieties. Predictive analytics suggest a particular course of action, such as investing in a small business, for example, because it is likely to go bankrupt. In-platform AI solutions, such as chatbots, are designed to make transactions from start to finish, without human intervention. Optimization AI is designed to find the best solution to a particular problem within a set of constraints. An example of such a system might decide at each intersection which of several different vehicles to allow through based on changing transportation metrics such as traffic conditions and the number of passengers in the vehicles. In general, and increasingly, AI is becoming synonymous with business and IT—helping optimize operations and generate leads, as well as provide better customer and user experiences. Many AI services are pre-packaged and available to use directly via the cloud. AI and ML technologies are increasingly interwoven into digital businesses, which is changing the core of business competition and innovation. In fact, enterprises' next big leap will be in creating intelligent systems. Conversely, providers of powerful AI tools are making it easier than ever to get started using these systems and are providing numerous user guides and training materials that companies can use to scale up AI expertise within their teams. These resources help businesses to face these four core challenges in transitioning to a digital business. Instilling an intuitive understanding of the components of AI is important, but solving these problems is more important and also more challenging. In combination, these challenges indicate that the AI transition is so difficult and has a high potential for failure. After all, in the end, it doesn’t matter how intuitive an AI overview is if you can’t implement it or translate it into a success story.
4.1. AI in Healthcare
AI has the potential to revolutionize the practice of medicine and improve patient outcomes through the development and application of AI-powered algorithms and solutions across hospitals and health systems. The increasing volume of work required of healthcare staff to manually analyze and act on information has driven the need for more advanced technology such as AI. Such algorithms have been used in diagnostic imaging to increase the accuracy of identifying and diagnosing pathologies as well as to predict the onset of disease in asymptomatic patients. However, from the data perspective, AI-based platforms are usually trained using datasets that include a combination of genetic data, electronic medical records, and other complex and unstructured data sources. The use of pre-trained convolutional neural networks in various disease states has shown superior diagnostic performance in low-contrast lesions in CT imaging and reduced reader variability. Providing a neural network-based second reader has been used to display the results of both readers. The benefit of using such a network is its potential use as a fully independent reader without the need for a second reader.
One of the main appeals of AI in healthcare is its ability to identify pathology far earlier than conventional methods, especially in the developing oncology field. AI has also been used in chest CT scans, for instance, to exclude coronavirus with high accuracy. Several other potentially life-saving use cases are in development—some already implemented—such as personalized medicine in oncology. The ability to analyze images and other diagnostic data with AI significantly increases the speed and accuracy of reporting, improves patient outcomes, and enables collaboration within and across professional groups. From a hospital perspective, AI has been found to free up staff time spent on scanning, imaging, or reporting by providing optimal lists, queuing of work, and advice on prioritization. All of these lead to increased patient outcomes and, at the same time, reduce waiting lists, improving the safety of healthcare delivery in countries where diagnostic waiting lists are prioritized. There are also potential large savings and added value to be made from time saved in each stage of the patient pathway. Initially, in each hospital, the cost of installation, licensure, and upkeep of an AI solution is offset by the reduction in workload placed upon radiologists.
4.2. ML in Finance
4.2. Machine Learning in Finance
ML and DL: Applications in Finance Technological advances have brought about new patterns of digital data flow that require radical innovation in digital and financial services. Rapid development in supervised and unsupervised pattern recognition is increasing the potential for radical innovation in financial services. There are various applications in several financial services: robo-advice, algorithmic trading platforms, financial behavioral analysis, personal financial management, social trading, marketing services, mortgage lending, investment recommendation, personal spending categorization, investment portfolio choice, financial news recommendation system, customer service, customer relationship management, personal financial coaching, financial planning, crowdfunding platforms, credit and counterparty risk default prediction, etc. Algorithmic trading uses vast data trading and social media data to optimize strategies and understand market trends. Natural language understanding helps to improve language services for finance, helping financial analysts to understand textual sentiment and author identity to improve their stock picking.
Algorithmic trading using machine learning has been reviewed and is combined with the best heuristic algorithms and performs better than these algorithms alone. For example, various types of supervised and unsupervised machine learning algorithms in financial markets such as Back-Propagation Neural Networks, General Regression Neural Networks, Genetic Algorithms, and Support Vector Machines false discovery rates. Credit scoring has also been one of the important areas for financial engineering. Lending criteria and predictions are so important that big companies rely on machine learning models to predict debt group membership, completion time of loan agreements, and loan approval processes by credit rating agencies. Because it can address data scarcity, overfitting, learning most of the information about the sample at hand, prompt or feature selection, non-orthogonality and noise, black box modeling, and missing information. The application of ANN in personal loans was compared with the application of a microfinance institution. Data was collected from 2014 to 2019. The 11 input parameters used for ANN were gender, age, occupation, work experience, education level, family size, income, ethnic acceptance, types of business, number of loans taken, and purpose of loan. The objective function in ANN was to develop a credit scoring model to predict good and bad clients using the label method. The model used a supervised backpropagation training algorithm to minimize the mean square error. The problem with ANN is that when run with sufficient data and iterations, it can provide the output with minimal overfitting. The model analyzes the input and output pairs and minimizes possible output variance. Applying ANNs to categorize credit groups according to the data, class labels good and bad were chosen.
5. Ethical Considerations and Future Trends
These discussions are also critical as AI and ML are developed further. Already, we have seen concerns regarding the lack of explanation in how AI decision-making processes reach their conclusions, termed "algorithmic bias," and worries about how quantum advances might enable new avenues for mass surveillance and large-scale computation, opening the door to the potential misuse of AI. Ethical discussions are also beginning around AI and ML, whether they are used for low-level tasks, such as computer vision or fast decision-making processes, or because of policy translation, where they become open to privacy issues, ethics concerns, and fears about accountability and the lack of a human setting to make or enforce a decision.
Moving forward, regulatory considerations are becoming increasingly important as AI and ML applications become more embedded within our societies. In essence, discussions around AI and ML are of two varieties: the technical aspects and the ethical considerations. While the technical side offers ways to guide ethical innovation, we remain culpable regarding the global – or at least, societal – solution to the ethical thresholds we set. Ideally, discussions around the ethical considerations should help shape the technical theory and applications in the construction of innovative research. To make technological innovations comprehensible for larger swaths of stakeholders, it is vital to consider knowledge systems, task allocation, and ways to engage with the myriad social, ethical, and legal considerations that are central to their development. Technological development does not, in vacuous form, equate to social good. To bring us back to the discourse around AI and ML, the suggestion abounds that the future is one in conjunction with human-robot interaction or human-robot collaboration. It is important to ensure that this is an inclusive conversation. Moreover, we continue to note the importance of including scholars and stakeholders who have not traditionally engaged with technology, as a means to discern the distribution of wealth or access, to stratify participation in concentric communities, and ultimately bolster a trajectory of RRI.