1. Difference between Weak AI and Strong AI
Weak AI, also known as Narrow AI, is designed to perform a specific task or solve a particular problem without any understanding or consciousness. Examples include virtual assistants like Siri, voice recognition systems, and spam filters. These systems cannot operate beyond their programmed capabilities.
Strong AI, on the other hand, refers to systems that possess human-like cognitive abilities, including self-awareness, understanding, learning, and reasoning across a wide range of tasks. Strong AI is still theoretical and has not yet been achieved. The goal of Strong AI is to replicate the full range of human intelligence.
2. Intelligent Agents and Their Components
An intelligent agent is any entity that perceives its environment through sensors and acts upon that environment through actuators to achieve specific goals. It uses an agent function that maps perceptual inputs to actions.
The main components of an intelligent agent are:
- Sensors: Used to perceive the environment (e.g., cameras, microphones).
- Actuators: Mechanisms for interacting with the environment (e.g., wheels, arms, speakers).
- Agent Function: The logic or algorithm that maps from percepts to actions.
- Performance Measure: A criterion that evaluates the agent's success in achieving its goal.
3. Difference Between Informed and Uninformed Search
Uninformed search strategies, also known as blind search strategies, do not have any domain-specific knowledge. They explore the search space without guidance, using methods like Breadth-First Search (BFS) and Depth-First Search (DFS).
In contrast, informed search strategies use problem-specific knowledge or heuristics to guide the search towards the goal more efficiently. Examples include A* Search and Greedy Best-First Search. These algorithms estimate the cost to reach the goal and use this estimate to make smarter decisions.
4. First-Order Logic (FOL) and Its Difference from Propositional Logic
First-Order Logic is a formal system used to represent knowledge using quantifiers, variables, predicates, and logical connectives. It allows reasoning about objects and their relationships.
Unlike propositional logic, which treats entire statements as indivisible units and assigns them true or false values, FOL allows for more expressive representations. For example, FOL can represent “All humans are mortal” using quantifiers, which is not possible in propositional logic.
5. Constraint Satisfaction Problem (CSP) Explained with an Example
A Constraint Satisfaction Problem is defined by three components:
- A set of variables.
- A domain of values for each variable.
- A set of constraints that restrict the values the variables can simultaneously take.
An example of a CSP is the Sudoku puzzle. Each cell is a variable, and the domain is the numbers 1–9. The constraints ensure that no number repeats in any row, column, or 3x3 grid. Solving a CSP involves assigning values to variables in a way that satisfies all the constraints.
6. Probabilistic Reasoning with Example
Probabilistic reasoning is a method used to deal with uncertainty in knowledge representation and decision-making. It uses probability theory to predict the likelihood of different outcomes.
For example, in medical diagnosis, probabilistic reasoning can help determine the probability that a patient has a certain disease given observed symptoms. Bayesian Networks are commonly used tools in probabilistic reasoning to compute conditional probabilities.
7. Role of Reinforcement Learning in AI
Reinforcement Learning (RL) is a type of machine learning where an agent learns to make decisions by performing actions and receiving rewards or penalties. Unlike supervised learning, RL does not require labeled input/output pairs. Instead, the agent learns through trial and error.
Over time, the agent develops a policy that maximizes cumulative reward. RL is used in robotics, game playing (e.g., AlphaGo), and autonomous driving systems. Key concepts include reward functions, value functions, and exploration vs. exploitation.
8. Key Issues in Safe AI Development
Safe AI focuses on ensuring that artificial intelligence systems are reliable, ethical, and aligned with human values. Major issues in safe AI include:
- Bias and fairness: Ensuring AI does not reinforce societal or data biases.
- Robustness: Protecting AI from adversarial inputs or unexpected failures.
- Transparency and explainability: Making AI decisions understandable to humans.
- Control and alignment: Ensuring AI systems do what we intend without unintended consequences.
- Privacy and security: Safeguarding data and ensuring ethical use of information.
9. Importance of Knowledge Representation in AI
Knowledge Representation (KR) is the process of structuring and storing information in a way that a machine can understand, reason about, and use to make decisions. It acts as the bridge between raw data and intelligent action.
KR is crucial because it allows AI systems to simulate human understanding and reasoning. Common KR methods include semantic networks, frames, ontologies, and logic-based systems. Proper representation allows for efficient inference and learning.
10. Significance and Applications of Expert Systems
Expert systems are computer programs that simulate the decision-making ability of a human expert in a specific domain. They consist of a knowledge base (containing domain-specific facts and rules) and an inference engine (that applies logical rules to the knowledge base to deduce new information or make decisions).
These systems are used in areas such as medical diagnosis (e.g., MYCIN), financial analysis, customer support, and fault diagnosis in machines. Expert systems help in areas where human expertise is scarce or needs to be automated for efficiency and consistency.
1. Components and Architecture of an Expert System
An expert system is a computer program that simulates the decision-making ability of a human expert in a particular domain. Its core components include:
- Knowledge Base: Stores facts and heuristics (rules) gathered from human experts. This is the heart of the expert system and is built through knowledge engineering.
- Inference Engine: Applies logical rules to the knowledge base to infer new facts or reach conclusions. It uses techniques such as forward chaining (data-driven) and backward chaining (goal-driven).
- User Interface: Allows the user to interact with the system. It helps input queries and view explanations or recommendations.
- Explanation Facility: Justifies the reasoning or decision made by the system, which builds user trust.
- Knowledge Acquisition Module: Assists in updating and refining the knowledge base with new information.
Expert systems are used in areas such as medical diagnosis, financial forecasting, and equipment troubleshooting.
2. The Turing Test and Its Significance in AI
The Turing Test, proposed by Alan Turing in 1950, is a benchmark for measuring machine intelligence. It involves a human evaluator who interacts with both a machine and another human via text communication. If the evaluator cannot reliably tell which is which, the machine is said to exhibit intelligent behavior.
This test is significant because it shifts the focus from how machines work internally to how well they mimic human-like responses. Although it has limitations (e.g., it measures only linguistic ability, not reasoning or emotional intelligence), it remains a foundational concept in AI and continues to influence AI development and philosophy.
3. Types of Environments in Artificial Intelligence
An AI agent operates in various types of environments, which impact its design and functioning:
- Fully Observable vs. Partially Observable: In fully observable environments, the agent has access to complete information (e.g., chess). In partially observable environments, information may be hidden or uncertain (e.g., poker).
- Deterministic vs. Stochastic: Deterministic environments have predictable outcomes, while stochastic ones involve randomness or uncertainty.
- Episodic vs. Sequential: In episodic environments, decisions are made per episode independently (e.g., image classification). Sequential environments require past decisions to influence future outcomes (e.g., driving).
- Static vs. Dynamic: Static environments do not change while the agent is deliberating, unlike dynamic ones.
- Discrete vs. Continuous: Based on the state and time space—chess is discrete, while robot movement is continuous.
Understanding the environment type helps design the right agent and strategy.
4. Steps in Building a Machine Learning Model
Building a machine learning model typically involves these sequential steps:
- Problem Definition: Understand the business or technical problem you want the AI to solve.
- Data Collection: Gather sufficient data that is relevant to the problem.
- Data Preprocessing: Clean the data (handle missing values, normalize features, etc.).
- Feature Selection/Engineering: Choose or create meaningful input variables.
- Model Selection: Choose the type of ML algorithm (e.g., decision trees, SVM, neural networks).
- Training: Feed the model training data so it learns patterns.
- Evaluation: Test the model using validation or test data to measure accuracy.
- Tuning: Adjust hyperparameters to improve performance.
- Deployment: Integrate the model into a production environment.
- Monitoring and Maintenance: Continuously evaluate and improve the model with new data.
5. Role of Natural Language Processing (NLP) in AI
Natural Language Processing (NLP) is a subfield of AI focused on enabling machines to understand, interpret, and generate human language. It combines linguistics, computer science, and machine learning.
NLP applications include:
- Machine Translation (e.g., Google Translate)
- Chatbots and Virtual Assistants
- Text Summarization and Sentiment Analysis
- Speech Recognition and Voice Interfaces
NLP involves tasks like tokenization, part-of-speech tagging, parsing, named entity recognition, and semantic analysis. It is vital for making AI systems communicative and accessible to humans.
6. Applications of AI in Real Life
AI has broad applications across industries:
- Healthcare: Diagnosis support systems, drug discovery, and personalized medicine.
- Finance: Fraud detection, algorithmic trading, and customer service chatbots.
- Transportation: Self-driving cars, traffic management systems.
- Agriculture: Crop monitoring, yield prediction using drones and sensors.
- Manufacturing: Predictive maintenance, quality control via computer vision.
- Retail: Customer recommendations, inventory forecasting.
- Entertainment: Personalized streaming suggestions, AI-generated content.
These examples show how AI is transforming sectors by automating complex tasks and making decision-making faster and more accurate.
7. Difference Between Rule-Based and Learning-Based Systems
Rule-Based Systems work using a fixed set of pre-programmed rules. These rules are manually defined by experts and do not adapt to new data unless changed explicitly. They are used in systems like traditional expert systems and are easy to understand and debug.
Learning-Based Systems, on the other hand, learn from data automatically. Machine learning models adjust their parameters based on input-output examples and can improve over time without manual intervention.
Rule-based systems are static and deterministic, whereas learning systems are dynamic and probabilistic. Learning-based systems are more suited for environments with high variability and data abundance.
8. Working and Layers of a Neural Network
A neural network is a machine learning model inspired by the human brain’s structure. It is composed of interconnected nodes (neurons) arranged in layers:
- Input Layer: Receives raw data.
- Hidden Layers: Perform transformations and extract features. These layers apply activation functions like ReLU or sigmoid to introduce non-linearity.
- Output Layer: Produces the final prediction or classification result.
Each connection between neurons has an associated weight that is adjusted during training using algorithms like backpropagation and gradient descent. Deeper networks (deep learning) with many hidden layers are used in tasks like image recognition and language modeling.
9. What is Heuristic Function? Importance in AI Search
A heuristic function is a method of estimating how close a current state is to the goal in a search problem. It provides domain-specific knowledge that can significantly reduce the search space and improve the efficiency of algorithms like A* and Greedy Best-First Search.
For example, in a pathfinding problem on a grid, a common heuristic is the Manhattan distance (sum of absolute differences in x and y coordinates). A good heuristic makes the search faster without compromising accuracy. Poor heuristics can mislead the algorithm and increase computational cost.
10. Importance of Evaluation Metrics in AI Models
After training an AI model, it must be evaluated to determine how well it performs. Different tasks require different evaluation metrics:
- Accuracy: Percentage of correctly predicted instances.
- Precision and Recall: Especially important in imbalanced datasets.
- F1 Score: Harmonic mean of precision and recall.
- Confusion Matrix: Provides insight into types of errors made.
- ROC-AUC: Used in classification to measure trade-off between true positive rate and false positive rate.
Choosing the right evaluation metric ensures the model is not only accurate but also suitable for the specific problem context.
Explain categories of general expert systems.
Expert systems are AI programs that simulate the decision-making abilities of human experts. They are designed to solve complex problems by reasoning through bodies of knowledge, expressed mainly as “if-then” rules.
The general categories of expert systems include:
1. Rule-Based Expert Systems
These systems use a set of predefined logical rules (if-then rules) and a knowledge base. The inference engine applies these rules to the known facts to derive new facts. For example, MYCIN (for medical diagnosis) is a classic rule-based system.
2. Frame-Based Expert Systems
These use frames (data structures) for representing stereotypical situations. Each frame represents an object or a concept and contains slots (attributes) and their values. It’s particularly effective for handling hierarchical and structured knowledge.
3. Fuzzy Expert Systems
These systems handle uncertainty and imprecision by using fuzzy logic instead of traditional Boolean logic. For example, rather than saying “temperature > 100”, it may say “temperature is high”. These are useful in control systems and real-world uncertain environments.
4. Neural Network-Based Expert Systems
Instead of using explicit rules, these systems learn patterns from data using artificial neural networks. They are highly effective in pattern recognition tasks like image or voice recognition. However, they often lack explainability.
5. Case-Based Expert Systems
These systems solve new problems by comparing them to previously solved cases stored in a case library. When a new problem arises, the system retrieves a similar case and adapts its solution to the current context. CBR (Case-Based Reasoning) is commonly used in law, customer support, and diagnostics.
6. Hybrid Expert Systems
These systems combine multiple approaches (e.g., rule-based with neural networks or fuzzy logic) to leverage the strengths of each method. They are more flexible and efficient in handling complex, real-world problems.
Compare contrast advantages limitations of Bi-Directional Search compared to Uni-Directional Search. Under what conditions is Bi-Directional Search particularly advantageous?
Uni-Directional (Traditional) Search
Uni-directional search algorithms like Breadth-First Search (BFS), Depth-First Search (DFS), and A* search start at the initial node and explore the search space until the goal node is found. They operate in a single direction and usually explore a large number of nodes, especially in vast or unstructured spaces.
Advantages:
- Simple to implement.
- Works well in small to medium search spaces.
- Algorithms like A* are optimal and complete if heuristics are admissible.
Limitations:
- Can be inefficient in large search spaces.
- May explore unnecessary nodes far from the goal.
Bi-Directional Search
Bi-directional search runs two simultaneous searches: one forward from the initial node and the other backward from the goal node. The searches meet in the middle, significantly reducing the number of nodes explored.
Advantages:
- Much faster than uni-directional search in many cases.
- The effective branching factor is reduced, leading to faster results.
- Search space is approximately the square root of the original space (O(b^(d/2)) instead of O(b^d)).
Limitations:
- Backward search can be difficult if the goal state or path to it is not easily reversible.
- Requires extra memory and careful implementation to manage two frontiers and detect intersection.
- Not always applicable when the goal state is not explicitly defined or is dynamic.
When is Bi-Directional Search Advantageous?
Bi-directional search is particularly beneficial when:
- The initial and goal states are clearly defined.
- The cost of traversing from goal to start is equal or similar to start to goal.
- The search space is large, and performance/memory efficiency is critical.
- The problem space is symmetrical and the operations are reversible.
A good example is pathfinding on a map, where both the start and end locations are known, and the map allows movement in both directions.
Q.No.8: Write short notes on the following:
a) Intelligent Agent:
An intelligent agent perceives its environment through sensors and acts upon it using actuators. It uses AI techniques to choose actions that maximize goal achievement.
b) Natural Language Processing (NLP):
NLP enables machines to process human language for tasks like translation, summarization, chatbot interactions, and speech recognition.
c) Robots and Its Applications:
Robots are programmable machines that can carry out tasks autonomously. Applications include manufacturing, space exploration, military drones, healthcare (surgery robots), and domestic assistants.
d) Knowledge Engineering:
Knowledge engineering involves acquiring expert knowledge and encoding it into a system’s knowledge base. It's essential for building expert systems that simulate human reasoning.
a) Planning Agent:
A planning agent not only selects actions but also sequences them over time, considering preconditions and effects. It constructs a plan before acting.
Similarity with Problem-Solving Agent:
Both agents use search and reasoning. However, a planning agent is more complex—it handles multiple steps, constraints, and future consequences.
b) Handling Incomplete/Incorrect Information:
- Probabilistic Reasoning: Using probability to estimate outcomes.
- Default Assumptions: Acting with best guesses.
- Sensor Feedback: Continuously updating knowledge.
- Belief State Maintenance: Tracking multiple possible states.
- Fuzzy Logic: Handling vague inputs or classifications.
What is Machine Learning? Discuss components of a Machine Learning System.
Machine Learning (ML) is a subfield of artificial intelligence that focuses on the development of algorithms that enable computers to learn from and make predictions or decisions based on data. Rather than being explicitly programmed to perform every task, machine learning systems learn patterns and relationships from examples and improve their performance over time.
In a typical machine learning system, the core components include:
- Data: Data is the foundational component. A machine learning system learns from historical or real-time data. The quality and quantity of this data directly impact the system's performance.
- Features: These are measurable properties or characteristics extracted from raw data. Feature engineering involves selecting and transforming data into useful inputs for learning.
- Learning Algorithm: This is the mathematical or statistical method used to discover patterns in the data. Examples include decision trees, support vector machines, k-nearest neighbors, or neural networks.
- Model: The output of the learning process is a model—a representation of the patterns learned from the data. Once trained, this model can be used to make predictions on new data.
- Training and Testing Phases: During training, the algorithm adjusts its parameters to reduce errors. Once trained, the model is tested on unseen data to evaluate its accuracy and generalization.
- Evaluation Metrics: Accuracy, precision, recall, F1-score, and mean squared error are used to measure how well the model performs.
- Feedback Mechanism: In some systems (especially online learning), the model continues to update and improve as it receives new data.
Machine learning plays a crucial role in applications like recommendation systems, fraud detection, predictive analytics, image recognition, and autonomous vehicles.
Define Speech Recognition. What are the Applications of Speech Recognition?
Speech Recognition refers to the capability of a machine or program to identify and process human speech into a written format. It involves converting spoken language into text using computational linguistics and machine learning algorithms.
The process of speech recognition includes several stages such as acoustic modeling (to recognize phonemes), language modeling (to interpret word sequences), and signal processing (to handle background noise and speaker variation). Modern systems use deep learning, particularly recurrent neural networks (RNNs) or transformer-based models, to achieve high accuracy.
Applications of speech recognition are wide-ranging and growing rapidly in both consumer and enterprise domains. Popular applications include:
- Virtual Assistants like Google Assistant, Alexa, and Siri that accept voice commands and perform tasks such as setting reminders, searching the internet, or controlling smart home devices.
- Voice Typing and Dictation tools that help users write texts and emails hands-free.
- Customer Support Automation, where speech-to-text technology is used in call centers to transcribe customer conversations and generate responses.
- Medical Transcription, where doctors dictate patient reports that are automatically transcribed.
- Voice Authentication, where user identity is verified through unique voice patterns.
Speech recognition is particularly valuable for accessibility, providing input methods for users with physical disabilities or literacy limitations.
Discuss the Differences Between Expert Systems and Conventional Computer Programs.
An Expert System is a computer program designed to simulate the reasoning and decision-making capabilities of a human expert. It uses a knowledge base of facts and rules and an inference engine to solve problems in a specific domain, such as medical diagnosis or legal advice.
In contrast, conventional computer programs are built using traditional programming techniques where every possible rule and outcome must be explicitly defined by the developer using procedural or object-oriented code.
The key differences lie in design philosophy, flexibility, and purpose. Expert systems are knowledge-driven and can be updated by modifying their knowledge base, without changing the underlying program logic. They operate on logical inference, often using "if-then" rules and backward/forward chaining mechanisms to deduce conclusions.
Conventional programs, on the other hand, are algorithm-driven and typically operate with fixed logic paths. They are less flexible when it comes to adapting to new rules or knowledge unless reprogrammed.
Another difference is that expert systems often include explanation facilities, which justify their conclusions by tracing the reasoning process—something conventional systems typically lack.
In summary, expert systems aim to mimic expert human reasoning in specialized domains, while conventional programs solve well-defined problems using predefined logic and rigid instructions.
Differentiate Between Single-State and Multiple-State Problems.
In artificial intelligence, single-state and multiple-state problems refer to the agent’s understanding of the world at any given time during problem-solving.
A single-state problem assumes that the agent has complete and accurate knowledge of the environment. At every step, it knows exactly which state it is in and what the outcome of its actions will be. This type of problem is simpler to solve because the search space is clearly defined and deterministic. Classical examples include board games like chess or puzzles like the 8-puzzle, where the current configuration is always fully known.
A multiple-state problem, in contrast, involves uncertainty or incomplete information about the current state. The agent cannot be sure of its exact position or the effects of its actions. This might be due to noisy sensors, unpredictable environments, or limited visibility. Such problems require the agent to maintain a belief state—representing all possible actual states it could be in—and update it as new information becomes available.
Real-world applications like navigating in unknown terrain, autonomous driving in foggy conditions, or robotic exploration on Mars often fall under multiple-state problem scenarios.
The complexity and design of the agent’s reasoning system must account for this uncertainty, often using probabilistic or fuzzy logic methods to make decisions.
What is an Agent? Explain the Structure of Intelligent Agents.
An agent in artificial intelligence is any entity that can perceive its environment through sensors and act upon that environment using actuators. Agents can range from simple software bots to complex physical robots or autonomous vehicles.
An intelligent agent is one that can make rational decisions based on its perceptions and built-in goals. It selects the most appropriate action to maximize its performance over time, even in complex or uncertain environments.
The structure of an intelligent agent typically includes:
- Sensors: These detect changes in the environment. For example, a robot may use cameras and infrared sensors, while a software agent may gather inputs from APIs or databases.
- Actuators: These are the components that carry out actions. For robots, this could mean motors or mechanical limbs; for software agents, it could be sending messages or updating files.
- Performance Measure: This defines what counts as success for the agent. It guides decision-making and may include efficiency, speed, accuracy, or safety.
- Agent Function: This is the core logic that maps percept histories to actions. It can be implemented using rules, machine learning models, or planning algorithms.
Based on this structure, intelligent agents can be further classified as simple reflex agents, model-based agents, goal-based agents, utility-based agents, or learning agents, depending on their internal complexity and decision strategies.
Artificial Intelligence
Artificial Intelligence (AI) is the simulation of human intelligence in machines.
It involves creating agents that can reason, learn, and act autonomously.
AI ranges from simple automation to complex systems with decision-making abilities.
AI Paradigms and Hypothesis
AI paradigms include symbolic (logic-based), sub-symbolic (neural networks), and evolutionary methods.
Hypotheses in AI are assumptions made to simplify reasoning and learning.
These paradigms guide how AI systems are designed and function.
Intelligent Agents
An intelligent agent perceives its environment and takes actions to achieve goals.
It includes sensors, actuators, and decision logic (agent function).
The goal is to act rationally to maximize performance.
Difference Between Cybernetic Intelligence and Artificial Intelligence
Cybernetic intelligence focuses on control systems and feedback in machines or organisms.
Artificial Intelligence emphasizes reasoning, learning, and decision-making in software.
Cybernetics is more biological/mechanical, while AI is cognitive and algorithmic.
Objectives and Scope of Weak AI and Strong AI
Weak AI aims at specific tasks and lacks consciousness.
Strong AI seeks to replicate general human intelligence and awareness.
The scope of AI includes automation, decision-making, robotics, and more.
Solving Problems by Searching
Search is a fundamental technique for navigating problem spaces.
It involves finding a sequence of actions that leads to the goal.
Algorithms include Breadth-First, Depth-First, and A*.
Informed Search and Exploration
Informed search uses heuristics to make search efficient.
It estimates the best path to the goal using domain knowledge.
Examples include Greedy Best-First Search and A* Search.
Constraint Satisfaction Problems
CSPs require assigning values to variables under constraints.
Solutions must satisfy all constraints without conflicts.
Examples include puzzles, scheduling, and map coloring.
Adversarial Search
Used in games and competitive scenarios with two or more agents.
Involves maximizing one’s gain while minimizing the opponent’s.
Example: Minimax algorithm in chess.
Logical Agents
Logical agents use logic to represent and reason about knowledge.
They apply inference to determine actions.
Often based on propositional or first-order logic.
First-Order Logic
FOL extends propositional logic with objects, functions, and quantifiers.
It allows detailed representation of relationships and rules.
Widely used in knowledge representation and AI reasoning.
Inference in First-Order Logic
Inference is the process of deriving new knowledge from known facts.
In FOL, it involves rules like Modus Ponens and Unification.
Automated inference enables intelligent decision-making.
Knowledge Representation
It's the method of encoding information for AI to reason and act.
Forms include semantic networks, frames, and logic.
Effective representation is key for intelligent behavior.
Planning and Acting in the Real World
AI systems must plan sequences of actions and adapt to real environments.
Planning involves goal setting and strategy development.
Action includes executing plans and handling uncertainty or change.
Uncertainty
AI often operates with incomplete or ambiguous data.
Uncertainty is handled using probabilities and fuzzy logic.
This enables robust decision-making under real-world conditions.
Probabilistic Reasoning
This approach uses statistical methods to handle uncertainty.
It calculates the likelihood of events or outcomes.
Common models include Bayesian Networks and Markov models.
Probabilistic Reasoning Over Time
Extends probabilistic models to dynamic, time-based scenarios.
Uses tools like Hidden Markov Models (HMMs) and Dynamic Bayesian Networks.
Useful in speech recognition, tracking, and temporal predictions.
Making Simple Decisions
Involves choosing the best action based on known outcomes.
Often uses decision trees or utility-based models.
Suitable for structured, low-uncertainty problems.
Making Complex Decisions
Deals with multiple variables, outcomes, and uncertainties.
Requires advanced models like influence diagrams and game theory.
Common in strategic planning and AI planning systems.
Learning from Observations
Machines learn patterns from data without explicit programming.
Observation-based learning forms the basis of machine learning.
Examples include supervised, unsupervised, and reinforcement learning.
Knowledge in Learning
AI systems acquire knowledge from experience or training data.
Knowledge helps improve predictions, decisions, and adaptability.
It can be represented symbolically or statistically.
Learning Methods
Includes supervised, unsupervised, semi-supervised, and reinforcement learning.
Each method fits different data types and learning goals.
They form the foundation of modern AI systems.
Reinforcement Learning
An agent learns through trial-and-error interactions with an environment.
It receives rewards or penalties based on actions taken.
Used in robotics, game AI, and autonomous systems.
Communicating, Perceiving, and Acting
AI must interpret inputs (perception), make decisions, and act accordingly.
Communication includes natural language, signals, or gestures.
These are key for interactive and robotic systems.
Probabilistic Language Processing
Uses statistical models to understand and generate language.
Includes NLP tasks like translation, sentiment analysis, and speech recognition.
Relies on models like HMMs and neural networks.
Perception and Robotics
Perception involves interpreting sensory data (vision, audio, etc.).
Robotics combines perception with action in physical environments.
AI in robotics enables autonomous navigation and interaction.
Introduction to LISP/PROLOG
LISP and PROLOG are AI-focused programming languages.
LISP is symbolic and functional, while PROLOG is logic-based.
Both are used for knowledge representation and AI research.
Expert Systems (ES) and Applications
Expert systems mimic human expert decision-making using rules.
They include a knowledge base and inference engine.
Used in medicine, finance, diagnostics, etc.
Artificial General Intelligence (AGI)
AGI refers to machines with human-level reasoning and understanding.
It contrasts with narrow AI, which is task-specific.
AGI is still theoretical and a major goal of AI research.
Issues in Safe AI
Concerns include bias, control, ethics, and unintended outcomes.
Safe AI ensures reliable, aligned, and harm-free behavior.
Involves transparency, robustness, and human oversight.
Introduction to Cognitive and Conscious Systems
Focuses on systems that mimic human thought and awareness.
Includes models of perception, memory, and consciousness.
Part of advanced AI exploring human-like understanding.