KR LEVEL Flashcards
Perceiving, thinking, acting – what did we set
out to learn in this lecture?
Perceive
We didn‘t talk about this yet: Data and natural language as input
to artificial intelligence-enabled systems, or artificial intelligence
analytics methods
Think
Rules, logic, graphs: Different knowledge representations and
reasoning mechanism
We didn’t talk about this yet: machine learning.
Act
This comes up very indirectly throughout this course, in the sense
that system/algorithm output informs human users or is consumed
by other systems.
Knowledge Representation
Express and differentiate between:
▪ Facts (assertional axioms - logic, graphs)
▪ General knowledge (rules; terminological axioms – predicate
logic,graphs)
Reasoning
Logic, graph-based procedures: Infer new knowledge, check for
the correctness of specific statements, can find related entities
(spreading activation, shortest path), and characterize entities
with different graph measures
Learning
(no learning yet)
Problems with knowledge representations
from symbolic AI
Symbolic AI – relies on expressing knowledge in terms of human-readable, high-level
symbolic expressions and concepts – in our lecture: rules, logic, and semantic
networks
Suitability to real-world problems:
: It is difficult to describe complex
entities in logic-based, symbolic KR formalisms (graph representations
help to some degree)
Modelling effort:
Symbolic knowledge representations are typically
manually engineered (rules, object-oriented KR, knowledge graphs)
➢ Vocabulary and formal KR needs to be modelled, is difficult to
derive automatically (loosening requirements on formality helps)
➢ Suitable for metadata and common characteristics (attributes,
properties, …)
➢ Suitable if formal description is sufficiently important – e.g., formal
verification of important systems.
Reasoning:
New inferences, validation, consistency checking.
Scalability and decidability can be an issue.
Representation
Goal: Computationally
represent books, e.g., Harry
Potter and the Half-Blood
Prince.
Definition: A representation
Y conforms in a systematic
manner to X , preserving
pre-selected characteristics
of X.
A representation always
loses something, it is an
approximation.
Represent this book in predicate logic:
Book(Harry-potter-hbp).
▪ hasCoverColor(Harry-potter-hbp, green)
▪ isGenre(Harry-potter-hbp, fantasy-novel)
▪ isGenre(Harry-potter-hbp, young-adult-novel)
▪ …
▪ Fictional-Character(Harry)
▪ Fictional-Character(Hermione)
▪ Fictional-Character(Ron).
▪ appearsIn(Harry,Harry-potter-hbp)
▪ …
▪ is-friend-of(Harry,Hermione)
▪ is-friend-of(Harry,Ron)
▪ ..
How suitable is predicate logic as a
knowledge representation formalism
to represent books?
+ Suitable for expressing meta-data, and
describing characteristics of the book that
are similar across books – facts and
knowledge ABOUT the book
▪ Imagine expressing the complete story of
Harry Potter in predicate logic!
▪ Yahoo! initially categorised web sites in a
web directory – required editors
~ Metadata could be extracted from a digital
version of the book with some heuristics, if
reasonable metadata schema exists a
priori (-> typical research/engineering goal
in NLP „fact extraction“, „slot filling“,
„named entity recognition“)
Idea: Vectors as numeric, non-symbolic
representation of complex entities
From items to vectors, and how to reason
using vectors
1. Choose knowledge representation (KR) formalism:
Vectors
2. Choose how to represent items into vectors (~feature
engineering).
▪ Choose relevant characteristics of the item!
KR perspective: An entity has a set of characteristics; in
machine learning these correspond approximately to
„features“; in statistics the term „variables“ is more typical.
▪ Choose easily computable characteristics of the item!
▪ …which are representable as real numbers.
3. Choose operations on the represented items
▪ Vector mathematics has a broad range of operations –
identify what do they mean in a given use case, and which
are useful
Choosing operations on vectors for reasoning
- What questions do we want to ask?
For a given entity:
▪ What are similar entities?
▪ To which groups does an entity belong? (classification)
▪ What will we be able to observe about this entity in the future? (prediction)
Over a set of entities:
▪ What are meaningful sub-groups? (clustering)
▪ Is there a correlation between different entity characteristics?
▪ Does one characteristic cause another one? (attention – typically needs
specific study set-up to assert!)
▪ Is there a more compact representation – which variables carry most
information? (factor analysis)
Why is similarity interesting?
Recommendation: If you like book A, and book B is
similar to A, then you will probably also like book B.
▪ Information retrieval: If your question (query) is q,
then the answer to your question (a document) should
be similar to q.
▪ Classification: All members of a class are similar
w.r.t. to specific features
▪ Clustering: Build groups of entities that are more
similar to each other than they are to members of
other groups
Vector Representations in this lecture
In this lecture, we will
▪ Compute similarities (today)
▪ Application example: information retrieval and
how to „translate“ natural language into vectors
(next lecture)
▪ Classification and clustering