100% satisfaction guarantee Immediately available after payment Both online and in PDF No strings attached
logo-home
College notes Deep Learning part 2 (XM_0083), Master VU AI $4.74   Add to cart

Class notes

College notes Deep Learning part 2 (XM_0083), Master VU AI

 28 views  0 purchase
  • Course
  • Institution

Those are the notes I made during the whole course, it is not a summary! As long as you understand this document you do not need to watch any lecture. With those notes I have passed the course with a 7.5.

Preview 4 out of 68  pages

  • February 23, 2022
  • 68
  • 2021/2022
  • Class notes
  • Jakub tomczak
  • 9-14
avatar-seller
Lecture 9
Learning with Graphs

introduction - graphs

for given graph you should know whether it has:

undirected, simple graph

undirected, multigraph: more lines between two nodes

directed

self loop possible

node/edge labels are they unique

properties on edge:

edge weight

RDF (unique id/link for each node)

any comb is possible

introduction - embeddings

embedding = low dimensional representations of objects

low dimension = much lower as the original size

representation = there is some meaning to it, a representation corresponds to sth (objects in real world)

objects = words, sentences, images, audio, graphs,...




this output is the representation of the input object, and so it the embedding of it



some embedding spaces have an useful structure in which you can navigate in certain directions with
changing features according to direction = navigable space, so direction corresponds to certain semantic
features in object




Deep Learning 98

, embedding of images

embeddings of words

distributional hypothesis

"you shall know a word by the company it keeps"

"if units of text have similar vectors in a text freq matrix, then they tend to have similar meaning"

⇒ we can use context information to define embeddings for words,
one vector representation for each word

generally, we can train them (see lecture on word2vec)

illustration idea




why use graphs as input to ML?

classification, regression, clustering of nodes/edges/whole graphs

recommender systems

document modeling: entity and document similarity (use concepts from a graph)

alignment of graphs (which nodes are similar)

link prediction and error detection

linking text and semi-structured knowledge to graphs

graph embedding techniques

one big challenge with embedding graphs

we can not straightforwardly put graph into neural network because of its structure, not a linear
structure

there are traditional ML on graphs

often having problems with scalability

often need manual feature engineering: task specific

embedding knowledge graphs in vector space

for each node in graph → create vector in vector space → easily feed into ML algo

preserve info

unsupervised: task and dataset independent

efficient computation

low dimensional representation




Deep Learning 99

, 2 major visions on how this should be done

Preserve topology

keeping neighbors close in embedding space: e.g. europe-germany

Preserve similarity

keeping similar nodes close together: e.g. europe-africa

2 major targets

improve original data

knowledge graph completion

link prediction

anomaly detection

downstream ML tasks

classification/regression/clustering/K-NN

then used as part of a larger process

QA/dialog systems

Translation

image segmentation

how to go from graph to embedding? 3 major approaches for propositionalization

translation: any relation in knowledge graph is directly mapped to what should happen in the
embedded space. So we want to see the same relations in the embedded space

has a lot of approaches, but all come down to same thing

TransE is the basic one, and all the other ones are the complex version of it

= translation embedding

take every edge of graph → if this exists in knowledge graph → we want to have it in our
embedded space with edge embedding, vector as translation form that one to that one




so adding embedding to head has to give tail




we want to minimalize distance between sum of the h and l (as close as possible) to t in
knowledge graph S




Deep Learning 100

, this is not completely sufficient, seems to work but the problem is that we only get positive
information ⇒ over optimized

so not only minimizing but also penalizing when wrong relations are given: (maximize bad
triple)




become less scalable because they do more and more stuff
one hop away from each other = current entity → one hop → next entity

tensor/matrix factorization

make a 3d matrix of all relations in the graph and factorize it [factorizing = make a lower
dimensional representation of the 3d matrix also with lower matrix dimensionality]




Deep Learning 101

The benefits of buying summaries with Stuvia:

Guaranteed quality through customer reviews

Guaranteed quality through customer reviews

Stuvia customers have reviewed more than 700,000 summaries. This how you know that you are buying the best documents.

Quick and easy check-out

Quick and easy check-out

You can quickly pay through credit card or Stuvia-credit for the summaries. There is no membership needed.

Focus on what matters

Focus on what matters

Your fellow students write the study notes themselves, which is why the documents are always reliable and up-to-date. This ensures you quickly get to the core!

Frequently asked questions

What do I get when I buy this document?

You get a PDF, available immediately after your purchase. The purchased document is accessible anytime, anywhere and indefinitely through your profile.

Satisfaction guarantee: how does it work?

Our satisfaction guarantee ensures that you always find a study document that suits you well. You fill out a form, and our customer service team takes care of the rest.

Who am I buying these notes from?

Stuvia is a marketplace, so you are not buying this document from us, but from seller MeldaMalkoc. Stuvia facilitates payment to the seller.

Will I be stuck with a subscription?

No, you only buy these notes for $4.74. You're not tied to anything after your purchase.

Can Stuvia be trusted?

4.6 stars on Google & Trustpilot (+1000 reviews)

66579 documents were sold in the last 30 days

Founded in 2010, the go-to place to buy study notes for 14 years now

Start selling
$4.74
  • (0)
  Add to cart