AI Creativity Flashcards
AI Creativity Case Study
What are the three types of art described in the taxonomy of computer-generated art?
C-art, G-art, and CG-art
What is C-art?
Art that uses computers as part of the art-making process
What is G-art?
Art that is generated, at least in part, by some process not under the artist’s direct control
What is CG-art?
Art produced by leaving a computer program to run by itself, with minimal or zero human interference
What are passive tools in the context of generative systems?
Tools that make no attempt to alter the user’s input, e.g., Microsoft Paint
What are active tools in the context of generative systems?
Tools that actively interpret and process user inputs, adding things to it, e.g., the sketch pad mentioned in the lecture
What is a first-order Markov model in text generation?
A model where the next state depends only on the current state (word)
What is a second-order Markov model in text generation?
A model where the next state depends on the two previous states (words)
What are the three parts of the example AI system for generating pop music?
- Lyric generator using GPT-2 transformer model, 2. Music generator using Music-VAE auto-encoder model, 3. Singing voice synthesis using the DiffSinger model
What is the basic process for generating text using a Markov model?
- Pick a random initial state, 2. Select from possible next states, 3. If no possible next state, go back to step 1
Name three examples of more complex generative text models.
Variable order Markov model, Long short-term memory network (LSTM), Transformer network
What are some features to consider when describing a generative system?
System architecture, Number of agents, Roles, Environment, Corpus, Input, Output, Communication, Human interactive modality, Task, Evaluation
How does a second-order Markov model differ from a first-order model?
It considers the two previous states (words) instead of just the current state
What is the main difference between more complex generative models and simple Markov models?
They have a more complex method for picking the next state/output, including more complex state representations
In the context of generative systems, what is a corpus?
The collection of data (e.g., text, images, music) that the system uses to learn and generate new content
What is the main advantage that transformers add to language models?
They add context awareness to embeddings, allowing for a combination of contextual and sequential data
How does self-attention work in transformer networks?
It creates a new type of embedding that incorporates information about other words in the context, not just the current word
What are the two main ways of encoding input in recurrent neural networks?
- One-hot encoded vectors, 2. Embeddings
What is the ‘bag of words’ approach good for, and what is its limitation?
It’s good for sentiment analysis, but not great for generating text as it ignores sequence
When was the transformer architecture first reported?
2017
What are the two main components that transformers model?
Sequence and context via self-attention
How many parameters does GPT-2 have?
1.5 billion parameters
How many layers does GPT-2 have?
48 layers
What is meant by ‘zero-shot’ concept in relation to GPT-2?
GPT-2 can perform tasks it wasn’t specifically trained for, outperforming some specialized models
What is Huggingface?
A platform on a mission to democratize good machine learning, providing tools and models for NLP
Why was GPT-2 initially considered ‘too dangerous to release’?
Due to concerns about potential malicious applications of the technology
What is the ‘auto-regressive mode’ of GPT-2?
It can generate an endless stream of words based on previous output
How much text data was GPT-2 trained on?
40G of text
What are ‘attention heads’ in the context of transformers?
Multiple projections of attention, allowing the model to focus on different aspects of the input simultaneously