first half
- What we’re doing now.
-
- Step 5 is done, 8 sheets of paper are created, and the encoding of the first two sheets of it has just been finished.
- 27 sticky notes and 42 slides.
Relating to Step 6. - path recommendation
Step 6 focuses on one sheet at a time
- This gives them a narrower perspective than when they were thinking about the overall composition.
- Allows you to think more deeply instead of narrowly focused
- There is a tradeoff between breadth and depth because of the limited capacity of the human brain.
Associate the next slide with the most recent slide as input
- Think about how to make this happen.
- If we map “slides” to “words”, this would be a sentence generation by Markov model, which generates a sentence (word sequence) by selecting the next word from the previous N wordsMarkov Sentence Generation
- In a naive Markov sentence generator, the probability of generating the next word is given by the frequency of occurrence in the original data.
- As N grows, most of the word sequences have zero number of observations.
- Small N only preserves relations between close words.
- Sequence transformation model by LSTM followed up on this point.
- dimensionality reduction caution
- Initially, we thought it would be a random Dropout.
- Is it appropriate to think of this type as creating a “vector to ignore” based on t and t+1?
- With this mechanism, you can only mask against key.
- I can multiply query and key by a rotation such that the (t - (t-1)) vector is (1, 0, 0, … , 0), and then multiply by (0, 1, 1, … , 0). If we apply a rotation to query and key such that the (1, 0, 0, … , 0) vector is (0, 1, 1, … , 1), and then apply a mask of (0, 1, 1, … , 1), we can perform a 1D reduction in a specific direction. 1), then we can reduce one dimension in a specific direction.
- Is one dimension all you need?
- This system, which only allows axial reductions, is better because it is easier to create meaning on the axis.
- I had a feeling that a dimensionality reduction caution on the back end of LSTM would be a good idea.
One sticky corresponds to multiple slides in the past
- A human can’t pull all the past slides from memory and make them sticky notes.
- Inevitably, a sticky note is created with only keyword X written on it.
- This is the same state as implicitly grouping “past slides” with title X
- One slide can contain multiple stickies, and one sticky can point to multiple slides
This page is auto-translated from /nishio/講演スライド作成プロセス using DeepL. If you looks something interesting but the auto-translated English is not good enough to understand it, feel free to let me know at @nishio_en. I’m very happy to spread my thought to non-Japanese readers.