What is the importance of the log2(k) for calculating sigma? #920
Unanswered
CRISTIANJULIOCESAR
asked this question in
Q&A
Replies: 1 comment 1 reply
-
I am sure I also asked @lmcinnes this several years ago, but I searched my emails and can't find an answer. However, my recollection was that the answer was that it is empirical (although bear in mind I might simply have imagined this). There is a connection the entropic affinity calculation in t-SNE, where the Shannon entropy of the input probabilities is calibrated to reproduce the log of the input perplexity. I suspect it is in analogy to that procedure. The important thing is to do some kind of normalization step to get the different density around each item in the dataset on a similar scale. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
To find the sigma using log2(k), in this video (1:11:00), it is mentioned that log2(k) has beneficial properties but i didn't find that information in the original article. This is the video: https://www.youtube.com/watch?v=G9s3cE8TNZo&t=4514s&ab_channel=MachineLearningDojowithTimScarfe
I fully understand that a binary search is used to satisfy an equation to find sigma, but I do not understand the reason for log2(k), was it empirical as she say in the video due to experiments carried out by the original authors??
Thank you in advance :)
Beta Was this translation helpful? Give feedback.
All reactions