Sentence Similarity
http://www.semanlink.net/tag/sentence_similarity
Documents tagged with Sentence SimilarityAdvance BERT model via transferring knowledge from Cross-Encoders to Bi-Encoders | by Chien Vu | Towards Data Science
http://www.semanlink.net/doc/2021/12/advance_bert_model_via_transfer
Data Augmentation Method to improve SBERT Bi-Encoders for Pairwise Sentence Scoring Tasks (Semantic sentence tasks)
2021-12-17T00:26:39Z[1908.10084] Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks
http://www.semanlink.net/doc/2019/08/_1908_10084_sentence_bert_sen
> Sentence-BERT
(SBERT), a modification of the pretrained
BERT network that use siamese and triplet network
structures to derive **semantically meaningful
sentence embeddings** that can be compared
using cosine-similarity.
Important because
- BERT ist unsuitable for semantic similarity
search as well as for unsupervised tasks
like clustering.
- simple methods such as using the CLS token give low quality sentence embeddings
However, the purpose of SBERT sentence embeddings
are **not to be used for transfer learning for other
tasks**.
[Related blog post](/doc/2020/01/richer_sentence_embeddings_usin); [Github](https://github.com/UKPLab/sentence-transformers)
2019-08-28T22:41:55Znlp-notebooks/Simple Sentence Similarity.ipynb at master ยท nlptown/nlp-notebooks
https://github.com/nlptown/nlp-notebooks/blob/master/Simple%20Sentence%20Similarity.ipynb
[blog post](/doc/?uri=http%3A%2F%2Fnlp.town%2Fblog%2Fsentence-similarity%2F)
2019-02-07T00:59:11Z[1704.05358] Representing Sentences as Low-Rank Subspaces
https://arxiv.org/abs/1704.05358
> We observe a simple geometry of sentences -- the word representations of a given sentence roughly lie in a low-rank subspace (roughly, rank 4). Motivated by this observation, we represent a sentence by the low-rank subspace spanned by its word vectors.
A sentence of N words is a matrix (300, N) (if 300 is the dim of the word embeddings space). We take the eg. 4 (hyperparam) heaviest singular values -> a subspace with dim 4
Similarity between docs: principal angle between the subspaces (reminiscent of cosine similarity)
2018-10-06T11:22:58ZComparing Sentence Similarity Methods
http://nlp.town/blog/sentence-similarity/
[notebook](/doc/?uri=https%3A%2F%2Fgithub.com%2Fnlptown%2Fnlp-notebooks%2Fblob%2Fmaster%2FSimple%2520Sentence%2520Similarity.ipynb)
2018-05-25T23:53:41Z