The problems with Swagger - NovaTec Blog(About) Swagger imposes some constraints, like the lack of hypermedia... if you are using swagger, you are probably giving up one of the most powerful feature of RESTful APIs. You are giving up evolvability!
Translating Embeddings for Modeling Multi-relational Data (2013)(About) This work focuses on modeling multi-relational
data from KBs (Wordnet and Freebase in this paper), with the goal of providing an efficient
tool to complete them by automatically adding new facts, without requiring extra knowledge.
**Embedding entities and relationships of multirelational
data**: a method which **models relationships by interpreting them as translations** operating on the
low-dimensional embeddings of the entities. Motivation:
- hierarchical relationships are extremely common in KBs and translations are the natural transformations for representing them.
- cf. word embeddings and the “capital of” relationship between countries and cities, which are (coincidentally rather than willingly) represented by the model as translations in the embedding space. This suggests that there may exist embedding spaces in which 1-to-1 relationships between entities of different types may, as well, be represented by translations. The intention of our model is to enforce such a structure of the embedding space.
[Good blog post by PY Vandenbussche](http://pyvandenbussche.info/2017/translating-embeddings-transe/)
Learning Deep Structured Semantic Models for Web Search using Clickthrough Data - Microsoft Research (2013)(About) we strive to develop a series of **new latent semantic models with a deep structure that project queries and documents into a common low-dimensional space** where the relevance of a document given a query is readily computed as the distance between them. The proposed deep structured semantic models are discriminatively trained by maximizing the conditional likelihood of the clicked documents given a query using the clickthrough data. To make our models applicable to large-scale Web search applications, we also use a technique called word hashing
Combining word and entity embeddings for entity linking (ESWC 2017)(About) The general approach for the entity linking task is to generate, for a given mention, a set of candidate entities from the base and, in a second step, determine which is the best
one. This paper proposes a novel method for the second step which is
based on the joint learning of embeddings for the words in the text and
the entities in the knowledge base.
[1712.09405] Advances in Pre-Training Distributed Word Representations(About) > we show how to train high-quality word vector representations by using a combination of known tricks that are however rarely used together. The main result of our work is the new set of publicly available pre-trained models that outperform the current state of the art by a large margin on a number of tasks