{"id":2764,"date":"2016-06-05T03:47:14","date_gmt":"2016-06-05T03:47:14","guid":{"rendered":"https:\/\/notebooks.dataone.org\/?p=2764"},"modified":"2016-06-05T03:47:49","modified_gmt":"2016-06-05T03:47:49","slug":"week-2-exploring-linkipedia","status":"publish","type":"post","link":"https:\/\/notebooks.dataone.org\/semantic-evolution\/week-2-exploring-linkipedia\/","title":{"rendered":"Week 2: Exploring Linkipedia"},"content":{"rendered":"
This week I explored the existing code for Linkipedia. Linkipedia consists of several endpoints that can be used to run it and I tracked down the workflow through all of them. The main objectives of this analysis were:<\/p>\n
Currently, the “annotate” endpoint segments the query into sentences, extracts n-grams (n=1 to 5) from the sentences, and queries the index to find matches for any of the extracted terms. To this list, I plan to add the normalized version of existing terms; along with the word pairs that are seen to be dependent upon each other in a dependency parse tree. Even though the python code for testing Linkipedia already has a component that does parsing; it is unclear as to whether Linkipedia receives the original query or the terms obtained after parsing. Hence, the first task for next week is to determine the kind of input sent by the python code to the “annotate” endpoint of Linkipedia. Depending upon this, I will apply different approaches to refine the existing list of terms and evaluate the performance of the system for each approach.<\/p>\n","protected":false},"excerpt":{"rendered":"
This week I explored the existing code for Linkipedia. Linkipedia consists of several endpoints that can be used to run it and I tracked down the workflow through all of them. The main objectives of this analysis were: \u00a0to understand the functioning behind each module of the system \u00a0to identify Continue reading Week 2: Exploring Linkipedia<\/span>