Rating is available when the video has been rented.
This feature is not available right now. Please try again later.
Published on Jun 1, 2016
PyData Berlin 2016
The development of large-scale Knowledge Base (KB) has drawn lots of attentions and efforts from both academy and industries recently . In this talk I will introduce how to use keywords and public available data to build our structural KB, and build knowledge retrieval system for different languages using python.
Many large-scale Knowledge Bases (KB), such as Yago, Wikidata, Freebase, and Google’s Knowledge Graph, have been build by extracting facts fro structural Wikipedia and/or natural language Web documents.
The main observation of using knowledge base is that not all facts are useful and have enough information. To tackle this problem I will introduce how we build various data sources to help facts and keywords selection. We will also discuss important questions of KB applications including, - architecture of a KB processing and extraction system using Wikipedia and two public available KB including Wikidata and Yago; - method for calculating contextual relevance between facts. - how to present different facts to users.