Show simple item record

dc.contributor.authorCochez, Michael
dc.date.accessioned2016-05-16T10:05:44Z
dc.date.available2016-05-16T10:05:44Z
dc.date.issued2016
dc.identifier.isbn978-951-39-6649-2
dc.identifier.otheroai:jykdok.linneanet.fi:1541145
dc.identifier.urihttps://jyx.jyu.fi/handle/123456789/49793
dc.description.abstractInformation and its derived knowledge are not static. Instead, information is changing over time and our understanding of it evolves with our ability and willingness to consume the information. When compared to humans, current computer systems seem very limited in their ability to really understand the meaning of things. On the other hand, they are very powerful when it comes down to performing exact computations. One aspect which sets humans apart from machines when trying to understand the world is that we will often make mistakes, forget information, or choose what to focus on. To put this in another perspective, it seems like humans can behave somehow more randomly and still outperform machines in knowledge related tasks. In computer science there is a branch of research concerned with allowing randomness or inaccuracy in algorithms, which are then called approximate algorithms. The main benefit of using these algorithms is that they are often much faster than their exact counterparts, at the cost of producing wrong or inexact results, once in a while. So, these algorithms could be used in contexts where erring once in while does not harm. If the chance of making a mistake is very slim, say lower than the chance of a memory error, then the expected precision will rival their exact counterparts. Furthermore, the input data to the algorithms often already contains a fair amount of uncertainty, such that the small error which the approximate algorithm introduces becomes more or less insignificant. In this dissertation, the author investigates the use of familiar and new approximate algorithms to knowledge discovery and evolution. The main contributions of the dissertation are a) an abstract formulation of what it means for an ontology to be and stay optimal over time, b) a contribution to a vision paper regarding the future of evolving knowledge ecosystems, c) an investigation of the application of locality-sensitive hashing (LSH) in the context of ontology matching and semantic search, d) the twister tries algorithm which is a novel approximate hierarchical clustering approach with linear space and time constraints, and e) an extension on the twister tries algorithm which trades a longer, but adaptable running time for a likely improvement of the clustering result.
dc.format.extent1 verkkoaineisto (56 sivua, 85 sivua useina numerointijaksoina)
dc.language.isoeng
dc.publisherUniversity of Jyväskylä
dc.relation.ispartofseriesJyväskylä studies in computing
dc.subject.otherknowledge evolution
dc.subject.otherhierarchial clustering
dc.subject.otherinformation retrieval
dc.titleTaming big knowledge evolution
dc.typeDiss.
dc.identifier.urnURN:ISBN:978-951-39-6649-2
dc.type.dcmitypeTexten
dc.type.ontasotVäitöskirjafi
dc.type.ontasotDoctoral dissertationen
dc.contributor.tiedekuntaInformaatioteknologian tiedekuntafi
dc.contributor.yliopistoUniversity of Jyväskyläen
dc.contributor.yliopistoJyväskylän yliopistofi
dc.contributor.oppiaineTietotekniikkafi
dc.relation.issn1456-5390
dc.relation.numberinseries237
dc.rights.accesslevelopenAccessfi
dc.subject.ysotiedonlouhinta
dc.subject.ysobig data
dc.subject.ysotiedonhaku
dc.subject.ysotiedonhakujärjestelmät
dc.subject.ysoontologiat
dc.subject.ysosemanttinen web
dc.subject.ysoalgoritmit
dc.subject.ysooptimointi
dc.subject.ysomatemaattinen optimointi
dc.subject.ysogeneettiset algoritmit
dc.subject.ysoklusterianalyysi


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record