Full metadata record
DC poleHodnotaJazyk
dc.contributor.authorBrychcín, Tomáš
dc.date.accessioned2016-06-21T06:45:43Z
dc.date.available2016-06-21T06:45:43Z
dc.date.issued2012
dc.identifier.urihttp:// www.kiv.zcu.cz/publications/
dc.identifier.urihttp://hdl.handle.net/11025/21549
dc.format10 s.cs
dc.format.mimetypeapplication/pdf
dc.language.isoenen
dc.publisherUniversity of West Bohemia in Pilsenen
dc.rights© University of West Bohemia in Pilsenen
dc.subjectjazykový modelcs
dc.subjectn-gramcs
dc.titleUnsupervised methods for language modeling: technical report no. DCSE/TR-2012-03en
dc.typezprávacs
dc.typereporten
dc.rights.accessopenAccessen
dc.type.versionpublishedVersionen
dc.description.abstract-translatedLanguage models are crucial for many tasks in NLP and N-grams are the best way to build them. Huge e ort is being invested in improving n-gram language models. By introducing external information (morphology, syntax, partitioning into documents, etc.) into the models a signi cant improvement can be achieved. The models can however be improved with no external information and smoothing is an excellent example of such an improvement. Thesis summarizes the state-of-the-art approaches to unsupervised language modeling with emphases on the in ectional languages, which are particularly hard to model. It is focused on methods that can discover hidden patterns that are already in a training corpora. These patterns can be very useful for enhancing the performance of language modeling, moreover they do not require additional information sources.en
dc.subject.translatedlanguage modelen
dc.subject.translatedn-gramen
Vyskytuje se v kolekcích:Zprávy / Reports (KIV)

Soubory připojené k záznamu:
Soubor Popis VelikostFormát 
Brychcin.pdfPlný text425,44 kBAdobe PDFZobrazit/otevřít


Použijte tento identifikátor k citaci nebo jako odkaz na tento záznam: http://hdl.handle.net/11025/21549

Všechny záznamy v DSpace jsou chráněny autorskými právy, všechna práva vyhrazena.