=== Corpora used by HaBiT system === * [http://corpora.fi.muni.cz/habit/run.cgi/first_form?corpname=wic&reload=1 Amharic WIC corpus], 200 thousand tokens Amharic WIC corpus (News from Walta Information Center), manually tagged. [[BR]] [https://nlp.fi.muni.cz/projects/habit/download/wic.vert.gz download corpus] ([https://nlp.fi.muni.cz/en/LicenceWebCorpus NLP Centre Web Corpus license]) * [http://corpora.fi.muni.cz/habit/run.cgi/first_form?corpname=amwac16&reload=1 amWaC16 corpus], 20 million tokens Amharic Web corpus. Crawled by !SpiderLing in August 2013, October 2015 and January 2016. Cleaned, de-duplicated. Tagged by !TreeTagger trained on Amharic WiC. [[BR]] [AmharicCorpus Corpus deliverable/technical report][[BR]] [https://nlp.fi.muni.cz/projects/habit/download/am131516.vert.gz download corpus] ([https://nlp.fi.muni.cz/en/LicenceWebCorpus NLP Centre Web Corpus license]) * [http://corpora.fi.muni.cz/habit/run.cgi/first_form?corpname=orwac16 orWaC16 corpus], 5.1 million tokens. Oromo Web corpus crawled by !SpiderLing in January 2016. Cleaned, de-duplicated. [[BR]] [OromoCorpus Corpus deliverable/technical report][[BR]] [https://nlp.fi.muni.cz/projects/habit/download/or16.tag.vert.gz download corpus] ([https://nlp.fi.muni.cz/en/LicenceWebCorpus NLP Centre Web Corpus license]) * [http://corpora.fi.muni.cz/habit/run.cgi/first_form?corpname=sowac16 soWaC16 corpus], 80 million tokens. Somali Web corpus crawled by !SpiderLing in January 2016. Cleaned, de-duplicated. [[BR]] [SomaliCorpus Corpus deliverable/technical report][[BR]] [https://nlp.fi.muni.cz/projects/habit/download/so16.tag.vert.gz download corpus] ([https://nlp.fi.muni.cz/en/LicenceWebCorpus NLP Centre Web Corpus license]) * [http://corpora.fi.muni.cz/habit/run.cgi/first_form?corpname=tiwac16 tiWaC16 corpus], 2.5 million tokens. Tigrinya Web corpus crawled by !SpiderLing in January 2016. Cleaned, de-duplicated. [[BR]] [TigrinyaCorpus Corpus deliverable/technical report][[BR]] [https://nlp.fi.muni.cz/projects/habit/download/ti16.tag.vert.gz download corpus] ([https://nlp.fi.muni.cz/en/LicenceWebCorpus NLP Centre Web Corpus license]) === Software === * WebBootCaT, a search engine querying and corpus building tool. It queries a search engine with user specified words or phrases, obtains URLs of relevant documents fount by the search engine and downloads the documents. * !SpiderLing, a web spider for linguistics — is software for obtaining text from the web useful for building text corpora. Many documents on the web only contain material not suitable for text corpora, such as site navigation, lists of links, lists of products, and other kind of text not comprised of full sentences. In fact such pages represent the vast majority of the web. Therefore, by doing unrestricted web crawls, we typically download a lot of data which gets filtered out during post-processing. This makes the process of web corpus collection inefficient. * Chared is a tool for detecting the character encoding of a text in a known language. The language of the text has to be specified as an input parameter so that correspondent language model can be used. The package contains models for a wide range of languages. In general, it should be more accurate than character encoding detection algorithms with no language constraints. * Justext is a tool for removing boilerplate content, such as navigation links, headers, and footers from HTML pages. It is designed to preserve mainly text containing full sentences and it is therefore well suited for creating linguistic resources such as Web corpora. A HTML page is split to paragraphs and a context sensitive heuristic algorithm is employed to separate content from boilerplate. * Onion is a tool for removing duplicate parts from large collections of texts. One instance of duplicate parts is kept, others are marked or removed. The tool allows removing both identical and similar documents, on any level (document, paragraph, sentence — if present in the data). The de-duplication algorithm is based on comparing hashes of n-grams of words. * Unitok is a universal text tokeniser with specific settings for many languages. It can turn plain text into a sequence of newline-separated tokens (vertical format), while preserving XML-like tags containing metadata. * [https://nlp.fi.muni.cz/projekty/habit/geez2sera/index.cgi geez2sera] is a tool for converting (transliterating) [https://en.wikipedia.org/wiki/Ge%27ez_script Geez] script into [http://abyssiniagateway.net/fidel/sera-faq_1.html SERA], System for Ethiopic Representation using Latin ASCII characters. It is based on [https://github.com/fgaim/HornMorpho L3 Python library] (GPL licence). * [https://nlp.fi.muni.cz/projekty/habit/amtag/index.cgi AmTag] is a tagger for Amharic. It is [http://www.cis.uni-muenchen.de/~schmid/tools/TreeTagger/ TreeTagger] with an Amharic model. * [http://code.activestate.com/recipes/326576-language-detection-using-character-trigrams/ Language identification using trigrams] (licensed under the [https://docs.python.org/3/license.html PSF licence])