Canonical Text Service (CTS) Endpoint
https://digilab.serbski-institut.de/cts/plain/editions.php
or
https://digilab.serbski-institut.de/cts/?request=GetCapabilities
Voller Korpus
DSB Tekstmining [baseurl] = https://digilab.serbski-institut.de/dsbtekstmining/
( External Backup with old data ) [baseurl] = https://tiepilab.de/dsbtekstmining/
Subkorpora
Casnik Tekstmining [baseurl] = https://digilab.serbski-institut.de/casniktm/
| Name | Beispiel | Ausgabeformat | Optionen ( URL + &... ) | Beispiel URL |
|---|---|---|---|---|
| Autocomplete (Prefixsearch) word | 🔗 | Token [NL] | sortby=[frequency|alphabet] limit=[1,2,...] | [baseurl] php/prefixsearch.php?limit=20&word= |
| Autocomplete (Prefixsearch) lemmatisation | 🔗 | Lemma [NL] | sortby=[frequency|alphabet] limit=[1,2,...] | [baseurl] php/prefixlemmasearch.php?limit=20&lemma= |
| Autocomplete (Prefixsearch) normalisation | 🔗 | Norm [NL] | sortby=[frequency|alphabet] limit=[1,2,...] | [baseurl] php/prefixnormsearch.php?limit=20&norm= |
| Documents including a character | 🔗 | Doc URN [TAB] Year [NL] | [baseurl] php/urnbychar.php?char= | |
| Documents including a lemma | 🔗 | Doc URN [TAB] Year [NL] | [baseurl] php/urnbylemma.php?lemma= | |
| Documents including a normalisation | 🔗 | Doc URN [TAB] Year [NL] | [baseurl] php/urnbynorm.php?norm= | |
| Documents including a word | 🔗 | Doc URN [TAB] Year [NL] | [baseurl] php/urnbyword.php?word= | |
| Lemmatisation of word | 🔗 | Lemma [TAB] Token [TAB] Type [NL] | [baseurl] php/token2lemma.php?token= | |
| Length examples for lemma | 🔗 | Lemma [NL] | [baseurl] php/wordbylength.php?random&length= | |
| Length examples for normalisation | 🔗 | Norm [NL] | [baseurl] php/wordbylength.php?random&length= | |
| Length examples for words | 🔗 | Word [NL] | [baseurl] php/wordbylength.php?random&length= | |
| Normalisation of word | 🔗 | Token [TAB] Norm [TAB] Type [NL] | [baseurl] php/token2norm.php?token= | |
| Ratio of Types / Tokens per Year | 🔗 | Year [TAB] TTR [NL] | [baseurl] data/stats/_typetokenratioperyear.txt | |
| Regex search word (max. 2.1M) | 🔗 | Word [TAB] Year [TAB] Frequency [NL] | [baseurl] php/regexsearch.php?token= | |
| Regex search lemma (max. 2.1M) | 🔗 | Lemma [TAB] Year [TAB] Frequency [TAB] Word [NL] | [baseurl] php/lemmaregexsearch.php?lemma= | |
| Regex search norm (max. 2.1M) | 🔗 | Norm [TAB] Year [TAB] Word [TAB] Frequency [NL] | [baseurl] php/normregexsearch.php?norm= | |
| Suffixsearch word | 🔗 | Token [NL] | [baseurl] php/suffixsearch.php?limit=20&word= | |
| Suffixsearch lemmatisation | 🔗 | Lemma [NL] | [baseurl] php/suffixlemmasearch.php?limit=20&token= | |
| Suffixsearch normalisation | 🔗 | Norm [NL] | [baseurl] php/suffixnormsearch.php?limit=20&token= | |
| Sum of Tokens per Year | 🔗 | Year [TAB] Sum [NL] | [baseurl] data/stats/_typesumperyear.txt | |
| Sum of Types per Year | 🔗 | Year [TAB] Sum [NL] | [baseurl] data/stats/_tokensumperyear.txt | |
| Token per normalisation | 🔗 | Norm [TAB] Token [TAB] Frequency [NL] | [baseurl] php/normtoken.php?norm= | |
| Token per lemmatisation | 🔗 | Lemma [TAB] Token [TAB] Frequency [NL] | [baseurl] php/lemmatoken.php?lemma= |