Cross-Linguistic Data Formats
See also this article describing CLDF:
Forkel, R. et al. Cross-Linguistic Data Formats, advancing data sharing and reuse in comparative linguistics. Sci. Data. 5:180205 doi: 10.1038/sdata.2018.205 (2018).
To allow exchange of cross-linguistic data and decouple development of tools and methods from that of databases, standardized data formats are necessary.
Once established, these dataformats could become a foundation not only for tools but also for instruction material in the spirit of Data Carpentry for historical linguistics and linguistic typology.
The main types of cross-linguistic data we are concerned with here are any tabular data which is typically analysed using quantitative (automated) methods or made accessible using software tools like the `clld` framework, such as
- wordlists (or more complex lexical data including e.g. cognate judgements),
- structure datasets (e.g. WALS features),
- simple dictionaries.
- Data should be both editable "by hand" and amenable to reading and writing by software (preferably software the typical linguist can be expected to use correctly).
- Data should be encoded as UTF-8 text files.
- If entities can be referenced, e.g. languages through their Glottocode, this should be done rather than duplicating information like language names.
- Compatibility with existing tools, standards and practice should always be kept in mind.
Automated re-use requires that the standard specifies not just the structure, but also the semantics of the data stored. Thus, the CLDF specification should be as rigid as possible. Of course, new types of data cannot be immediately compatible with independently developed tools; so the CLDF standard should also provide mechanisms to let data types evolve well understood semantics, while being syntactically compatible from the start.
Since we are concerned with tabular data here, CLDF is built on W3C's Model for Tabular Data and Metadata on the Web and Metadata Vocabulary for Tabular Data. This model - by virtue of being a JSON-LD dialect - is ideally suited to be combined with an ontology to specify syntax as well as semantics of a data serialization format. Much like MDF - SIL's Multi-Dictionary Formatter - adds a hierarchical data model on top of Toolbox' standard format markers to support a data reuse scenario, CLDF structures cross-linguistic data to make automated reuse in typical analyses in historical linguistics possible.
One of the main goals of the CLDF specification is a useful delineation of data and tools. Using a CSV based format makes it really easy to use this data in a UNIX-style pipeline of data transformation commands. This pipeline-style of data transformation and analysis seems to be at the core of typical workflows e.g. in historical linguistics, e.g. LingPy or QLC.
If suitable text- and line-based formats are available, this pipeline-style does also allow for easy extensibility; E.g. a workflow for automatic cognate judgements based on LingPy functionality could be extended with phylogenetic analysis and post-processing via phyltr, which processes sets of phylogenetic trees represented in the newick format, or ete.
If cross-linguistic comparisons procede in the footsteps of bioinformatics, workflows based UNIX pipelines may at some point be formalized using a common workflow language.
While data formats to exchange linguistic data have been around for some time, e.g. the SFM or Standard Format used by Toolbox, new developments in the area of language diversity research have motivated this push for a new set of formats:
- A new interest in standardizing tabular data on the web, with a particular focus on CSV
- A trend towards using computational methods to analyse large scale cross-linguistic data.
- The clld framework, developed within the CLLD project has shown that many different cross-linguistic databases can be built on top of the same core data model. CLDF is an attempt to externalise this data model.
Thus, following up discussions from the first workshop on Language Comparison with Linguistic Databases a second workshop in Leipzig focused on the idea of a very simple CSV based format to exchange very simple cross-linguistic data.
Simplicity was the main design goal from the start, so the formats under consideration will evolve starting out as simple as possible. With CLDF 1.0 we provide a stable baseline for further evolution.