As we may link: time-aligned concordances of field recordings. A working model
Access status:
Open Access
Type
Conference paperAuthor/s
Thieberger, NickAbstract
It should be easy to link text and audio and retrieve audio based on the textual representation or transcript. The potential has been there for some time (since digitised sound) and is realised in several ways by currently available software tools. However, none of the current ...
See moreIt should be easy to link text and audio and retrieve audio based on the textual representation or transcript. The potential has been there for some time (since digitised sound) and is realised in several ways by currently available software tools. However, none of the current solutions allows you to simply amass your field tapes and produce a text-based interface to them so that you can click on a sentence anywhere in your transcript and hear it. Further, having established such links, it would make sense to be able to have a concordance of the transcripts, providing a live link between the concordance and the text/audio. In this demonstration we will see an implementation for my linguistic data of the concordance/ text/ audio linkage based on output from LACITO's SoundIndex and using HyperCard as the modeling tool.
See less
See moreIt should be easy to link text and audio and retrieve audio based on the textual representation or transcript. The potential has been there for some time (since digitised sound) and is realised in several ways by currently available software tools. However, none of the current solutions allows you to simply amass your field tapes and produce a text-based interface to them so that you can click on a sentence anywhere in your transcript and hear it. Further, having established such links, it would make sense to be able to have a concordance of the transcripts, providing a live link between the concordance and the text/audio. In this demonstration we will see an implementation for my linguistic data of the concordance/ text/ audio linkage based on output from LACITO's SoundIndex and using HyperCard as the modeling tool.
See less
Date
2001-01-01Publisher
Research Institute for Humanities and Social Sciences (RIHSS), the University of Sydney.Licence
Copyright the University of SydneyCitation
Computing Arts 2001 : digital resources for research in the humanities : 26th-28th September 2001, Veterinary Science Conference Centre, the University of Sydney / hosted by the Scholarly Text and Imaging Service (SETIS), the University of Sydney Library, and the Research Institute for Humanities and Social Sciences (RIHSS), the University of SydneyShare