Abstract
AbstractThere is broad consensus that audio description (AD) is a modality of intersemiotic translation, but there are different views in relation to how AD can be more precisely conceptualised. While Benecke (Audiodeskription als partielle Translation. Modell und Methode, LIT, Berlin, 2014) characterises AD as ‘partial translation’, Braun (T 28: 302–313, 2016) hypothesises that what audio describers appear to ‘omit’ from their descriptions can normally be inferred by the audience, drawing on narrative cues from dialogue, mise-en-scène, kinesis, music or sound effects. The study reported in this paper tested this hypothesis using a corpus of material created during the H2020 MeMAD project. The MeMAD project aimed to improve access to audiovisual (AV) content through a combination of human and computer-based methods of description. One of the MeMAD workstreams addressed human approaches to describing visually salient cues. This included an analysis of the potential impact of omissions in AD, which is the focus of this paper. Using a corpus of approximately 500 audio described film extracts we identified the visual elements that can be considered essential for the construction of the filmic narrative and then performed a qualitative analysis of the corresponding audio descriptions to determine how these elements are verbally represented and whether any omitted elements could be inferred from other cues that are accessible to visually impaired audiences. We then identified the most likely source of these inferences and the conditions upon which retrieval could be predicated, preparing the ground for future reception studies to test our hypotheses with target audiences. In this paper, we discuss the methodology used to determine where omissions occur in the analysed audio descriptions, consider worked examples from the MeMAD500 film corpus, and outline the findings of our study namely that various strategies are relevant to inferring omitted information, including the use of proximal and distal contextual cues, and reliance on the application of common knowledge and iconic scenarios. To conclude, consideration is given to overcoming significant omissions in human-generated AD, such as using extended AD formats, and mitigating similar gaps in machine-generated descriptions, where incorporating dialogue analysis and other supplementary data into the computer model could resolve many omissions.
Funder
H2020 Industrial Leadership
Publisher
Springer Science and Business Media LLC
Subject
Computer Networks and Communications,Human-Computer Interaction,Information Systems,Software
Reference29 articles.
1. Abdel-Raheem, A.: Mental model theory as a model for analysing visual and multimodal discourse. J. Pragmat. 155, 303–320 (2020). https://doi.org/10.1016/j.pragma.2019.09.012
2. Audio Description Coalition (2009) ‘Standards for audio description and code of professional conduct for describers’. Available at: https://adp.acb.org/docs/ADP_Standards.doc.
3. Benecke, B.: Audiodeskription als partielle Translation. Modell und Methode. LIT, Berlin (2014)
4. Bernabé, R., Orero, P.: Easier audio description: Exploring the potential of easy-to-read principles in simplifying AD. In: Innovation in Audio Description Research, pp. 55–77. Routledge, Abingdon (2021)
5. Braun, S.: Audio description research: state of the art and beyond. Trans. Stud. New Millenn. 6, 14–30 (2008)
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献