Research Publications

2016

Greene GJ, Fischer B. CVExplorer: Identifying Candidate Developers by Mining and Exploring Their Open Source Contributions. In: Automated Software Engineering. ; 2016.

No Abstract

@{141,
  author = {G.J. Greene and Bernd Fischer},
  title = {CVExplorer: Identifying Candidate Developers by Mining and Exploring Their Open Source Contributions},
  abstract = {No Abstract},
  year = {2016},
  journal = {Automated Software Engineering},
  pages = {804-809},
  month = {03/09-07/09},
  isbn = {978-1-4503-3845-5},
}
Rens G. A Stochastic Belief Change Framework with an Observation Stream and Defaults as Expired Observations. 2016.

A framework for an agent to change its probabilistic beliefs after a stream of noisy observations is received is proposed. Observations which are no longer relevant, become default assumptions until overridden by newer, more prevalent observations. A distinction is made between background and foreground beliefs. Agent actions and environment events are distinguishable and form part of the agent model. It is left up to the agent designer to provide an environment model; a submodel of the agent model. An example of an environment model is provided in the paper, and an example scenario is based on it. Given the particular form of the agent model, several ‘patterns of cognition’ can be identified. An argument is made for four particular patterns.

@misc{140,
  author = {Gavin Rens},
  title = {A Stochastic Belief Change Framework with an Observation Stream and Defaults as Expired Observations},
  abstract = {A framework for an agent to change its probabilistic beliefs after a stream of noisy observations is received is proposed. Observations which are no longer relevant, become default assumptions until overridden by newer, more prevalent observations. A distinction is made between background and foreground beliefs. Agent actions and environment events are distinguishable and form part of the agent model. It is left up to the agent designer to provide an environment model; a submodel of the agent model. An example of an environment model is provided in the paper, and an example scenario is based on it. Given the particular form of the agent model, several ‘patterns of cognition’ can be identified. An argument is made for four particular patterns.},
  year = {2016},
}
Rens G, Kern-Isberner G. An Approach to Qualitative Belief Change Modulo Ontic Strength. 2016.

Sometimes, strictly choosing between belief revision and belief update is inadequate in a dynamical, uncertain environment. Boutilier combined the two notions to allow updates in response to external changes to inform an agent about its prior beliefs. His approach is based on ranking functions. Rens proposed a new method to trade off probabilistic revision and update, in proportion to the agent’s confidence for whether to revise or update. In this paper, we translate Rens’s approach from a probabilistic setting to a setting with ranking functions. Given the translation, we are able to compare Boutilier’s and Rens’s approaches.We found that Rens’s approach is an extension of Boutilier’s.

@misc{139,
  author = {Gavin Rens and Gabriele Kern-Isberner},
  title = {An Approach to Qualitative Belief Change Modulo Ontic Strength},
  abstract = {Sometimes, strictly choosing between belief revision and belief update is inadequate in a dynamical, uncertain environment. Boutilier combined the two notions to allow updates in response to external changes to inform an agent about its prior beliefs. His approach is based on ranking functions. Rens proposed a new method to trade off probabilistic revision and update, in proportion to the agent’s confidence for whether to revise or update. In this paper, we translate Rens’s approach from a probabilistic setting to a setting with ranking functions. Given the translation, we are able to compare Boutilier’s and Rens’s approaches.We found that Rens’s approach is an extension of Boutilier’s.},
  year = {2016},
}
Greene GJ, Fischer B. Single-Focus Broadening Navigation in Concept Lattices. 2016. https://cla2016.hse.ru/data/2016/07/24/1119022942/CDUD2016.pdf#page=39.

No Abstract

@misc{138,
  author = {G.J. Greene and Bernd Fischer},
  title = {Single-Focus Broadening Navigation in Concept Lattices},
  abstract = {No Abstract},
  year = {2016},
  url = {https://cla2016.hse.ru/data/2016/07/24/1119022942/CDUD2016.pdf#page=39},
}
Britz K, Varzinczak I. Preferential modalities revisited. 2016. http://nmr2016.cs.uct.ac.za/proceedings_nmr2016_online.pdf.

We venture beyond the customary semantic approach in NMR, namely that of placing orderings on worlds (or valuations). In a modal-logic setting, we motivate and investigate the idea of ordering elements of the accessibility relations in Kripke frames, i.e, world pairs (w,w') (or arrows). The underlying intuition is that some world pairs may be seen as more normal (or typical, or expected) than others. We show this delivers an elegant and intuitive semantic construction, which gives a new perspective on present notions of defeasible necessity. From a modeler's perspective, the new framework we propose is more intuitively appealing. Technically, though, the revisited logic happens to not substantively increase the expressive power of the previously defined preferential modalities. This conclusion follows from an analysis of both semantic constructions via a generalisation of bisimulations to the preferential case. Lest this be seen as a negative result, it essentially means that reasoners based on the previous semantics (which have been shown to preserve the computational complexity of the underlying classical modal language) suffice for reasoning over the new semantics. Finally, we show that the kind of construction we here propose has many fruitful applications, notably in a description-logic context, where it provides the foundations on which to ground useful notions of defeasibility in ontologies yet to be explored.

@misc{135,
  author = {Katarina Britz and Ivan Varzinczak},
  title = {Preferential modalities revisited},
  abstract = {We venture beyond the customary semantic approach in NMR, namely that of placing orderings on worlds (or valuations). In a modal-logic setting, we motivate and investigate the idea of ordering elements of the accessibility relations in Kripke frames, i.e, world pairs (w,w') (or arrows). The underlying intuition is that some world pairs may be seen as more normal (or typical, or expected) than others. We show this delivers an elegant and intuitive semantic construction, which gives a new perspective on present notions of defeasible necessity. From a modeler's perspective, the new framework we propose is more intuitively appealing. Technically, though, the revisited logic happens to not substantively increase the expressive power of the previously defined preferential modalities. This conclusion follows from an analysis of both semantic constructions via a generalisation of bisimulations to the preferential case. Lest this be seen as a negative result, it essentially means that reasoners based on the previous semantics (which have been shown to preserve the computational complexity of the underlying classical modal language) suffice for reasoning over the new semantics. Finally, we show that the kind of construction we here propose has many fruitful applications, notably in a description-logic context, where it provides the foundations on which to ground useful notions of defeasibility in ontologies yet to be explored.},
  year = {2016},
  url = {http://nmr2016.cs.uct.ac.za/proceedings_nmr2016_online.pdf},
}
Rens G. On Stochastic Belief Revision and Update and their Combination. 2016.

I propose a framework for an agent to change its probabilistic beliefs when a new piece of propositional information alpha is observed. Traditionally, belief change occurs by either a revision process or by an update process, depending on whether the agent is informed with alpha in a static world or, respectively, whether alpha is a ‘signal’ from the environment due to an event occurring. Boutilier suggested a unified model of qualitative belief change, which “combines aspects of revision and update, providing a more realistic characterization of belief change.” In this paper, I propose a unified model of quantitative belief change, where an agent’s beliefs are represented as a probability distribution over possible worlds. As does Boutilier, I take a dynamical systems perspective. The proposed approach is evaluated against several rationality postulated, and some properties of the approach are worked out.

@misc{132,
  author = {Gavin Rens},
  title = {On Stochastic Belief Revision and Update and their Combination},
  abstract = {I propose a framework for an agent to change its probabilistic beliefs when a new piece of propositional information alpha is observed. Traditionally, belief change occurs by either a revision process or by an update process, depending on whether the agent is informed with alpha in a static world or, respectively, whether alpha is a ‘signal’ from the environment due to an event occurring. Boutilier suggested a unified model of qualitative belief change, which “combines aspects of revision and update, providing a more realistic characterization of belief change.” In this paper, I propose a unified model of quantitative belief change, where an agent’s beliefs are represented as a probability distribution over possible worlds. As does Boutilier, I take a dynamical systems perspective. The proposed approach is evaluated against several rationality postulated, and some properties of the approach are worked out.},
  year = {2016},
  isbn = {ISSN 0933-6192},
}
Rens G, Meyer T, Casini G. Revising Incompletely Specified Convex Probabilistic Belief Bases. 2016.

We propose a method for an agent to revise its incomplete probabilistic beliefs when a new piece of propositional information is observed. In this work, an agent’s beliefs are represented by a set of probabilistic formulae – a belief base. The method involves determining a representative set of ‘boundary’ probability distributions consistent with the current belief base, revising each of these probability distributions and then translating the revised information into a new belief base. We use a version of Lewis Imaging as the revision operation. The correctness of the approach is proved. The expressivity of the belief bases under consideration are rather restricted, but has some applications. We also discuss methods of belief base revision employing the notion of optimum entropy, and point out some of the benefits and difficulties in those methods. Both the boundary distribution method and the optimum entropy method are reasonable, yet yield different results.

@misc{131,
  author = {Gavin Rens and Thomas Meyer and Giovanni Casini},
  title = {Revising Incompletely Specified Convex Probabilistic Belief Bases},
  abstract = {We propose a method for an agent to revise its incomplete probabilistic beliefs when a new piece of propositional information is observed. In this work, an agent’s beliefs are represented by a set of probabilistic formulae – a belief base. The method involves determining a representative set of ‘boundary’ probability distributions consistent with the current belief base, revising each of these probability distributions and then translating the revised information into a new belief base. We use a version of Lewis Imaging as the revision operation. The correctness of the approach is proved. The expressivity of the belief bases under consideration are rather restricted, but has some applications. We also discuss methods of belief base revision employing the notion of optimum entropy, and point out some of the benefits and difficulties in those methods. Both the boundary distribution method and the optimum entropy method are reasonable, yet yield different results.},
  year = {2016},
  isbn = {ISSN 0933-6192},
}

2015

Booth R, Casini G, Meyer T, Varzinczak I. On the Entailment Problem for a Logic of Typicality. In: Twenty-Fourth International Joint Conference on Artificial Intelligence (IJCAI). AAAI Press; 2015. https://www.semanticscholar.org/paper/On-the-Entailment-Problem-for-a-Logic-of-Typicality-Booth-Casini/47bb88f18d591df1355c7ad53593a1f1b46522f0.

Propositional Typicality Logic (PTL) is a recently proposed logic, obtained by enriching classical propositional logic with a typicality operator. In spite of the non-monotonic features introduced by the semantics adopted for the typicality operator, the obvious Tarskian definition of entailment for PTL remains monotonic and is therefore not appropriate. We investigate different (semantic) versions of entailment for PTL, based on the notion of Rational Closure as defined by Lehmann and Magidor for KLM-style conditionals, and constructed using minimality. Our first important result is an impossibility theorem showing that a set of proposed postulates that at first all seem appropriate for a notion of entailment with regard to typicality cannot be satisfied simultaneously. Closer inspection reveals that this result is best interpreted as an argument for advocating the development of more than one type of PTL entailment. In the spirit of this interpretation, we define two primary forms of entailment for PTL and discuss their advantages and disadvantages

@{363,
  author = {Richard Booth and Giovanni Casini and Thomas Meyer and Ivan Varzinczak},
  title = {On the Entailment Problem for a Logic of Typicality},
  abstract = {Propositional Typicality Logic (PTL) is a recently proposed logic, obtained by enriching classical propositional logic with a typicality operator. In spite of the non-monotonic features introduced by the semantics adopted for the typicality operator, the obvious Tarskian definition of entailment for PTL remains monotonic and is therefore not appropriate. We investigate different (semantic) versions of entailment for PTL, based on the notion of Rational Closure as defined by Lehmann and Magidor for KLM-style conditionals, and constructed using minimality. Our first important result is an impossibility theorem showing that a set of proposed postulates that at first all seem appropriate for a notion of entailment with regard to typicality cannot be satisfied simultaneously. Closer inspection reveals that this result is best interpreted as an argument for advocating the development of more than one type of PTL entailment. In the spirit of this interpretation, we define two primary forms of entailment for PTL and discuss their advantages and disadvantages},
  year = {2015},
  journal = {Twenty-Fourth International Joint Conference on Artificial Intelligence (IJCAI)},
  pages = {2805-2811},
  month = {25/07 - 31/07},
  publisher = {AAAI Press},
  url = {https://www.semanticscholar.org/paper/On-the-Entailment-Problem-for-a-Logic-of-Typicality-Booth-Casini/47bb88f18d591df1355c7ad53593a1f1b46522f0},
}
Rens G, Meyer T. A New Approach to Probabilistic Belief Change. In: International Florida AI Research Society Conference (FLAIRS). AAAI Press; 2015. https://www.researchgate.net/publication/277004602_A_New_Approach_to_Probabilistic_Belief_Change.

One way for an agent to deal with uncertainty about its beliefs is to maintain a probability distribution over the worlds it believes are possible. A belief change operation may recommend some previously believed worlds to become impossible and some previously disbelieved worlds to become possible. This work investigates how to redistribute probabilities due to worlds being added to and removed from an agent’s belief-state. Two related approaches are proposed and analyzed.

@{362,
  author = {Gavin Rens and Thomas Meyer},
  title = {A New Approach to Probabilistic Belief Change},
  abstract = {One way for an agent to deal with uncertainty about its beliefs is to maintain a probability distribution over the worlds it believes are possible. A belief change operation may recommend some previously believed worlds to become impossible and some previously disbelieved worlds to become possible. This work investigates how to redistribute probabilities due to worlds being added to and removed from an agent’s belief-state. Two related approaches are proposed and analyzed.},
  year = {2015},
  journal = {International Florida AI Research Society Conference (FLAIRS)},
  pages = {582-587},
  month = {18/05 - 20/05},
  publisher = {AAAI Press},
  url = {https://www.researchgate.net/publication/277004602_A_New_Approach_to_Probabilistic_Belief_Change},
}
Wissing D, Pienaar W, Van Niekerk DR. Palatalisation of /s/ in Afrikaans. Stellenbosch Papers in Linguistics Plus. 2015;48. doi:10.5842/48-0-688.

This article reports on the investigation of the acoustic characteristics of the Afrikaans voiceless alveolar fricative /s/[1]. As yet, a palatal [ʃ] for /s/ has been reported only in a limited case, namely where /s/ is followed by palatal /j/, for example in the phrase is jy (‘are you’), pronounced as [ə-ʃəi]. This seems to be an instance of regressive coarticulation, resulting in coalescence of basic /s/ and /j/. The present study revealed that, especially in the pronunciation of young, white Afrikaans-speakers, /s/ is also palatalised progressively when preceded by /r/ in the coda cluster /rs/, and, to a lesser extent, also in other contexts where /r/ is involved, for example across syllable and word boundaries. Only a slight presence of palatalisation was detected in the production of /s/ in the speech of the white, older speakers of the present study. This finding might be indicative of a definite change in the Afrikaans consonant system. A post hoc reflection is offered here on the possible presence of /s/-fronting, especially in the speech of the younger females. Such pronunciation could very well be a prestige marker for affluent speakers of Afrikaans.

@article{293,
  author = {Daan Wissing and Wikus Pienaar and Daniel Van Niekerk},
  title = {Palatalisation of /s/ in Afrikaans},
  abstract = {This article reports on the investigation of the acoustic characteristics of the Afrikaans voiceless alveolar fricative /s/[1]. As yet, a palatal [ʃ] for /s/ has been reported only in a limited case, namely where /s/ is followed by palatal /j/, for example in the phrase is jy (‘are you’), pronounced as [ə-ʃəi]. This seems to be an instance of regressive coarticulation, resulting in coalescence of basic /s/ and /j/. The present study revealed that, especially in the pronunciation of young, white Afrikaans-speakers, /s/ is also palatalised progressively when preceded by /r/ in the coda cluster /rs/, and, to a lesser extent, also in other contexts where /r/ is involved, for example across syllable and word boundaries. Only a slight presence of palatalisation was detected in the production of /s/ in the speech of the white, older speakers of the present study. This finding might be indicative of a definite change in the Afrikaans consonant system. A post hoc reflection is offered here on the possible presence of /s/-fronting, especially in the speech of the younger females. Such pronunciation could very well be a prestige marker for affluent speakers of Afrikaans.},
  year = {2015},
  journal = {Stellenbosch Papers in Linguistics Plus},
  volume = {48},
  pages = {137-158},
  publisher = {Stellenbosch University},
  doi = {10.5842/48-0-688},
}
Modipa T, Davel MH. Predicting vowel substitution in code-switched speech. In: Pattern Recognition Association of South Africa (PRASA). Port Elizabeth, South Africa; 2015. doi:10.1109/RoboMech.2015.7359515.

The accuracy of automatic speech recognition (ASR) systems typically degrades when encountering code-switched speech. Some of this degradation is due to the unexpected pronunciation effects introduced when languages are mixed. Embedded (foreign) phonemes typically show more variation than phonemes from the matrix language: either approximating the embedded language pronunciation fairly closely, or realised as any of a set of phonemic counterparts from the matrix language. In this paper we describe a technique for predicting the phoneme substitutions that are expected to occur during code-switching, using non-acoustic features only. As case study we consider Sepedi/English code switching and analyse the different realisations of the English schwa. A code-switched speech corpus is used as input and vowel substitutions identified by auto-tagging this corpus based on acoustic characteristics. We first evaluate the accuracy of our auto-tagging process, before determining the predictability of our auto-tagged corpus, using non-acoustic features.

@{292,
  author = {Thipe Modipa and Marelie Davel},
  title = {Predicting vowel substitution in code-switched speech},
  abstract = {The accuracy of automatic speech recognition (ASR) systems typically degrades when encountering code-switched speech. Some of this degradation is due to the unexpected pronunciation effects introduced when languages are mixed. Embedded (foreign) phonemes typically show more variation than phonemes from the matrix language: either approximating the embedded language pronunciation fairly closely, or realised as any of a set of phonemic counterparts from the matrix language. In this paper we describe a technique for predicting the phoneme substitutions that are expected to occur during code-switching, using non-acoustic features only. As case study we consider Sepedi/English code switching and analyse the different realisations of the English schwa. A code-switched speech corpus is used as input and vowel substitutions identified by auto-tagging this corpus based on acoustic characteristics. We first evaluate the accuracy of our auto-tagging process, before determining the predictability of our auto-tagged corpus, using non-acoustic features.},
  year = {2015},
  journal = {Pattern Recognition Association of South Africa (PRASA)},
  pages = {154-159},
  month = {26/11-27/11},
  address = {Port Elizabeth, South Africa},
  isbn = {978-1-4673-7450-7, 978-1-4673-7449-1},
  doi = {10.1109/RoboMech.2015.7359515},
}
Kleynhans N, Barnard E. Efficient data selection for ASR. Language Resources and Evaluation. 2015;49(2). doi:10.1007/s10579-014-9285-0.

Automatic speech recognition (ASR) technology has matured over the past few decades and has made significant impacts in a variety of fields, from assistive technologies to commercial products. However, ASR system development is a resource intensive activity and requires language resources in the form of text annotated audio recordings and pronunciation dictionaries. Unfortunately, many languages found in the developing world fall into the resource-scarce category and due to this resource scarcity the deployment of ASR systems in the developing world is severely inhibited. One approach to assist with resource-scarce ASR system development, is to select 'useful' training samples which could reduce the resources needed to collect new corpora. In this work, we propose a new data selection framework which can be used to design a speech recognition corpus. We show for limited data sets, independent of language and bandwidth, the most effective strategy for data selection is frequency-matched selection and that the widely-used maximum entropy methods generally produced the least promising results. In our model, the frequency-matched selection method corresponds to a logarithmic relationship between accuracy and corpus size; we also investigated other model relationships, and found that a hyperbolic relationship (as suggested from simple asymptotic arguments in learning theory) may lead to somewhat better performance under certain conditions.

@article{291,
  author = {Neil Kleynhans and Etienne Barnard},
  title = {Efficient data selection for ASR},
  abstract = {Automatic speech recognition (ASR) technology has matured over the past few decades and has made significant impacts in a variety of fields, from assistive technologies to commercial products. However, ASR system development is a resource intensive activity and requires language resources in the form of text annotated audio recordings and pronunciation dictionaries. Unfortunately, many languages found in the developing world fall into the resource-scarce category and due to this resource scarcity the deployment of ASR systems in the developing world is severely inhibited. One approach to assist with resource-scarce ASR system development, is to select 'useful' training samples which could reduce the resources needed to collect new corpora. In this work, we propose a new data selection framework which can be used to design a speech recognition corpus. We show for limited data sets, independent of language and bandwidth, the most effective strategy for data selection is frequency-matched selection and that the widely-used maximum entropy methods generally produced the least promising results. In our model, the frequency-matched selection method corresponds to a logarithmic relationship between accuracy and corpus size; we also investigated other model relationships, and found that a hyperbolic relationship (as suggested from simple asymptotic arguments in learning theory) may lead to somewhat better performance under certain conditions.},
  year = {2015},
  journal = {Language Resources and Evaluation},
  volume = {49},
  pages = {327-353},
  issue = {2},
  publisher = {Springer Science+Business Media},
  address = {Dordrecht},
  doi = {10.1007/s10579-014-9285-0},
}
Kleynhans N, De Wet F, Barnard E. Unsupervised acoustic model training: comparing South African English and isiZulu. In: Pattern Recognition Association of South Africa (PRASA). Port Elizabeth, South Africa; 2015. doi: 10.1109/RoboMech.2015.7359512.

Large amounts of untranscribed audio data are generated every day. These audio resources can be used to develop robust acoustic models that can be used in a variety of speech-based systems. Manually transcribing this data is resource intensive and requires funding, time and expertise. Lightly-supervised training techniques, however, provide a means to rapidly transcribe audio, thus reducing the initial resource investment to begin the modelling process. Our findings suggest that the lightly-supervised training technique works well for English but when moving to an agglutinative language, such as isiZulu, the process fails to achieve the performance seen for English. Additionally, phone-based performances are significantly worse when compared to an approach using word-based language models. These results indicate a strong dependence on large or well-matched text resources for lightly-supervised training techniques.

@{290,
  author = {Neil Kleynhans and Febe De Wet and Etienne Barnard},
  title = {Unsupervised acoustic model training: comparing South African English and isiZulu},
  abstract = {Large amounts of untranscribed audio data are generated every day. These audio resources can be used to develop robust acoustic models that can be used in a variety of speech-based systems. Manually transcribing this data is resource intensive and requires funding, time and expertise. Lightly-supervised training techniques, however, provide a means to rapidly transcribe audio, thus reducing the initial resource investment to begin the modelling process. Our findings suggest that the lightly-supervised training technique works well for English but when moving to an agglutinative language, such as isiZulu, the process fails to achieve the performance seen for English. Additionally, phone-based performances are significantly worse when compared to an approach using word-based language models. These results indicate a strong dependence on large or well-matched text resources for lightly-supervised training techniques.},
  year = {2015},
  journal = {Pattern Recognition Association of South Africa (PRASA)},
  pages = {136-141},
  address = {Port Elizabeth, South Africa},
  isbn = {978-1-4673-7450-7, 978-1-4673-7449-1},
  doi = {10.1109/RoboMech.2015.7359512},
}
Giwa O, Davel MH. Text-based Language Identification of Multilingual Names. In: Pattern Recognition Association of South Africa (PRASA). Port Elizabeth, South Africa; 2015. doi: 10.1109/RoboMech.2015.7359517.

Text-based language identification (T-LID) of isolated words has been shown to be useful for various speech processing tasks, including pronunciation modelling and data categorisation. When the words to be categorised are proper names, the task becomes more difficult: not only do proper names often have idiosyncratic spellings, they are also often considered to be multilingual. We, therefore, investigate how an existing T-LID technique can be adapted to perform multilingual word classification. That is, given a proper name, which may be either mono- or multilingual, we aim to determine how accurately we can predict how many possible source languages the word has, and what they are. Using a Joint Sequence Model-based approach to T-LID and the SADE corpus - a newly developed proper names corpus of South African names - we experiment with different approaches to multilingual T-LID. We compare posterior-based and likelihood-based methods and obtain promising results on a challenging task.

@{289,
  author = {Oluwapelumi Giwa and Marelie Davel},
  title = {Text-based Language Identification of Multilingual Names},
  abstract = {Text-based language identification (T-LID) of isolated words has been shown to be useful for various speech processing tasks, including pronunciation modelling and data categorisation. When the words to be categorised are proper names, the task becomes more difficult: not only do proper names often have idiosyncratic spellings, they are also often considered to be multilingual. We, therefore, investigate how an existing T-LID technique can be adapted to perform multilingual word classification. That is, given a proper name, which may be either mono- or multilingual, we aim to determine how accurately we can predict how many possible source languages the word has, and what they are. Using a Joint Sequence Model-based approach to T-LID and the SADE corpus - a newly developed proper names corpus of South African names - we experiment with different approaches to multilingual T-LID. We compare posterior-based and likelihood-based methods and obtain promising results on a challenging task.},
  year = {2015},
  journal = {Pattern Recognition Association of South Africa (PRASA)},
  pages = {166-171},
  address = {Port Elizabeth, South Africa},
  isbn = {978-1-4673-7450-7, 978-1-4673-7449-1},
  doi = {10.1109/RoboMech.2015.7359517},
}
Davel MH, Barnard E, Van Heerden CJ, et al. Exploring minimal pronunciation modeling for low resource languages. In: Interspeech. Dresden, Germany; 2015.

Pronunciation lexicons can range from fully graphemic (modeling each word using the orthography directly) to fully phonemic (first mapping each word to a phoneme string). Between these two options lies a continuum of modeling options. We analyze techniques that can improve the accuracy of a graphemic system without requiring significant effort to design or implement. The analysis is performed in the context of the IARPA Babel project, which aims to develop spoken term detection systems for previously unseen languages rapidly, and with minimal human effort. We consider techniques related to letter-to-sound mapping and language-independent syllabification of primarily graphemic systems, and discuss results obtained for six languages: Cebuano, Kazakh, Kurmanji Kurdish, Lithuanian, Telugu and Tok Pisin.

@{288,
  author = {Marelie Davel and Etienne Barnard and Charl Van Heerden and William Hartman and Damianos Karakos and Richard Schwartz and Stavros Tsakalidis},
  title = {Exploring minimal pronunciation modeling for low resource languages},
  abstract = {Pronunciation lexicons can range from fully graphemic (modeling each word using the orthography directly) to fully phonemic (first mapping each word to a phoneme string). Between these two options lies a continuum of modeling options. We analyze techniques that can improve the accuracy of a graphemic system without requiring significant effort to design or implement. The analysis is performed in the context of the IARPA Babel project, which aims to develop spoken term detection systems for previously unseen languages rapidly, and with minimal human effort. We consider techniques related to letter-to-sound mapping and language-independent syllabification of primarily graphemic systems, and discuss results obtained for six languages: Cebuano, Kazakh, Kurmanji Kurdish, Lithuanian, Telugu and Tok Pisin.},
  year = {2015},
  journal = {Interspeech},
  pages = {538-542},
  address = {Dresden, Germany},
}
Badenhorst J, Davel MH. Synthetic triphones from trajectory-based feature distributions. In: Pattern Recognition Association of South Africa (PRASA). Port Elizabeth, South Africa; 2015. doi:10.1109/RoboMech.2015.7359509.

We experiment with a new method to create synthetic models of rare and unseen triphones in order to supplement limited automatic speech recognition (ASR) training data. A trajectory model is used to characterise seen transitions at the spectral level, and these models are then used to create features for unseen or rare triphones. We find that a fairly restricted model (piece-wise linear with three line segments per channel of a diphone transition) is able to represent training data quite accurately. We report on initial results when creating additional triphones for a single-speaker data set, finding small but significant gains, especially when adding additional samples of rare (rather than unseen) triphones.

@{287,
  author = {Jaco Badenhorst and Marelie Davel},
  title = {Synthetic triphones from trajectory-based feature distributions},
  abstract = {We experiment with a new method to create synthetic models of rare and unseen triphones in order to supplement limited automatic speech recognition (ASR) training data. A trajectory model is used to characterise seen transitions at the spectral level, and these models are then used to create features for unseen or rare triphones. We find that a fairly restricted model (piece-wise linear with three line segments per channel of a diphone transition) is able to represent training data quite accurately. We report on initial results when creating additional triphones for a single-speaker data set, finding small but significant gains, especially when adding additional samples of rare (rather than unseen) triphones.},
  year = {2015},
  journal = {Pattern Recognition Association of South Africa (PRASA)},
  pages = {118-122},
  address = {Port Elizabeth, South Africa},
  isbn = {978-1-4673-7450-7, 978-1-4673-7449-1},
  doi = {10.1109/RoboMech.2015.7359509},
}
Newman G, Fischer B. Language Fuzzing with Name Binding. 2015;Honours.

The language fuzzing with name binding project generates syntactically valid test programs that exercise the name binding semantics of a language processor. We introduce generation algorithm and a tool, NameFuzz, for the test suite generation. It achieves this by by parsing in a ANTLR grammar representing the context free grammar of the language, along with the language’s name binding rules in the NaBL meta-language. The test sentences are intended to be either accepted (positive test cases) or rejected (failing test cases) by the language processor. The intention is to promote confidence in the language processor as far as semantic correctness is concerned. The generated test suite in syntactically correct, but the limitations of not taking type checking into account or having a method to evaluate expressions lead to a large number of test sentences that are semantically incorrect. To a degree these limitations are overcome by the combinatorial nature of the generation algorithm that ensures that each possible type correct sentence is generated as well.

@phdthesis{137,
  author = {G. Newman and Bernd Fischer},
  title = {Language Fuzzing with Name Binding},
  abstract = {The language fuzzing with name binding project generates syntactically valid test programs that exercise the name binding semantics of a language processor. We introduce generation algorithm and a tool, NameFuzz, for the test suite generation. It achieves this by by parsing in a ANTLR grammar representing the context free grammar of the language, along with the language’s name binding rules in the NaBL meta-language. The test sentences are intended to be either accepted (positive test cases) or rejected (failing test cases) by the language processor. The intention is to promote confidence in the language processor as far as semantic correctness is concerned. The generated test suite in syntactically correct, but the limitations of not taking type checking into account or having a method to evaluate expressions lead to a large number of test sentences that are semantically incorrect. To a degree these limitations are overcome by the combinatorial nature of the generation algorithm that ensures that each possible type correct sentence is generated as well.},
  year = {2015},
  volume = {Honours},
}
Breytenbach JA, Fischer B. Progressive Software Design Tool. 2015;Honours.

Visualising software can be a tedious and cluttered affair with the design process and development often being out of sync. Some development methodologies even largely do away with the design entirely and focus on short bursts of coding and validation to make sure the project is still on the right track. This document focuses on deriving a methodology and subsequent tool to iteratively and progressively expand concepts, the understanding of the project and development cycles. An existing visualisation is adapted to better suit the needs of the designer by providing the ability to view the project from different layers of abstraction in one concise visualisation.

@phdthesis{136,
  author = {J. Breytenbach and Bernd Fischer},
  title = {Progressive Software Design Tool},
  abstract = {Visualising software can be a tedious and cluttered affair with the design process and development often being out of sync. Some development methodologies even largely do away with the design entirely and focus on short bursts of coding and validation to make sure the project is still on the right track. This document focuses on deriving a methodology and subsequent tool to iteratively and progressively expand concepts, the understanding of the project and development cycles. An existing visualisation is adapted to better suit the needs of the designer by providing the ability to view the project from different layers of abstraction in one concise visualisation.},
  year = {2015},
  volume = {Honours},
}
Britz K, Casini G, Meyer T, Moodley K, Sattler U, Varzinczak I. Rational Defeasible Reasoning for Expressive Description Logics. 2015.

In this paper, we enrich description logics (DLs) with non-monotonic reasoning features in a number of ways. We start by investigating a notion of defeasible conditional in the spirit of KLM-style defeasible consequence. In particular, we consider a natural and intuitive semantics for defeasible subsumption in terms of DL interpretations enriched with a preference relation. We propose and investigate syntactic properties (à la Gentzen) for both preferential and rational conditionals and prove representation results for the description logic ALC. This representation result paves the way for more effective decision procedures for defeasible reasoning in DLs. We then move to non-monotonicity in DLs at the level of entailment. We investigate versions of entailment in the context of both preferential and rational subsumption, relate them to preferential and rational closure, and show that computing them can be reduced to classical ALC entailment. This provides further evidence that our semantic constructions are appropriate in a non-monotonic DL setting. One of the barriers to evaluating performance scalability of rational closure is the abscence of naturally occurring DL-based ontologies with defeasible features. We overcome this barrier by devising an approach to introduce defeasible subsumption into classical real world ontologies. This culminates in a set of semi-natural defeasible ontologies that is used, together with a purely artificial set, to test our rational closure algorithms. We found that performance is scalable on the whole with no major bottlenecks.

@misc{130,
  author = {Katarina Britz and Giovanni Casini and Thomas Meyer and Kody Moodley and U. Sattler and Ivan Varzinczak},
  title = {Rational Defeasible Reasoning for Expressive Description Logics},
  abstract = {In this paper, we enrich description logics (DLs) with non-monotonic reasoning features in a number of ways. We start by investigating a notion of defeasible conditional in the spirit of KLM-style defeasible consequence. In particular, we consider a natural and intuitive semantics for defeasible subsumption in terms of DL interpretations enriched with a preference relation. We propose and investigate syntactic properties (à la Gentzen) for both preferential and rational conditionals and prove representation results for the description logic ALC. This representation result paves the way for more effective decision procedures for defeasible reasoning in DLs. We then move to non-monotonicity in DLs at the level of entailment. We investigate versions of entailment in the context of both preferential and rational subsumption, relate them to preferential and rational closure, and show that computing them can be reduced to classical ALC entailment. This provides further evidence that our semantic constructions are appropriate in a non-monotonic DL setting. One of the barriers to evaluating performance scalability of rational closure is the abscence of naturally occurring DL-based ontologies with defeasible features. We overcome this barrier by devising an approach to introduce defeasible subsumption into classical real world ontologies. This culminates in a set of semi-natural defeasible ontologies that is used, together with a purely artificial set, to test our rational closure algorithms. We found that performance is scalable on the whole with no major bottlenecks.},
  year = {2015},
}
Kroon S, Nienaber S, Booysen MJ. A Comparison of Low-Cost Monocular Vision Techniques for Pothole Distance Estimation. In: IEEE Symposium Series on Computational Intelligence: IEEE Symposium on Computational Intelligence in Vehicles and Transportation Systems. ; 2015.

No Abstract

@{129,
  author = {Steve Kroon and S. Nienaber and M.J. Booysen},
  title = {A Comparison of Low-Cost Monocular Vision Techniques for Pothole Distance Estimation},
  abstract = {No Abstract},
  year = {2015},
  journal = {IEEE Symposium Series on Computational Intelligence: IEEE Symposium on Computational Intelligence in Vehicles and Transportation Systems},
  pages = {419-426},
  month = {08/12-10/12},
}
van der Merwe B, Visser WC, van der Merwe H, Nel SEA, Tkachuk O. Environment Modeling Using Runtime Values for JPF-Android. ACM SIGSOFT Software Engineering Notes. 2015;40(6). http://dx.doi.org/10.1145/2830719.2830727.

No Abstract

@article{128,
  author = {Brink van der Merwe and W.C. Visser and Heila van der Merwe and S.E.A. Nel and O. Tkachuk},
  title = {Environment Modeling Using Runtime Values for JPF-Android},
  abstract = {No Abstract},
  year = {2015},
  journal = {ACM SIGSOFT Software Engineering Notes},
  volume = {40},
  pages = {1-5},
  issue = {6},
  publisher = {ACM},
  address = {New York, USA},
  url = {http://dx.doi.org/10.1145/2830719.2830727},
}
Adeleke JA, Moodley D. An Ontology for Proactive Indoor Environmental Quality Monitoring and Control. In: The 2015 Annual Conference of the South African Institute of Computer Scientists and Information Technologists (SAICSIT '15). New York, NY, USA ©2015; 2015.

Proactive monitoring and control of indoor air quality in homes where there are pregnant mothers and infants is essential for healthy development and well-being of children. This is especially true in low income households where cooking practices and exposure to harmful pollutants produced by nearby industries can negatively impact on a healthy home environment. Interdisciplinary expert knowledge is required to make sense of dynamic and complex environmental phenomena from multivariate low level sensor observations and high level human activities to detect health risks and enact decisions about control. We have developed an ontology for indoor environmental quality monitoring and control based on an ongoing real world case study in Durban, South Africa. We implemented an Indoor Air Quality Index and a thermal comfort index which can be automatically determined by reasoning on the ontology. We evaluated the ontology by populating it with test sensor data and showing how it can be queried to analyze health risk situations and determine control actions. Our evaluation shows that the ontology can be used for real world indoor monitoring and control applications in resource constrained settings.

@{127,
  author = {Jude Adeleke and Deshen Moodley},
  title = {An Ontology for Proactive Indoor Environmental Quality Monitoring and Control},
  abstract = {Proactive monitoring and control of indoor air quality in homes where there are pregnant mothers and infants is essential for healthy development and well-being of children. This is especially true in low income households where cooking practices and exposure to harmful pollutants produced by nearby industries can negatively impact on a healthy home environment. Interdisciplinary expert knowledge is required to make sense of dynamic and complex environmental phenomena from multivariate low level sensor observations and high level human activities to detect health risks and enact decisions about control. We have developed an ontology for indoor environmental quality monitoring and control based on an ongoing real world case study in Durban, South Africa. We implemented an Indoor Air Quality Index and a thermal comfort index which can be automatically determined by reasoning on the ontology. We evaluated the ontology by populating it with test sensor data and showing how it can be queried to analyze health risk situations and determine control actions. Our evaluation shows that the ontology can be used for real world indoor monitoring and control applications in resource constrained settings.},
  year = {2015},
  journal = {The 2015 Annual Conference of the South African Institute of Computer Scientists and Information Technologists (SAICSIT '15)},
  month = {28/09-30/09},
  address = {New York, NY, USA ©2015},
  isbn = {978-1-4503-3683-3},
}
  • CSIR
  • DSI
  • Covid-19