<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://projectswiki.eleceng.adelaide.edu.au/projects/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=A1687658</id>
	<title>Projects - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://projectswiki.eleceng.adelaide.edu.au/projects/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=A1687658"/>
	<link rel="alternate" type="text/html" href="https://projectswiki.eleceng.adelaide.edu.au/projects/index.php/Special:Contributions/A1687658"/>
	<updated>2026-05-04T09:26:37Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.31.4</generator>
	<entry>
		<id>https://projectswiki.eleceng.adelaide.edu.au/projects/index.php?title=Projects:2020s1-1410_Speech_Enhancement_for_Automatic_Speech_Recognition&amp;diff=14201</id>
		<title>Projects:2020s1-1410 Speech Enhancement for Automatic Speech Recognition</title>
		<link rel="alternate" type="text/html" href="https://projectswiki.eleceng.adelaide.edu.au/projects/index.php?title=Projects:2020s1-1410_Speech_Enhancement_for_Automatic_Speech_Recognition&amp;diff=14201"/>
		<updated>2020-04-27T01:32:07Z</updated>

		<summary type="html">&lt;p&gt;A1687658: Added newline after sponsorship message&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Category:Projects]]&lt;br /&gt;
[[Category:Final Year Projects]]&lt;br /&gt;
[[Category:2020s1|1410]]&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;&amp;#039;&amp;#039;This project is sponsored by DST Group&amp;#039;&amp;#039;&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
&lt;br /&gt;
Speech recognition is becoming more and more widely used, though the input audio to these systems is rarely clean. A number of techniques&lt;br /&gt;
&amp;lt;ref name=&amp;quot;SEGAN&amp;quot;&amp;gt;Pascual, S., Bonafonte, A. and Serra, J., 2017. &amp;#039;&amp;#039;SEGAN: Speech enhancement generative adversarial network&amp;#039;&amp;#039;. arXiv preprint arXiv:1703.09452&amp;lt;/ref&amp;gt;&lt;br /&gt;
&amp;lt;ref name=&amp;quot;Wave-U-Net&amp;quot;&amp;gt;Stoller, D., Ewert, S. and Dixon, S., 2018. &amp;#039;&amp;#039;Wave-u-net: A multi-scale neural network for end-to-end audio source separation&amp;#039;&amp;#039;. arXiv preprint arXiv:1806.03185&amp;lt;/ref&amp;gt;&lt;br /&gt;
have been developed to reduce the background noise of speech clips, both using deep neural networks, and more traditional filters.&lt;br /&gt;
&lt;br /&gt;
The overall objective of this project is to compare a number of speech enhancement techniques in a fair environment, and to also compare the results of each technique after its output is fed through an automatic speech recogniser.&lt;br /&gt;
&lt;br /&gt;
== Introduction ==&lt;br /&gt;
&amp;#039;&amp;#039;This project follows from work done previously by University of Adelaide students Jordan Parker, Shalin Shah, and Nha Nam (Harry) Nguyen as a summer scholarship project.&amp;#039;&amp;#039;&lt;br /&gt;
&lt;br /&gt;
=== Project team ===&lt;br /&gt;
==== Project students ====&lt;br /&gt;
* Patrick Gregory&lt;br /&gt;
* Zachary Knopoff&lt;br /&gt;
==== Supervisors ====&lt;br /&gt;
* Dr. Said Al-Sarawi&lt;br /&gt;
* Dr. Ahmad Hashemi-Sakhtsari (DST Group)&lt;br /&gt;
* Mr. Paul Jager (DST Group)&lt;br /&gt;
==== Advisors ====&lt;br /&gt;
* Ms. Wei Gao (Emily)&lt;br /&gt;
&lt;br /&gt;
=== Objectives ===&lt;br /&gt;
==== Obtain a dataset ====&lt;br /&gt;
&lt;br /&gt;
Each speech enhancement method has been demonstrated by using different audio datasets depending on the creator(s). Despite this, the general concept is very similar:&lt;br /&gt;
* Collect a large amount of &amp;quot;noise&amp;quot; audio&lt;br /&gt;
* Collect a large amount of clean speech audio - if a transcription exists too, this is called a &amp;#039;&amp;#039;&amp;#039;corpus&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* Combine the two datasets to synthesise noisy speech audio&lt;br /&gt;
&lt;br /&gt;
The goal for this objective is to generate the means of creating a very large (approx 1000hrs) dataset of mixed audio, while maintaining a record of the original clean and noise files - as some methods use these during training. This dataset / generation methodology can then be used by all methods for a fair comparison.&lt;br /&gt;
&lt;br /&gt;
==== Train and optimise ====&lt;br /&gt;
&lt;br /&gt;
A number of promising techniques are selected, and their models trained on the dataset from the previous objective. For non-learning methods, their algorithms may be optimised or altered in some small manner to generate the best results.&lt;br /&gt;
&lt;br /&gt;
==== Compare methods ====&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Background ==&lt;br /&gt;
=== Topic 1 ===&lt;br /&gt;
&lt;br /&gt;
== Method ==&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
== Conclusion ==&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>A1687658</name></author>
		
	</entry>
	<entry>
		<id>https://projectswiki.eleceng.adelaide.edu.au/projects/index.php?title=Projects:2020s1-1410_Speech_Enhancement_for_Automatic_Speech_Recognition&amp;diff=14200</id>
		<title>Projects:2020s1-1410 Speech Enhancement for Automatic Speech Recognition</title>
		<link rel="alternate" type="text/html" href="https://projectswiki.eleceng.adelaide.edu.au/projects/index.php?title=Projects:2020s1-1410_Speech_Enhancement_for_Automatic_Speech_Recognition&amp;diff=14200"/>
		<updated>2020-04-27T01:26:32Z</updated>

		<summary type="html">&lt;p&gt;A1687658: Added sponsorship message to top of page&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Category:Projects]]&lt;br /&gt;
[[Category:Final Year Projects]]&lt;br /&gt;
[[Category:2020s1|1410]]&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;&amp;#039;&amp;#039;This project is sponsored by DST Group&amp;#039;&amp;#039;&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
Speech recognition is becoming more and more widely used, though the input audio to these systems is rarely clean. A number of techniques&lt;br /&gt;
&amp;lt;ref name=&amp;quot;SEGAN&amp;quot;&amp;gt;Pascual, S., Bonafonte, A. and Serra, J., 2017. &amp;#039;&amp;#039;SEGAN: Speech enhancement generative adversarial network&amp;#039;&amp;#039;. arXiv preprint arXiv:1703.09452&amp;lt;/ref&amp;gt;&lt;br /&gt;
&amp;lt;ref name=&amp;quot;Wave-U-Net&amp;quot;&amp;gt;Stoller, D., Ewert, S. and Dixon, S., 2018. &amp;#039;&amp;#039;Wave-u-net: A multi-scale neural network for end-to-end audio source separation&amp;#039;&amp;#039;. arXiv preprint arXiv:1806.03185&amp;lt;/ref&amp;gt;&lt;br /&gt;
have been developed to reduce the background noise of speech clips, both using deep neural networks, and more traditional filters.&lt;br /&gt;
&lt;br /&gt;
The overall objective of this project is to compare a number of speech enhancement techniques in a fair environment, and to also compare the results of each technique after its output is fed through an automatic speech recogniser.&lt;br /&gt;
&lt;br /&gt;
== Introduction ==&lt;br /&gt;
&amp;#039;&amp;#039;This project follows from work done previously by University of Adelaide students Jordan Parker, Shalin Shah, and Nha Nam (Harry) Nguyen as a summer scholarship project.&amp;#039;&amp;#039;&lt;br /&gt;
&lt;br /&gt;
=== Project team ===&lt;br /&gt;
==== Project students ====&lt;br /&gt;
* Patrick Gregory&lt;br /&gt;
* Zachary Knopoff&lt;br /&gt;
==== Supervisors ====&lt;br /&gt;
* Dr. Said Al-Sarawi&lt;br /&gt;
* Dr. Ahmad Hashemi-Sakhtsari (DST Group)&lt;br /&gt;
* Mr. Paul Jager (DST Group)&lt;br /&gt;
==== Advisors ====&lt;br /&gt;
* Ms. Wei Gao (Emily)&lt;br /&gt;
&lt;br /&gt;
=== Objectives ===&lt;br /&gt;
==== Obtain a dataset ====&lt;br /&gt;
&lt;br /&gt;
Each speech enhancement method has been demonstrated by using different audio datasets depending on the creator(s). Despite this, the general concept is very similar:&lt;br /&gt;
* Collect a large amount of &amp;quot;noise&amp;quot; audio&lt;br /&gt;
* Collect a large amount of clean speech audio - if a transcription exists too, this is called a &amp;#039;&amp;#039;&amp;#039;corpus&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* Combine the two datasets to synthesise noisy speech audio&lt;br /&gt;
&lt;br /&gt;
The goal for this objective is to generate the means of creating a very large (approx 1000hrs) dataset of mixed audio, while maintaining a record of the original clean and noise files - as some methods use these during training. This dataset / generation methodology can then be used by all methods for a fair comparison.&lt;br /&gt;
&lt;br /&gt;
==== Train and optimise ====&lt;br /&gt;
&lt;br /&gt;
A number of promising techniques are selected, and their models trained on the dataset from the previous objective. For non-learning methods, their algorithms may be optimised or altered in some small manner to generate the best results.&lt;br /&gt;
&lt;br /&gt;
==== Compare methods ====&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Background ==&lt;br /&gt;
=== Topic 1 ===&lt;br /&gt;
&lt;br /&gt;
== Method ==&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
== Conclusion ==&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>A1687658</name></author>
		
	</entry>
	<entry>
		<id>https://projectswiki.eleceng.adelaide.edu.au/projects/index.php?title=Projects:2020s1-1410_Speech_Enhancement_for_Automatic_Speech_Recognition&amp;diff=14072</id>
		<title>Projects:2020s1-1410 Speech Enhancement for Automatic Speech Recognition</title>
		<link rel="alternate" type="text/html" href="https://projectswiki.eleceng.adelaide.edu.au/projects/index.php?title=Projects:2020s1-1410_Speech_Enhancement_for_Automatic_Speech_Recognition&amp;diff=14072"/>
		<updated>2020-04-21T00:23:55Z</updated>

		<summary type="html">&lt;p&gt;A1687658: Filled in template a little, should be enough for the time being.&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Category:Projects]]&lt;br /&gt;
[[Category:Final Year Projects]]&lt;br /&gt;
[[Category:2020s1|1410]]&lt;br /&gt;
Speech recognition is becoming more and more widely used, though the input audio to these systems is rarely clean. A number of techniques&lt;br /&gt;
&amp;lt;ref name=&amp;quot;SEGAN&amp;quot;&amp;gt;Pascual, S., Bonafonte, A. and Serra, J., 2017. &amp;#039;&amp;#039;SEGAN: Speech enhancement generative adversarial network&amp;#039;&amp;#039;. arXiv preprint arXiv:1703.09452&amp;lt;/ref&amp;gt;&lt;br /&gt;
&amp;lt;ref name=&amp;quot;Wave-U-Net&amp;quot;&amp;gt;Stoller, D., Ewert, S. and Dixon, S., 2018. &amp;#039;&amp;#039;Wave-u-net: A multi-scale neural network for end-to-end audio source separation&amp;#039;&amp;#039;. arXiv preprint arXiv:1806.03185&amp;lt;/ref&amp;gt;&lt;br /&gt;
have been developed to reduce the background noise of speech clips, both using deep neural networks, and more traditional filters.&lt;br /&gt;
&lt;br /&gt;
The overall objective of this project is to compare a number of speech enhancement techniques in a fair environment, and to also compare the results of each technique after its output is fed through an automatic speech recogniser.&lt;br /&gt;
&lt;br /&gt;
== Introduction ==&lt;br /&gt;
&amp;#039;&amp;#039;This project follows from work done previously by University of Adelaide students Jordan Parker, Shalin Shah, and Nha Nam (Harry) Nguyen as a summer scholarship project.&amp;#039;&amp;#039;&lt;br /&gt;
&lt;br /&gt;
=== Project team ===&lt;br /&gt;
==== Project students ====&lt;br /&gt;
* Patrick Gregory&lt;br /&gt;
* Zachary Knopoff&lt;br /&gt;
==== Supervisors ====&lt;br /&gt;
* Dr. Said Al-Sarawi&lt;br /&gt;
* Dr. Ahmad Hashemi-Sakhtsari (DST Group)&lt;br /&gt;
* Mr. Paul Jager (DST Group)&lt;br /&gt;
==== Advisors ====&lt;br /&gt;
* Ms. Wei Gao (Emily)&lt;br /&gt;
&lt;br /&gt;
=== Objectives ===&lt;br /&gt;
==== Obtain a dataset ====&lt;br /&gt;
&lt;br /&gt;
Each speech enhancement method has been demonstrated by using different audio datasets depending on the creator(s). Despite this, the general concept is very similar:&lt;br /&gt;
* Collect a large amount of &amp;quot;noise&amp;quot; audio&lt;br /&gt;
* Collect a large amount of clean speech audio - if a transcription exists too, this is called a &amp;#039;&amp;#039;&amp;#039;corpus&amp;#039;&amp;#039;&amp;#039;&lt;br /&gt;
* Combine the two datasets to synthesise noisy speech audio&lt;br /&gt;
&lt;br /&gt;
The goal for this objective is to generate the means of creating a very large (approx 1000hrs) dataset of mixed audio, while maintaining a record of the original clean and noise files - as some methods use these during training. This dataset / generation methodology can then be used by all methods for a fair comparison.&lt;br /&gt;
&lt;br /&gt;
==== Train and optimise ====&lt;br /&gt;
&lt;br /&gt;
A number of promising techniques are selected, and their models trained on the dataset from the previous objective. For non-learning methods, their algorithms may be optimised or altered in some small manner to generate the best results.&lt;br /&gt;
&lt;br /&gt;
==== Compare methods ====&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Background ==&lt;br /&gt;
=== Topic 1 ===&lt;br /&gt;
&lt;br /&gt;
== Method ==&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
== Conclusion ==&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;/div&gt;</summary>
		<author><name>A1687658</name></author>
		
	</entry>
	<entry>
		<id>https://projectswiki.eleceng.adelaide.edu.au/projects/index.php?title=Projects:2020s1-1410_Speech_Enhancement_for_Automatic_Speech_Recognition&amp;diff=14067</id>
		<title>Projects:2020s1-1410 Speech Enhancement for Automatic Speech Recognition</title>
		<link rel="alternate" type="text/html" href="https://projectswiki.eleceng.adelaide.edu.au/projects/index.php?title=Projects:2020s1-1410_Speech_Enhancement_for_Automatic_Speech_Recognition&amp;diff=14067"/>
		<updated>2020-04-20T22:14:04Z</updated>

		<summary type="html">&lt;p&gt;A1687658: Used skeleton template&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[Category:Projects]]&lt;br /&gt;
[[Category:Final Year Projects]]&lt;br /&gt;
[[Category:2020s1|1410]]&lt;br /&gt;
Abstract here&lt;br /&gt;
== Introduction ==&lt;br /&gt;
This project is a continuation of work done previously by University of Adelaide students Jordan Parker, Shalin Shah, and Nha Nam (Harry) Nguyen.&lt;br /&gt;
&lt;br /&gt;
=== Project team ===&lt;br /&gt;
==== Project students ====&lt;br /&gt;
* Patrick Gregory&lt;br /&gt;
* Zachary Knopoff&lt;br /&gt;
==== Supervisors ====&lt;br /&gt;
* Dr. Said Al-Sarawi&lt;br /&gt;
* Dr. Ahmad Hashemi-Sakhtsari (DST Group)&lt;br /&gt;
* Mr. Paul Jager (DST Group)&lt;br /&gt;
==== Advisors ====&lt;br /&gt;
* Ms. Wei Gao (Emily)&lt;br /&gt;
&lt;br /&gt;
=== Objectives ===&lt;br /&gt;
Set of objectives&lt;br /&gt;
&lt;br /&gt;
== Background ==&lt;br /&gt;
=== Topic 1 ===&lt;br /&gt;
&lt;br /&gt;
== Method ==&lt;br /&gt;
&lt;br /&gt;
== Results ==&lt;br /&gt;
&lt;br /&gt;
== Conclusion ==&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
[1] a, b, c, &amp;quot;Simple page&amp;quot;, In Proceedings of the Conference of Simpleness, 2010.&lt;br /&gt;
&lt;br /&gt;
[2] ...&lt;/div&gt;</summary>
		<author><name>A1687658</name></author>
		
	</entry>
	<entry>
		<id>https://projectswiki.eleceng.adelaide.edu.au/projects/index.php?title=Projects:2020s1-1410_Speech_Enhancement_for_Automatic_Speech_Recognition&amp;diff=13955</id>
		<title>Projects:2020s1-1410 Speech Enhancement for Automatic Speech Recognition</title>
		<link rel="alternate" type="text/html" href="https://projectswiki.eleceng.adelaide.edu.au/projects/index.php?title=Projects:2020s1-1410_Speech_Enhancement_for_Automatic_Speech_Recognition&amp;diff=13955"/>
		<updated>2020-03-25T03:20:12Z</updated>

		<summary type="html">&lt;p&gt;A1687658: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This project is a continuation of work done previously by University of Adelaide students Jordan Parker, Shalin Shah, and Nha Nam (Harry) Nguyen.&lt;br /&gt;
&lt;br /&gt;
Honours project 1410 is being undertaken by Patrick Gregory and Zachary Knopoff, and supervised by Dr Said Al-Sarawi (the University of Adelaide), Mr Paul Jager (DST Group), and Dr Ahmad Hashemi-Sakhtsari (DST Group) with help from Ms Wei ”Emily” Gao.&lt;/div&gt;</summary>
		<author><name>A1687658</name></author>
		
	</entry>
	<entry>
		<id>https://projectswiki.eleceng.adelaide.edu.au/projects/index.php?title=Projects:2020s1-1410_Speech_Enhancement_for_Automatic_Speech_Recognition&amp;diff=13944</id>
		<title>Projects:2020s1-1410 Speech Enhancement for Automatic Speech Recognition</title>
		<link rel="alternate" type="text/html" href="https://projectswiki.eleceng.adelaide.edu.au/projects/index.php?title=Projects:2020s1-1410_Speech_Enhancement_for_Automatic_Speech_Recognition&amp;diff=13944"/>
		<updated>2020-03-23T06:53:15Z</updated>

		<summary type="html">&lt;p&gt;A1687658: Created page with &amp;quot;This project is a continuation of work done previously by University of Adelaide students Jordan Parker, Shalin Shah, and Nha Nam (Harry) Nguyen.  Honours project 1410 is bein...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This project is a continuation of work done previously by University of Adelaide students Jordan Parker, Shalin Shah, and Nha Nam (Harry) Nguyen.&lt;br /&gt;
&lt;br /&gt;
Honours project 1410 is being undertaken by Patrick Gregory and Zachary Knopoff, and supervised by Dr Said Al-Sarawi (the University of Adelaide) and Dr Ahmad Hashemi-Sakhtsari (DST Group) with help from Ms Wei ”Emily” Gao.&lt;/div&gt;</summary>
		<author><name>A1687658</name></author>
		
	</entry>
</feed>