<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://wiki.soundscapes.nuclio.org:443/w/index.php?action=history&amp;feed=atom&amp;title=Translations%3ASonification_in_practice%2F44%2Fen</id>
	<title>Translations:Sonification in practice/44/en - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.soundscapes.nuclio.org:443/w/index.php?action=history&amp;feed=atom&amp;title=Translations%3ASonification_in_practice%2F44%2Fen"/>
	<link rel="alternate" type="text/html" href="https://wiki.soundscapes.nuclio.org:443/w/index.php?title=Translations:Sonification_in_practice/44/en&amp;action=history"/>
	<updated>2026-04-13T06:08:36Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.41.0</generator>
	<entry>
		<id>https://wiki.soundscapes.nuclio.org:443/w/index.php?title=Translations:Sonification_in_practice/44/en&amp;diff=1963&amp;oldid=prev</id>
		<title>FuzzyBot: Importing a new version from external source</title>
		<link rel="alternate" type="text/html" href="https://wiki.soundscapes.nuclio.org:443/w/index.php?title=Translations:Sonification_in_practice/44/en&amp;diff=1963&amp;oldid=prev"/>
		<updated>2026-04-06T14:46:26Z</updated>

		<summary type="html">&lt;p&gt;Importing a new version from external source&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 14:46, 6 April 2026&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l1&quot;&gt;Line 1:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 1:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Our sense of hearing is able to focus on a particular sound in between many others (see the “cocktail party effect”) &lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;(&lt;/del&gt;Arons, B., &lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;1992&lt;/del&gt;) based on timbre. Our auditory system can process information at a far higher rate than our visual system. For example, while video typically updates at 60 frames per second (60 Hz), standard audio is sampled at 44,100 times per second (44.1 kHz). This means that even a single, brief spike in an audio signal—lasting just one sample—is instantly perceived as a distinct &quot;click.&quot; As a result, hearing allows us to monitor multiple layers of information simultaneously, often more efficiently than through visual perception alone&lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;. (&lt;/del&gt;Kramer, G. et al., &lt;del style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;1999&lt;/del&gt;).&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Our sense of hearing is able to focus on a particular sound in between many others (see the “cocktail party effect”) &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&amp;lt;ref&amp;gt;&lt;/ins&gt;Arons, B. &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;(1992). A review of the cocktail party effect. Journal of the American Voice I/O society&lt;/ins&gt;, &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;12(7&lt;/ins&gt;)&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;, 35-50.&amp;lt;/ref&amp;gt; &lt;/ins&gt;based on timbre. Our auditory system can process information at a far higher rate than our visual system. For example, while video typically updates at 60 frames per second (60 Hz), standard audio is sampled at 44,100 times per second (44.1 kHz). This means that even a single, brief spike in an audio signal—lasting just one sample—is instantly perceived as a distinct &quot;click.&quot; As a result, hearing allows us to monitor multiple layers of information simultaneously, often more efficiently than through visual perception alone &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&amp;lt;ref&amp;gt;&lt;/ins&gt;Kramer, G.&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;, Walker, B. N., Bonebright, T., Cook, P., Flowers, J., Miner, N., &lt;/ins&gt;et al. &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;(1999). The Sonification Report: Status of the Field and Research Agenda. Report prepared for the National Science Foundation by members of the International Community for Auditory Display. Santa Fe&lt;/ins&gt;, &lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;NM: International Community for Auditory Display (ICAD&lt;/ins&gt;)&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;.&amp;lt;/ref&amp;gt;&lt;/ins&gt;.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>FuzzyBot</name></author>
	</entry>
	<entry>
		<id>https://wiki.soundscapes.nuclio.org:443/w/index.php?title=Translations:Sonification_in_practice/44/en&amp;diff=1524&amp;oldid=prev</id>
		<title>FuzzyBot: Importing a new version from external source</title>
		<link rel="alternate" type="text/html" href="https://wiki.soundscapes.nuclio.org:443/w/index.php?title=Translations:Sonification_in_practice/44/en&amp;diff=1524&amp;oldid=prev"/>
		<updated>2026-04-06T12:01:57Z</updated>

		<summary type="html">&lt;p&gt;Importing a new version from external source&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;Our sense of hearing is able to focus on a particular sound in between many others (see the “cocktail party effect”) (Arons, B., 1992) based on timbre. Our auditory system can process information at a far higher rate than our visual system. For example, while video typically updates at 60 frames per second (60 Hz), standard audio is sampled at 44,100 times per second (44.1 kHz). This means that even a single, brief spike in an audio signal—lasting just one sample—is instantly perceived as a distinct &amp;quot;click.&amp;quot; As a result, hearing allows us to monitor multiple layers of information simultaneously, often more efficiently than through visual perception alone. (Kramer, G. et al., 1999).&lt;/div&gt;</summary>
		<author><name>FuzzyBot</name></author>
	</entry>
</feed>