<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://mw.hh.se/caisr/index.php?action=history&amp;feed=atom&amp;title=Evolving_Kolmogorov-Arnold_Networks</id>
	<title>Evolving Kolmogorov-Arnold Networks - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://mw.hh.se/caisr/index.php?action=history&amp;feed=atom&amp;title=Evolving_Kolmogorov-Arnold_Networks"/>
	<link rel="alternate" type="text/html" href="https://mw.hh.se/caisr/index.php?title=Evolving_Kolmogorov-Arnold_Networks&amp;action=history"/>
	<updated>2026-04-04T17:23:11Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.35.13</generator>
	<entry>
		<id>https://mw.hh.se/caisr/index.php?title=Evolving_Kolmogorov-Arnold_Networks&amp;diff=5437&amp;oldid=prev</id>
		<title>Islab at 14:32, 23 September 2024</title>
		<link rel="alternate" type="text/html" href="https://mw.hh.se/caisr/index.php?title=Evolving_Kolmogorov-Arnold_Networks&amp;diff=5437&amp;oldid=prev"/>
		<updated>2024-09-23T14:32:01Z</updated>

		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left diff-editfont-monospace&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 14:32, 23 September 2024&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l7&quot; &gt;Line 7:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 7:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;}}&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;}}&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Kolmogorov-Arnold Networks (KANs), recently proposed by researchers at MIT, present a promising alternative to traditional Multi-Layer Perceptrons (MLPs), demonstrating superior performance in terms of both accuracy and interpretability. The goal of this project is to further advance the architecture of KANs by enhancing their computational graph through various research directions. These include optimizing loss functions, refining activation functions, developing more effective initialization schemes, and improving learning processes.&lt;/div&gt;&lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt; &lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Kolmogorov-Arnold Networks (KANs), recently proposed by researchers at MIT, present a promising alternative to traditional Multi-Layer Perceptrons (MLPs), demonstrating superior performance in terms of both accuracy and interpretability. The goal of this project is to further advance the architecture of KANs by enhancing their computational graph through various research directions. These include optimizing loss functions, refining activation functions, developing more effective initialization schemes, and improving learning processes.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;https://arxiv.org/abs/2404.19756&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;https://arxiv.org/abs/2407.13044&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt; &lt;/td&gt;&lt;td class=&#039;diff-marker&#039;&gt;+&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;https://arxiv.org/abs/2407.20667&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Islab</name></author>
	</entry>
	<entry>
		<id>https://mw.hh.se/caisr/index.php?title=Evolving_Kolmogorov-Arnold_Networks&amp;diff=5436&amp;oldid=prev</id>
		<title>Islab: Created page with &quot;{{StudentProjectTemplate |Summary=This project aims to enhance the architecture of Kolmogorov-Arnold Networks (KANs) by optimizing key components such as loss functions, activ...&quot;</title>
		<link rel="alternate" type="text/html" href="https://mw.hh.se/caisr/index.php?title=Evolving_Kolmogorov-Arnold_Networks&amp;diff=5436&amp;oldid=prev"/>
		<updated>2024-09-23T14:29:28Z</updated>

		<summary type="html">&lt;p&gt;Created page with &amp;quot;{{StudentProjectTemplate |Summary=This project aims to enhance the architecture of Kolmogorov-Arnold Networks (KANs) by optimizing key components such as loss functions, activ...&amp;quot;&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;{{StudentProjectTemplate&lt;br /&gt;
|Summary=This project aims to enhance the architecture of Kolmogorov-Arnold Networks (KANs) by optimizing key components such as loss functions, activation functions, initialization methods, and learning processes to improve their performance and interpretability.&lt;br /&gt;
|TimeFrame=Fall 2024&lt;br /&gt;
|Supervisor=Mohammed Ghaith Altarabichi&lt;br /&gt;
|Level=Master&lt;br /&gt;
|Status=Open&lt;br /&gt;
}}&lt;br /&gt;
Kolmogorov-Arnold Networks (KANs), recently proposed by researchers at MIT, present a promising alternative to traditional Multi-Layer Perceptrons (MLPs), demonstrating superior performance in terms of both accuracy and interpretability. The goal of this project is to further advance the architecture of KANs by enhancing their computational graph through various research directions. These include optimizing loss functions, refining activation functions, developing more effective initialization schemes, and improving learning processes.&lt;/div&gt;</summary>
		<author><name>Islab</name></author>
	</entry>
</feed>