<?xml version="1.0" encoding="ISO-8859-1"?><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<front>
<journal-meta>
<journal-id>2518-4431</journal-id>
<journal-title><![CDATA[Investigación & Desarrollo]]></journal-title>
<abbrev-journal-title><![CDATA[Inv. y Des.]]></abbrev-journal-title>
<issn>2518-4431</issn>
<publisher>
<publisher-name><![CDATA[UNIVERSIDAD PRIVADA BOLIVIANA]]></publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id>S2518-44312019000100008</article-id>
<title-group>
<article-title xml:lang="en"><![CDATA[TEMPORAL DICTIONARY LEARNING FOR TIME-SERIES DECOMPOSITION]]></article-title>
<article-title xml:lang="es"><![CDATA[APRENDIZAJE DE DICCIONARIOS TEMPORALES PARA LA DESCOMPOSICIÓN DE SERIES DE TIEMPOS]]></article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname><![CDATA[Bürger]]></surname>
<given-names><![CDATA[Jens]]></given-names>
</name>
<xref ref-type="aff" rid="A01"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname><![CDATA[Calvimontes]]></surname>
<given-names><![CDATA[Jorge]]></given-names>
</name>
<xref ref-type="aff" rid="A01"/>
</contrib>
</contrib-group>
<aff id="A01">
<institution><![CDATA[,Universidad Privada Boliviana Institute for Computational Intelligence (ICI) ]]></institution>
<addr-line><![CDATA[ ]]></addr-line>
</aff>
<pub-date pub-type="pub">
<day>00</day>
<month>00</month>
<year>2019</year>
</pub-date>
<pub-date pub-type="epub">
<day>00</day>
<month>00</month>
<year>2019</year>
</pub-date>
<volume>19</volume>
<numero>1</numero>
<fpage>105</fpage>
<lpage>112</lpage>
<copyright-statement/>
<copyright-year/>
<self-uri xlink:href="http://www.scielo.org.bo/scielo.php?script=sci_arttext&amp;pid=S2518-44312019000100008&amp;lng=en&amp;nrm=iso"></self-uri><self-uri xlink:href="http://www.scielo.org.bo/scielo.php?script=sci_abstract&amp;pid=S2518-44312019000100008&amp;lng=en&amp;nrm=iso"></self-uri><self-uri xlink:href="http://www.scielo.org.bo/scielo.php?script=sci_pdf&amp;pid=S2518-44312019000100008&amp;lng=en&amp;nrm=iso"></self-uri><abstract abstract-type="short" xml:lang="en"><p><![CDATA[Dictionary Learning (DL) is a feature learning method that derives a finite collection of dictionary elements (atoms) from a given dataset. These atoms are small characteristic features representing recurring patterns within the data. A dictionary therefore is a compact representation of complex or large scale datasets. In this paper we investigate DL for temporal signal decomposition and reconstruction. Decomposition is a common method in time-series forecasting to separate a complex composite signal into different frequency components as to reduce forecasting complexity. By representing characteristic features, we consider dictionary elements to function as filters for the decomposition of temporal signals. Rather than simple filters with clearly defined frequency spectra, we hypothesize for dictionaries and the corresponding reconstructions to act as more complex filters. Training different dictionaries then permits to decompose the original signal into different components. This makes it a potential alternative to existing decomposition methods. We apply a known sparse DL algorithm to a wind speed dataset and investigate decomposition quality and filtering characteristics. Reconstruction accuracy serves as a proxy for evaluating the dictionary quality and a coherence analysis is performed to analyze how different dictionary configurations lead to different filtering characteristics. The results of the presented work demonstrate how learned features of different dictionaries represent transfer functions corresponding to frequency components found in the original data. Based on finite sets of atoms, dictionaries provide a deterministic mechanism to decompose a signal into various reconstructions and their respective remainders. These insights have direct application to the investigation and development of advanced signal decomposition and forecasting techniques.]]></p></abstract>
<abstract abstract-type="short" xml:lang="es"><p><![CDATA[Dictionary Learning (DL) es un método de aprendizaje de características que deriva una colección finita de elementos del diccionario (átomos) de un conjunto de datos determinado. Estos átomos son pequeños rasgos característicos que representan patrones recurrentes dentro de los datos. Por lo tanto, un diccionario es una representación compacta de conjuntos de datos complejos o de gran escala. En este trabajo investigamos DL para la descomposición y reconstrucción de señales temporales. La descomposición es un método común en el pronóstico de series de tiempo para separar una señal compuesta compleja en diferentes componentes de frecuencia para reducir la complejidad del pronóstico. Al representar los rasgos característicos, consideramos que los elementos del diccionario funcionan como filtros para la descomposición de las señales temporales. En lugar de filtros simples con espectros de frecuencia claramente definidos, planteamos la hipótesis de que los diccionarios y las reconstrucciones correspondientes actúen como filtros más complejos. La capacitación de diferentes diccionarios permite luego descomponer la señal original en diferentes componentes. Esto lo convierte en una alternativa potencial a los métodos de descomposición existentes. Aplicamos un algoritmo de DL disperso conocido a un conjunto de datos de velocidad del viento e investigamos la calidad de descomposición y las características de filtrado. La precisión de la reconstrucción sirve como un proxy para evaluar la calidad del diccionario y se realiza un análisis de coherencia para analizar cómo diferentes configuraciones de diccionarios llevan a diferentes características de filtrado. Los resultados del trabajo presentado demuestran cómo las características aprendidas de diferentes diccionarios representan funciones de transferencia correspondientes a los componentes de frecuencia encontrados en los datos originales. Basados &#8203;&#8203;en conjuntos finitos de átomos, los diccionarios proporcionan un mecanismo determinista para descomponer una señal en varias reconstrucciones y sus respectivos residuos. Estos conocimientos tienen una aplicación directa en la investigación y el desarrollo de técnicas avanzadas de descomposición de señales y pronóstico.]]></p></abstract>
<kwd-group>
<kwd lng="en"><![CDATA[Dictionary Learning]]></kwd>
<kwd lng="en"><![CDATA[SAILnet]]></kwd>
<kwd lng="en"><![CDATA[Time-Series]]></kwd>
<kwd lng="en"><![CDATA[Decomposition]]></kwd>
<kwd lng="es"><![CDATA[Dictionary Learning]]></kwd>
<kwd lng="es"><![CDATA[SAILnet]]></kwd>
<kwd lng="es"><![CDATA[Series de Tiempo]]></kwd>
<kwd lng="es"><![CDATA[Descomposición]]></kwd>
</kwd-group>
</article-meta>
</front><body><![CDATA[ <p align=left><font color="#800000" size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>DOI:</b> 10.23881/idupbo.019.1-7i</font></p>     <p align=right><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>ART&Iacute;CULOS &ndash; INGENIER&Iacute;AS</b></font></p>     <p align=center><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>&nbsp;</b></font></p>     <p align=center><font size="4" face="Verdana, Arial, Helvetica, sans-serif"><b>TEMPORAL DICTIONARY LEARNING FOR TIME-SERIES   DECOMPOSITION</b></font></p>     <p>&nbsp;</p>     <p align=center><font size="3" face="Verdana, Arial, Helvetica, sans-serif"><b>APRENDIZAJE DE DICCIONARIOS TEMPORALES PARA LA   DESCOMPOSICI&Oacute;N DE SERIES DE TIEMPOS</b></font></p>     <p align=center>&nbsp;</p>     <p align=center>&nbsp;</p>     <p align=center><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>Jens B&uuml;rger and Jorge Calvimontes</b></font></p>     <p align=center><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><i>Institute for Computational Intelligence </i>(ICI)</font>    ]]></body>
<body><![CDATA[<br> <font size="2" face="Verdana, Arial, Helvetica, sans-serif"><i>Universidad Privada Boliviana</i></font>    <br> <font size="2" face="Verdana, Arial, Helvetica, sans-serif"><a href="mailto:jensburger@upb.edu">jensburger@upb.edu</a></font></p>     <p align=center><font size="2" face="Verdana, Arial, Helvetica, sans-serif">(Recibido el 04 junio 2019,   aceptado para publicaci&oacute;n el 26 junio 2019)</font></p>     <p align=center>&nbsp;</p>     <p>&nbsp;</p> <hr noshade>     <p><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>ABSTRACT&nbsp;</b></font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><i>Dictionary Learning</i> (DL) is a feature learning method that derives a   finite collection of dictionary elements (atoms) from a given dataset. These   atoms are small characteristic features representing recurring patterns within   the data. A dictionary therefore is a compact representation of complex or   large scale datasets. In this paper we investigate DL for temporal signal   decomposition and reconstruction. Decomposition is a common method in   time-series forecasting to separate a complex composite signal into different   frequency components as to reduce forecasting complexity. By representing   characteristic features, we consider dictionary elements to function as filters   for the decomposition of temporal signals. Rather than simple filters with   clearly defined frequency spectra, we hypothesize for dictionaries and the   corresponding reconstructions to act as more complex filters. Training   different dictionaries then permits to decompose the original signal into   different components. This makes it a potential alternative to existing decomposition   methods. We apply a known sparse DL algorithm to a wind speed dataset and   investigate decomposition quality and filtering characteristics. Reconstruction   accuracy serves as a proxy for evaluating the dictionary quality and a   coherence analysis is performed to analyze how different dictionary   configurations lead to different filtering characteristics. The results of the   presented work demonstrate how learned features of different dictionaries   represent transfer functions corresponding to frequency components found in the   original data. Based on finite sets of atoms, dictionaries provide a   deterministic mechanism to decompose a signal into various reconstructions and   their respective remainders. These insights have direct application to the investigation   and development of advanced signal decomposition and forecasting techniques.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>Keywords:</b> Dictionary Learning, SAILnet, Time-Series, Decomposition.</font></p> <hr align="JUSTIFY" noshade>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>RESUMEN&nbsp;</b></font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">Dictionary   Learning (DL) es un m&eacute;todo de aprendizaje de caracter&iacute;sticas que deriva una colecci&oacute;n   finita de elementos del diccionario (&aacute;tomos) de un conjunto de datos   determinado. Estos &aacute;tomos son peque&ntilde;os rasgos caracter&iacute;sticos que representan   patrones recurrentes dentro de los datos. Por lo tanto, un diccionario es una   representaci&oacute;n compacta de conjuntos de datos complejos o de gran escala. En   este trabajo investigamos DL para la descomposici&oacute;n y reconstrucci&oacute;n de se&ntilde;ales   temporales. La descomposici&oacute;n es un m&eacute;todo com&uacute;n en el pron&oacute;stico de series de   tiempo para separar una se&ntilde;al compuesta compleja en diferentes componentes de   frecuencia para reducir la complejidad del pron&oacute;stico. Al representar los   rasgos caracter&iacute;sticos, consideramos que los elementos del diccionario   funcionan como filtros para la descomposici&oacute;n de las se&ntilde;ales temporales. En   lugar de filtros simples con espectros de frecuencia claramente definidos,   planteamos la hip&oacute;tesis de que los diccionarios y las reconstrucciones   correspondientes act&uacute;en como filtros m&aacute;s complejos. La capacitaci&oacute;n de   diferentes diccionarios permite luego descomponer la se&ntilde;al original en   diferentes componentes. Esto lo convierte en una alternativa potencial a los   m&eacute;todos de descomposici&oacute;n existentes. Aplicamos un algoritmo de DL disperso   conocido a un conjunto de datos de velocidad del viento e investigamos la   calidad de descomposici&oacute;n y las caracter&iacute;sticas de filtrado. La precisi&oacute;n de la   reconstrucci&oacute;n sirve como un proxy para evaluar la calidad del diccionario y se   realiza un an&aacute;lisis de coherencia para analizar c&oacute;mo diferentes configuraciones   de diccionarios llevan a diferentes caracter&iacute;sticas de filtrado. Los resultados   del trabajo presentado demuestran c&oacute;mo las caracter&iacute;sticas aprendidas de   diferentes diccionarios representan funciones de transferencia correspondientes   a los componentes de frecuencia encontrados en los datos originales. Basados   &#8203;&#8203;en conjuntos finitos de &aacute;tomos, los diccionarios proporcionan un mecanismo   determinista para descomponer una se&ntilde;al en varias reconstrucciones y sus   respectivos residuos. Estos conocimientos tienen una aplicaci&oacute;n directa en la   investigaci&oacute;n y el desarrollo de t&eacute;cnicas avanzadas de descomposici&oacute;n de   se&ntilde;ales y pron&oacute;stico.</font></p>     ]]></body>
<body><![CDATA[<p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>Palabras   Clave:</b> Dictionary Learning, SAILnet,   Series de Tiempo, Descomposici&oacute;n.</font></p> <hr align="JUSTIFY" noshade>     <p align="justify">&nbsp;</p>     <p align="justify">&nbsp;</p>     <p align="justify"><font size="3" face="Verdana, Arial, Helvetica, sans-serif"><b>1. INTRODUCTION</b></font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">Mammalian visual and auditory processing, located in the visual and   auditory cortices, rely on a layered reconstruction of complex sensory   information based on a finite set of simple features. In the primary visual   cortex (V1) features such as edges are often represented by Gabor filters.   Likely the most well-known example for such a layered sensory information   processing structure is deep convolutional neural networks (deep learning) [1].   Another class of algorithms for feature learning is known as <i>Dictionary     Learning</i> (DL) [2]. While deep learning is most commonly used as a   supervised learning scheme for classification tasks, DL is a mostly   unsupervised learning approach that aims to find a set of features permitting   sparse representation of the data at hand. Such a sparse representation (or   sparse code) allows reconstructing, with some degree of error <i>&epsilon;</i>, any   complex signal with a linear combination of few features. Such a sparse   representation of complex signals holds significant potential in signal   decomposition where one wants to find a small set of signal components that   describe underlying patterns and therefore permit separate analysis and   processing of different types of patterns. Especially in time-series processing   decomposition methods are standard tools in analysis and forecasting [3]. However,   many decomposition methods are rather simplistic in relation to the signals   they are trying to decompose. For example, STL decomposition (<i>Seasonal and     Trend decomposition using Loess</i>) separates a signal into trend, seasonality   and a remainder. Trend is an extremely low-frequency component, seasonality a   periodic signal of frequency <i>f</i>, and the remainder everything else. For   complex signals the remainder often represents a significant part of the   overall signals amplitude and it is not obvious if the remainder simply   represents noise or more complex patterns.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">As mentioned above, DL has found applications in visual as well as   auditory signal processing [4,5]. Due to DL's close relation to biological   information processing principles much of the work has focused on applications   related to human sensory information such as images and speech. However,   auditory signals more broadly understood as (multi-channel) temporal signals   also opens up the possibility to apply auditory (or temporal) dictionary   learning to other signal classes. One example of temporal dictionary learning   applied to <i>electroencephalogram</i> (EEG) data has been demonstrated to   outperform dictionaries based on formal Gabor functions in their representative   power [6]. As was argued by the authors, the flexibility of capturing diverse   patterns directly from data holds advantages over formal mathematical   definitions of dictionary atoms. This observation motivates the application of   temporal DL to other complex time-series data.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">The problem of DL is typically understood as the reconstruction of an   original signal <i>X</i> so that <i>X = DR + &epsilon;</i>, with <i>D</i> being the   dictionary consisting of a finite set of atoms, <i>R</i> a sparse activity   pattern corresponding atoms in the dictionary <i>D</i>, and <i>&epsilon;</i> an error   term representing the reconstruction error. If we reinterpret this function in   light of decomposition, we can assert that <i>&epsilon;</i> represents a signal   component that has been filtered out by the reconstruction process. If we   assume that any reconstructed signal <img width=72 height=15 src="/img/revistas/riyd/v19n1/a08_image001.gif">, with <img width=17 height=13 src="/img/revistas/riyd/v19n1/a08_image002.gif">&nbsp;being the filter transfer function of related to a particular   dictionary, then we can assert that the reconstruction of a dictionary acts as   a decomposition through extracting a specific component from the original   signal. Rather than being a decomposition by means of a clearly defined   mathematical function, the dictionary poses a complex filter based on   particular patterns present in the actual data. </font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">In this paper we adopt the <i>SAILnet</i> dictionary learning algorithm   [7], originally used for learning V1 features, to learn temporal patterns of a   wind speed dataset [8]. Precisely, we are interested in temporal DL as a   decomposition method which will be analyzed based on a set of dictionaries   trained with different parameters. The underlying hypothesis is that, in   comparison to standard decomposition methods, reconstruction through   dictionaries will represent more complex (or less obvious) decompositions of   the original signal. We will analyze the quality of the dictionaries through an   evaluation of their reconstruction errors and the decomposition characteristics   by means of coherence analysis that describe the different dictionaries through   their filtering characteristics. </font></p>     <p align="justify">&nbsp;</p>     <p align="justify"><font size="3" face="Verdana, Arial, Helvetica, sans-serif"><b>2. METHODOLOGY</b></font></p>     ]]></body>
<body><![CDATA[<p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">As a general concept, DL has been implemented through a variety of   algorithms. An important distinction to make across these algorithms is whether   the dictionary is created based on formal mathematical functions or directly   learned from the data. [6] pointed out that atoms learned from data can exhibit   a higher representative power. With our intention being to extract relevant   patterns, rather than minimizing a reconstruction error, high representative   power of the dictionary seems central. Another relevant aspect concerns the   optimization of the dictionary atoms with respect to obtaining a diverse and   complete representation of patterns found in the data. Various methods apply a   global reconstruction error minimization for the definition of a (near-)   optimal dictionary (i.e., decorrelation of atoms' receptive fields). While such   approaches are mathematically justified they do not comply with biologically   plausible principles of learning. Instead, learning in visual and auditory   cortices is not mediated by a global error minimization, but rather by an activity   regulation [9, 7]. For reasons of achieving high representative power of our   dictionaries following biologically plausible learning rules, we adopt the <i>SAILnet</i> algorithm for temporal dictionary learning [7] (for any specific details of <i>SAILnet</i> we refer the reader to the corresponding publication).</font></p>  <table width="600" border="0" align="center">   <tr>     <td><a name="f1"></a><img src="/img/revistas/riyd/v19n1/a08_figure_01.gif" width="748" height="481"></td>   </tr>   <tr>     <td valign="top">    <p align="center"><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>Figure 1:</b></font>      <font size="2" face="Verdana, Arial, Helvetica, sans-serif">Dictionary Learning and reconstruction         architecture. (a) Training phase with random selection of patches from the         training data, whitening of the patches, and training of the dictionary         according to the <i>SAILnet</i> algorithm. (b) Reconstruction phase taking whitened,         non-overlapping patches from an original time series, retrieving a sparse         activity representation <i>R</i> from <i>D</i>, creating the linear         combinations of the active atoms, applying original mean and standard         deviation, and concatenating the individually reconstructed patches into the     reconstructed time series.</font></p>     </td>   </tr> </table>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">The most common interpretation of the performance of DL is its ability   to reduce a reconstruction error <i>&epsilon;</i> according to <img width=95 height=14 src="/img/revistas/riyd/v19n1/a08_image005.gif">. Here, <i>X</i> is the input signal, <i>D</i> the dictionary comprised   of a set of atoms and <i>R</i> the activity representation as result of <i>X</i> applied to <i>D</i>. For the case where&nbsp; <i>&epsilon; &gt; 0</i>, it follows that <img width=36 height=11 src="/img/revistas/riyd/v19n1/a08_image006.gif">. Instead of interpreting this difference as a reconstruction error, we   can assume that <img src="/img/revistas/riyd/v19n1/a08_image006.gif" width=36 height=11 align="absmiddle">, because <img src="/img/revistas/riyd/v19n1/a08_image007.gif" width=56 height=15 align="absmiddle">. <i>h</i> represents a transfer function that extracts specific   components from <i>x</i>. With a specific dictionary <i>D</i> being a direct   result of the applied training data (<i>D = f(X)</i>), a representation <i>R</i> being a function of a dictionary <i>D</i>, and input signal <i>x</i> (<i>R= f(x,     D)</i>) we can then assert that <img src="/img/revistas/riyd/v19n1/a08_image001.gif" width=72 height=15 align="absmiddle">&nbsp;with <img src="/img/revistas/riyd/v19n1/a08_image008.gif" width=86 height=14 align="absmiddle">. Here, <img width=16 height=10 src="/img/revistas/riyd/v19n1/a08_image009.gif">&nbsp;represents a dictionary-dependent   reconstruction of <i>x</i>, assuming that <img width=17 height=13 src="/img/revistas/riyd/v19n1/a08_image002.gif">performed some filtering function on <i>x</i>. We will therefore   investigate the impact of different dictionaries <img src="/img/revistas/riyd/v19n1/a08_image010.gif" width=14 height=13 align="absmiddle">&nbsp;with respect to their ability to decompose a   complex composite signal into different components.</font></p>  <table width="600" border="0" align="center">   <tr>     <td><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><img width=724 height=352 src="/img/revistas/riyd/v19n1/a08_figure_02.gif" align=left></font></td>   </tr>   <tr>     <td valign="top">    <p align="center"><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>Figure 2:</b></font>      <font size="2" face="Verdana, Arial, Helvetica, sans-serif">Dictionary elements.         From top to bottom, each row shows 10 randomly selected atoms from dictionaries         trained with patch size 6, 12, 18, 24, 36, 48, and 60, respectively. It can be         observed how atoms corresponding to shorter patch sizes adopted to higher         frequency components, while the other dictionaries' atoms capture predominantly         lower frequency components. (Note that the time-scale is different for each row     with number of time steps according to patch size.)</font></p>     </td>   </tr> </table>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">Related to the DL algorithm a design decision had to be made with   respect to the representation of the time-series signals. The original <i>SAILnet</i> implementation represented a 2D visual receptive field as a reshaped single   column vector within the excitatory weight matrix. To represent single-channel   auditory signals one can choose either 1D or 2D representations of the signal.   The 1D representation, referred to as the temporal receptive fields (TRF),   maintains the time-domain representation of the signal. [10] have demonstrated   how TRF can be used for efficient encoding of naturalistic sounds. The 2D   representation of temporal signals is known as spectro-temporal receptive   fields (STRF). STRF represent the spectral (frequency) domain behavior as a   function of time. In [5, 11] it has been argued that STRF are the biological   bases for sound representation in the primary auditory cortex. In this work we   will adopt TRF as they reduce model complexity and still allow for efficient   encoding in terms of sparse dictionary learning. We therefore directly apply 1D   time-series signals to the DL architecture as shown in <a href="#f1">Figure 1</a>. The length of   sub-sequences extracted from the data and applied to the dictionary for   learning is referred to as patch size. In this project we trained different   dictionaries varying in the patch sizes they are implemented and trained with,   hypothesizing that this will determine the filtering characteristics relevant   for signal decomposition.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">In this work we are using a dataset containing wind speed time series   from Northeast Brazil extracted from the NCEP climate forecast system [8]. The   dataset contains 5544 independent time series, with 8784 measurements each,   recorded in hourly intervals.&nbsp; As wind speed is to a significant degree related   to daylight and temperature patterns, we use patch sizes of 6h to 60h in order   to train dictionaries with wind patterns of meaningful lengths. The dataset was   split along the time dimension, taking the first 80% as training data and the   remaining 20% as test data. Additionally, time series are identified by the   geographical location of the measurement point. Geographically nearby time   series do exhibit high degrees of correlation. For the purpose of learning   general dictionary atoms, we did not take the spatial information into account.</font><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>&nbsp;</b></font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>&nbsp;</b></font></p> <table width="600" border="0" align="center">   <tr>     <td><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b><img border=0 width=708 height=365 src="/img/revistas/riyd/v19n1/a08_figure_03.png"></b></font></td>   </tr>   <tr>     <td valign="top">    <p align="center"><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>Figure 3:</b></font>      <font size="2" face="Verdana, Arial, Helvetica, sans-serif">Reconstructions         of original signal. Top plot shows reconstruction with patch size 12 and close         approximation of original signal. Only high-frequency fluctuations are not         captured. Bottom plot shows low-frequency component reconstruction resulting     from patch size 60.</font></p>     </td>   </tr> </table>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>&nbsp;</b></font></p>     <p align="justify"><font size="3" face="Verdana, Arial, Helvetica, sans-serif"><b>3. RESULTS</b></font></p>     ]]></body>
<body><![CDATA[<p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">As shown in <a href="#f1">Figure 1a</a> we start by training multiple dictionaries with   random samples (patches) drawn from the wind speed dataset. Dictionaries differ   with respect to the patch size they are trained with. We assume that these   different dictionaries will therefore vary in the temporal receptive fields   that the atoms adopt. Given that the wind speed data represents hourly   measurements, and that wind is partially effected by natural daylight and   temperature cycles, we define patch sizes between 6 hours and 60 hours,   representing quarter day to two and a half day wind cycles. In <a href="#f2">Figure 2</a> we   demonstrate a set of randomly chosen atoms for all trained dictionaries (rows).   As anticipated, the different dictionaries contain atoms that have adopted to   different patterns of wind cycles. Shorter patches (i.e., 6h, 12h) clearly show   adaptations to basic features such as peaks, rising or falling edges   (corresponding to high frequency components within the data). Longer patches   (i.e., 48h, 60h) led the dictionaries to learn atoms that respond to more   smooth recurring patterns, indicative of lower frequency components. </font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">Based on the trained dictionaries, we reconstructed the original data   by running the dictionaries in inference mode (see <a href="#f1">Figure 1b</a>). Each patch   applied to the dictionary resulted in a sparse spiking activity (see [7] for   details on leaky-integrate and fire neuron implemented in <i>SAILnet</i>). The   reconstructed signal is the weighted sum of all dictionary elements with the   weight being determined by the number of spikes per dictionary element. After   the reconstruction, we performed unwhitening to transform the reconstruction to   the mean and standard deviation of the original patch. For sake of   demonstration we are showing two reconstructed signals in <a href="#f3">Figure 3</a> (sequence   zoomed in to first 300 hours). For the dictionary trained with patch size 12 we   can observe a fairly close reconstruction of the original signal, concluding   that the dictionary learning algorithm is capable of capturing a diverse set of   descriptive features from the applied training data. For the dictionary trained   with patch size 60 we can observe that higher frequency components are not   present in the reconstruction as the corresponding atoms adopted to lower   frequency patterns.</font></p>     <p align=center><a name="t1"></a><img src="/img/revistas/riyd/v19n1/a08_table_01.png" width="754" height="164"></p>      <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">To evaluate the different dictionaries more quantitatively, and to   better understand qualitative differences between the different dictionaries,   we are comparing the respective reconstruction errors in Table 1. For this we   randomly chose 1000 signals from the test set and calculated the normalized   root mean square errors (NRMSE) for all dictionaries. As expected, shorter   patch sizes maintaining higher frequency components allow for more accurate   reconstructions of the original signals, while longer patch sizes exhibit   higher reconstruction errors. Interestingly though, patch size is not directly   linearly related to reconstruction accuracy. As can be observed, patch size of   18h does not provide improvement over patch size of 24h. For all measures it   exhibits inferior reconstruction accuracies.</font></p> <table width="600" border="0" align="center">   <tr>     <td><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><img border=0 width=642 height=332 src="/img/revistas/riyd/v19n1/a08_image013.gif"></font></td>   </tr>   <tr>     <td valign="top">    <p align="center"><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>Figure 4:</b></font>      <font size="2" face="Verdana, Arial, Helvetica, sans-serif">Coherence analysis of different dictionaries         (performed with scipy.signal.coherence function and parameter <i>nperseg</i>=120;         all other parameters used with their default values). Shorter patch sizes allow         reconstructions maintaining coherence over a wider frequency spectrum, while         longer patch sizes cut off higher frequencies more quickly. It can be observed         that the filter behaviors of most dictionaries are tuned to specific wind         cycles like 24h or 12h (see vertical lines indicating &ldquo;wave length'&rdquo;). All dictionaries     have low coherence for rapid changes in wind speed (high frequencies).</font></p>     </td>   </tr> </table>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">We further investigated the filtering behavior of dictionaries by   performing a coherence analysis between the original and the reconstructed   signals. Coherence is a similarity measure of two signals as a function of   their frequency components. Here, coherence expresses the similarity in the   power spectrum between the original and the reconstructed signals. Coherence is   bound in the interval [0, 1] from zero to full similarity. We selected 1000   random signals from the test set and performed the coherence analysis for all   dictionaries. Coherence results for each dictionary are expressed as the   respective average across all 1000 signals. Results of the coherence analysis   can be seen in <a href="#f4">Figure 4</a>. In a general sense, the shown coherence plot confirms   two initial assumptions: first, different dictionaries lead to different   reconstructions (with specific frequency components filtered out) and second, the transfer functions <img width=17 height=13 src="/img/revistas/riyd/v19n1/a08_image002.gif">&nbsp;are more complex than simple low or band pass   filters. While detecting a clear drop in coherence past <i>0.2</i> cycles per   hour, short patch sizes maintain some coherence with the original signal   implying the theoretical possibility to reconstruct higher frequency behavior.   This was already qualitatively observed in <a href="#f3">Figure 3</a>. Longer patch sizes already   lead to a clear and consistent drop in coherence past <i>0.1</i> cycles per   hour. The second aspect, the transfer function of a dictionary, shows some   multi-band pass behavior for most dictionaries. While patch size 6 still   somewhat resembles a low pass filter, all other dictionaries more clearly   exhibit selectivity to a set of frequency components. As indicated in the   coherence plot, most notable frequencies captured are with wavelengths of 24h, 12h, 8h and 6h.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">As our dictionaries were learned directly from data we verify if the   coherence patterns are related to frequencies found in the actual data. In     <a href="#f5">Figure 5</a> we demonstrate the Fast Fourier Transform (FFT) of a time series from   the original dataset. It can be observed that the coherences from <a href="#f4">Figure 4</a> are   closely related to the actual frequency components found in the data. This   indicates the ability of dictionaries to be learned as filters tuned to the   most prevalent frequencies within the data. </font></p> <table width="600" border="0" align="center">   <tr>     <td><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><img border=0 width=642 height=332 id="0 Imagen" src="/img/revistas/riyd/v19n1/a08_image014.gif"></font></td>   </tr>   <tr>     <td valign="top">    <p align="center"><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>Figure 5:</b></font>      <font size="2" face="Verdana, Arial, Helvetica, sans-serif">Fast Fourier Transform of a wind speed time         series. Vertical lines indicate wavelengths. Peaks indicate strong daily and     half-daily wind patterns.</font></p>     </td>   </tr> </table>     <p align="justify">&nbsp;</p>     <p align="justify"><font size="3" face="Verdana, Arial, Helvetica, sans-serif"><b>4. DISCUSSION</b></font></p>     ]]></body>
<body><![CDATA[<p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">In the first part of the results section we have shown dictionary atoms   corresponding to different patch sizes. Akin to what <i>SAILnet</i> was   learning as V1 features, dictionary with patch size 6 has learned the most   basic features that would allow reconstruction of more complex signals. Indeed,   this was demonstrated by the low reconstruction errors. Dictionaries trained   with larger patches then adopted their atoms to learn receptive fields that   capture not just lower frequency components, but specifically these frequencies   found in the data. The counter example to this was the dictionary trained with   patch size 18. The lower reconstruction accuracy obtained can in fact be   related to the patch size not matching wavelengths within the data. This raises   the question of how critical the selection of patch size is in relation to a   given dataset. In [6] authors used an adaptive patch size that converged to   some specific length as training progressed.</font><font size="2" face="Verdana, Arial, Helvetica, sans-serif"><b>&nbsp;</b></font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">In the introduction of this paper we have briefly discussed STL   decomposition and its limited ability to detect more complex patterns.   Decomposition has as its goal to extract deterministic patterns and separate   random fluctuations from the data. However, for complex signals the remainder   of the decomposition often represents a major part. DL allows to closely reconstructing   complex signals with minimum error. This has two implications. First, by choice   of the dictionary the reconstruction error (or remainder) can be lowered in   magnitude in comparison to the overall signal amplitude. And second,   reconstructing signals is a deterministic process based on linear combinations   of known features. Utilizing multiple dictionaries with different filter   characteristics (similar to ensemble learning methods) has the theoretical   potential to further improve decomposition results. An application of ensemble   methods to time series forecasting has recently been presented in [12].</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">Another potential application of decomposition through dictionaries is   forecasting. Decomposition is often performed to simplify forecasting. DL   offers the possibility to transform the challenge into forecasting a sparse   signal representation. Rather than forecasting the actual signal, one could   forecast the sparse spike patterns generated by the dictionaries. [13]   described that through such a neocortical encoding of activity forecasting   becomes linearly decodable. A problem as pointed out in [14] might arise from   the sparse, seemingly random activity patterns of individual neurons. Known   activity correlations between neurons over a relevant time span might alleviate   this potential problem. </font></p>     <p align="justify">&nbsp;</p>     <p align="justify"><font size="3" face="Verdana, Arial, Helvetica, sans-serif"><b>5. CONCLUSION</b></font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">Despite decades of research, across various engineering and scientific   disciplines, time series analysis often remains a daunting task. This can be   reasoned about from various perspectives. For one, it is a question of   separating real data from noise. Decomposition aims at separating composite   time series into more easily analyzable components on the one side and noise on   the other. However, it is rarely obvious to affirm that remainders of   decompositions are purely noise and do not contain any more information that   could aid in the overall analysis. Another perspective is that of   representation. Depending on the domain one either decomposes or transforms in,   i.e., the frequency domain to better analyze a signal. But every decomposition   or transformation provides some insights while it obscures others. Ultimately   it is a question of finding a good match between representation and application   challenges. Here we have presented temporal DL as a method for creating   different decompositions from time series signals. We have adopted an existing   DL algorithm to learn temporal receptive fields. By training a variety of   dictionaries with different TRF lengths, we have shown the ability for a given   dictionary to learn specific patterns inherent in the data. The ability to   directly learn features from data, contrary to using fixed mathematical kernel   functions, provides the theoretical possibility to better separate noise from   data. This hypothesis was supported by performing a coherence analysis between   original signals and their dictionary dependent reconstructions. The coherence   analysis has shown how shorter patch sizes maintain coherence over a wider   frequency range while longer patch sizes function more selectively to a set of   lower frequencies found in the data. This was shown exemplary on   reconstructions with two different dictionaries. Relevant for future research,   DL allows creating ensembles of deterministic and sparse representations of   dense time series signals with, depending on the patch size, very low reconstruction   errors. Such deterministic, sparse representations might open up new directions   for analysis and processing of time series signals. We therefore stress the   significance of this paper as contributing to new directions for methodological   development in time series analysis.</font></p>     <p align="justify">&nbsp;</p>     <p align="justify"><font size="3" face="Verdana, Arial, Helvetica, sans-serif"><b>6. REFERENCES</b></font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[1] Y. LeCun, Y. Bengio, and   G. Hinton, &ldquo;Deep learning,&rdquo; Nature, vol. 521, no. 7553, p. 436, 2015.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[2] M. S. Lewicki and T. J.   Sejnowski, &ldquo;Learning Overcomplete Representations,&rdquo; Neural Computation, &nbsp; vol.   12, no. 2, pp. 337&ndash;365, 2000.</font></p>     ]]></body>
<body><![CDATA[<p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[3] R. J. Hyndman and G.   Athanasopoulos, &ldquo;Forecasting: Principles and practice.&rdquo; OTexts, 2018.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[4] J. Mairal, F. Bach, J.   Ponce, and G. Sapiro, &ldquo;Online Dictionary Learning for Sparse Coding,&rdquo; in   Proceedings of the 26th Annual International Conference on Machine Learning,   ACM, 2009, pp. 689&ndash;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp; 696.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[5] J. Fritz, S. Shamma, M.   Elhilali, and D. Klein, &ldquo;Rapid task-related plasticity of spectrotemporal   receptive fields in primary auditory cortex,&rdquo; Nature Neuroscience, vol. 6, no.   11, p. 1216, 2003.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[6] Q. Barth&eacute;lemy, C.   Gouy-Pailler, Y. Isaac, A. Souloumiac, A. Larue, and J. I. Mars, &ldquo;Multivariate   Temporal Dictionary Learning for EEG,&rdquo; Journal of Neuroscience Methods, vol.   215, no. 1, pp. 19&ndash;28, 2013.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[7] J. Zylberberg, J. T.   Murphy, and M. R. DeWeese, &ldquo;A Sparse Coding Model with Synaptically Local   Plasticity and Spiking Neurons Can Account for the Diverse Shapes of V1 Simple   Cell Receptive Fields,&rdquo; PLoS Computational Biology, vol. 7, no. 10, e1002250,   2011.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[8] S. Saha, S. Moorthi, X.   Wu, J. Wang, S. Nadiga, P. Tripp, D. Behringer, Y.-T. Hou, H.-y. Chuang, M.   Iredell, et al., &ldquo;The NCEP Climate Forecast System Version 2,&rdquo; Journal of   Climate, vol. 27, no. 6, pp. 2185&ndash;2208, 2014.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[9] S. Dasgupta, F.   W&ouml;rg&ouml;tter, and P. Manoonpong, &ldquo;Information Theoretic Self-organised Adaptation   in Reservoirs for Temporal Memory Tasks,&rdquo; in International Conference on   Engineering Applications of&nbsp; Neural Networks, Springer, 2012, pp. 31&ndash;40.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[10] N. A. Lesica and B.   Grothe, &ldquo;Efficient Temporal Processing of Naturalistic Sounds,&rdquo; PloS One, vol.   3, no.2, e1655, 2008.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[11] K. Patil, D. Pressnitzer,   S. Shamma, and M. Elhilali, &ldquo;Music in Our Ears: The Biological Bases of Musical   Timbre Perception,&rdquo; PLoS Computational Biology, vol. 8, no. 11, e1002759, 2012.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[12] P. Laurinec, M. L&oacute;derer,   M. Luck&aacute;, and V. Rozinajov&aacute;, &ldquo;Density-based unsupervised ensemble learning   methods for time series forecasting of aggregated or clustered electricity   consumption,&rdquo;&nbsp; Journal of Intelligent Information Systems, pp. 1&ndash;21, 2019.</font></p>     ]]></body>
<body><![CDATA[<p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[13] W. B. Levy, A. B. Hocking,   and X. Wu, &ldquo;Interpreting hippocampal function as recoding and  forecasting,&rdquo;   Neural Networks, vol. 18, no. 9, pp. 1242&ndash;1264, 2005.</font></p>     <p align="justify"><font size="2" face="Verdana, Arial, Helvetica, sans-serif">[14] A. Longtin, &ldquo;Nonlinear   Forecasting of Spike Trains from Sensory Neurons,&rdquo; International Journal of   Bifurcation and Chaos, vol. 3, no. 03, pp. 651&ndash;661, 1993.</font></p>     <p align="justify">&nbsp;</p>      ]]></body><back>
<ref-list>
<ref id="B1">
<label>1</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[LeCun]]></surname>
<given-names><![CDATA[Y.]]></given-names>
</name>
<name>
<surname><![CDATA[Bengio]]></surname>
<given-names><![CDATA[Y.]]></given-names>
</name>
<name>
<surname><![CDATA[Hinton]]></surname>
<given-names><![CDATA[G.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Deep learning]]></article-title>
<source><![CDATA[Nature]]></source>
<year>2015</year>
<volume>521</volume>
<numero>7553</numero>
<issue>7553</issue>
<page-range>436</page-range></nlm-citation>
</ref>
<ref id="B2">
<label>2</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Lewicki]]></surname>
<given-names><![CDATA[M.S.]]></given-names>
</name>
<name>
<surname><![CDATA[Sejnowski]]></surname>
<given-names><![CDATA[T.J.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Learning Overcomplete Representations]]></article-title>
<source><![CDATA[Neural Computation]]></source>
<year>2000</year>
<volume>12</volume>
<numero>2</numero>
<issue>2</issue>
<page-range>337-365</page-range></nlm-citation>
</ref>
<ref id="B3">
<label>3</label><nlm-citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Hyndman]]></surname>
<given-names><![CDATA[R.J.]]></given-names>
</name>
<name>
<surname><![CDATA[Athanasopoulos]]></surname>
<given-names><![CDATA[G.]]></given-names>
</name>
</person-group>
<source><![CDATA[Forecasting: Principles and practice]]></source>
<year>2018</year>
<publisher-name><![CDATA[OTexts]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B4">
<label>4</label><nlm-citation citation-type="confpro">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Mairal]]></surname>
<given-names><![CDATA[J.]]></given-names>
</name>
<name>
<surname><![CDATA[Bach]]></surname>
<given-names><![CDATA[F.]]></given-names>
</name>
<name>
<surname><![CDATA[Ponce]]></surname>
<given-names><![CDATA[J.]]></given-names>
</name>
<name>
<surname><![CDATA[Sapiro]]></surname>
<given-names><![CDATA[G.]]></given-names>
</name>
</person-group>
<source><![CDATA[Online Dictionary Learning for Sparse Coding]]></source>
<year>2009</year>
<conf-name><![CDATA[ 26th Annual International Conference on Machine Learning]]></conf-name>
<conf-loc> </conf-loc>
<page-range>689-696</page-range><publisher-name><![CDATA[ACM]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B5">
<label>5</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Fritz]]></surname>
<given-names><![CDATA[J.]]></given-names>
</name>
<name>
<surname><![CDATA[Shamma]]></surname>
<given-names><![CDATA[S.]]></given-names>
</name>
<name>
<surname><![CDATA[Elhilali]]></surname>
<given-names><![CDATA[M.]]></given-names>
</name>
<name>
<surname><![CDATA[Klein]]></surname>
<given-names><![CDATA[D.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Rapid task-related plasticity of spectrotemporal receptive fields in primary auditory cortex]]></article-title>
<source><![CDATA[Nature Neuroscience]]></source>
<year>2003</year>
<volume>6</volume>
<numero>11</numero>
<issue>11</issue>
<page-range>1216</page-range></nlm-citation>
</ref>
<ref id="B6">
<label>6</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Barthélemy]]></surname>
<given-names><![CDATA[Q.]]></given-names>
</name>
<name>
<surname><![CDATA[Gouy-Pailler]]></surname>
<given-names><![CDATA[C.]]></given-names>
</name>
<name>
<surname><![CDATA[Isaac]]></surname>
<given-names><![CDATA[Y.]]></given-names>
</name>
<name>
<surname><![CDATA[Souloumiac]]></surname>
<given-names><![CDATA[A.]]></given-names>
</name>
<name>
<surname><![CDATA[Larue]]></surname>
<given-names><![CDATA[A.]]></given-names>
</name>
<name>
<surname><![CDATA[Mars]]></surname>
<given-names><![CDATA[J.I.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Multivariate Temporal Dictionary Learning for EEG]]></article-title>
<source><![CDATA[Journal of Neuroscience Methods]]></source>
<year>2013</year>
<volume>215</volume>
<numero>1</numero>
<issue>1</issue>
<page-range>19-28</page-range></nlm-citation>
</ref>
<ref id="B7">
<label>7</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Zylberberg]]></surname>
<given-names><![CDATA[J.]]></given-names>
</name>
<name>
<surname><![CDATA[Murphy]]></surname>
<given-names><![CDATA[J.T.]]></given-names>
</name>
<name>
<surname><![CDATA[DeWeese]]></surname>
<given-names><![CDATA[M.R.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[A Sparse Coding Model with Synaptically Local Plasticity and Spiking Neurons Can Account for the Diverse Shapes of V1 Simple Cell Receptive Fields]]></article-title>
<source><![CDATA[PLoS Computational Biology]]></source>
<year>2011</year>
<volume>7</volume>
<numero>10</numero>
<issue>10</issue>
<page-range>e1002250</page-range></nlm-citation>
</ref>
<ref id="B8">
<label>8</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Saha]]></surname>
<given-names><![CDATA[S.]]></given-names>
</name>
<name>
<surname><![CDATA[Moorthi]]></surname>
<given-names><![CDATA[S.]]></given-names>
</name>
<name>
<surname><![CDATA[Wu]]></surname>
<given-names><![CDATA[X.]]></given-names>
</name>
<name>
<surname><![CDATA[Wang]]></surname>
<given-names><![CDATA[J.]]></given-names>
</name>
<name>
<surname><![CDATA[Nadiga]]></surname>
<given-names><![CDATA[S.]]></given-names>
</name>
<name>
<surname><![CDATA[Tripp]]></surname>
<given-names><![CDATA[P.]]></given-names>
</name>
<name>
<surname><![CDATA[Behringer]]></surname>
<given-names><![CDATA[D.]]></given-names>
</name>
<name>
<surname><![CDATA[Hou]]></surname>
<given-names><![CDATA[Y.-T.]]></given-names>
</name>
<name>
<surname><![CDATA[Chuang]]></surname>
<given-names><![CDATA[H.-y.]]></given-names>
</name>
<name>
<surname><![CDATA[Iredell]]></surname>
<given-names><![CDATA[M.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[The NCEP Climate Forecast System Version 2]]></article-title>
<source><![CDATA[Journal of Climate]]></source>
<year>2014</year>
<volume>27</volume>
<numero>6</numero>
<issue>6</issue>
<page-range>2185-2208</page-range></nlm-citation>
</ref>
<ref id="B9">
<label>9</label><nlm-citation citation-type="confpro">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Dasgupta]]></surname>
<given-names><![CDATA[S.]]></given-names>
</name>
<name>
<surname><![CDATA[Wörgötter]]></surname>
<given-names><![CDATA[F.]]></given-names>
</name>
<name>
<surname><![CDATA[Manoonpong]]></surname>
<given-names><![CDATA[P.]]></given-names>
</name>
</person-group>
<source><![CDATA[Information Theoretic Self-organised Adaptation in Reservoirs for Temporal Memory Tasks]]></source>
<year>2012</year>
<conf-name><![CDATA[ International Conference on Engineering Applications of Neural Networks]]></conf-name>
<conf-loc> </conf-loc>
<page-range>31-40</page-range><publisher-name><![CDATA[Springer]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B10">
<label>10</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Lesica]]></surname>
<given-names><![CDATA[N.A.]]></given-names>
</name>
<name>
<surname><![CDATA[Grothe]]></surname>
<given-names><![CDATA[B.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Efficient Temporal Processing of Naturalistic Sounds]]></article-title>
<source><![CDATA[PloS One]]></source>
<year>2008</year>
<volume>3</volume>
<numero>2</numero>
<issue>2</issue>
<page-range>e1655</page-range></nlm-citation>
</ref>
<ref id="B11">
<label>11</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Patil]]></surname>
<given-names><![CDATA[K.]]></given-names>
</name>
<name>
<surname><![CDATA[Pressnitzer]]></surname>
<given-names><![CDATA[D.]]></given-names>
</name>
<name>
<surname><![CDATA[Shamma]]></surname>
<given-names><![CDATA[S.]]></given-names>
</name>
<name>
<surname><![CDATA[Elhilali]]></surname>
<given-names><![CDATA[M.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Music in Our Ears: The Biological Bases of Musical Timbre Perception]]></article-title>
<source><![CDATA[PLoS Computational Biology]]></source>
<year>2012</year>
<volume>8</volume>
<numero>11</numero>
<issue>11</issue>
<page-range>e1002759</page-range></nlm-citation>
</ref>
<ref id="B12">
<label>12</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Laurinec]]></surname>
<given-names><![CDATA[P.]]></given-names>
</name>
<name>
<surname><![CDATA[Lóderer]]></surname>
<given-names><![CDATA[M.]]></given-names>
</name>
<name>
<surname><![CDATA[Lucká]]></surname>
<given-names><![CDATA[M.]]></given-names>
</name>
<name>
<surname><![CDATA[Rozinajová]]></surname>
<given-names><![CDATA[V.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Density-based unsupervised ensemble learning methods for time series forecasting of aggregated or clustered electricity consumption]]></article-title>
<source><![CDATA[Journal of Intelligent Information Systems]]></source>
<year>2019</year>
<page-range>1-21</page-range></nlm-citation>
</ref>
<ref id="B13">
<label>13</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Levy]]></surname>
<given-names><![CDATA[W.B.]]></given-names>
</name>
<name>
<surname><![CDATA[Hocking]]></surname>
<given-names><![CDATA[A.B.]]></given-names>
</name>
<name>
<surname><![CDATA[Wu]]></surname>
<given-names><![CDATA[X.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Interpreting hippocampal function as recoding and forecasting]]></article-title>
<source><![CDATA[Neural Networks]]></source>
<year>2005</year>
<volume>18</volume>
<numero>9</numero>
<issue>9</issue>
<page-range>1242-1264</page-range></nlm-citation>
</ref>
<ref id="B14">
<label>14</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Longtin]]></surname>
<given-names><![CDATA[A.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Nonlinear Forecasting of Spike Trains from Sensory Neurons]]></article-title>
<source><![CDATA[International Journal of Bifurcation and Chaos]]></source>
<year>1993</year>
<volume>3</volume>
<numero>03</numero>
<issue>03</issue>
<page-range>651-661</page-range></nlm-citation>
</ref>
</ref-list>
</back>
</article>
