EnK: Encoding time-information in convolution
- Publication Type:
- Journal Article
Recently Added
Filename | Description | Size | |||
---|---|---|---|---|---|
2006.04198v1.pdf | Submitted version | 2.24 MB |
Copyright Clearance Process
- Recently Added
- In Progress
- Open Access
This item is new to OPUS and is not currently available.
Recent development in deep learning techniques has attracted attention in
decoding and classification in EEG signals. Despite several efforts utilizing
different features of EEG signals, a significant research challenge is to use
time-dependent features in combination with local and global features. There
have been several efforts to remodel the deep learning convolution neural
networks (CNNs) to capture time-dependency information by incorporating
hand-crafted features, slicing the input data in a smaller time-windows, and
recurrent convolution. However, these approaches partially solve the problem,
but simultaneously hinder the CNN's capability to learn from unknown
information that might be present in the data. To solve this, we have proposed
a novel time encoding kernel (EnK) approach, which introduces the increasing
time information during convolution operation in CNN. The encoded information
by EnK lets CNN learn time-dependent features in-addition to local and global
features. We performed extensive experiments on several EEG datasets: cognitive
conflict (CC), physical-human robot collaboration (pHRC), P300 visual-evoked
potentials, movement-related cortical potentials (MRCP). EnK outperforms the
state-of-art by 12\% (F1 score). Moreover, the EnK approach required only one
additional parameter to learn and can be applied to a virtually any CNN
architectures with minimal efforts. These results support our methodology and
show high potential to improve CNN performance in the context of time-series
data in general.
Please use this identifier to cite or link to this item: