[go: up one dir, main page]

CN110687999B - A method and device for semantic processing of electroencephalogram signals - Google Patents

A method and device for semantic processing of electroencephalogram signals Download PDF

Info

Publication number
CN110687999B
CN110687999B CN201810728403.6A CN201810728403A CN110687999B CN 110687999 B CN110687999 B CN 110687999B CN 201810728403 A CN201810728403 A CN 201810728403A CN 110687999 B CN110687999 B CN 110687999B
Authority
CN
China
Prior art keywords
feature
semantic
alphabet
matrix
information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810728403.6A
Other languages
Chinese (zh)
Other versions
CN110687999A (en
Inventor
刘彬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN201810728403.6A priority Critical patent/CN110687999B/en
Publication of CN110687999A publication Critical patent/CN110687999A/en
Application granted granted Critical
Publication of CN110687999B publication Critical patent/CN110687999B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/015Input arrangements based on nervous system activity detection, e.g. brain waves [EEG] detection, electromyograms [EMG] detection, electrodermal response detection
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/02Preprocessing
    • G06F2218/04Denoising
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/08Feature extraction

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Software Systems (AREA)
  • Biophysics (AREA)
  • Dermatology (AREA)
  • Neurology (AREA)
  • Neurosurgery (AREA)
  • Human Computer Interaction (AREA)
  • Measurement And Recording Of Electrical Phenomena And Electrical Characteristics Of The Living Body (AREA)

Abstract

The invention discloses a semantic processing method and a corresponding device for electroencephalogram signals, and relates to a data mining technology for electroencephalogram signals. Current BCI studies on brain electrical signals rely on a few classical brain electrical paradigms that make use of only a few of the brain electrical signals, with little improvement to BCI methodologies, most of which do not introduce semantic concepts. A number of potential paradigms in the actual measurement have not been discovered yet and have not been summarized as valid semantics and new paradigms. The invention realizes a new semantic paradigm framework: the electroencephalogram data is taken as a research object, related grammar, semantic processing rules and algorithms are designed, grammar and semantic connotation are given to complex signals, and semantic interpretation is constructed for the deep neural network. The electroencephalogram signals are analyzed through the paradigm framework, researchers are helped to find out information blocks with specific semantics and semantic combinations among the information blocks, and efficient filters are automatically learned.

Description

一种对脑电信号进行语义处理的方法和装置A method and device for semantic processing of electroencephalogram signals

技术领域Technical Field

本发明涉及对脑电信号(brain electrophysiological signals)的数据挖掘技术,尤指一种从脑电信号(EEG/ECoG)中挖掘并分析出其中蕴含的语义信息的处理方法和装置。The present invention relates to a data mining technology for brain electrophysiological signals, and in particular to a processing method and device for mining and analyzing semantic information contained in brain electrophysiological signals (EEG/ECoG).

背景技术Background Art

2008年第65届美国金球奖最佳外语片《潜水钟与蝴蝶》的原著是法国ELLE杂志的总编辑、尚·多明尼克·鲍比。他曾在患急性中风后陷入深度昏迷达20多天。从昏迷中醒来后,除了左眼皮外的任何身体部位都不能动,无法与人说话沟通,没有任何表情,就连他是否还有意识旁人都无从得知;但他没有气馁,在他的语言治疗师的帮助下,鲍比学会了用字母表交流,用左眼皮“写”出了同名小说。在鲍比看来,他的肉体如同沉重的潜水钟,但内心却如同蝴蝶般自由飞翔。鲍比的故事其实是“闭锁综合症”的一个典型案例。此外,根据统计资料显示,我国有8502万残疾人[1],其中很多人处于生活难于自理的状态,这种状况在全世界广泛存在,绝非极少数。不久前离世的著名理论物理学家斯蒂芬·威廉·霍金,以及1995年从马背上跌落的克里斯托弗.里夫都患有葛雷克氏症(Lou Gehrig),克里斯托弗甚至不能通过眨眼的方式与人交流。这一切都提醒了医生和相关领域的科研人员要尽力帮助这些严重伤残的病人,找到并推广更好的治疗或康复技术,为那些伤残程度不像他们那么严重的病人提供帮助,以提高他们的生活质量[2]。很重要的一条背景信息是,克里斯托弗.里夫以这种状态直至2004年离世,他本人其实相当富有,他享受着全世界最好的医疗保险,接受全世界最好的科研医疗团队的研究,还有一家大力推动研究这一问题解决的基金会持续投入大量资金,尽管如此,可他个人仍然在承受了难言的痛苦直至去世也未能治愈。对此类问题的解决之所以如此困难重重,从康复的角度来说,神经细胞的再生一直是一个世界级难题,直到2012年诺贝尔生理学或医学奖授予在“体细胞重编程技术”领域做出革命性贡献的英国科学家约翰·格登和日本医学教授山中伸弥,才出现一些曙光,但iPS细胞(Induced Pluripotent Stem Cells)植入人体可能会导致基因突变和癌症的产生,因此依然任重而道远;从体现终极关怀、提高生活质量的角度来说,用于帮助他的脑机接口技术在当时也面临诸多挑战,不论是对数据采集的可靠性、稳定性、长期性、信号来源的定位,还是对复杂脑电信号解析的完美性、实时性、准确性都存在诸多技术难题需要解决。最近几年,随着科技的发展,人类在脑机接口领域已经逐步具备了接受上述挑战的能力。脑机接口技术未来一旦得到成熟的应用,将极大地造福人类,不但将让使用者在人生的后期能够有尊严地活着,从科学的角度来说,也将进一步加深我们对大脑的生理机能,以及其与思维互动的原理的理解,推动新的科学技术进步。The original novel of The Diving Bell and the Butterfly, which won the 65th Golden Globe Award for Best Foreign Language Film in 2008, was written by Jean-Dominique Bobby, the editor-in-chief of French ELLE magazine. He fell into a deep coma for more than 20 days after suffering an acute stroke. After waking up from the coma, he could not move any part of his body except his left eyelid, could not talk to people, had no expression, and no one could tell whether he was still conscious. But he did not give up. With the help of his speech therapist, Bobby learned to communicate using the alphabet and "wrote" the novel of the same name with his left eyelid. In Bobby's view, his body was like a heavy diving bell, but his heart was like a butterfly flying freely. Bobby's story is actually a typical case of "locked-in syndrome". In addition, according to statistics, there are 85.02 million disabled people in my country[1], many of whom are in a state of difficulty in taking care of themselves. This situation is widespread around the world and is by no means a rare occurrence. Stephen William Hawking, a famous theoretical physicist who passed away not long ago, and Christopher Reeve, who fell off a horse in 1995, both suffered from Lou Gehrig's disease. Christopher could not even communicate with others by blinking. All of this reminds doctors and researchers in related fields to try their best to help these severely disabled patients, find and promote better treatment or rehabilitation technologies, and provide help to patients whose disabilities are not as severe as theirs to improve their quality of life [2]. An important piece of background information is that Christopher Reeve lived in this state until his death in 2004. He was actually quite wealthy. He enjoyed the best medical insurance in the world, received research from the best scientific research and medical teams in the world, and a foundation that vigorously promoted research on this problem continued to invest a lot of money. Despite this, he still suffered unspeakable pain and was not cured until his death. The reason why it is so difficult to solve such problems is that from the perspective of rehabilitation, the regeneration of nerve cells has always been a world-class problem. It was not until the 2012 Nobel Prize in Physiology or Medicine was awarded to British scientist John Gurdon and Japanese medical professor Shinya Yamanaka, who made revolutionary contributions in the field of "somatic cell reprogramming technology", that some hope appeared. However, the implantation of iPS cells (Induced Pluripotent Stem Cells) into the human body may cause gene mutations and cancer, so there is still a long way to go; from the perspective of embodying ultimate care and improving the quality of life, the brain-computer interface technology used to help him also faced many challenges at the time, whether it was the reliability, stability, long-term nature of data acquisition, the location of the signal source, or the perfection, real-time nature, and accuracy of the analysis of complex EEG signals. There are many technical problems that need to be solved. In recent years, with the development of science and technology, humans have gradually acquired the ability to accept the above challenges in the field of brain-computer interfaces. Once brain-computer interface technology is maturely applied in the future, it will greatly benefit mankind. It will not only allow users to live with dignity in the later stages of their lives, but from a scientific point of view, it will also further deepen our understanding of the physiological functions of the brain and the principles of its interaction with the mind, and promote new scientific and technological advances.

目前对BCI的研究,往往依赖于以特定的脑电范式为识别依据,然后将识别出来的特定的范式作为输入,触发专门的BCI系统,比如基于P300的字符拼写输入BCI系统中,使用了有单词联想功能的、改进的T9方式减少输入时间[15];还有的方案通过稳态视觉诱发电位[Steady-State Visual EvokedPotential(SSVEP),诱发电位]实现[88],这些方法只利用了脑电信号中很少的一部分,而且都是通过已有脑电范式进行转译实现,对BCI方法论的改进很少。而人的大脑使用大脑皮层的感觉运动和语言功能与外界交换信息[88],尤其是具有语义信息的语言功能起着关键作用,然而截至目前仍然只有很少的研究考虑了直接分析含有语义信息的信号,比如在将语义运用于BCI技术的探索中,有研究通过让两个健康的被试者进行实验[93],以真实和虚假的陈述作为条件刺激,以检查可区分的条件反应。在获得条件反应之后,这些条件反应就对应真实和虚假的语义信息。该研究使用支持向量机(一种径向基函数分类器)进行离线分类。在区分条件“是”与“否”的思想,单个试验分类准确率的平均值达到了50.5%,显示其分类准确率很低,而且该方法仍属于传统方法范畴。而另外一些研究显示能单个概念的语义可以从某个单独的脑电数据试验中解码出来,使用CSP方法提取全头皮同步活动成分,使用通用的SVM机器学习算法,并结合谱功率的测量结果来预测每个试验的语义类别。还有一些研究[10]中,研究了在句子阅读过程中,语义引起的伽马脑电波活动,针对由语义引起的脑电图功率谱的变化审查了语句子中的违规行为与句子正确性有关的重要影响,使用FFT方法进行处理。调查显示,目前针对脑电中语义的研究还处于探索阶段,使用传统的机器学习方法,和本研究中的语义范式模型完全不同。Current research on BCI often relies on specific EEG patterns as the basis for recognition, and then uses the recognized specific patterns as input to trigger a special BCI system. For example, in the BCI system based on P300 character spelling input, an improved T9 method with word association function is used to reduce input time [15]; there are also schemes that use steady-state visual evoked potential (SSVEP, evoked potential) [88]. These methods only use a small part of the EEG signal, and are all translated through existing EEG paradigms. There is little improvement in BCI methodology. The human brain uses the sensory motor and language functions of the cerebral cortex to exchange information with the outside world [88], especially the language function with semantic information plays a key role. However, to date, there are still very few studies that consider directly analyzing signals containing semantic information. For example, in the exploration of applying semantics to BCI technology, a study conducted an experiment with two healthy subjects [93], using true and false statements as conditioned stimuli to examine distinguishable conditioned responses. After obtaining the conditioned responses, these conditioned responses correspond to true and false semantic information. This study used support vector machines (a radial basis function classifier) for offline classification. In the idea of distinguishing between the conditions "yes" and "no", the average classification accuracy of a single trial reached 50.5%, indicating that its classification accuracy is very low and the method still belongs to the category of traditional methods. Other studies have shown that the semantics of a single concept can be decoded from a single EEG data trial. The CSP method is used to extract the whole scalp synchronous activity component, and the general SVM machine learning algorithm is used, combined with the measurement results of spectral power to predict the semantic category of each trial. In some other studies [10], the semantically induced gamma EEG activity during sentence reading was studied. The changes in the EEG power spectrum caused by semantics were examined to examine the important influence of irregularities in sentences on sentence correctness, and the FFT method was used for processing. The survey shows that the current research on semantics in EEG is still in the exploratory stage, using traditional machine learning methods, which is completely different from the semantic paradigm model in this study.

考虑到目前的许多针对脑机接口(BCI)中脑电信号的模型方法主要是应对若干常规的目标任务,比如将某种已知范式转译成“是与否”的选择问题,但这些范式中多数没有引入语义的概念;实测中大量潜在的范式还未被发现,尚未被梳理总结成有效的语义和新的范式。此外,大多数针对运动相关电位的脑机接口的处理方法主要借助经验来寻找特征信息,较少借助工具和算法自动的搜寻到有效特征。在过去很长一段时间,导致这种现状的原因可能是相关的数据体量还不够大、数据处理过程复杂、方法效率不高等制约了上述目标的达成。近年来,深度学习的发展和硬件技术的进步,结合本文所提的语义范式模型,有潜力将这些新模型应用于对脑机接口的信号处理、特征自动搜寻和数据挖掘中。本研究对上述目标做了尝试和探索,并使用了若干数据集进行了测试,输出了呈现结果。语义范式模型所具有的开放性、扩展性和标准化的特征,为将来构建可重复使用的、基于本语义范式模型的脑机接口的标记特征数据库,提供了有效的解决方案。Considering that many current models and methods for EEG signals in brain-computer interfaces (BCI) are mainly used to deal with several conventional target tasks, such as translating a known paradigm into a "yes or no" choice problem, most of these paradigms do not introduce the concept of semantics; a large number of potential paradigms have not been discovered in actual measurements, and have not yet been sorted out and summarized into effective semantics and new paradigms. In addition, most of the processing methods for brain-computer interfaces of movement-related potentials mainly rely on experience to find feature information, and rarely use tools and algorithms to automatically search for effective features. For a long time in the past, the reasons for this situation may be that the relevant data volume is not large enough, the data processing process is complicated, and the method is not efficient, which restricts the achievement of the above goals. In recent years, the development of deep learning and the advancement of hardware technology, combined with the semantic paradigm model proposed in this article, have the potential to apply these new models to signal processing, automatic feature search and data mining of brain-computer interfaces. This study has made attempts and explorations on the above goals, and used several data sets for testing, and output presentation results. The openness, extensibility and standardization of the semantic paradigm model provide an effective solution for building a reusable brain-computer interface marker feature database based on this semantic paradigm model in the future.

目前BCI领域的范式模型有限,接口命令类型少,只有少量研究针对实际运动/运动想象的类型和参数做了分析,而且没有满意的结果。为了从脑电信号(包括头皮脑电图信号EEG,脑皮层脑电图信号ECoG,脑磁图信号MEG等多种电生理信号)中提取出更多具有明确语义的信息,本发明将跨度较大的三个领域(模糊性的电生理信号、兼容多种信号的深度学习和明确的语义模型)衔接,实现了一个新的语义范式框架:以脑电信号数据为研究对象,设计了相关的语法、语义处理规则和算法,赋予复杂信号以文法、语法和语义内涵,为深度神经网络构筑了语义解释。通过该范式框架对脑电信号进行分析,帮助研究者找出其中特定语义的信息块、以及这些信息块之间的语义组合,自动学习出高效的滤波器。如果用新采集的、体量更大的脑电信号的数据集来开展训练和测试,就能从脑电信号中提取出更多具有明确意义的信息,并结合自反馈机制来调整模型参数、更新学习模型,对不同大小的脑电数据集达到准确率高、传输通量大、拟合度优、普适性强的效果。At present, the paradigm models in the field of BCI are limited, and the types of interface commands are few. Only a small number of studies have analyzed the types and parameters of actual movements/motor imagination, and there are no satisfactory results. In order to extract more information with clear semantics from EEG signals (including scalp EEG signals, cortical EEG signals, ECoG signals, magnetoencephalogram signals, MEG signals and other electrophysiological signals), the present invention connects three fields with a large span (fuzzy electrophysiological signals, deep learning compatible with multiple signals and clear semantic models) to achieve a new semantic paradigm framework: taking EEG signal data as the research object, designing relevant grammar, semantic processing rules and algorithms, giving complex signals grammar, syntax and semantic connotations, and constructing semantic interpretations for deep neural networks. Through this paradigm framework, EEG signals are analyzed to help researchers find information blocks with specific semantics and semantic combinations between these information blocks, and automatically learn efficient filters. If a newly collected, larger data set of EEG signals is used for training and testing, more meaningful information can be extracted from the EEG signals, and the self-feedback mechanism can be combined to adjust the model parameters and update the learning model, achieving high accuracy, large transmission flux, excellent fitting, and strong universality for EEG data sets of different sizes.

发明内容Summary of the invention

为了解决上述问题、达到所述目标,本发明提供了一种对脑电信号进行文法、语法和语义范式处理的方法和装置,包括:In order to solve the above problems and achieve the above objectives, the present invention provides a method and device for processing grammar, syntax and semantic paradigms of EEG signals, including:

对脑电信号数据进行处理,得到特征信息;Process the EEG signal data to obtain feature information;

将特征信息继续分成有不同语义的组,每一组转换成特征字符串;The feature information is further divided into groups with different semantics, and each group is converted into a feature string;

对脑电信号信息进行处理,得到语法树;Process the EEG signal information to obtain a syntax tree;

由生成的语法树(数量可以为1或大于1的值)形成文法G[s]。The generated syntax trees (the number of which can be 1 or greater) form a grammar G[s].

所述脑电信号包括头皮脑电图信号,脑皮层电图信号,本方法可以用于其他电生理信号处理。The electroencephalogram (EEG) signals include scalp EEG signals and cortical EEG signals. The method can be used for processing other electrophysiological signals.

将脑电信号信息通过机器学习模型转化成相应的决策树;Convert EEG signal information into corresponding decision trees through machine learning models;

将决策树转化成语法树。Convert the decision tree into a syntax tree.

所述对脑电信号进行处理,其中包括对数据进行译码,就是,输入采集来的原始脑电信号数据,译码模块把经过预处理的脑电数据处理后得到离散的“标签序列”矩阵,每个标签有对应的语义块,如式4.2所示。机器学习模块在处理数据时,会把局部特征进行特征提取、降维和滤波。The processing of EEG signals includes decoding the data, that is, the raw EEG signal data collected is input, and the decoding module processes the pre-processed EEG data to obtain a discrete "label sequence" matrix, and each label has a corresponding semantic block, as shown in Formula 4.2. When processing data, the machine learning module will extract, reduce the dimension and filter the local features.

所述译码处理的具体实现,分为基于CNN网络模式、基于CTC+LSTM网络模式、其他RNN模式、小波变换模式等各种机器学习模式;对于基于CNN网络模型,该矩阵就相当于图7中的α2,矩阵元素就相当于α2中的灰色小块;于是得到了标签序列,以及标签序列所对应的特征数据,以及这两者之间的关系。如果是基于CTC+LSTM网络模型,该矩阵就对应于标签序列矩阵;就得到了标签序列,以及标签序列所对应的特征数据,这两者之间的关系,以及序列长度数据。序列长度是一维数据,其表现形式可以是:[max_time_step,…,max_time_step],其长度为batch_size,具体的值为max_time_step。还可以是通过小波变换形成的输出构成特征矩阵。The specific implementation of the decoding process is divided into various machine learning modes such as CNN network mode, CTC+LSTM network mode, other RNN mode, wavelet transform mode, etc.; for the CNN network model, the matrix is equivalent to α2 in Figure 7, and the matrix elements are equivalent to the gray blocks in α2; thus, the label sequence, the feature data corresponding to the label sequence, and the relationship between the two are obtained. If it is based on the CTC+LSTM network model, the matrix corresponds to the label sequence matrix; the label sequence, the feature data corresponding to the label sequence, the relationship between the two, and the sequence length data are obtained. The sequence length is one-dimensional data, and its expression can be: [max_time_step,…,max_time_step], its length is batch_size, and the specific value is max_time_step. It can also be a feature matrix formed by the output formed by wavelet transform.

所述母矩阵(式4.2)中每一个矩阵元素值的来源,除了可以直接来自把对局部特征进行特征提取、降维和滤波后的特征信息之外,该母矩阵中的元素信息,还可以来自机器学习网络或深度学习神经网络模型中的权重矩阵。The source of each matrix element value in the mother matrix (Formula 4.2) can come directly from the feature information after feature extraction, dimensionality reduction and filtering of local features. In addition, the element information in the mother matrix can also come from the weight matrix in the machine learning network or deep learning neural network model.

可选地,在进行数据译码之前,可以对脑电信号进行预处理;Optionally, the EEG signal may be preprocessed before data decoding;

所述机器学习模型包括卷积神经网络(CNN)模型,长短期记忆网络(LSTM)模型,其他RNN模型,小波变换模型,等各种机器学习模型;The machine learning models include convolutional neural network (CNN) models, long short-term memory network (LSTM) models, other RNN models, wavelet transform models, and other machine learning models;

所述对脑电信号数据译码,就是为后续的语义分析处理做准备,目的是为了把数据按不同层次处理成有语义特征、而且被标签化的有序形式。同时得到对应的决策树和语法树。语义分析处理需要将得到的标签化的特征和树(决策树和语法树)进行映射。The decoding of EEG signal data is to prepare for the subsequent semantic analysis and processing, the purpose is to process the data into an ordered form with semantic features and labels at different levels. At the same time, the corresponding decision tree and syntax tree are obtained. The semantic analysis and processing requires mapping the obtained labeled features and trees (decision trees and syntax trees).

按语法树生成句型、短语,用于高层次的文法、语法和语义分析处理。Generate sentence patterns and phrases according to the syntax tree for high-level grammar, syntax and semantic analysis.

文法还可以有另外一种产生方式:将得到的各特征字符串按文法的形式写成产生式,构成文法G[s];There is another way to generate grammar: write each characteristic string obtained into production rules in the form of grammar to form grammar G[s];

按文法生成特征语法树;Generate feature syntax tree according to grammar;

一种构造语法树的方法,其特征在于,所述构造语法树的处理,就是将输入的信息按照当前实验分析的语义关系结果,将降维后的关键特征信息块之间的关系,形成特征字母表,再通过特征字母表形成语义字母表,形成特征文法,形成特征生成式集(是一个由生成式或规则组成的非空有限集合),并进而形成特征语法树。根据特征语法树构造句型,以便算法在后续处理中使用。有了所述新形成的文法和特征语法树,对新采集的脑电信号形成句型、判定其短语、句柄和语法,理解其中的语义。A method for constructing a syntax tree, characterized in that the process of constructing the syntax tree is to form a feature alphabet according to the semantic relationship results of the current experimental analysis of the input information, and the relationship between the key feature information blocks after dimensionality reduction, and then form a semantic alphabet through the feature alphabet to form a feature grammar, form a feature generation formula set (a non-empty finite set composed of generation formulas or rules), and then form a feature syntax tree. Construct a sentence pattern according to the feature syntax tree so that the algorithm can be used in subsequent processing. With the newly formed grammar and feature syntax tree, the sentence pattern of the newly collected EEG signal is formed, its phrases, handles and grammar are determined, and the semantics therein is understood.

不但可以通过机器学习模型得到语法树,还可以通过实验构造、修改语法树,该处理就是:将输入的信息按照当前实验分析的语义关系结果,将降维后的关键特征信息块之间的关系,形成文法的产生式,形成特征信号文法,并进而形成特征语法树。根据特征语法树构造句型,以便算法在后续处理中使用。有了所述新形成的文法和特征语法树,对新采集的脑电信号形成句型、判定其短语、句柄和语法,理解其中的语义。Not only can the grammar tree be obtained through the machine learning model, but the grammar tree can also be constructed and modified through experiments. The process is: the input information is analyzed according to the semantic relationship results of the current experiment, and the relationship between the key feature information blocks after dimensionality reduction is formed into the production of grammar, forming a feature signal grammar, and then forming a feature grammar tree. Construct a sentence pattern according to the feature grammar tree so that the algorithm can be used in subsequent processing. With the newly formed grammar and feature grammar tree, the sentence pattern of the newly collected EEG signal is formed, its phrases, handles and grammar are determined, and the semantics therein is understood.

按前述两种方法得到的语法树,都能够生成句型、短语,用于高层次的文法、语法和语义分析处理。The syntax trees obtained by the above two methods can generate sentence patterns and phrases for high-level grammar, syntax and semantic analysis.

上面所述两种形成文法和语法树的不同路径,所形成的特征信号文法,都依赖于下列12个概念定义和规则定义:The two different paths for forming grammar and syntax tree described above, and the characteristic signal grammar formed by them, all rely on the following 12 concept definitions and rule definitions:

定义1:所述特征就是特征矩阵字母表,其定义为:.特征矩阵字母表:所有的输入信息被函数f处理成特征矩阵M1,M1的每一个元素同时也是一个矩阵M2,M2的元素是由二进制或十六进制形式的数值表示,所有M1的元素都属于同一个特征矩阵字母表,该字母表由不重复的元素构成,用AM表示。Definition 1: The feature is the feature matrix alphabet, which is defined as: . Feature matrix alphabet: All input information is processed by function f into a feature matrix M1 . Each element of M1 is also a matrix M2 . The elements of M2 are represented by numerical values in binary or hexadecimal form. All elements of M1 belong to the same feature matrix alphabet, which is composed of non-repeating elements and is represented by A M.

定义2.语义字母表:所有的输入信息M2的元素在语义上可以同时被函数fM->a转换映射成特定的符号,这些符号可以是被计算机系统识别的字符集中的元素,这些符号和运算符、分隔符构成一个有限的非空集合A,其中每一个元素是可以被识别(如计算机系统)的字符,也可以是多个字符,因此在本文讨论范围内,也可以用V表示,也就是说语义字母表和词汇表被视作相同,简称为字母表。Definition 2. Semantic alphabet: All elements of the input information M2 can be semantically mapped into specific symbols by the function fM->a . These symbols can be elements of the character set recognized by the computer system. These symbols, operators and separators constitute a finite non-empty set A, in which each element is a character that can be recognized (such as a computer system) or multiple characters. Therefore, within the scope of this article, it can also be represented by V. That is to say, the semantic alphabet and the vocabulary are regarded as the same, and are referred to as the alphabet for short.

定义3.特征符号串:经由定义1得到特定符号所组成的任何有限序列称之为符号串∑,又名字典。Definition 3. Characteristic symbol string: Any finite sequence of specific symbols obtained by Definition 1 is called a symbol string ∑, also known as a dictionary.

定义4.特征符号矩阵:由多个符号串以行排的方式组成二维符号矩阵,同理,可以扩展成三维的符号矩阵。Definition 4. Characteristic symbol matrix: A two-dimensional symbol matrix composed of multiple symbol strings arranged in rows. Similarly, it can be expanded into a three-dimensional symbol matrix.

定义5.特征符号串前缀:假设p是一特征符号串,从p的末尾删掉n个(n是小于p长度的自然数)符号后剩下的部分就是特征符号串p的前缀。Definition 5. Prefix of a characteristic symbol string: Assume that p is a characteristic symbol string. The remaining part after deleting n (n is a natural number less than the length of p) symbols from the end of p is the prefix of the characteristic symbol string p.

定义6.特征符号串后缀:假设p是一特征符号串,从p的首部开始删掉n个(n是小于p的长度的自然数)符号后剩下的部分就是特征符号串p的后缀。Definition 6. Characteristic symbol string suffix: Assume p is a characteristic symbol string, and the remaining part after deleting n (n is a natural number less than the length of p) symbols from the beginning of p is the suffix of the characteristic symbol string p.

定义7.特征符号串的子串:指从特征符号串p中删掉它的一个特征符号串前缀和一个特征符号串后缀之后,剩下的那部分称符号串被称为原特征符号串p的子串。Definition 7. Substring of a characteristic symbol string: refers to the part of the characteristic symbol string that remains after deleting a characteristic symbol string prefix and a characteristic symbol string suffix from the characteristic symbol string p.

定义8.特征符号串的连接:假设p和q各是两个特征符号串,那么合并得到的pq是将特征符号串q接续在特征符号串p的后面。Definition 8. Connection of feature symbol strings: Assuming that p and q are two feature symbol strings respectively, then the combined pq is the feature symbol string q connected to the feature symbol string p.

定义9.特征文法G的定义:G=(Vn,Vt,P,S)Definition 9. Definition of feature grammar G: G = (V n , V t , P, S)

Vn(非终结符号集)是一个由非终结符号(比如大写字母或<汉字>,或其他特定符号)组成的非空有限符号集合,所述的非终结符号指的是包括有限个须定义的数学运算规则的符号,在语义层次上就是语法范畴。V n (non-terminal symbol set) is a non-empty finite symbol set consisting of non-terminal symbols (such as uppercase letters or <Chinese characters>, or other specific symbols). The non-terminal symbols refer to symbols that include a finite number of mathematical operation rules that need to be defined, which is the grammatical category at the semantic level.

Vt(终结符号集)是一个由终结符号(比如小写字母、数字、标点符号,或其他特定符号)组成的非空有限符号集合,所述的终结符号是不需要进一步在文法中定义的基本符号,具有明确的数学或语义层次上的运算符含义。V=Vt∪Vn=A,V是该特征文法G的字母表A或词汇表。V t (terminal symbol set) is a non-empty finite symbol set consisting of terminal symbols (such as lowercase letters, numbers, punctuation marks, or other specific symbols). The terminal symbols are basic symbols that do not need to be further defined in the grammar and have clear operator meanings at the mathematical or semantic level. V = V t ∪ V n = A, where V is the alphabet A or vocabulary of the feature grammar G.

定义10.P(特征生成式集)是一个由生成式或规则组成的非空有限集合。Definition 10. P (feature generation set) is a non-empty finite set of generation rules or rules.

特征生成式的形式为: The form of the feature generator is: or

特征生成式的左边元素所属集合表达为α∈V+,因此α不能为空,生成式的右边元素所属集合表达为β∈V*,上述表达式中的的内涵一致,其内涵代表“由……构成”或“定义为”。The set to which the elements on the left of the feature generator belong is expressed as α∈V+, so α cannot be empty. The set to which the elements on the right of the generator belong is expressed as β∈V * . or is consistent with the connotation of, which means "composed of..." or "defined as".

S是特征文法的起始符号,S∈Vn,且必要条件是:S必须在某个生成式的左边至少出现1次。S is the starting symbol of the feature grammar, S∈V n , and the necessary condition is that S must appear at least once on the left side of a production.

定义11.特征语法分析树:由于前述定义1和定义2的映射逻辑,语法分析树的叶子节点既可以是A的元素,也可以是AM的元素。设有特征文法G=(Vn,Vt,P,S),对于特征文法G的任一句型均能找到一棵相应的特征语法树,它支持以下4个前提:Definition 11. Feature grammar parse tree: Due to the mapping logic of Definition 1 and Definition 2 above, the leaf nodes of the grammar parse tree can be either elements of A or elements of AM . Suppose there is a feature grammar G = ( Vn , Vt , P, S). For any sentence type of the feature grammar G, a corresponding feature grammar tree can be found, which supports the following four premises:

语法树上的每个node都有一个标识,它是V集合中的一个特征符号。Each node in the syntax tree has an identifier, which is a characteristic symbol in the set V.

语法树的根node是特征文法的起始符号S。The root node of the syntax tree is the start symbol S of the feature grammar.

如果语法树上的一个node至少有一个直接后继node,那么该node必为一非终结特征符号。若语法树上的一个node N有若干个直接后继node,则按自左向右/自右向左的顺序进行排序。If a node on the syntax tree has at least one direct successor node, then the node must be a non-terminal feature symbol. If a node N on the syntax tree has several direct successor nodes, they are sorted from left to right/right to left.

定义12.如果特征文法中的规则表达式均拥有下面的表达方式:Definition 12. If the regular expressions in the feature grammar have the following expressions:

设Cn为非终结特征符号,且所属集合为U∈Vn,u∈V*,即该特征文法里的规则的左边必为一Cn符号,规则表达式的右边u是V上的特征符号序列表达式。Let Cn be a non-terminal feature symbol, and its set is U∈V n , u∈V * , that is, the left side of the rule in the feature grammar must be a Cn symbol, and the right side u of the rule expression is a feature symbol sequence expression on V.

与现有技术相比,本申请技术方案包括:将脑电信息转换成有明确语义的信息后,文法中就新增了明确的语义,将得到的各个语义特征块组成语义范式。Compared with the prior art, the technical solution of the present application includes: after converting EEG information into information with clear semantics, clear semantics are added to the grammar, and the various semantic feature blocks obtained are combined into a semantic paradigm.

本发明实现了一个新的语义范式框架:以脑电数据为研究对象,设计了相关的语法、语义处理规则和算法,赋予复杂信号以文法、语法和语义内涵,为深度神经网络构筑了语义解释。通过该范式框架对脑电信号进行分析,帮助研究者找出其中特定语义的信息块、以及这些信息块之间的语义组合,自动学习出高效的滤波器。如果用更多更大新采集的脑机接口数据集来开展训练和测试,就能从脑电信号中提取出更多具有明确意义的信息,并结合自反馈机制来调整模型参数、更新学习模型,对不同大小的脑电数据集达到准确率高、传输通量大、拟合度优、普适性强的效果。The present invention implements a new semantic paradigm framework: taking EEG data as the research object, relevant grammar, semantic processing rules and algorithms are designed, complex signals are given grammar, syntax and semantic connotations, and semantic interpretation is constructed for deep neural networks. Through the paradigm framework, EEG signals are analyzed to help researchers find information blocks with specific semantics and semantic combinations between these information blocks, and automatically learn efficient filters. If more and larger newly collected brain-computer interface data sets are used for training and testing, more information with clear meaning can be extracted from EEG signals, and the self-feedback mechanism can be combined to adjust the model parameters and update the learning model, so as to achieve high accuracy, large transmission flux, good fit and strong universality for EEG data sets of different sizes.

所述脑电信号包括头皮脑电图信号,大脑皮层电图信号。本发明方法可以用于其他电生理信号处理。The electroencephalogram signal includes scalp electroencephalogram signal and cerebral cortex electroencephalogram signal. The method of the present invention can be used for other electrophysiological signal processing.

相应地,本发明提供一种对脑电信号、大脑皮层电信号及其他电生理信号进行处理的装置,其特征在于,所述装置包括(图17所示):Accordingly, the present invention provides a device for processing electroencephalogram signals, cerebral cortex electrical signals and other electrophysiological signals, characterized in that the device comprises (as shown in FIG. 17 ):

信号采集单元;Signal acquisition unit;

对脑电信号数据进行处理,得到特征信息的特征信号处理单元;A feature signal processing unit that processes the EEG signal data to obtain feature information;

将特征信息分成有不同语义的组,每一组转换成特征字符串的特征字符串形成单元;The feature information is divided into groups with different semantics, and each group is converted into a feature string forming unit of the feature string;

从特征字符串形成字母表的形成字母表单元;forming alphabet units of the alphabet from the characteristic string;

对脑电信号信息进行处理,得到语法树的形成语法树单元;Processing the EEG signal information to obtain a syntax tree unit for forming a syntax tree;

由生成的语法树(数量可以为1或大于1的值)形成文法G[s]的形成文法单元;The generated syntax trees (the number of which can be 1 or greater) form grammar units of the grammar G[s];

如图3所示,被虚线框框住的部分是译码处理,被实线框框住的部分是文法、语法和语义范式分析处理;译码处理之后,原始数据被简化,数据维度被降低,参数被减少,得到关键信息被突出的特征信息;图中的解码处理包括,在输入给机器学习模型之前的、由本发明方法所列出的各种对脑电信号的专门预处理。As shown in Figure 3, the part enclosed by the dotted box is the decoding process, and the part enclosed by the solid box is the grammar, syntax and semantic paradigm analysis process; after the decoding process, the original data is simplified, the data dimension is reduced, the parameters are reduced, and the feature information with key information highlighted is obtained; the decoding process in the figure includes various special pre-processing of the EEG signal listed by the method of the present invention before inputting it into the machine learning model.

本发明的其它特征和优点将在随后的说明书中阐述,并且,部分地从说明书中变得显而易见,或者通过实施本发明而了解。本发明的目的和其他优点可通过在说明书、权利要求书以及附图中所特别指出的结构来实现和获得。Other features and advantages of the present invention will be described in the following description, and partly become apparent from the description, or understood by practicing the present invention. The purpose and other advantages of the present invention can be realized and obtained by the structures particularly pointed out in the description, claims and drawings.

附图说明BRIEF DESCRIPTION OF THE DRAWINGS

此处所说明的附图用来提供对本发明的进一步理解,构成本申请的一部分,本发明的示意性实施例及其说明用于解释本发明,并不构成对本发明的不当限定。在附图中:The drawings described herein are used to provide a further understanding of the present invention and constitute a part of this application. The exemplary embodiments of the present invention and their descriptions are used to explain the present invention and do not constitute an improper limitation of the present invention. In the drawings:

图1为现有传统处理流程Figure 1 shows the existing traditional processing flow

图2为本发明含语义解析的处理流程FIG. 2 is a processing flow of the present invention including semantic analysis

图3为总体工作流程Figure 3 shows the overall workflow

图4为降维后的类别1的索引的三维可视化呈现效果图Figure 4 is a three-dimensional visualization of the index of category 1 after dimensionality reduction

图5为句型ddefdd的特征语义语法树Figure 5 is the feature semantic syntax tree of the sentence pattern ddefdd

图6为通过语法树求一个句型的短语Figure 6 is a phrase for finding a sentence pattern through a syntax tree

图7为卷积层滤波Figure 7 shows the convolutional layer filtering

图8为电位序列形成矩阵字母表和标签化Figure 8 shows the potential sequence forming the matrix alphabet and labeling

图9为在字母表上的前缀搜索译码Figure 9 shows the prefix search decoding on the alphabet

图10决策树形成Figure 10 Decision tree formation

图11为LSTM的Block中要更新的权重Figure 11 shows the weights to be updated in the LSTM Block

图12为语义分析和解释流程Figure 12 shows the semantic analysis and interpretation process

图13为特征信息矩阵Figure 13 is the feature information matrix

图14为特征矩阵Figure 14 is the feature matrix

图15为标注了终结符r的特征矩阵Figure 15 is a feature matrix with the terminal symbol r marked

图16为从可视化图获取原始特征Figure 16 shows how to obtain the original features from the visualization graph.

图17为本发明的装置构成FIG. 17 is a diagram showing the structure of the device of the present invention.

具体实施方式DETAILED DESCRIPTION

为使本发明的目的、技术方案和优点更加清楚明白,下文中将结合附图对本发明的实施例进行详细说明。需要说明的是,在不冲突的情况下,本申请中的实施例及实施例中的特征可以相互任意组合。In order to make the purpose, technical solution and advantages of the present invention more clear, the embodiments of the present invention will be described in detail with reference to the accompanying drawings. It should be noted that the embodiments and features in the embodiments of the present application can be combined with each other arbitrarily without conflict.

目前的BCI模型处理只解决了对若干简单目标任务的是与否问题,这些任务对应着一些明确的范式,而且只能用概率来表明准确率,没有语义的概念模型,实际测试中的大量潜在范式无法总结出有效的语义和新范式,此外,寻找特征信息模型基本全凭借已有经验,难以借助工具和算法进行自动化寻找。导致这种问题的原因之一是数据处理的复杂,进而导致效率低下,制约了上述任务目标的达成。与此同时,近年深度学习的发展和硬件的进步迅速,然而这些技术尚未有明确的可解释的模型用于BCI领域。本研究考虑用明确的文法模型连接两个领域的最新研究成果,以提高对BCI数据模型和语义模型的提升,提高对数据的利用效率,为创造新的模型和范式提供思路和工具。The current BCI model processing only solves the yes or no problem of several simple target tasks. These tasks correspond to some clear paradigms, and the accuracy can only be indicated by probability. There is no semantic conceptual model. A large number of potential paradigms in actual tests cannot summarize effective semantics and new paradigms. In addition, the search for feature information models is basically based on existing experience, and it is difficult to automate the search with the help of tools and algorithms. One of the reasons for this problem is the complexity of data processing, which leads to low efficiency and restricts the achievement of the above task goals. At the same time, the development of deep learning and hardware progress have been rapid in recent years, but these technologies have not yet had clear and interpretable models for use in the BCI field. This study considers using a clear grammar model to connect the latest research results in the two fields to improve the BCI data model and semantic model, improve the efficiency of data utilization, and provide ideas and tools for creating new models and paradigms.

虽然在当下,深度神经网络在许多视觉任务中取得了很好的成果,但是仍未在基于脑机接口的脑电信号的数据挖掘中得到较多的运用,仍然面临一些问题需要解决。本发明的目标就是要将基于传统机器学习的改进方法、深度学习方法和语义范式模型等几种方法应用于对脑电信号的数据挖掘,以应对上述挑战和问题。目前BCI领域的实验范式模型不仅有限,而且其接口命令的类型少,使得BCI技术始终要依靠典型范式进行转译,如何直接理解人类思维并构造仿真模型一直是一个难题,囿于原始意义解读的困难,以及分析手段的低下,使得这个问题一直没有进展。为了探索该问题的解决,研究人员将跨度较大的三个领域(模糊性的电生理信号、兼容各种异构信号的深度学习和明确的语义模型)衔接,将最新的深度学习技术应用到BCI数据挖掘领域,通过对脑电信号进行译码,把脑电数据处理成能寻址到对应语义块的离散的标签序列,同时通过对深度神经网络的语义解析,将深度神经网络转化成相应的决策树和语法树,由此我们实现了一个新的以脑电信号数据为研究对象的语义范式框架,为脑电信息的语义分析做了初步但十分有意义的探索:设计了相关的语法、语义处理规则和算法,赋予复杂信号以文法、语法和语义内涵,为深度神经网络构筑了语义解释,对脑电信息的研究从简单的命令范式提升到语义和语法的高度,提升了对脑电信号的复杂模式研究效率。通过该范式框架能够对脑电信号中的信息进行语义范式分析,帮助研究者找出其中包含的特定语义的信息块、以及这些信息块之间的语义组合,自动学习出高效的滤波器,达到准确率高、传输通量大、普适性强的效果。可以利用该框架可以降维并获取关键特征信息,高效率地总结出新的范式特征,优化学习模型;同时利用一系列的可视化功能,观察对数据的训练情况,并能够通过对数据降维后的结果获取对应的原始特征,提高了分析脑电信号特征的效率。Although deep neural networks have achieved good results in many visual tasks at present, they have not been widely used in data mining of EEG signals based on brain-computer interfaces, and there are still some problems to be solved. The goal of the present invention is to apply several methods such as improved methods based on traditional machine learning, deep learning methods and semantic paradigm models to data mining of EEG signals to meet the above challenges and problems. At present, the experimental paradigm models in the field of BCI are not only limited, but also have few types of interface commands, which makes BCI technology always rely on typical paradigms for translation. How to directly understand human thinking and construct simulation models has always been a difficult problem. Due to the difficulty of interpreting the original meaning and the low level of analysis methods, this problem has not made any progress. In order to explore the solution to this problem, the researchers connected three fields with a large span (fuzzy electrophysiological signals, deep learning compatible with various heterogeneous signals, and clear semantic models), applied the latest deep learning technology to the field of BCI data mining, decoded EEG signals, processed EEG data into discrete label sequences that can address corresponding semantic blocks, and converted deep neural networks into corresponding decision trees and syntax trees through semantic analysis of deep neural networks. As a result, we realized a new semantic paradigm framework with EEG signal data as the research object, and made preliminary but very meaningful explorations for the semantic analysis of EEG information: relevant grammar, semantic processing rules and algorithms were designed, complex signals were given grammar, syntax and semantic connotations, and semantic interpretations were constructed for deep neural networks. The research on EEG information was elevated from a simple command paradigm to the height of semantics and syntax, and the efficiency of complex pattern research on EEG signals was improved. Through this paradigm framework, semantic paradigm analysis of information in EEG signals can be performed to help researchers find information blocks with specific semantics contained therein and the semantic combinations between these information blocks, and automatically learn efficient filters to achieve high accuracy, large transmission flux and strong universality. This framework can be used to reduce the dimension and obtain key feature information, efficiently summarize new paradigm features, and optimize the learning model. At the same time, a series of visualization functions can be used to observe the training status of the data, and the corresponding original features can be obtained through the results of data dimensionality reduction, thereby improving the efficiency of analyzing EEG signal features.

如图1和图2所示,传统处理流程之外,可以新增语义处理模块,对语义的解析结果可以反馈给语义模块和机器学习模块,以帮助改进语义模型和内部模型,语义模型的改进可以新增特征信息和语义的映射关系,内部模型的改进可以突出关键特征,发掘新的特征和范式。As shown in Figures 1 and 2, in addition to the traditional processing flow, a semantic processing module can be added. The semantic analysis results can be fed back to the semantic module and the machine learning module to help improve the semantic model and the internal model. The improvement of the semantic model can add a mapping relationship between feature information and semantics, and the improvement of the internal model can highlight key features and discover new features and paradigms.

得到新的语义模型(含特征位置信息)之后,对新的数据可以直接使用:对新数据的特征位置不再做重复的学习,而是发掘新的时空过滤器模型,或者发掘不同特征之间的潜在逻辑关联,对新数据先用现有训练结果预测其标签类型,如果计算发现准确概率达到认可标准(该标准可以自己定义),就不必修正模型,也不必新增语义推导式;如果准确率不够,就将该新数据作为训练数据,未来可以重新训练模型。这种算法将使得过拟合和欠拟合问题得到修正,更重要的是使得每一次计算得到的结果成为有特定语义的结果,并被记录,研究人员可凭借此记录进行演绎,制定下一步滤波器设计方案,而且形成的语义关系树可以方便地采用依赖关系特征处理以进行更好的预处理。这种方案使得过去的计算结果有了特定意义,不会随着新增数据量的增加而重新做大量训练,学习结果难以收敛,相反,会大大加快研究进度,使特定问题的分析能够以更快的方式收敛,计算量逐步下降,越到后期将越使得对硬件的要求降低,对在线分析的速度要求有了保障。After obtaining the new semantic model (including feature position information), the new data can be used directly: the feature position of the new data is no longer repeatedly learned, but a new spatiotemporal filter model is discovered, or the potential logical association between different features is discovered. The existing training results are used to predict the label type of the new data. If the calculation finds that the accuracy probability reaches the recognition standard (the standard can be defined by oneself), there is no need to modify the model or add a new semantic derivation formula; if the accuracy is not enough, the new data is used as training data, and the model can be retrained in the future. This algorithm will correct the overfitting and underfitting problems. More importantly, it makes the results obtained from each calculation become results with specific semantics and recorded. Researchers can use this record to deduce and formulate the next filter design plan. In addition, the semantic relationship tree formed can easily use dependency feature processing for better preprocessing. This solution makes the past calculation results have specific meanings. It will not re-train a lot with the increase of the amount of new data, and the learning results will be difficult to converge. On the contrary, it will greatly speed up the research progress, so that the analysis of specific problems can converge in a faster way, and the amount of calculation will gradually decrease. The later the hardware requirements will be reduced, and the speed requirements for online analysis will be guaranteed.

虽然神经网络在许多视觉任务中取得了很好的结果,但是模型本身的可解释性仍然面临诸多挑战。一个模型可解释性的高低决定了其究竟可以被视为黑盒还是白盒,显然,作为白盒的倾向性越明显,则该模型能够被应用的场合越明确,能够被优化的程度就越高。然而在现实中,往往由于模型的可解释性尚不明确,导致许多开发者不敢大胆使用,更谈不上如何对模型进行优化和扩展。在此,我们把神经网络对脑电信号的处理抽象成人类语言语义的形成过程,而对人类语言文法进行了数学建模的形式化方法是乔姆斯基创立的,我们将提出一种既能适配脑电信号及神经网络处理、又符合文法规则的一种具体的形式化描述。该方法适用于对头皮脑电图信号EEG,脑皮层脑电图信号ECoG,脑磁图信号MEG处理(可以推广到其他各种电生理信号),能够从这些信号中解读出更多的语义信息,并形成可供更高层次应用使用的文法。而且本发明方法可以用于其他电生理信号处理。Although neural networks have achieved good results in many visual tasks, the interpretability of the model itself still faces many challenges. The interpretability of a model determines whether it can be regarded as a black box or a white box. Obviously, the more obvious the tendency to be a white box is, the clearer the occasions where the model can be applied and the higher the degree of optimization. However, in reality, many developers dare not use it boldly because the interpretability of the model is still unclear, let alone how to optimize and expand the model. Here, we abstract the processing of EEG signals by neural networks into the formation process of human language semantics, and the formal method of mathematical modeling of human language grammar was created by Chomsky. We will propose a specific formal description that can adapt to EEG signals and neural network processing and conform to grammar rules. This method is suitable for processing scalp EEG signals, cortical EEG signals, and magnetoencephalogram signals (which can be extended to various other electrophysiological signals). It can interpret more semantic information from these signals and form grammars that can be used for higher-level applications. Moreover, the method of the present invention can be used for other electrophysiological signal processing.

从BCI研究的历史进程来看,脑电信号由于被淹没在各种噪声之中,以及人类对大脑电生理机制研究的缓慢,使得对BCI的研究还停留在初级阶段,需要有新的工具和思路帮助提高研究效率,发现新的语义模型或范式。Judging from the historical process of BCI research, EEG signals are submerged in various noises, and human research on the electrophysiological mechanisms of the brain is slow, so research on BCI is still in its early stages. New tools and ideas are needed to help improve research efficiency and discover new semantic models or paradigms.

经过验证可以知道,同一个神经网络模型对同一个数据集的降维后得到的数据是稳定不变的,这使得我们可以基于已有数据集计算得到其对应的语法和语义的有效集合;利用这个已有的有效集合作为模板,我们可以去对新的测试数据集进行有针对性的降维去噪,达到大幅度降维的目的,同时不会影响判决结果。考虑到被试在不同测试批次之间的信号波动,对新引入的信号不是做完全匹配,而是允许一定比例的误差,这种运行逻辑和全数据集的计算相同,由于数据维度已经大幅度降低,此时的匹配运算速度会非常快。It has been verified that the data obtained by the same neural network model after dimensionality reduction of the same data set is stable and unchanged, which allows us to calculate the corresponding grammatical and semantic valid sets based on the existing data set; using this existing valid set as a template, we can carry out targeted dimensionality reduction and denoising on the new test data set to achieve the purpose of substantial dimensionality reduction without affecting the judgment results. Taking into account the signal fluctuations of the subjects between different test batches, the newly introduced signals are not fully matched, but a certain proportion of errors are allowed. This operation logic is the same as the calculation of the full data set. Since the data dimension has been greatly reduced, the matching operation speed at this time will be very fast.

所述机器学习模型包括卷积神经网络(CNN)模型,长短期记忆网络(LSTM)模型,其他RNN模型,小波变换模型等等各种机器学习模型。The machine learning models include various machine learning models such as convolutional neural network (CNN) models, long short-term memory network (LSTM) models, other RNN models, wavelet transform models, etc.

总体工作流程如图3所示。图的左侧是流程图,右侧是对应于当前左侧步骤时的数据表征、数据的内部形态和数据结构,图3的右侧部分单独放大之后就是图8。The overall workflow is shown in Figure 3. The left side of the figure is a flowchart, and the right side is the data representation, internal form and data structure corresponding to the current left step. The right part of Figure 3 is enlarged separately to become Figure 8.

图中被虚线框框住的部分是译码处理,被实线框框住的部分是文法、语法和语义范式处理。译码处理是为后续的语义分析处理做准备,目的是为了把数据按不同层次处理成有语义特征、而且被标签化的有序形式。语义分析处理是将得到的标签化的特征构造决策树和语法树,并对这两者进行映射。下面将图中关键模块及其使用的概念进行详细的阐述。The part enclosed by the dotted line in the figure is the decoding process, and the part enclosed by the solid line is the grammar, syntax and semantic paradigm processing. The decoding process is to prepare for the subsequent semantic analysis process, and the purpose is to process the data into an ordered form with semantic features and labels according to different levels. The semantic analysis process is to construct a decision tree and a syntax tree with the obtained labeled features, and map the two. The key modules in the figure and the concepts used are explained in detail below.

本发明涉及如下概念和规则定义。The present invention involves the following concepts and rule definitions.

由于脑电的信号本身是无法象计算机信号那样有明确的值,因此,在目前的计算机体系中是无法用明确的无法来表达具有一定随机性和概率性质的脑电信号的语义的,所以,我们必须在通过某种转换方式得到脑电信息的语义特征之前,必须有相应的对脑电信号进行译码的模型,将这种模糊的信息转换成明确的的信息。因此,我们必须为文法语义和语义范式模型定义一套独有的概念体系,以便和明确的计算机体系的功能进行区分。因此,有以下一系列定义。Since EEG signals themselves cannot have clear values like computer signals, it is impossible to express the semantics of EEG signals with certain randomness and probability in the current computer system. Therefore, before we can obtain the semantic features of EEG information through some conversion method, we must have a corresponding model to decode EEG signals and convert this vague information into clear information. Therefore, we must define a unique concept system for grammatical semantics and semantic paradigm models in order to distinguish them from the functions of clear computer systems. Therefore, there are the following series of definitions.

定义1.特征矩阵字母表:所有的输入信息被函数f处理成特征矩阵M1,M1的每一个元素同时也是一个矩阵M2,M2的元素是由二进制或十六进制形式的数值表示,所有M1的元素都属于同一个特征矩阵字母表,该字母表由不重复的元素构成,用AM表示。Definition 1. Feature matrix alphabet: All input information is processed by function f into a feature matrix M1 . Each element of M1 is also a matrix M2 . The elements of M2 are represented by binary or hexadecimal numbers. All elements of M1 belong to the same feature matrix alphabet, which consists of non-repeating elements and is represented by A M.

定义2.语义字母表:所有的输入信息M2的元素在语义上可以同时被函数fM->a转换映射成特定的符号,这些符号可以是被计算机系统识别的字符集中的元素,这些符号和运算符、分隔符构成一个有限的非空集合A,其中每一个元素是可以被识别(如计算机系统)的字符,也可以是多个字符组成的字符串,因此在本文讨论范围内,也可以用V表示,也就是说语义字母表和语义词汇表被视作相同,简称为字母表。Definition 2. Semantic alphabet: All elements of the input information M2 can be semantically mapped into specific symbols by the function fM->a . These symbols can be elements of the character set recognized by the computer system. These symbols, operators and separators constitute a finite non-empty set A, in which each element is a character that can be recognized (such as a computer system) or a string of multiple characters. Therefore, within the scope of this article, it can also be represented by V. That is to say, the semantic alphabet and the semantic vocabulary are regarded as the same, and are referred to as the alphabet for short.

字母表的元素以字符或字符串表示,所有的元素都属于同一个字母表,字母表由不重复的元素构成。字母表的元素由字符集定义该元素中的字符,所述字符集可以是ASCII字符集,或者双字节字符集DBCS,或者Unicode字符集。The elements of the alphabet are represented by characters or strings, all elements belong to the same alphabet, and the alphabet is composed of non-repeating elements. The elements of the alphabet are defined by a character set, and the character set can be an ASCII character set, a double-byte character set DBCS, or a Unicode character set.

例如,V中的成员是0xA,0xB,0x1A2B,0x1A2B3C等等。这些二进制串都是被神经网络降维处理之后的有效特征的二进制表达。For example, the members of V are 0xA, 0xB, 0x1A2B, 0x1A2B3C, etc. These binary strings are binary expressions of effective features after dimensionality reduction by the neural network.

所述运算符有特定的数学定义,含有限个须定义的数学运算规则的符号,在语义层次上就是语法范畴。The operators have specific mathematical definitions, contain symbols with a finite number of mathematical operation rules that must be defined, and are grammatical categories at the semantic level.

所述分隔符用于分隔不同的语义单元,可以用空格表示,也可以是其他无歧义的定义。例如,在脑电信号中,可以将当前关注的有明确语义的脑电特征信号之外的译作分隔符。The separator is used to separate different semantic units, and can be represented by a space or other unambiguous definitions. For example, in an EEG signal, the signal other than the EEG characteristic signal with clear semantics that is currently being focused on can be translated as a separator.

定义3.特征符号串:经由定义1得到特定符号所组成的任何有限序列称之为符号串∑,又名字典。Definition 3. Characteristic symbol string: Any finite sequence of specific symbols obtained by Definition 1 is called a symbol string ∑, also known as a dictionary.

例如,假设V={1,a,b},则1,a,b,11,1a,1b,a1,ab,aa,b1,ba,111,aaa,bbb都是V上的符号串。符号串1ab的长度为3,记为|1ab|=3。For example, assuming V = {1, a, b}, then 1, a, b, 11, 1a, 1b, a1, ab, aa, b1, ba, 111, aaa, bbb are all symbol strings on V. The length of the symbol string 1ab is 3, denoted by |1ab| = 3.

定义4.特征符号矩阵:由多个符号串以行排的方式组成二维符号矩阵,同理,可以扩展成三维的符号矩阵。Definition 4. Characteristic symbol matrix: A two-dimensional symbol matrix composed of multiple symbol strings arranged in rows. Similarly, it can be expanded into a three-dimensional symbol matrix.

以输入神经网络的信息为例,假设某一个通道的信息被经过CNN/LSTM网络处理之后,得到包含关键特征信息的一个二进制串,如果把该串转化成十六进制表达,就可以得到一个十六进制的字符串,该字符串中可以由字母表中的若干个元素组合而成,而具有关键特征信息的该十六进制串就是定义3的符号串,或者定义4的符号矩阵。以左右手运动的二分类问题为例,经过神经网络处理之后,得到降维后的类别1的索引的相关可视化结果如图4所示:Taking the information input into the neural network as an example, suppose that the information of a certain channel is processed by the CNN/LSTM network to obtain a binary string containing key feature information. If the string is converted into hexadecimal expression, a hexadecimal string can be obtained. The string can be composed of several elements in the alphabet, and the hexadecimal string with key feature information is the symbol string of Definition 3 or the symbol matrix of Definition 4. Taking the binary classification problem of left and right hand movements as an example, after the neural network processing, the relevant visualization results of the index of category 1 after dimensionality reduction are shown in Figure 4:

上述索引对应的原始信号组成了该类别的一个符号串,假设是0x1a2b34567890,则可以表示为:L->1a2b34567890;我们同样可以知道某个索引信息对应的附近的原始信号,并且将这些原始信号组成的符号串与该索引对应:22->0d265b8792c;这样的好处是可以定义某个特定信号的局部特征,以P300信号来举例,某个P300信号可以由若干个索引组合而成,比如52、57、7(假设图4的子图b中相邻的三个索引构成了降维后的P300信号的典型特征值)。不同被试的P300信号的脑电信息在经过神经网络模型处理之后,可以被转化为多种可能的字符串,这些字符串之间具有一定的相似性,但都能表达P300的关键特征。为了更加方便理解,可以将L视作一个句子,22视作一个单词,而->右侧的字符都是字母表中的元素。符号矩阵在进行语法语义表述时,可以转化成符号串形式以统一处理,而在进行其他处理时仍然按张量进行存储和运算。The original signals corresponding to the above indexes constitute a symbol string of this category. Assuming it is 0x1a2b34567890, it can be expressed as: L->1a2b34567890; we can also know the original signals nearby corresponding to a certain index information, and correspond the symbol string composed of these original signals to the index: 22->0d265b8792c; the advantage of this is that the local features of a specific signal can be defined. Taking the P300 signal as an example, a P300 signal can be composed of several indexes, such as 52, 57, and 7 (assuming that the three adjacent indexes in sub-graph b of Figure 4 constitute the typical feature values of the P300 signal after dimensionality reduction). After being processed by the neural network model, the EEG information of the P300 signals of different subjects can be converted into a variety of possible character strings. These character strings have certain similarities, but they can all express the key features of P300. For easier understanding, L can be regarded as a sentence, 22 as a word, and the characters on the right side of -> are all elements in the alphabet. When expressing grammatical semantics, the symbol matrix can be converted into a symbol string for unified processing, while it is still stored and calculated as a tensor when performing other processing.

由上面的举例可以知道,为了方便对对字符串及其子集串进行操作,我们还需要定义一些新概念,以及这些概念之间进行运算的规则。From the above examples, we can see that in order to facilitate operations on strings and their subsets, we also need to define some new concepts and rules for operations between these concepts.

定义5.特征符号串前缀:假设p是一特征符号串,从p的末尾删掉n个(n是小于p长度的自然数)符号后剩下的部分就是特征符号串p的前缀。Definition 5. Prefix of a characteristic symbol string: Assume that p is a characteristic symbol string. The remaining part after deleting n (n is a natural number less than the length of p) symbols from the end of p is the prefix of the characteristic symbol string p.

例如,1,1a,1a2b34567890都是1a2b34567890的前缀。在本文的语义中,特征符号串可以简称为符号串。For example, 1, 1a, 1a2b34567890 are all prefixes of 1a2b34567890. In the semantics of this article, the characteristic symbol string can be referred to as the symbol string.

定义6.特征符号串后缀:假设p是一特征符号串,从p的首部开始删掉n个(n是小于p的长度的自然数)符号后剩下的部分就是特征符号串p的后缀。Definition 6. Characteristic symbol string suffix: Assume that p is a characteristic symbol string. The remaining part after deleting n (n is a natural number less than the length of p) symbols from the beginning of p is the suffix of the characteristic symbol string p.

定义7.特征符号串的子串:指从特征符号串p中删掉它的一个特征符号串前缀和一个特征符号串后缀之后,剩下的那部分称符号串被称为原特征符号串p的子串。Definition 7. Substring of a characteristic symbol string: refers to the part of the characteristic symbol string that remains after deleting a characteristic symbol string prefix and a characteristic symbol string suffix from the characteristic symbol string p.

定义8.特征符号串的连接:假设p和q各是两个特征符号串,那么合并得到的pq是将特征符号串q接续在特征符号串p的后面。Definition 8. Connection of feature symbol strings: Assuming that p and q are two feature symbol strings respectively, then the combined pq is the feature symbol string q connected to the feature symbol string p.

定义9.特征文法G的定义:G=(Vn,Vt,P,S)Definition 9. Definition of feature grammar G: G = (V n , V t , P, S)

Vn(非终结符号集)是一个由非终结符号(比如大写字母或<汉字>,或其他特定符号)组成的非空有限符号集合,所述的非终结符号指的是包括有限个须定义的数学运算规则的符号,在语义层次上就是语法范畴。V n (non-terminal symbol set) is a non-empty finite symbol set consisting of non-terminal symbols (such as uppercase letters or <Chinese characters>, or other specific symbols). The non-terminal symbols refer to symbols that include a finite number of mathematical operation rules that need to be defined, which is the grammatical category at the semantic level.

Vt(终结符号集)是一个由终结符号(比如小写字母、数字、标点符号,或其他特定符号)组成的非空有限符号集合,所述的终结符号是不需要进一步在文法中定义的基本符号,具有明确的数学或语义层次上的运算符含义。V=Vt∪Vn=A,V是该特征文法G的字母表A或词汇表。V t (terminal symbol set) is a non-empty finite symbol set consisting of terminal symbols (such as lowercase letters, numbers, punctuation marks, or other specific symbols). The terminal symbols are basic symbols that do not need to be further defined in the grammar and have clear operator meanings at the mathematical or semantic level. V = V t ∪ V n = A, where V is the alphabet A or vocabulary of the feature grammar G.

定义10.P(特征生成式集)是一个由生成式或规则组成的非空有限集合。Definition 10. P (feature generation set) is a non-empty finite set of generation rules or rules.

特征生成式的形式为:或α→βThe form of the feature generator is: or α→β

特征生成式的左边元素所属集合表达为α∈V+,因此α不能为空,生成式的右边元素所属集合表达为β∈V*,上述表达式中的的内涵一致,其内涵代表“由……构成”或“定义为”。The set to which the elements on the left of the feature generator belong is expressed as α∈V+, so α cannot be empty. The set to which the elements on the right of the generator belong is expressed as β∈V * . or is consistent with the connotation of, which means "composed of..." or "defined as".

S是特征文法的起始符号,S∈Vn,且必要条件是:S必须在某个生成式的左边至少出现1次。S is the starting symbol of the feature grammar, S∈V n , and the necessary condition is that S must appear at least once on the left side of a production.

例如,终结符可以由对脑电信号进行语义分析之后,根据语义块之间的关系,定义专门的语义运算符作为终结符。For example, after semantic analysis of EEG signals, a special semantic operator can be defined as the terminal symbol according to the relationship between semantic blocks.

定义11.特征语法分析树:由于前述定义1和定义2的映射逻辑,语法分析树的叶子节点既可以是A的元素,也可以是AM的元素。设有特征文法G=(Vn,Vt,P,S),对于特征文法G的任一句型均能找到一棵相应的特征语法树,它支持以下4个前提:Definition 11. Feature grammar parse tree: Due to the mapping logic of Definition 1 and Definition 2 above, the leaf nodes of the grammar parse tree can be either elements of A or elements of AM . Suppose there is a feature grammar G = ( Vn , Vt , P, S). For any sentence type of the feature grammar G, a corresponding feature grammar tree can be found, which supports the following four premises:

特征语法树上的每个node都有一个标识,它是V集合中的一个特征符号。Each node in the feature syntax tree has an identifier, which is a feature symbol in the set V.

特征语法树的根node是特征文法的起始符号S。The root node of the feature syntax tree is the start symbol S of the feature grammar.

特征如果语法树上的一个node至少有一个直接后继node,那么该node必为一非终结特征符号。If a node in the syntax tree has at least one direct successor node, then the node must be a non-terminal feature symbol.

若特征语法树上的一个node N有若干个直接后继node,则按自左向右/自右向左的顺序进行排序。If a node N on the feature syntax tree has several direct successor nodes, they are sorted from left to right/right to left.

举例:依据对脑电信号分析的结果,将关键二进制特征与自定义的符号映射之后,生成如下特征文法:For example, based on the results of EEG signal analysis, the key binary features are mapped to custom symbols to generate the following feature grammar:

G[S]:G[S]:

S→dDSS→dDS

D→SDD→SD

S→dS→d

D→efdD→efd

根据已有的特征语法树,有两种推导方法:自上而下和自下而上;可以任选一种方法执行推导。比如,采用自上而下对特征语法树建立推导时,构造出的推导就是最右特征推导,参见图5所示,利用文法G[S]的产生式推导得出特征句型。According to the existing feature syntax tree, there are two derivation methods: top-down and bottom-up; you can choose any one of these methods to perform derivation. For example, when the feature syntax tree is derivation is established from top to bottom, the derivation constructed is the rightmost feature derivation, as shown in Figure 5, the feature sentence pattern is derived using the production rules of the grammar G[S].

求一个特征句型的短语的方法是运用语法树,由文法的开始符号开始,通过产生式来构造与该句型相对应的特征语法树。如图5中,d,efd,defd,ddefdd都是句型ddefdd的短语。只要赋予这些符号特定数学意义,就能够使得二进制特征之间形成特定的语义,比如图6。参见图6,可以推导出来,特征句型的短语为: (A×S)、A×S、k。图5和图6中的短语、非叶子节点,可以用来作为大幅度降维后的关键特征信息,而且在脑电事件中都对应某种事件或情况发生的概率,因此这些表达形式可以在决策树理论中找到对应的概念和形式,从而将看似无规律的脑电信号与文法语义和决策树关联了起来。The method to find a phrase of a characteristic sentence pattern is to use a syntax tree, starting from the start symbol of the grammar, and construct a characteristic syntax tree corresponding to the sentence pattern through production rules. As shown in Figure 5, d, efd, defd, and ddefdd are all phrases of the sentence pattern ddefdd. As long as these symbols are given specific mathematical meanings, specific semantics can be formed between binary features, such as Figure 6. Referring to Figure 6, it can be deduced that the characteristic sentence pattern The phrase is: (A×S), A×S, k. The phrases and non-leaf nodes in Figures 5 and 6 can be used as key feature information after a large-scale dimensionality reduction, and they all correspond to the probability of a certain event or situation in the EEG event. Therefore, these expressions can find corresponding concepts and forms in the decision tree theory, thus associating the seemingly irregular EEG signals with grammar semantics and decision trees.

总结:Summarize:

推导就好比卷积操作,各个子树相当于各个卷积函数,句柄相当于一个有序的子图。由于语法推导树,要求短语是非常容易的。这就好比根据已知图像,求出P300这个短语。而文法中的各个“→”表达式相当于多种可能的波形,不同被试的波形,我们要做的就是找出各种可能性,就相当于神经网络训练出的各种结果。Derivation is like a convolution operation, each subtree is equivalent to each convolution function, and the handle is equivalent to an ordered subgraph. Due to the grammar derivation tree, it is very easy to require a phrase. This is like finding the phrase P300 based on a known image. Each "→" expression in the grammar is equivalent to multiple possible waveforms, waveforms of different subjects. What we need to do is to find various possibilities, which are equivalent to various results of neural network training.

定义12.如果特征文法中的规则表达式均拥有下面的表达方式:Definition 12. If the regular expressions in the feature grammar have the following expressions:

设Cn为非终结特征符号,且所属集合为U∈Vn,u∈V*,即该特征文法里的规则的左边必为一Cn符号,规则表达式的右边u是V上的特征符号序列表达式。Let Cn be a non-terminal feature symbol, and its set is U∈V n , u∈V * , that is, the left side of the rule in the feature grammar must be a Cn symbol, and the right side u of the rule expression is a feature symbol sequence expression on V.

文法可以描述在表达式中出现了变量名,并描述可能出现的位置。对应于LSTM的记忆输出。The grammar can describe the variable names that appear in the expression and the possible positions. It corresponds to the memory output of LSTM.

LSTM的语义解释Semantic interpretation of LSTM

不同于树形LSTM,本研究以传统LSTM模型为研究对象,树形LSTM由于其自带决策树特征,在语义模型本质上和传统模型没有不同。LSTM模型在处理序列化的信息时,不仅可以对未来时间步的信息进行预测,同样可以对一个完整的时间序列进行分类识别,这两种方式对输入信息的识别的颗粒度不同,以及输出形式不同之外,内部逻辑流程相同。Different from the tree-shaped LSTM, this study uses the traditional LSTM model as the research object. The tree-shaped LSTM has its own decision tree features, and its semantic model is essentially the same as the traditional model. When processing serialized information, the LSTM model can not only predict the information of future time steps, but also classify and identify a complete time series. These two methods have different granularity of input information recognition and different output forms, but the internal logical process is the same.

设xi∈RH×W×D表示输入LSTM网络的矩阵数据,H代表矩阵的行数,W代表矩阵的列数,D代表矩阵的深度。对应于脑电数据集,H是采集的每一通道的数据长度,W是采集数据的通道数,D是实验次数。x(d)∈RH×W表示第d个输入给LSTM的cell的特征数据。当新的输入张量出现在x(d)上H×W个不同的候选位置时,我们设计H×W个正模板Let x iRH×W×D represent the matrix data of the input LSTM network, H represents the number of rows, W represents the number of columns, and D represents the depth of the matrix. Corresponding to the EEG dataset, H is the data length of each channel collected, W is the number of channels for collecting data, and D is the number of experiments. x (d)RH×W represents the feature data of the dth input cell to the LSTM. When the new input tensor appears at H×W different candidate positions on x (d) , we design H×W positive templates.

Τ+={T1,1,T1,2......,TH,W}为正采样表示理想的激励特征;同理,负模板Τ-={T1,1,T1,2......,TH,W}也用于描述负采样时的特征。神经网络的评估函数表示为Τ + = {T 1,1 ,T 1,2 ......, TH,W } represents the ideal excitation feature for positive sampling; similarly, the negative template Τ - = {T 1,1 ,T 1,2 ......,TH ,W } is also used to describe the features of negative sampling. The evaluation function of the neural network is expressed as

f(x)c=p(y=c|x)(4.6)f(x) c = p(y = c|x)(4.6)

其中c表示标签类别。为了使f(x)c尽量大,就要让它的相反数尽量最小,并避免多分类情况下接近0,这就是loss函数Where c represents the label category. In order to make f(x) c as large as possible, its inverse should be minimized and avoid being close to 0 in the case of multi-classification. This is the loss function

loss(f(x),y)=-Σc1(y=c)logf(x)c=-logf(x)y=-lnf(x)y (4.7)loss(f(x),y)=-Σ c 1 (y=c) logf(x) c =-logf(x) y =-lnf(x) y (4.7)

该表达式本质是表示负的互信息,因此可以表示为:This expression essentially represents negative mutual information, so it can be expressed as:

其中X表示训练集上的所有的输入张量,先验概率是常量1,p(x(d)|T)是后验概率。该loss值确保第d个输入给cell的特征数据表征某一个输入张量被按LSTM的方式处理,不会出现梯度弥散。因我们的LSTM模型的最终输出是用softmax方程计算结果属于某一类的概率(设总共k个类别),因此Where X represents all input tensors in the training set, the prior probability is a constant 1, and p(x (d) |T) is the posterior probability. This loss value ensures that the feature data of the dth input to the cell represents a certain input tensor and is processed in the LSTM way without gradient diffusion. Because the final output of our LSTM model is the probability of the result belonging to a certain category (assuming a total of k categories) calculated by the softmax equation,

此时,LSTM模型的预测如表述,全连接层为第i个输入的决策模式可以大致描述如下:At this point, the prediction of the LSTM model is as described, and the decision mode of the fully connected layer as the i-th input can be roughly described as follows:

其中表示把模型中的各个门处理映射为卷积处理,其中的可以通过梯度反向传播进行计算。LSTM模型内含的全连接层与CNN中的全连接层本质相同,所以也可以构建一个决策树,提取编码在最后的全连接层中的决策模式,构建如下决策树,从顶部节点到到终端节点:每个决策树的节点c表示一个共同的决策模式,相邻的节点合并,延伸出子节点,这些节点的值来自权重值,图10的矩阵就是LSTM网络的权重值矩阵。in Indicates that each gate processing in the model is mapped to convolution processing, where It can be calculated by gradient back propagation. The fully connected layer contained in the LSTM model is essentially the same as the fully connected layer in CNN, so a decision tree can also be constructed to extract the decision mode encoded in the last fully connected layer and construct the following decision tree, from the top node to the terminal node: Each node c of the decision tree represents a common decision mode, and adjacent nodes are merged to extend sub-nodes. The values of these nodes come from the weight values. The matrix in Figure 10 is the weight value matrix of the LSTM network.

在图10中,节点值的来源是神经网络的权重矩阵(例如图中右下角的两个矩阵),我们构建了一个初始树T1,其中顶部节点把所有梯度的正样本g作为终端节点,于是形成了图10中最左边的树T1,随后将每两个节点如c和c’合并并得到一个新的节点p,c和c’成为p的子节点。重复上述操作进行合并,最后把初始树修改为最后的决策树,如图10所示。In Figure 10, the source of the node value is the weight matrix of the neural network (such as the two matrices in the lower right corner of the figure). We construct an initial tree T1, in which the top node takes all the positive samples g of the gradient as the terminal node, thus forming the leftmost tree T1 in Figure 10. Then, every two nodes such as c and c' are merged to obtain a new node p, and c and c' become the child nodes of p. Repeat the above operation to merge, and finally modify the initial tree to the final decision tree, as shown in Figure 10.

但不同于CNN的地方是,LSTM的w涉及多处(图中w标注处),如图11所示。However, unlike CNN, the w of LSTM involves multiple places (marked with w in the figure), as shown in Figure 11.

接下来,只要将w和文法中的短语进行映射,就能实现从决策树到语法树的转换。由于深度学习模型通过调整参数能够产生不同的训练结果,因此也就可以产生不同的语法树,通过产生的多棵语法树就能逐步完善针对某个脑电数据的文法G[s]。至此,完成对采样脑电信号的语义分析和解释(概要流程如图12所示)。Next, as long as w is mapped to the phrases in the grammar, the conversion from decision tree to grammar tree can be achieved. Since the deep learning model can produce different training results by adjusting parameters, different grammar trees can be generated. Through the generated multiple grammar trees, the grammar G[s] for a certain EEG data can be gradually improved. At this point, the semantic analysis and interpretation of the sampled EEG signal is completed (the outline process is shown in Figure 12).

具体实施例1Specific embodiment 1

对脑电信号数据译码Decoding EEG data

所依据的脑电数据集是国际BCI竞赛的数据集。The EEG dataset used is the dataset of the International BCI Competition.

总体流程图3显示,首先要进行降噪处理,输入通过电极采集的原始脑电信息矩阵,在进行对不定长的脑电数据生成输出标签的预处理之后,译码模块需要把经过预处理的脑电数据处理后,得到离散的“标签序列”矩阵,每个标签有对应的语义块,如式4.2所示。因此,输入的特征数据和输出标签之间对齐关系不确定,允许模型在输入序列的任何点进行标签预测。机器学习模块在处理数据时,会把局部特征进行特征提取、降维和滤波。The overall flow chart 3 shows that the first step is to perform noise reduction processing, input the original EEG information matrix collected by the electrodes, and after preprocessing the EEG data of indefinite length to generate output labels, the decoding module needs to process the preprocessed EEG data to obtain a discrete "label sequence" matrix, where each label has a corresponding semantic block, as shown in Formula 4.2. Therefore, the alignment relationship between the input feature data and the output label is uncertain, allowing the model to predict labels at any point in the input sequence. When processing data, the machine learning module will extract, reduce dimensionality, and filter local features.

图8中母矩阵的每一个矩阵元素,就是把局部特征进行特征提取、降维和滤波后的结果(而每一个矩阵元素中的元素,就是4.2.2小节中定义1.特征矩阵字母表所定义的M2的元素,这些元素被定义2所述的fM->a函数处理之后,就生成了定义2所述的语义字母表A中的元素,这些A中的元素就是本小节第一段第一句话所述的“标签序列”,之所以叫序列,是因为这些元素由字符或字符串组成,这些有意义的字符或字符串就是语义单位,能支持文法、语法和语义范式处理),可以是各种不同的方法,如小波,平均值,最大值,功率谱,滤波器。比如,对于基于CNN网络模型,该矩阵就相当于图7中的α2,矩阵元素就相当于α2中的灰色小块;这些局部特征的初始划分大小和起始位置,既可以是无特殊意义的初始值,可以也可以根据已知的BCI范式进行自定义。如果初始域随机,可以根据最终的语义分析结果和评估效果情况进行反馈更新。至此,对于基于CNN网络的译码模型,就得到了标签序列,以及标签序列所对应的特征数据,以及这两者之间的关系。Each matrix element of the mother matrix in Figure 8 is the result of feature extraction, dimensionality reduction and filtering of local features (and the elements in each matrix element are the elements of M2 defined in Definition 1. Feature matrix alphabet in Section 4.2.2. After these elements are processed by the fM->a function described in Definition 2, the elements in the semantic alphabet A described in Definition 2 are generated. These elements in A are the "label sequence" described in the first sentence of the first paragraph of this section. The reason why it is called a sequence is that these elements are composed of characters or strings. These meaningful characters or strings are semantic units that can support grammar, syntax and semantic paradigm processing). It can be a variety of different methods, such as wavelet, average, maximum, power spectrum, filter. For example, for the CNN network model, the matrix is equivalent to α2 in Figure 7, and the matrix elements are equivalent to the gray blocks in α2; the initial partition size and starting position of these local features can be either initial values without special meaning or customized according to the known BCI paradigm. If the initial domain is random, it can be updated based on the final semantic analysis results and evaluation results. At this point, for the decoding model based on the CNN network, the label sequence, the feature data corresponding to the label sequence, and the relationship between the two are obtained.

如果是基于LSTM网络模型,该矩阵就对应于标签序列矩阵(若用小波变换方法生成给CNN网络的输入数据,矩阵元素对应于使用小波分解之后的系数)。标签序列是一个稀疏矩阵,假设我们生成的batch_size是32的样本,每个样本是长度为1~30的数字串,则生成一个(32,30)的矩阵,矩阵有数字的是非零元素,无数字的是零元素,而且该标签是不定长的,非零元素分布随机,标签在存储数字串的同时还要存储位置信息。至此,对于基于LSTM网络的译码模型,就得到了标签序列,以及标签序列所对应的特征数据,这两者之间的关系,以及序列长度数据。序列长度是一维数据,其表现形式可以是:[max_time_step,…,max_time_step],其长度为batch_size,具体的值为max_time_step。If it is based on the LSTM network model, the matrix corresponds to the label sequence matrix (if the wavelet transform method is used to generate the input data for the CNN network, the matrix elements correspond to the coefficients after wavelet decomposition). The label sequence is a sparse matrix. Assuming that the batch_size we generate is 32 samples, each sample is a digital string of length 1 to 30, then a (32, 30) matrix is generated. The matrix has non-zero elements with numbers and zero elements without numbers. Moreover, the label is of indefinite length, and the non-zero elements are randomly distributed. The label stores the position information while storing the digital string. So far, for the decoding model based on the LSTM network, the label sequence, the feature data corresponding to the label sequence, the relationship between the two, and the sequence length data are obtained. The sequence length is one-dimensional data, which can be expressed as: [max_time_step,…,max_time_step], its length is batch_size, and the specific value is max_time_step.

此外,上述母矩阵中每一个矩阵元素值的来源,除了可以直接来自把对局部特征进行特征提取、降维和滤波后的特征信息之外,该母矩阵中的元素信息,还可以来自权重矩阵。相比而言,后者比原始特征对输出的影响更大。对于决策树中同一节点下的图像集合而言,权重值具有通用属性,它解释了一组图像而不是单个图像,因此使用权重值形成特征母矩阵中的值更具有优势;此外,来自原始特征的信息与特定图像相关,不能用作一组相似图像的通用属性。In addition, the source of each matrix element value in the above mother matrix can come directly from the feature information after feature extraction, dimensionality reduction and filtering of local features, and the element information in the mother matrix can also come from the weight matrix. In comparison, the latter has a greater impact on the output than the original feature. For the image set under the same node in the decision tree, the weight value has a universal property, which explains a group of images rather than a single image, so it is more advantageous to use the weight value to form the value in the feature mother matrix; in addition, the information from the original feature is related to a specific image and cannot be used as a universal attribute for a group of similar images.

图8显示了脑电电位序列形成矩阵字母表和标签化的过程。图中的a代表原始脑电信号波形图,b代表输入样本矩阵映射成一张图(部分),c代表式(4.2)母矩阵,该母矩阵既可以直接来自降维后的特征信息,也可以来自权重矩阵,d代表4.2.5节阐述的特征矩阵。其中的子图c显示了每一个矩阵元素由4个成员组成,涉及c1和c2两个通道,所涉及的通道数量可以根据需要进行设计,比如,也可以只由一个通道的部分序列组成,也可以由两个甚至多个通道的部分序列组成。Figure 8 shows the process of forming a matrix alphabet and labeling the EEG potential sequence. In the figure, a represents the original EEG signal waveform, b represents the input sample matrix mapped into a picture (part), c represents the mother matrix of formula (4.2), which can be directly derived from the feature information after dimensionality reduction or from the weight matrix, and d represents the feature matrix described in Section 4.2.5. Sub-figure c shows that each matrix element consists of 4 members, involving two channels c1 and c2. The number of channels involved can be designed as needed. For example, it can be composed of only a partial sequence of one channel, or it can be composed of partial sequences of two or even more channels.

图8中显示将离散标签存储于一个矩阵,矩阵的行数与数据流的多少有关。该矩阵的元素以字符或字符串表示,所有的元素都属于同一个字母表,字母表由不重复的元素构成。字母表的元素由字符集定义该元素中的字符,所述字符集可以是ASCII字符集,或者双字节字符集DBCS,或者Unicode字符集。FIG8 shows that discrete labels are stored in a matrix, and the number of rows of the matrix is related to the number of data streams. The elements of the matrix are represented by characters or strings, and all elements belong to the same alphabet, which is composed of non-repeating elements. The elements of the alphabet are defined by a character set, which can be an ASCII character set, a double-byte character set DBCS, or a Unicode character set.

为了实现评估当前输入的脑电信号序列属于A中相应标签的概率,需要在机器学习模型的softmax输出层增加相应的处理单元,而且该单元必须要能够估算出现未知意义、噪音、分隔区域标签的出现概率,所有的这些信息就给出了所有时间步的所有标签的联合条件概率,然后可以通过对相应的联合概率求和来找到任何一个标签序列的条件概率。In order to evaluate the probability that the current input EEG signal sequence belongs to the corresponding label in A, it is necessary to add a corresponding processing unit to the softmax output layer of the machine learning model, and this unit must be able to estimate the probability of the occurrence of unknown meanings, noise, and separation area labels. All this information gives the joint conditional probability of all labels at all time steps. Then the conditional probability of any label sequence can be found by summing the corresponding joint probabilities.

对于某个脑电信号输入序列,经处理后输出单元n在时间t处的表示经过LSTM模型处理之后,在时间t处观察其属于A中标签k的概率,给定长度L的输入序列x和训练集合Tr,下面的概率评估在π上的分布,π∈AL,其中AL是在字母表A上的长度为L的序列的集合,A=Vn∪Vt For a certain EEG signal input sequence, after processing, the output unit n at time t represents the probability of an observation at time t belonging to label k in A after being processed by the LSTM model. Given an input sequence x of length L and a training set Tr, the following probability evaluation is distributed over π, π∈AL , where AL is the set of sequences of length L over the alphabet A, A= Vn∪Vt

π是随机变量,其值可取A中的任何一个符号。由于脑电信号中存在大量的背景噪音,还有躯体感觉皮层、心理活动等和个人特征强相关的信号,这些信号的特点之一是存在小范围的重复,因此,存在多个特征矩阵到A上标签的映射T:AM→A。由于这些映射路径之间是互斥的,因此可以通过将T上所有映射到其上的路径的概率相加来计算某个标签的条件概率π is a random variable whose value can take any sign in A. Since there is a lot of background noise in EEG signals, as well as signals that are strongly related to personal characteristics such as the somatosensory cortex and psychological activities, one of the characteristics of these signals is the presence of small-scale repetitions. Therefore, there are multiple mappings from feature matrices to labels on A, T: AM →A. Since these mapping paths are mutually exclusive, the conditional probability of a label can be calculated by adding the probabilities of all paths on T that are mapped to it.

至此,我们已经定义了输入的脑电数据到A的可能的标签序列的条件概率,接下来的处理,和现有的RNN/LSTM算法处理一致。So far, we have defined the conditional probability of the possible label sequence of the input EEG data to A. The subsequent processing is consistent with the existing RNN/LSTM algorithm processing.

对于A(或者词汇表V)中的标签化的label(label是标签序列),假设前向变量δt(s)是所有长度为t的路径前缀的概率之和,该前缀由T映射到s/2长度前缀;反向变量θt(s)是所有从t开始的路径后缀的概率之和,该后缀由T映射到标签label的s/2长度后缀。于是,依据相似推导过程对标签化了的脑电信息进行一系列推导,最终结果为:For the labeled label (label is a label sequence) in A (or vocabulary V), assume that the forward variable δ t (s) is the sum of the probabilities of all path prefixes of length t, which are mapped from T to s/2 length prefixes; the reverse variable θ t (s) is the sum of the probabilities of all path suffixes starting from t, which are mapped from T to s/2 length suffixes of label label. Therefore, a series of derivations are performed on the labeled EEG information based on a similar derivation process, and the final result is:

其中,v是标签变量,x是给定长度的输入的脑电信号的时序序列,指的是标签k在v’中的位置,而v’指的是,对于给定的s和t条件时的v值,向前和向后变量的乘积是所有穿过v′s的路径的总概率。Among them, v is the label variable, x is the time series of the input EEG signal of a given length, refers to the position of label k in v', while v' refers to the value of v for given conditions s and t. The product of the forward and backward variables is the total probability of all paths passing through v 's .

因此,对于LSTM网络的输出a1,a2,…ak…,我们可以对应得到p(C1|x),p(C2|x),p(Ck|x),即给定输入,输出类别为C1,C2…Ck的概率。Therefore, for the outputs a 1 , a 2 , … ak … of the LSTM network, we can correspondingly obtain p(C 1 |x), p(C 2 |x), p(C k |x), that is, the probability that the output category is C 1 , C 2 …C k given the input.

网络被训练好了之后,将通过选择最可能的标签l来标记某个未知的输入序列x,搜索到这个标签就是解码的任务。解码算法可以利用前述的前缀信息来解码,依赖于通过修改的前向变量,可以高效地计算前缀标签连续扩展的概率。该搜索解码是一种best-first搜索,形成图9所示的标签树,其中给定前缀(标签)节点的子节点之间共享该前缀。图b中当标签pr比任何前缀都更可能时(结束符e的概率最大)搜索结束。After the network is trained, it will label an unknown input sequence x by selecting the most likely label l. Searching for this label is the decoding task. The decoding algorithm can use the aforementioned prefix information to decode, relying on the modified forward variables to efficiently calculate the probability of continuous expansion of the prefix label. The search decoding is a best-first search, forming a label tree as shown in Figure 9, where the child nodes of a given prefix (label) node share the prefix. In Figure b, the search ends when the label pr is more likely than any prefix (the probability of the end symbol e is the largest).

该搜索树、标签树同时也就是决策树,同时可以通过定义文法G形成语法树。对于同一批数据,只要进一步采用Bagging策略和随机策略就可以形成随机森林。The search tree and label tree are also decision trees, and a syntax tree can be formed by defining a grammar G. For the same batch of data, a random forest can be formed by further adopting the bagging strategy and the random strategy.

至此,就完成了译码处理,完成了从原始输入脑电数据和各个语义逻辑实体(AM|A|语义流标签|决策树|文法语法树)之间的关联,实现了脑电数据使用各种机器学习算法,尤其是神经网络算法的语义范式解释。At this point, the decoding process is completed, and the association between the original input EEG data and various semantic logical entities ( AM |A|semantic flow label|decision tree|grammar syntax tree) is completed, realizing the semantic paradigm interpretation of EEG data using various machine learning algorithms, especially neural network algorithms.

具体实施例2Specific embodiment 2

语义范式过程实例Semantic Paradigm Process Example

给定一个被试的左右手运动的脑电信号的时间序列信息,根据前述构造的文法G,s能否从G推导出结果?Given the time series information of the EEG signals of a subject's left and right hand movements, according to the grammar G constructed above, can s deduce the results from G?

算法思想:从脑电信号的时间序列信息经过神经网络处理后形成降维后的矩阵,将该矩阵信息分解成两部分,一部分对应身份和动作信息,另一部分对应左手(或右手);于是形成了一个待分析的句子s,从G的开始符号出发,随意推导出某个句子t,比较t和s。Algorithm idea: The time series information of the EEG signal is processed by the neural network to form a reduced-dimensional matrix, and the matrix information is decomposed into two parts, one part corresponds to the identity and action information, and the other part corresponds to the left hand (or right hand); thus, a sentence s to be analyzed is formed, and starting from the starting symbol of G, a sentence t is randomly derived, and t and s are compared.

这是从开始符号出发推导出的句子,自顶向下分析,对应着分析树自顶向下的构造顺序。This is the sentence derived from the start symbol, parsed from top to bottom, corresponding to the top-down construction order of the parse tree.

用S来代表一个句子,→代表推出,文法G如下:Let S represent a sentence, → represent deduction, and the grammar G is as follows:

非终结符:{N,V}Non-terminal symbol: {N, V}

终结符:{s,r,l,p}Terminal symbol: {s, r, l, p}

开始符号:SStart symbol: S

上面文法中|代表“或”。In the above grammar, | represents "or".

数据处理和语义句子推导过程:Data processing and semantic sentence derivation process:

假设[原始信息矩阵]是从被试采集的3个通道的脑电信息,该信息被机器学习模型网络处理,信息被滤波、降维后的结果用下面的式(4.12)的矩阵来表达:Assume that [the original information matrix] is the EEG information of three channels collected from the subjects. This information is processed by the machine learning model network. The information is filtered and reduced in dimension, and the result is expressed by the matrix of the following formula (4.12):

在进行文法分析时可以写成一行:1a2b345678902a2b345678913a2b34567892When performing grammar analysis, it can be written in one line: 1a2b345678902a2b345678913a2b34567892

其中:in:

1a2b34567890对应通道1的信息(C1T)1a2b34567890 corresponds to the information of channel 1 (C1T)

2a2b34567891对应通道2的信息(C2T)2a2b34567891 corresponds to the information of channel 2 (C2T)

3a2b34567892对应通道3的信息(C3T)3a2b34567892 corresponds to the information of channel 3 (C3T)

由于所有的信号最终可以对应成两个要素:“身份信息+按”,“右(左)”,因此在形式上,可以将矩阵中的信息分为两部分,一部分代表“身份信息+按”,另外一部分代表“右(左)”,身份信息包括关联性负变(contingentnegativevariation,CNV)和事件相关去同步(ERD)等和个人相关的背景特征信号。于是矩阵中的信息分为两部分,如图13所示:Since all signals can be ultimately mapped to two elements: "identity information + press", "right (left)", the information in the matrix can be formally divided into two parts, one representing "identity information + press", and the other representing "right (left)". Identity information includes background feature signals related to individuals such as contingent negative variation (CNV) and event-related desynchronization (ERD). Therefore, the information in the matrix is divided into two parts, as shown in Figure 13:

被红色框住的部分代表“身份信息+按”,剩下的最后一列代表“右”,于是红色框住的部分就对应推导中终结符s和p,对应终结符r。The part framed in red represents "identity information + press", and the last column represents "right", so the part framed in red corresponds to the terminal symbols s and p in the derivation. Corresponding to the terminal symbol r.

基于脑电信号包含一定程度的随机性,这两部分信息位置也有一定程度的随机性,值的大小也有波动,当举例来说,解析得到的特征矩阵(图14)中,对应终结符r的信息位置可以夹在其他信号中,如图15所示的红色部分。Since EEG signals contain a certain degree of randomness, the positions of these two parts of information also have a certain degree of randomness, and the values also fluctuate. For example, in the feature matrix obtained by analysis (Figure 14), the information position corresponding to the terminal symbol r can be sandwiched between other signals, as shown in the red part of Figure 15.

在将脑电信息转换成有明确语义的信息后,文法中就新增了明确的语义,如下:After converting EEG information into information with clear semantics, clear semantics are added to the grammar, as follows:

S→N V NS→N V N

→s p r(4.13)→s p r (4.13)

式(4.13)就是新形成的文法,有了文法就能够构造语法树,然后对新采集的脑电信号形成句型、判定其短语,理解其中的语义。Formula (4.13) is the newly formed grammar. With the grammar, we can construct a syntax tree, and then form sentence patterns for the newly collected EEG signals, determine their phrases, and understand their semantics.

对应式(4.13)的动作,其实就是按下按键的动作,该动作可以通过EEGLAB工具捕捉到。该事件由按压按键动作引发,相比于之前和之后的波形明显不同,具有明显的语义特征。The action corresponding to equation (4.13) is actually the action of pressing a key, which can be captured by the EEGLAB tool. This event is triggered by the key pressing action, which is significantly different from the waveforms before and after, and has obvious semantic features.

由于脑电信号可能带有噪声,因此新的待测试信号经过降维得到的信号可能不能被精确匹配,如同图15和式(4.12)之间的差异,所以需要有一种手段能够将信息和已经训练的信息进行匹配,并且能够在确实无法完全匹配的情况下能够以概率的方法识别出来,而且这些处理原本就是神经网络已有的技术,因此神经网络的预测功能就恰好对应着语义分析中的语法分析树的相关处理流程。Since EEG signals may contain noise, the new test signal obtained after dimensionality reduction may not be accurately matched, just like the difference between Figure 15 and formula (4.12). Therefore, there needs to be a way to match the information with the trained information, and to be able to identify it probabilistically when it is indeed impossible to match it completely. Moreover, these processes are already existing technologies of neural networks, so the prediction function of neural networks corresponds exactly to the relevant processing flow of the grammatical analysis tree in semantic analysis.

由于CNN可以自动学习出对特定问题最有效的滤波器,利用CNN模型得出结果,如图16所示,然后根据可视化的图和数据反查有效的原始数据,找到特征数据。Since CNN can automatically learn the most effective filter for a specific problem, the CNN model is used to obtain the result, as shown in Figure 16. Then, the effective original data is checked based on the visualized graph and data to find the feature data.

图16中的点代表降维后的中间特征数据,这些数据的组合能够有效地进行分类,我们可以依据映射表发查原始特征数据。通过这种方式,可以方便地找到某个被试的特征数据模板。The points in Figure 16 represent the intermediate feature data after dimensionality reduction. The combination of these data can be effectively classified. We can check the original feature data based on the mapping table. In this way, it is easy to find the feature data template of a subject.

具体实施例3Specific embodiment 3

实现构造语法树的方法,所述构造语法树的处理,就是将输入的信息按照当前实验分析的语义关系结果,将降维后的关键特征信息块之间的关系,形成特征字母表,再通过特征字母表形成语义字母表,形成特征文法,形成特征生成式集(是一个由生成式或规则组成的非空有限集合),并进而形成特征语法树。根据特征语法树构造句型,以便算法在后续处理中使用。有了所述新形成的文法和特征语法树,对新采集的脑电信号形成句型、判定其短语、句柄和语法,理解其中的语义。A method for constructing a syntax tree is implemented, wherein the process of constructing a syntax tree is to form a feature alphabet according to the semantic relationship results of the current experimental analysis of the input information, and the relationship between the key feature information blocks after dimensionality reduction, and then form a semantic alphabet through the feature alphabet to form a feature grammar, form a feature generation formula set (a non-empty finite set composed of generation formulas or rules), and then form a feature syntax tree. Sentence patterns are constructed according to the feature syntax tree so that the algorithm can use it in subsequent processing. With the newly formed grammar and feature syntax tree, sentence patterns are formed for the newly collected EEG signals, their phrases, handles and grammar are determined, and the semantics therein is understood.

根据所述的方法,其特征还包括:According to the method, the characteristics also include:

特征信息形成特征字母表:所有的输入信息被函数处理成特征矩阵M1,M1的每一个元素同时也是一个矩阵M2,M2的元素是由二进制或十六进制形式的数值表示,所有M1的元素都属于同一个特征矩阵字母表,该字母表由不重复的元素构成。Feature information forms a feature alphabet: All input information is processed by the function into a feature matrix M 1 . Each element of M 1 is also a matrix M 2 . The elements of M 2 are represented by binary or hexadecimal values. All elements of M 1 belong to the same feature matrix alphabet, which consists of non-repeating elements.

根据所述的方法,其特征还包括:According to the method, the characteristics also include:

所述将特征信息继续分成有不同语义的组,每一组转换成特征字符串,指的是,将特征字符串构成语义字母表,该语义字母表指的是:所有的输入信息M2的元素在语义上可以同时被函数fM->a转换映射成特定的符号,这些符号可以是被计算机系统识别的字符集中的元素,这些符号和运算符、分隔符构成一个有限的非空集合A,其中每一个元素是可以被识别(如计算机系统)的字符,也可以是多个字符组成的字符串,因此在本文讨论范围内,也可以用V表示,也就是说语义字母表和语义词汇表被视作相同,简称为字母表。The further division of feature information into groups with different semantics and conversion of each group into a feature string refers to the formation of a semantic alphabet from the feature strings. The semantic alphabet means that all elements of the input information M2 can be semantically mapped into specific symbols by the function fM->a at the same time. These symbols can be elements in a character set recognized by a computer system. These symbols, operators and separators form a finite non-empty set A, in which each element is a character that can be recognized (such as a computer system) or a string of multiple characters. Therefore, within the scope of discussion in this article, it can also be represented by V. That is to say, the semantic alphabet and the semantic vocabulary are regarded as the same and are referred to as the alphabet for short.

后续特征可以参照前述其他10个定义。Subsequent features can refer to the other 10 definitions mentioned above.

具体实施例4Specific embodiment 4

通过本方法可以做出对应的处理装置,该装置包括:A corresponding processing device can be made by this method, and the device includes:

第一处理模块用于对脑电信号数据进行处理,得到特征信息;The first processing module is used to process the EEG signal data to obtain feature information;

第二处理模块用于将特征信息继续分成有不同语义的组,每一组转换成特征字符串;The second processing module is used to further divide the feature information into groups with different semantics, and convert each group into a feature string;

第三处理模块用于将脑电信号信息通过机器学习模型转化成相应的决策树和语法树;The third processing module is used to convert the EEG signal information into corresponding decision trees and syntax trees through a machine learning model;

第四处理模块用于由生成的多棵语法树形成文法G[s]。The fourth processing module is used to form a grammar G[s] from the generated multiple syntax trees.

第五处理模块用于将得到的各特征字符串按文法的形式写成产生式,构成文法G[s];也就是说:通过实验构造、修改语法树,该处理就是:将输入的信息按照当前实验分析的语义关系结果,将降维后的关键特征信息块之间的关系,形成文法的产生式,形成特征信号文法,并进而形成特征语法树。根据特征语法树构造句型,以便算法在后续处理中使用。有了所述新形成的文法和特征语法树,对新采集的脑电信号形成句型、判定其短语、句柄和语法,理解其中的语义。The fifth processing module is used to write each obtained characteristic character string into a production formula in the form of grammar to form a grammar G[s]; that is to say: construct and modify the grammar tree through experiments. The processing is to form the production formula of the grammar, the characteristic signal grammar, and then the characteristic grammar tree according to the semantic relationship results of the current experimental analysis of the input information and the relationship between the key characteristic information blocks after dimensionality reduction. Construct a sentence pattern according to the characteristic grammar tree so that the algorithm can be used in subsequent processing. With the newly formed grammar and characteristic grammar tree, the sentence pattern of the newly collected EEG signal is formed, its phrases, handles and grammar are determined, and the semantics therein is understood.

显然,本领域的技术人员应该明白,上述的本发明的各模块或各步骤可以用通用的计算装置来实现,它们可以集中在单个的计算装置上,或者分布在多个计算装置所组成的网络上,可选地,它们可以用计算装置可执行的程序代码来实现,从而,可以将它们存储在存储装置中由计算装置来执行,并且在某些情况下,可以以不同于此处的顺序执行所示出或描述的步骤,或者将它们分别制作成各个集成电路模块,或者将它们中的多个模块或步骤制作成单个集成电路模块来实现。这样,本发明不限制于任何特定的硬件和软件结合。Obviously, those skilled in the art should understand that the above modules or steps of the present invention can be implemented by a general computing device, they can be concentrated on a single computing device, or distributed on a network composed of multiple computing devices, and optionally, they can be implemented by a program code executable by a computing device, so that they can be stored in a storage device and executed by the computing device, and in some cases, the steps shown or described can be executed in a different order than here, or they can be made into individual integrated circuit modules, or multiple modules or steps therein can be made into a single integrated circuit module for implementation. Thus, the present invention is not limited to any specific combination of hardware and software.

以上所述仅为本发明的优选实施例而已,并不用于限制本发明,对于本领域的技术人员来说,本发明可以有各种更改和变化。凡在本发明的精神和原则之内,所作的任何修改、等同替换、改进等,均应包含在本发明的保护范围之内。The above description is only a preferred embodiment of the present invention and is not intended to limit the present invention. For those skilled in the art, the present invention may have various modifications and variations. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention shall be included in the protection scope of the present invention.

Claims (16)

1.一种对脑电信号进行文法、语法和语义范式处理的方法,其特征在于,包括:1. A method for processing grammar, syntax and semantic paradigms of EEG signals, comprising: 对脑电信号数据进行处理,得到特征信息,所述脑电信号包括头皮脑电图信号EEG,脑皮层脑电图信号ECoG,脑磁图信号MEG;Processing the EEG signal data to obtain feature information, wherein the EEG signal includes a scalp EEG signal, a cortical ECoG signal, and a magnetoencephalogram signal; 将特征信息分成有不同语义的组,每一组转换成特征字符串;Divide the feature information into groups with different semantics, and convert each group into a feature string; 根据所述特征信息形成特征字母表;具体地包括:将输入的信息按照当前实验分析的语义关系结果,将降维后的关键特征信息块之间的关系,形成特征字母表;Forming a feature alphabet according to the feature information; specifically comprising: forming a feature alphabet according to the semantic relationship result of the current experimental analysis of the input information and the relationship between the key feature information blocks after dimensionality reduction; 对脑电信号信息进行处理,得到语法树;具体地包括:通过特征字母表形成语义字母表,形成特征文法,形成特征生成式集,所述特征生成式集是一个由生成式或规则组成的非空有限集合,并进而形成语法树;Processing the EEG signal information to obtain a syntax tree; specifically, the process includes: forming a semantic alphabet through a feature alphabet, forming a feature grammar, forming a feature generation set, wherein the feature generation set is a non-empty finite set consisting of generation formulas or rules, and then forming a syntax tree; 由生成的语法树形成文法G[s],其中语法树数量可以为1或大于1的值。The generated syntax trees form a grammar G[s], where the number of syntax trees can be 1 or a value greater than 1. 2.根据权利要求1所述的方法,其特征在于,包括:2. The method according to claim 1, characterized in that it comprises: 将脑电信号信息通过机器学习模型转化成相应的决策树;Convert EEG signal information into corresponding decision trees through machine learning models; 将决策树转化成语法树。Convert the decision tree into a syntax tree. 3.根据权利要求1所述的方法,其特征在于,包括:3. The method according to claim 1, characterized in that it comprises: 所述对脑电信号进行处理,其中包括对数据进行译码,就是,输入采集来的原始脑电信号数据,译码模块把经过预处理的脑电数据处理后得到离散的“标签序列”矩阵,每个标签有对应的语义块,如公式1所示,机器学习模块在处理数据时,会把局部特征进行特征提取、降维和滤波;The processing of EEG signals includes decoding the data, that is, the raw EEG signal data collected is input, and the decoding module processes the pre-processed EEG data to obtain a discrete "label sequence" matrix, and each label has a corresponding semantic block, as shown in Formula 1. When processing the data, the machine learning module extracts, reduces the dimension and filters the local features; 4.根据权利要求3所述的译码方法,其特征在于,包括:4. The decoding method according to claim 3, characterized in that it comprises: 所述译码处理的具体实现,分为基于CNN网络模式、基于CTC+LSTM网络模式、其他RNN模式、小波变换模式,各种机器学习模式;对于基于CNN网络模式,该矩阵如公式1所示;于是得到了标签序列,以及标签序列所对应的特征数据,以及这两者之间的关系,如果是基于CTC+LSTM网络模式,该矩阵就对应于标签序列矩阵;就得到了标签序列,以及标签序列所对应的特征数据,这两者之间的关系,以及序列长度数据;序列长度是一维数据,其表现形式是:[max_time_step,…,max_time_step],其长度为batch_size,具体的值为max_time_step,或者,通过小波变换形成的输出构成特征矩阵。The specific implementation of the decoding process is divided into CNN network mode, CTC+LSTM network mode, other RNN mode, wavelet transform mode, and various machine learning modes; for the CNN network mode, the matrix is shown in Formula 1; thus, a label sequence and feature data corresponding to the label sequence, as well as the relationship between the two are obtained. If it is based on the CTC+LSTM network mode, the matrix corresponds to the label sequence matrix; thus, a label sequence and feature data corresponding to the label sequence, the relationship between the two, and sequence length data are obtained; the sequence length is one-dimensional data, and its expression is: [max_time_step,…,max_time_step], and its length is batch_size, and the specific value is max_time_step, or, the output formed by wavelet transform constitutes a feature matrix. 5.根据权利要求3中所述的方法,其特征在于,包括:5. The method according to claim 3, characterized in that it comprises: 公式1的母矩阵中,其中每一个矩阵元素的来源,除了可以直接来自把对局部特征进行特征提取、降维和滤波后的特征信息之外,该母矩阵中的元素信息,还可以来自机器学习网络或深度学习神经网络模型中的权重矩阵。In the mother matrix of Formula 1, the source of each matrix element can come directly from the feature information after feature extraction, dimensionality reduction and filtering of local features. The element information in the mother matrix can also come from the weight matrix in the machine learning network or deep learning neural network model. 6.根据权利要求2所述的方法,其特征在于,包括:6. The method according to claim 2, characterized in that it comprises: 所述机器学习模型包括卷积神经网络CNN模型,长短期记忆网络LSTM模型,其他RNN模型,小波变换模型,各种机器学习模型。The machine learning models include convolutional neural network (CNN) models, long short-term memory (LSTM) models, other RNN models, wavelet transform models, and various machine learning models. 7.根据权利要求3所述的方法,其特征在于,包括:所述对脑电信号数据译码,就是为后续的语义分析处理做准备,目的是为了把数据按不同层次处理成有语义特征、而且被标签化的有序形式,同时得到对应的决策树和语法树,语义分析处理需要将得到的标签化的特征和决策树及语法树进行映射。7. The method according to claim 3 is characterized in that it includes: the decoding of the EEG signal data is to prepare for the subsequent semantic analysis and processing, the purpose of which is to process the data into an ordered form with semantic features and labels according to different levels, and to obtain corresponding decision trees and syntax trees at the same time. The semantic analysis and processing requires mapping the obtained labeled features to the decision tree and syntax tree. 8.根据权利要求1中所述的方法,其特征在于,包括:8. The method according to claim 1, comprising: 对脑电信号数据进行处理,得到特征信息,根据特征信息形成特征字母表:所有的输入信息被函数处理成特征矩阵M1,M1的每一个元素同时也是一个矩阵M2,M2的元素是由二进制或十六进制形式的数值表示,所有M1的元素都属于同一个特征字母表,该特征字母表由不重复的元素构成。The EEG signal data is processed to obtain feature information, and a feature alphabet is formed based on the feature information: all input information is processed by the function into a feature matrix M 1 , and each element of M 1 is also a matrix M 2 . The elements of M 2 are represented by binary or hexadecimal numerical values. All elements of M 1 belong to the same feature alphabet, which is composed of non-repeating elements. 9.根据权利要求8中所述的方法,其特征在于,包括:9. The method according to claim 8, characterized in that it comprises: 将所述特征信息继续分成有不同语义的组,每一组转换成特征字符串,指的是,将特征字符串构成语义字母表,该语义字母表指的是:所有的输入信息M2的元素在语义上可以同时被函数fM->a转换映射成特定的符号,这些符号可以是被计算机系统识别的字符集中的元素,这些符号和运算符、分隔符构成一个有限的非空集合A,其中每一个元素是被识别的字符,或者是多个字符组成的字符串,用V表示,也就是说语义字母表和语义词汇表被视作相同,简称为字母表。The feature information is further divided into groups with different semantics, and each group is converted into a feature string, which means that the feature strings constitute a semantic alphabet. The semantic alphabet means that all elements of the input information M2 can be semantically converted and mapped into specific symbols by the function fM->a at the same time. These symbols can be elements in the character set recognized by the computer system. These symbols and operators and separators constitute a finite non-empty set A, in which each element is a recognized character, or a string composed of multiple characters, represented by V. That is to say, the semantic alphabet and the semantic vocabulary are regarded as the same, and are referred to as the alphabet for short. 10.根据权利要求1所述的方法,其特征在于,包括:10. The method according to claim 1, characterized in that it comprises: 按语法树生成句型、短语,用于高层次的文法、语法和语义分析处理。Generate sentence patterns and phrases according to the syntax tree for high-level grammar, syntax and semantic analysis. 11.一种构造语法树的方法,其特征在于,所述构造语法树的处理,就是将输入的信息按照当前实验分析的语义关系结果,将降维后的关键特征信息块之间的关系,形成特征字母表,再通过特征字母表形成语义字母表,形成特征文法,形成特征生成式集,所述特征生成式集是一个由生成式或规则组成的非空有限集合,并进而形成语法树;根据语法树构造句型,以便算法在后续处理中使用;有了所述特征文法和所述语法树,对新采集的脑电信号形成句型、判定其短语、句柄和语法,理解其中的语义。11. A method for constructing a syntax tree, characterized in that the process of constructing the syntax tree is to form a feature alphabet according to the semantic relationship results of the current experimental analysis of the input information and the relationship between the key feature information blocks after dimensionality reduction, and then form a semantic alphabet through the feature alphabet to form a feature grammar and a feature generation formula set, wherein the feature generation formula set is a non-empty finite set composed of generation formulas or rules, and then form a syntax tree; construct sentence patterns according to the syntax tree so that the algorithm can be used in subsequent processing; with the feature grammar and the syntax tree, form sentence patterns for the newly collected EEG signals, determine their phrases, handles and grammar, and understand the semantics therein. 12.根据权利要求11所述的方法,其特征在于,包括:12. The method according to claim 11, characterized in that it comprises: 对脑电信号数据进行处理,得到特征信息,根据特征信息形成特征字母表:所有的输入信息被函数处理成特征矩阵M1,M1的每一个元素同时也是一个矩阵M2,M2的元素是由二进制或十六进制形式的数值表示,所有M1的元素都属于同一个特征字母表,该字母表由不重复的元素构成。The EEG signal data is processed to obtain feature information, and a feature alphabet is formed based on the feature information: all input information is processed by the function into a feature matrix M 1 , and each element of M 1 is also a matrix M 2 . The elements of M 2 are represented by values in binary or hexadecimal form. All elements of M 1 belong to the same feature alphabet, which consists of non-repeating elements. 13.根据权利要求12所述的方法,其特征在于,包括:13. The method according to claim 12, characterized in that it comprises: 将所述特征信息继续分成有不同语义的组,每一组转换成特征字符串,指的是,将特征字符串构成语义字母表,该语义字母表指的是:所有的输入信息M2的元素在语义上可以同时被函数fM->a转换映射成特定的符号,这些符号可以是被计算机系统识别的字符集中的元素,这些符号和运算符、分隔符构成一个有限的非空集合A,其中每一个元素是被识别的字符,或者是多个字符组成的字符串,用V表示,也就是说语义字母表和语义词汇表被视作相同,简称为字母表。The feature information is further divided into groups with different semantics, and each group is converted into a feature string, which means that the feature strings constitute a semantic alphabet. The semantic alphabet means that all elements of the input information M2 can be semantically converted and mapped into specific symbols by the function fM->a at the same time. These symbols can be elements in the character set recognized by the computer system. These symbols and operators and separators constitute a finite non-empty set A, in which each element is a recognized character, or a string composed of multiple characters, represented by V. That is to say, the semantic alphabet and the semantic vocabulary are regarded as the same, and are referred to as the alphabet for short. 14.一种脑电信号处理的方法,其特征在于,包括:14. A method for processing electroencephalogram signals, comprising: 对采集来的脑电信号进行预处理,并通过译码模块对所述经过预处理的脑电信号进行译码处理,通过译码模块把经过预处理的脑电数据处理后得到离散的“标签序列”矩阵,通过机器学习模块对局部特征进行特征提取、降维和滤波,通过对所述脑电信号进行处理,获得特征信息,其中所述脑电信号包括头皮脑电图信号EEG,脑皮层脑电图信号ECoG,脑磁图信号MEG;Preprocessing the collected EEG signals, and decoding the preprocessed EEG signals through a decoding module, obtaining a discrete "label sequence" matrix after processing the preprocessed EEG data through the decoding module, extracting, reducing and filtering local features through a machine learning module, and obtaining feature information by processing the EEG signals, wherein the EEG signals include scalp EEG signals, cerebral cortical EEG signals, and magnetoencephalogram signals; 对脑电信号信息进行处理,得到语法树,由所述语法树形成文法G[s],具体地包括:通过特征字母表形成语义字母表,形成特征文法,形成特征生成式集,所述特征生成式集是一个由生成式或规则组成的非空有限集合,并进而形成语法树。The EEG signal information is processed to obtain a syntax tree, and a grammar G[s] is formed from the syntax tree, specifically including: forming a semantic alphabet through a feature alphabet, forming a feature grammar, forming a feature generation formula set, wherein the feature generation formula set is a non-empty finite set composed of generation formulas or rules, and then forming a syntax tree. 15.根据权利要求14所述的方法,其特征在于,15. The method according to claim 14, characterized in that 所述“标签序列”矩阵,每个标签有对应的语义块,如公式1所示;In the “label sequence” matrix, each label has a corresponding semantic block, as shown in Formula 1; 公式1的母矩阵中,每一个矩阵元素的来源,直接来自对局部特征进行特征提取、降维和滤波后的特征信息,或者来自机器学习网络或深度学习神经网络模型中的权重矩阵;In the mother matrix of Formula 1, the source of each matrix element comes directly from the feature information after feature extraction, dimensionality reduction and filtering of local features, or from the weight matrix in the machine learning network or deep learning neural network model; 所述译码处理的具体实现分为基于CNN网络模式、基于CTC+LSTM网络模式、其他RNN模式、小波变换模式这些各种机器学习模式;The specific implementation of the decoding process is divided into various machine learning modes such as CNN network mode, CTC+LSTM network mode, other RNN mode, and wavelet transform mode; 所述基于CNN网络模式,该矩阵如公式1所示;于是得到了标签序列,以及标签序列所对应的特征数据,以及这两者之间的关系;Based on the CNN network model, the matrix is shown in Formula 1; thus, the label sequence, the feature data corresponding to the label sequence, and the relationship between the two are obtained; 所述基于CTC+LSTM网络模式,该矩阵就对应于标签序列矩阵;就得到了标签序列,以及标签序列所对应的特征数据,这两者之间的关系,以及序列长度数据;序列长度是一维数据,其表现形式是:[max_time_step,…,max_time_step],其长度为batch_size,具体的值为max_time_step,还可以是通过小波变换形成的输出构成特征矩阵。Based on the CTC+LSTM network model, the matrix corresponds to the label sequence matrix; the label sequence, the feature data corresponding to the label sequence, the relationship between the two, and the sequence length data are obtained; the sequence length is one-dimensional data, and its expression is: [max_time_step,…,max_time_step], its length is batch_size, the specific value is max_time_step, and it can also be the output formed by wavelet transform to form a feature matrix. 16.根据权利要求14所述的方法,其特征在于,16. The method according to claim 14, characterized in that 所述对脑电信号信息进行处理,得到语法树之前,将特征信息分成有不同语义的组,每一组转换成特征字符串;根据所述特征信息形成特征字母表形成字母表;Before the EEG signal information is processed to obtain the syntax tree, the feature information is divided into groups with different semantics, and each group is converted into a feature string; a feature alphabet is formed according to the feature information; 所述对脑电信号信息进行处理,得到语法树包括:将脑电信号信息通过机器学习模型转化成相应的决策树;将决策树转化成语法树;The processing of the EEG signal information to obtain a syntax tree includes: converting the EEG signal information into a corresponding decision tree through a machine learning model; converting the decision tree into a syntax tree; 所述机器学习模型包括卷积神经网络CNN模型,长短期记忆网络LSTM模型,其他RNN模型,小波变换模型,各种机器学习模型;The machine learning models include convolutional neural network (CNN) models, long short-term memory (LSTM) models, other RNN models, wavelet transform models, and various machine learning models; 所述对脑电信号数据译码,就是为后续的语义分析处理做准备,目的是为了把数据按不同层次处理成有语义特征、而且被标签化的有序形式,同时得到对应的决策树和语法树;语义分析处理需要将得到的标签化的特征和决策树及语法树进行映射;The decoding of EEG signal data is to prepare for the subsequent semantic analysis and processing, the purpose of which is to process the data into an ordered form with semantic features and labels according to different levels, and to obtain the corresponding decision tree and syntax tree at the same time; the semantic analysis and processing requires mapping the obtained labeled features with the decision tree and syntax tree; 特征信息形成特征字母表:所有的输入信息被函数处理成特征矩阵M1,M 1的每一个元素同时也是一个矩阵M 2,M 2的元素是由二进制或十六进制形式的数值表示,所有M1的元素都属于同一个特征字母表,该字母表由不重复的元素构成;Feature information forms a feature alphabet: All input information is processed by the function into a feature matrix M1. Each element of M1 is also a matrix M2. The elements of M2 are represented by binary or hexadecimal values. All elements of M1 belong to the same feature alphabet, which consists of non-repeating elements. 所述将特征信息继续分成有不同语义的组,每一组转换成特征字符串,指的是,将特征字符串构成语义字母表,该语义字母表指的是:所有的输入信息M 2的元素在语义上可以同时被函数f M->a转换映射成特定的符号,这些符号可以是被计算机系统识别的字符集中的元素,这些符号和运算符、分隔符构成一个有限的非空集合A,其中每一个元素是可以被识别的字符,也可以是多个字符组成的字符串,因此在本文讨论范围内,也可以用V表示,也就是说语义字母表和语义词汇表被视作相同,简称为字母表;The feature information is further divided into groups with different semantics, and each group is converted into a feature string, which means that the feature string constitutes a semantic alphabet. The semantic alphabet means that all elements of the input information M2 can be semantically converted and mapped into specific symbols by the function fM->a at the same time. These symbols can be elements in the character set recognized by the computer system. These symbols and operators and separators constitute a finite non-empty set A, in which each element is a recognizable character or a string composed of multiple characters. Therefore, within the scope of discussion in this article, it can also be represented by V, that is, the semantic alphabet and the semantic vocabulary are regarded as the same, and are referred to as the alphabet for short; 按语法树生成句型、短语,用于高层次的文法、语法和语义分析处理。Generate sentence patterns and phrases according to the syntax tree for high-level grammar, syntax and semantic analysis.
CN201810728403.6A 2018-07-04 2018-07-04 A method and device for semantic processing of electroencephalogram signals Active CN110687999B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810728403.6A CN110687999B (en) 2018-07-04 2018-07-04 A method and device for semantic processing of electroencephalogram signals

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810728403.6A CN110687999B (en) 2018-07-04 2018-07-04 A method and device for semantic processing of electroencephalogram signals

Publications (2)

Publication Number Publication Date
CN110687999A CN110687999A (en) 2020-01-14
CN110687999B true CN110687999B (en) 2024-11-05

Family

ID=69107313

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810728403.6A Active CN110687999B (en) 2018-07-04 2018-07-04 A method and device for semantic processing of electroencephalogram signals

Country Status (1)

Country Link
CN (1) CN110687999B (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111580643B (en) * 2020-04-10 2022-04-26 天津大学 A Brain-Computer Interface Method Based on Steady-State Asymmetric Visual Evoked Potentials
CN113269084B (en) * 2021-05-19 2022-11-01 上海外国语大学 Movie and television play market prediction method and system based on audience group emotional nerve similarity
CN113792588B (en) * 2021-08-05 2024-04-09 深兰科技(上海)有限公司 Brain wave processing device, brain wave processing method, computer equipment and storage medium
CN115192044B (en) * 2022-07-28 2024-08-20 西安交通大学 Single-target SSVEP recognition system and method based on data continuation
CN116049630B (en) * 2023-02-14 2025-08-29 北京理工大学 A deep learning decoding method and system for hand movements driven by neural representation
CN117272036A (en) * 2023-10-09 2023-12-22 哈尔滨理工大学 Spike wave identification model construction method, spike wave identification method, device and equipment
CN117784940B (en) * 2024-02-23 2024-06-14 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) High-dimensional tensor analysis method and device aiming at Chinese character stroke writing imagination

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
RU2415642C1 (en) * 2009-09-03 2011-04-10 Российская Федерация, в лице которой выступает Министерство образования и науки Российской Федерации Method of classification of electroencephalographic signals in interface brain-computer
US9477651B2 (en) * 2010-09-29 2016-10-25 International Business Machines Corporation Finding partition boundaries for parallel processing of markup language documents
JP6618702B2 (en) * 2015-04-06 2019-12-11 国立研究開発法人情報通信研究機構 Perceptual meaning content estimation device and perceptual meaning content estimation method by analyzing brain activity
CN106569604B (en) * 2016-11-04 2019-09-17 天津大学 Audiovisual bimodal semantic matches and semantic mismatch collaboration stimulation brain-machine interface method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
运动相关电位分类算法比较和语义范式分析;刘彬等;《计算机与现代化》;20181115(第11期);第88-102页 *

Also Published As

Publication number Publication date
CN110687999A (en) 2020-01-14

Similar Documents

Publication Publication Date Title
CN110687999B (en) A method and device for semantic processing of electroencephalogram signals
CN111414393B (en) Semantic similar case retrieval method and equipment based on medical knowledge graph
CN113724882B (en) Method, device, equipment and medium for constructing user portrait based on inquiry session
CN112863630A (en) Personalized accurate medical question-answering system based on data and knowledge
JP7464800B2 (en) METHOD AND SYSTEM FOR RECOGNITION OF MEDICAL EVENTS UNDER SMALL SAMPLE WEAKLY LABELING CONDITIONS - Patent application
CN111126065B (en) Information extraction method and device for natural language text
CN117854748A (en) Knowledge graph and generation-based large model inquiry method and system
CN117316466B (en) Clinical decision method, system and equipment based on knowledge graph and natural language processing technology
CN107833629A (en) Aided diagnosis method and system based on deep learning
KR102424085B1 (en) Machine-assisted conversation system and medical condition inquiry device and method
CN112541086A (en) Knowledge graph construction method for cerebral apoplexy
CN113764112A (en) Online medical question and answer method
CN111462841A (en) An intelligent diagnosis device and system for depression based on knowledge graph
CN109003677B (en) Medical record data structured analysis and processing method
Jin et al. Affective computing for healthcare: Recent trends, applications, challenges, and beyond
CN114420233A (en) A method for extracting structured information from Chinese electronic medical records
Pendyala et al. Automated medical diagnosis from clinical data
WO2025118396A1 (en) Method for training natural language processing model, and method for generating subsequent text of dialogue
Drakopoulos et al. Emotion Recognition from Speech: A Survey.
CN120015351A (en) Multimodal large model recognition and intelligent intervention methods and systems for mental disorders
CN117057350A (en) Chinese electronic medical record named entity recognition method and system
CN117457237A (en) Auxiliary inquiry method and system based on medical entity extraction
CN117009408A (en) Method and device for searching infectious diseases based on infectious disease knowledge graph
CN112466462A (en) EMR information association and evolution method based on deep learning of image
CN115691786A (en) Information extraction method and auxiliary diagnosis device for ophthalmic diseases based on electronic medical records

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant