WO2003071480A1 - Active learning system, active learning method used in the same, program for the same, and recording medium containing the program - Google Patents
Active learning system, active learning method used in the same, program for the same, and recording medium containing the program Download PDFInfo
- Publication number
- WO2003071480A1 WO2003071480A1 PCT/JP2003/001184 JP0301184W WO03071480A1 WO 2003071480 A1 WO2003071480 A1 WO 2003071480A1 JP 0301184 W JP0301184 W JP 0301184W WO 03071480 A1 WO03071480 A1 WO 03071480A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- data
- learning
- input
- function
- active learning
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Classifications
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B7/00—Electrically-operated teaching apparatus or devices working with questions and answers
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
Definitions
- the present invention relates to an active learning system, an active learning method used for the active learning system, a program for the active learning method, and an active learning method.
- the present invention relates to a method for developing a plan for efficiently executing a plan.
- a factor (which is called a factor) that has an effect on the purpose of improving the performance of a certain machine (this is called a characteristic) is extracted, and the factor is extracted.
- the purpose is to detect whether the characteristics can be improved by the possible values (this is called the level).
- the “experimental design method” analyzes the correlation between factors and its effect on characteristics by performing experiments with varying levels of factors.
- the conventional “experimental design method” does not always determine the level of factors to be used in the next experiment based on the experimental results obtained in the actual experiment.
- the traditional “design of experiment” is not a method of deciding the strategy of what experiment procedure to pursue in pursuit of characteristics.
- the problems arising in the above-mentioned conventional “experimental design method” will be described below with reference to “a specific problem of the binding ability of immunopeptides” as an example.
- the immune system recognizes antigens by binding cells in the body called MHC (Major Histocoatability Croples) cells to small fragments of proteins, called peptides, that the antigen has. It is considered. For this reason, the problem of identifying the binding ability of immunopeptides is one of the most important issues in medicine, especially in the field of immunology.
- a certain protein binds to an amino acid sequence having a length of 10 (hereinafter referred to as a peptide).
- a peptide an amino acid sequence having a length of 10
- the binding ability changes depending on the amino acid sequence. How If we try to find out if a peptide has a high binding ability to this protein by conducting biochemical experiments, we will produce all 10-length peptides and conduct experiments to measure each binding ability It is necessary to go through the procedure.
- An object of the present invention is to solve the above-described problem, and to perform an experiment in which it is practically impossible to perform all of the target trials when input / output data is a continuous value. It is an object of the present invention to provide an active learning system capable of estimating the value, an active learning method used therein, a program thereof, and a recording medium storing the program.
- An active learning system is an active learning system that executes an active learning method for estimating a functional relationship that at least approximately holds between input and output data based on selected input and output data, A function relationship is estimated from past input / output data using a certain expression, and a plurality of learning algorithms of any of a predetermined single type and a plurality of types are used as lower-level algorithms, and the lower-level algorithm is used as the lower-level algorithm.
- Another active learning system further improves the accuracy by repeating data sampling in combination with a boosting technique that improves the accuracy of a passive learning algorithm.
- a lower algorithm performs learning on a plurality of input candidate data selected at random.
- the input data designating means the data in which the variance of the predicted values of a plurality of hypotheses is the largest is selected as the input data.
- Another configuration example of the active learning system according to the present invention includes data updating means for adding input candidate data and function values to past data.
- Still another configuration example of the active learning system according to the present invention includes an experiment means for experimentally finding a function value for input candidate data.
- Another configuration example of the active learning system according to the present invention includes simulation means for obtaining a function value for input candidate data by simulation.
- any one of a conclusion according to an arbitrary regularity and a result obtained by a simulation calculation is used as input data of the lower algorithm.
- the active learning method is an active learning method for estimating a functional relationship that at least approximately holds between input and output data based on selected input and output data.
- a learning step of learning data as training data and a function prediction step of predicting a function value for a plurality of input candidate data using a final hypothesis obtained from a plurality of hypotheses obtained from the respective algorithms in the learning step.
- Another active learning method is to repeat data resampling.
- the method further includes a boosting step of improving the learning accuracy of the learning step by a boosting technique for improving the accuracy of a passive learning algorithm.
- a lower algorithm performs learning on a plurality of input candidate data selected at random.
- the data having the largest variance of the predicted values of a plurality of hypotheses is selected as input candidate data.
- a recording medium on which a program for an active learning method according to the present invention is recorded is an active learning system for executing an active learning method for estimating a functional relationship that at least approximately holds between input and output data based on selected input and output data.
- a computer-readable recording medium on which a program for causing a computer to function as a computer is recorded.
- Function prediction processing for predicting a function value for force candidate data; and input data for selecting input data from the input candidate data based on a function value using a variance of predicted values of the plurality of learned hypotheses.
- the program for executing the specified process is recorded.
- a recording medium on which another active learning method program according to the present invention is recorded is provided with a boosting technique for improving the accuracy of a passive learning algorithm by repeating the resampling of data by a computer.
- a program for further executing the boosting process for improving the learning accuracy of the learning process is recorded.
- One configuration example of a recording medium on which the program of the active learning method according to the present invention is recorded is a computer in which a lower algorithm learns a plurality of randomly selected input candidate data in a learning process and a function prediction process.
- It records a program to execute the process of performing Another example of the configuration of the recording medium is to allow a computer to execute a process of selecting data having a maximum variance of predicted values of a plurality of hypotheses as input candidate data in an input data designation process. Program Have recorded.
- a program for an active learning method causes a computer to function as an active learning system that executes an active learning method for estimating a functional relationship that at least approximately holds between input and output data based on selected input and output data.
- a function prediction process for predicting a function value for a plurality of input candidate data using a final hypothesis More it is made to execute the input data designation processing for selecting input data from the input candidate data distribution based on the value of functions with the predicted value of the hypothesis.
- Another active learning program according to the present invention is a boosting technique for improving the accuracy of a learning process by boosting a computer to improve the accuracy of a passive learning algorithm by repeating data resampling. The process is being executed further.
- One example of the configuration of the program of the active learning method according to the present invention causes a computer to execute a process in which a lower-level algorithm performs learning on a plurality of randomly selected input candidate data in a learning process and a function prediction process.
- Another configuration example of the program causes a computer to execute a process of selecting, as input candidate data, data in which a variance of predicted values of a plurality of hypotheses is maximum in an input data designation process.
- the active learning system of the present invention proposes an experiment design method using active learning for continuous input / output data in order to solve the above-mentioned problems.
- the “boosting” technology which is a method of improving the accuracy of the passive learning algorithm by repeating data sampling, is one of the leading methods of active learning.
- the accuracy between the input and the output is In a design of experiment, which estimates functional relationships that approximately hold based on input / output data obtained from experiments, a predetermined expression that estimates functional relationships from past input / output data using a certain expression
- a hypothesis is obtained from past data by passive learning. Estimating a hypothesis, estimating the next input candidate data based on the hypothesis, specifying input data based on the hypothesis, experimentally obtaining a function value for the input caught data by experiment, and calculating the input candidate data and the function value.
- An experimental design method is obtained, which includes a data updating step to be added to past data. In this case, the above steps are executed at each time of selecting the input data of the next experiment, that is, each time the input data of the next experiment is selected until the estimation of the functional relationship is completed.
- an experimental design program for estimating a functional relationship that is established accurately or approximately between an input and an output based on input / output data obtained by an experiment includes: An arbitrary learning algorithm that estimates a functional relationship from data using a certain expression is used as a lower algorithm, and the lower algorithm learns input / output data up to that point as training data. Boo boosted by boosting technology The computer performs a sting process and an input data designation process that selects input candidate data as input based on the value of a function that uses the variance of the predicted values of multiple hypotheses that have been improved by the boosting process. .
- a hypothesis is estimated from past data by passive learning in an experiment design program for estimating a functional relationship between input and output based on input / output data obtained by experiments.
- Hypothesis estimation processing input candidate data instruction processing to indicate the next input candidate data based on the hypothesis, experimental processing to experimentally determine the function value for the input candidate data, and past input data and function values to past data
- the data update process to be added to the computer is executed by the computer.
- each of the above steps is performed at each time of selecting the input point of the next experiment, that is, each time the input of the next experiment is selected until the estimation of the functional relationship is completed.
- the hypothesis estimation processing preferably includes a boosting processing.
- the input candidate data processing includes input candidate data generation processing for generating a plurality of input candidate data that are not included in past data, and input candidate data selection for selecting the next input candidate data from the plurality of input candidate data based on a hypothesis. And data selection processing.
- the above-described experiment design method is realized by active learning.
- Active learning refers to learning as described in, for example, Abe and Nakamura's commentary “Active Learning Overview” (“Information Processing”, Vol. 38, No. 55, pp. 58-56, 1997). Is a method by which a person selectively obtains information from the environment and learns.
- passive learning learning without active learning.
- the learner does not selectively obtain information from the environment, but estimates the rules (hypotheses) inherent in the data from a given large amount of data.
- ⁇ boosting '' a technique called ⁇ boosting '', which improves learning accuracy by repeating resampling so that data areas for which the hypothesis up to that point is not good at given data is trained, has been attracting attention in recent years. Has been confirmed to have high performance experimentally. Is coming.
- boosting techniques have only been used in passive learning and have not yet been applied to active learning.
- active learning there has never been a concept unique to “boosting” technology, such as repeating resampling, and no method of using this technology in the framework of active learning has been proposed.
- FIG. 1 is a diagram showing a flow of an active learning algorithm according to an embodiment of the present invention.
- FIG. 2 is a block diagram showing a configuration of the active learning machine according to the embodiment of the present invention.
- FIG. 3 is a block diagram showing the configuration of the active learning system according to the first embodiment of the present invention.
- FIG. 4 is a block diagram showing a configuration of an active learning system according to a second embodiment of the present invention.
- FIG. 5 is a block diagram showing the configuration of the active learning system according to the third embodiment of the present invention.
- FIG. 6 is a block diagram showing a configuration of an active learning system according to a fourth embodiment of the present invention.
- FIG. 7 is a block diagram showing the configuration of the active learning system according to the fifth embodiment of the present invention.
- FIG. 8 is a block diagram showing a configuration of an active learning system according to a sixth embodiment of the present invention.
- FIG. 9 is a block diagram showing the configuration of the active learning system according to the seventh embodiment of the present invention.
- FIG. 10 is a block diagram showing a configuration of an active learning machine according to an eighth embodiment of the present invention. Detailed description of the embodiment
- FIG. 1 is a diagram showing a flow of an active learning algorithm according to an embodiment of the present invention
- FIG. It is a block diagram showing composition of an active learning machine by an example.
- the boosting device 1 includes a learning prediction error calculation unit 11, a data distribution updating unit 12, and a judgment unit 13, a function calculation unit 21, and input candidate data selection unit 22.
- the experiment simulation 30 need not necessarily be included in the apparatus, and may be replaced with an external simulator / manual experiment / simulation calculation. It should be noted that it is also possible to realize by a program by replacing each of the above means with a step.
- the active learning machine 4 includes a preprocessing device 41, a lower learning machine device 42 composed of lower learning machines 4 2 1 to 4 2 N, and a hypothesis' predicted value storage area 4 3 1 to 4 3 N. It is composed of a prediction value storage device 43, a post-processing device 44, and a recording medium 45 for recording a program to be executed by a computer when each of the above devices is realized by a computer. I have.
- the active learning machine 4 includes a computer and software that causes the computer to function as the active learning machine 4.
- this program and the hardware resources of the computer cooperate with each other, and the preprocessing device 41 and the lower learning machine device 42 are connected to the computer.
- Hypothesis ⁇ The predicted value storage device 43 and the post-processing device 44 are realized.
- the experiment design method according to the embodiment of the present invention will be described with reference to FIGS. In this embodiment, first, initial data and initial distribution of the data are input to the input means 100.
- the questions in the active learning are repeated between the learning prediction error calculating means 11 and the judging means 25.
- the first part of the iteration is performed in the process of the boosting device 1. This processing is performed in the active learning machine 4 by the preprocessing device 41 and the lower learning machine 42.
- the learning prediction error calculation means 11 Learning from the current data is performed using a learning algorithm, and a hypothesis is obtained as the learning result.
- the lower learning algorithm may be any learning algorithm that handles continuous values, such as a regression tree.
- the result is output to the hypothesis / predicted value storage device 43.
- the learning prediction error calculation means 11 calculates a prediction value from the hypothesis to the current data, and calculates an error from the prediction value.
- the data distribution updating means 12 updates the data resampling distribution from the error obtained by the learning prediction error calculating means 11. '
- the number of repetitions of the processing of the boosting device 1 may be one.
- the data distribution updating means 12 updates the data distribution and the condition judgment statement of the judging means 13 and the loop returning from the judging means 13 to the learning prediction error calculating means 11 becomes unnecessary. In other words, lower learning processing without boosting processing is performed.
- the post-processing device 44 includes function calculation means 21, input candidate data selection means 22, function value input means 23, data update means 24, determination means 25, and an experimental simulation. 30.
- the function calculation means 21 calculates the final hypothesis using the function value using the variance of the predicted values of the plurality of learned hypotheses and the plurality of hypotheses, and outputs the final hypothesis.
- the input candidate data selection means 22 selects input candidate data based on the value of the function using the variance obtained from the calculation result by the function calculation means 21. Indicate as a new experiment for simulation 30).
- the function value input means 23 receives the function value of the input candidate data tested in the experiment (experimental simulation 30) by the experimenter.
- the data updating means 24 adds the obtained input candidate data and its function value to the past data.
- the determining means 25 determines whether or not the number of questions has reached a certain value (N). If not, the process returns to the boosting device 1 to continue active learning. If the number of questions reaches a certain value, the active learning ends.
- N certain value
- the lower-order learning machine device 42 and the hypothesis / prediction value storage device 4 3 are combined and operated as the learning prediction error calculation means 11, so that the time until the next input data is selected for the lower-order algorithm is obtained. It functions as learning means for learning input / output data as training data and function prediction means for predicting function values for a plurality of input candidate data.
- the lower learning machine device 42 also operates as the data distribution updating device 12 and the judging device 13 and acts as a boosting device for improving the learning accuracy of the learning device.
- the post-processing device 44 operates as the function calculating means 21 and the input candidate data selecting means 22, so that the input candidate data based on the variance of the predicted values of the plurality of learned hypotheses is used. Acts as input data designating means for selecting input data from Further, the pre-processing device 41 and the post-processing device 44 cooperate with each other to operate as the function value input means 23, the data updating means 24, and the experiment simulation 30, so that the input candidate data and the function value can be obtained. Acts as a data updating means to add to past data. As a result, it becomes possible to update the experimental result database 6 and input to the active learning machine 4 without manual intervention. In this way, it can be regarded as another learning system, and the user's labor can be greatly reduced.
- the embodiment of the active learning machine 4 according to the present invention has been described above, the present invention is not limited to this, and it is possible to make changes and improvements within the ordinary knowledge of those skilled in the art. it is obvious.
- the lower algorithm may perform learning on a plurality of input candidate data selected at random.
- the lower learning algorithms become independent of each other and parallelization becomes easier. Furthermore, since only random numbers are generated, the computation time is shorter than other methods.
- the input candidate data selecting means 22 acting as a part of the input data designating means may select, as the input candidate data, data in which the variance of the predicted values of a plurality of hypotheses is maximum. Data with the largest variance has the least amount of information available. Since it is large, the effect of learning can be maximized by selecting it as input candidate data.
- any one of a conclusion according to an arbitrary regularity and a result obtained by simulation calculation may be used as input data of the lower algorithm. Les ,.
- FIG. 3 is a block diagram showing the configuration of the active learning system according to the first embodiment of the present invention.
- the active learning system 5 includes an active learning machine 4 and a storage device [experiment result database (attribute values and results)] 6 for storing past inputs and outputs.
- the active learning machine 4 performs the same operation as the above-described embodiment of the active learning machine 4 according to the present invention.
- the active learning system 5 has the same input and output as the active learning machine 4 shown in FIG. 2, a system that does not require an external storage device can be constructed.
- FIG. 4 is a block diagram showing a configuration of an active learning system according to a second embodiment of the present invention.
- the active learning system according to the second embodiment of the present invention is an example of a system incorporating an active learning machine 4.
- the next input candidate output by the system of the active learning machine 4 is manually experimented by the experimenter and simulation calculation 7 by the experimenter.
- the experimenter actually performs the experiment, and the set of the input candidate and the experiment result is stored in the storage device 6. .
- the active learning machine 4 performs learning by using it as training data, and outputs the next input candidate.
- the experimenter can efficiently select an infinite number of experimental objects from the innumerable experimental objects according to the instructions of the active learning machine 4. Experiments can be performed efficiently.
- FIG. 5 is a block diagram showing a configuration of the Noh learning system according to the third embodiment of the present invention.
- an active learning system 5 is a system in which a storage device 6 is combined with an active learning machine 4.
- the storage device 6, which was required in the second embodiment of the present invention, is no longer required outside the active learning system 5, so that the space can be reduced and the data transfer time between the database and the active learning machine can be reduced. Can be reduced.
- FIG. 6 is a block diagram showing a configuration of an active learning system according to a fourth embodiment of the present invention.
- the active learning system according to the fourth embodiment of the present invention is an example of a system incorporating an active learning machine 4.
- the simulator 8 automatically performs a simulation corresponding to the next input candidate output by the system of the active learning machine 4, and stores a set of the input candidate and the simulation result in the storage device 6. Further, the active learning machine 4 performs learning using the training data as training data, and outputs the next input candidate.
- the experiment target can be efficiently selected from the myriad of experiment targets, and the user can select only the combination of the attribute of the input candidate data and the result value. If you do, you will be able to automatically carry out experiments that meet your purpose.
- FIG. 7 is a block diagram showing a configuration of an active learning system according to a fifth embodiment of the present invention.
- an active learning system 5 is configured by incorporating a storage device 6 into an active learning machine 4.
- the storage device 6, which was required in the above-described fourth embodiment of the present invention, is not required outside the system.
- the data transfer time between the two can be reduced.
- FIG. 8 is a block diagram showing a configuration of an active learning system according to a sixth embodiment of the present invention.
- an active learning system 9 is configured by incorporating a simulator 8 into an active learning machine 4.
- an active learning system including simulation can be constructed, and optimal prediction simulation can be automatically performed without human intervention. This In other words, the user can obtain the desired result automatically without any trouble if only the initial conditions are input, so that the burden on the user can be reduced.
- the system user only needs to prepare the storage device 6, and the simulator 8 and the like are different from those of the above-described fifth embodiment of the present invention. It is possible to reduce the trouble of preparing an external system.
- FIG. 9 is a block diagram showing a configuration of an active learning system according to a seventh embodiment of the present invention.
- the active learning system 10 the active learning system 10
- FIG. 10 is a block diagram showing a configuration of an active learning machine according to an eighth embodiment of the present invention.
- an active learning machine 40 according to an eighth embodiment of the present invention includes an input data generation machine 4 of a lower learning machine 42 between a preprocessing device 41 and a lower learning machine 42.
- the configuration is the same as that of the active learning machine 4 shown in FIG. 2 except that 6 is provided, and the same components are denoted by the same reference numerals.
- the input data generation machine 46 includes an input data generation machine 46 1 to 46 N.
- the data stored in the storage device 6 is processed and supplied to the lower learning machine 42, so that the lower learning is performed.
- the mechanical device 42 can learn from data provided from the input data generating mechanical device 46, and can be applied to a wider range of problems.
- the active learning machine 40 includes a computer and software that causes the computer to function as the active learning machine 40.
- the program recorded on the recording medium 45 is read into the computer, the program and the hardware resources of the computer cooperate with each other, and the preprocessing device 41 and the input data generation device 46 are connected to the computer.
- the lower learning machine device 42, the hypothesis 'predicted value storage device 43' and the post-processing device 44 are realized.
- Boosting technology that can improve the prediction accuracy to an arbitrary accuracy if it is a possible learning algorithm is combined with group query learning, which is an active learning method, by using an appropriate method to improve accuracy. It is possible to realize a high active learning method.
- the active learning system of the present invention is an active learning system that executes an active learning method for estimating a functional relationship that at least approximately holds between input and output data based on the selected input and output data.
- a functional relationship is estimated from past input / output data using a certain expression form, and a plurality of predetermined learning algorithms, one of a single type or a plurality of types, are used as lower-level algorithms, and the next lower-level algorithms are used.
- Function to make predictions Predictive means and input based on function values using the variance of predicted values of multiple learned hypotheses
- Another active learning system of the present invention further includes a boosting means for improving the learning accuracy of the learning means by a boosting technique for improving the accuracy of a passive learning algorithm by repeating resampling of data.
- a boosting technique that can improve the prediction accuracy to an arbitrary level as long as it is a learning algorithm that can achieve a moderate level of accuracy is called a group learning method that is an active learning method.
- the active learning system according to the present invention and the active learning method used in the active learning system perform an experiment in which it is practically impossible to perform all the trials when the input and output data are continuous values. It is useful as a system and method for estimating an objective function with a small number of experiments, and is particularly suitable for making a plan for efficiently performing experiments and designs in fields such as molecular biology.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Medical Informatics (AREA)
- Data Mining & Analysis (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Mathematical Physics (AREA)
- Artificial Intelligence (AREA)
- Business, Economics & Management (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
- Complex Calculations (AREA)
Abstract
Description
能動学習システム、 それに用いる能動学習法、 そのプログラム及びそのプログラ ムを記録した記録媒体 発明の背景 BACKGROUND OF THE INVENTION Active learning system, active learning method used therefor, program thereof, and recording medium storing the program
本発明は、 能動学習システム、 それに用いる能動学習法、 そのプログラム及び 明 The present invention relates to an active learning system, an active learning method used for the active learning system, a program for the active learning method, and an active learning method.
そのプログラムを記録した記録媒体に関し、 特に分子生物学等の分野の実験ゃ設 田 Regarding the recording medium on which the program is recorded, particularly experiments in fields such as molecular biology.
計を効率的に遂行する計画を立案する方法に関する。 The present invention relates to a method for developing a plan for efficiently executing a plan.
一般に、 従来の 「実験計画法」 においては、 例えば、 ある機械の性能向上とい つた目的 (これを特性と呼ぶ) に対し、 影響を及ぼすような要因 (これを因子と 呼ぶ) を抽出し、 因子のとり得る値 (これを水準と呼ぶ) によって、 特性が向上 するかどうかを検查することが目的となっている。 すなわち、 「実験計画法」 で は、 因子のとる水準を変動させて実験することによって、 因子同士の相関とその 特性への影響とを解析している。 In general, in the conventional “design of experiment”, for example, a factor (which is called a factor) that has an effect on the purpose of improving the performance of a certain machine (this is called a characteristic) is extracted, and the factor is extracted. The purpose is to detect whether the characteristics can be improved by the possible values (this is called the level). In other words, the “experimental design method” analyzes the correlation between factors and its effect on characteristics by performing experiments with varying levels of factors.
しかしながら、 従来の 「実験計画法」 では、 実際の実験において、 過去に得ら れた実験結果を踏まえて、 次に行う実験で使用するべき因子の水準を逐一決める わけではない。 つまり、 従来の 「実験計画法」 は特性の追求のために、 どのよう な実験手順を行うかという戦略を決定する方法ではない。 However, the conventional “experimental design method” does not always determine the level of factors to be used in the next experiment based on the experimental results obtained in the actual experiment. In other words, the traditional “design of experiment” is not a method of deciding the strategy of what experiment procedure to pursue in pursuit of characteristics.
上述した従来の 「実験計画法」 において生じる問題を、 一例として 「免疫ぺプ チドの結合能の特定問題」 を挙げて以下に説明する。 体内の MHC (Ma j o r H i s t o c omp a t i b i l i t y C omp l e ) 細胞と呼ばれる細 胞が、 抗原の持つぺプチドと呼ばれる小さい蛋白質のかけらと結合することによ つて、 免疫系は抗原を認識していると考えられている。 このため、 「免疫べプチ ドの結合能の特定問題」 は、 医学、 特に免疫学の分野における最重要問題の一つ である。 The problems arising in the above-mentioned conventional “experimental design method” will be described below with reference to “a specific problem of the binding ability of immunopeptides” as an example. The immune system recognizes antigens by binding cells in the body called MHC (Major Histocoatability Croples) cells to small fragments of proteins, called peptides, that the antigen has. It is considered. For this reason, the problem of identifying the binding ability of immunopeptides is one of the most important issues in medicine, especially in the field of immunology.
例えば、 ある蛋白質が長さ 10のアミノ酸配列 (以下、 ペプチドと呼ぶ) と結 合するとし、 この時、 アミノ酸配列に依存して結合能が変わるとする。 どのよう なぺプチドがこの蛋白質に対して結合能が高いのかを仮に生化学実験を行つて知 ろうとすれば、 長さ 1 0のペプチドを全種類生成し、 それぞれの結合能を測定す る実験を行うという手順を経る必要がある。 For example, suppose that a certain protein binds to an amino acid sequence having a length of 10 (hereinafter referred to as a peptide). At this time, it is assumed that the binding ability changes depending on the amino acid sequence. How If we try to find out if a peptide has a high binding ability to this protein by conducting biochemical experiments, we will produce all 10-length peptides and conduct experiments to measure each binding ability It is necessary to go through the procedure.
しかしながら、 アミノ酸の種類は 2 0なので、 長さ 1 0のペプチドに対しては、 2 0 1 Qの組合せが存在することになる。 これら全ての配列に対する結合能のデー タを生化学実験から得るのは到底不可能である。 However, since the number of amino acids is 20, there are 201 Q combinations for a peptide of length 10. It is almost impossible to obtain data on binding ability to all these sequences from biochemical experiments.
つまり、 より一般的に述べれば、 変数 Xと、 それに対応するラベル (出力) y のデータとから、 Xと yとの関係を表現する関数 f ( y = f ( χ ) ) を知りたい 時、 変数 Xの取り得る範囲が非常に広く、 全ての Xに対して f ( X ) の値を実験 的に求めることは現実的に不可能な場合が多々存在する。 このような場合、 従来 の 「実験計画法」 によれば、 現実には実行不可能な数の実験を行わなければ解析 を行うができない、 すなわち、 実際には解析することができないことになる。 このように、 従来、 過去に得られたデータ (X , y ) から次の実験すべき入力 候補データを決めながら関数 f を推定していくための精度の良い方法が知られて おらず、 そのような方法が待望されている。 発明の概要 In other words, more generally, when we want to know the function f (y = f (χ)) that expresses the relationship between X and y from the variable X and the data of the corresponding label (output) y, The range of the variable X is very wide, and it is often not practically possible to experimentally determine the value of f (X) for all X. In such a case, according to the conventional “experimental design method”, analysis cannot be performed unless a number of experiments that cannot be actually performed is performed, that is, analysis cannot be actually performed. As described above, there has not been known a highly accurate method for estimating the function f while deciding the next input candidate data to be tested from the data (X, y) obtained in the past. Such a method is long-awaited. Summary of the Invention
本発明の目的は上記の問題点を解消し、 入出力データが連続的な値の時に、 対 象となる全試行の遂行が事実上不可能な実験を行う際に、 少ない実験回数で目的 関数の推定を行うことができる能動学習システム、 それに用いる能動学習法、 そ のプログラム及びそのプログラムを記録した記録媒体を提供することにある。 本発明による能動学習システムは、 選択した入出力のデータを基に、 入出力デ ータ間に少なくとも近似的に成り立つ関数関係の推定を行う能動学習法を実行す る能動学習システムであって、 過去の入出力データから一定の表現形を用いて関 数関係を推定しかつ予め定められた単一及ぴ複数種類のうちのいずれかの複数の 学習アルゴリズムを下位アルゴリズムとして用いて前記下位アルゴリズムに次入 力データを選択する時点までの入出力データを訓練データとして学習させる学習 手段と、 前記学習手段各々のアルゴリズムから得られた複数の仮説から得られる 最終仮説を用いて複数の入力候補データに対する関数値の予測を行う関数予測手 段と、 前記学習された複数の仮説の予測値の分散を使った関数の値を基に前記入 力候補データの中から入力データを選択する入力データ指定手段とを備えている。 本発明による他の能動学習システムは、 受動的な学習アルゴリズムの精度を向 上させるブースティング技術と組合せ、 データのサンプリングを繰り返すことで、 精度をさらに向上させる。 An object of the present invention is to solve the above-described problem, and to perform an experiment in which it is practically impossible to perform all of the target trials when input / output data is a continuous value. It is an object of the present invention to provide an active learning system capable of estimating the value, an active learning method used therein, a program thereof, and a recording medium storing the program. An active learning system according to the present invention is an active learning system that executes an active learning method for estimating a functional relationship that at least approximately holds between input and output data based on selected input and output data, A function relationship is estimated from past input / output data using a certain expression, and a plurality of learning algorithms of any of a predetermined single type and a plurality of types are used as lower-level algorithms, and the lower-level algorithm is used as the lower-level algorithm. Learning means for learning input / output data up to the point of selecting the next input data as training data; and a plurality of input candidate data using a final hypothesis obtained from a plurality of hypotheses obtained from an algorithm of each of the learning means. Function predictor that predicts function values And input data designating means for selecting input data from the input candidate data based on a function value using the variance of the learned predicted values of the plurality of hypotheses. Another active learning system according to the present invention further improves the accuracy by repeating data sampling in combination with a boosting technique that improves the accuracy of a passive learning algorithm.
本発明による能動学習システムの一構成例は、 学習手段及び関数予測手段にお いて、 ランダムに選ばれた複数の入力候補データに対して下位アルゴリズムがそ れぞれ学習を行う。 また、 入力データ指定手段において、 複数の仮説の予測値の 分散が最大であるデータを入力候捕データとして選択する。 In one configuration example of the active learning system according to the present invention, in a learning unit and a function prediction unit, a lower algorithm performs learning on a plurality of input candidate data selected at random. In the input data designating means, the data in which the variance of the predicted values of a plurality of hypotheses is the largest is selected as the input data.
本発明による能動学習システムの別の構成例は、 入力候補データ及び関数値を 過去のデータに追加するデータ更新手段を含む。 Another configuration example of the active learning system according to the present invention includes data updating means for adding input candidate data and function values to past data.
本発明による能動学習システムのさらに別の構成例は、 入力候補データに対す る関数値を実験によつて求める実験手段を含む。 Still another configuration example of the active learning system according to the present invention includes an experiment means for experimentally finding a function value for input candidate data.
本発明による能動学習システムの他の構成例は、 入力候補データに対する関数 値をシミュレーションによって求めるシミュレーション手段を含む。 Another configuration example of the active learning system according to the present invention includes simulation means for obtaining a function value for input candidate data by simulation.
本発明による能動学習システムのさらに他の構成例は、 学習手段において、 下 位アルゴリズムの入力データとしてそれぞれ任意の規則性にしたがった結論及び シミュレーション計算によって求めた結果のいずれかを用いる。 In still another configuration example of the active learning system according to the present invention, in the learning means, any one of a conclusion according to an arbitrary regularity and a result obtained by a simulation calculation is used as input data of the lower algorithm.
本発明による能動学習法は、 選択した入出力のデータを基に、 入出力データ間 に少なくとも近似的に成り立つ関数関係の推定を行う能動学習法であって、 過去 の入出力データから一定の表現形を用!/、て関数関係を推定しかつ予め定められた 単一及び複数種類のうちのいずれかの複数の学習アルゴリズムを下位アルゴリズ ムとして用いて前記下位アルゴリズムに次入力データを選択する時点までの入出 力データを訓練データとして学習させる学習段階と、 前記学習段階各々のァルゴ リズムから得られた複数の仮説から得られる最終仮説を用いて複数の入力候補デ ータに対する関数値の予測を行う関数予測段階と、 前記学習された複数の仮説の 予測値の分散を使つた関数の値を基に前記入力候補データの中から入力データを 選択する入力データ指定段階とを備えている。 The active learning method according to the present invention is an active learning method for estimating a functional relationship that at least approximately holds between input and output data based on selected input and output data. Use shape! /, Input / output up to the point when the next input data is selected for the lower-level algorithm by estimating the functional relationship and using a plurality of learning algorithms of one or more of predetermined single types as the lower-level algorithm A learning step of learning data as training data; and a function prediction step of predicting a function value for a plurality of input candidate data using a final hypothesis obtained from a plurality of hypotheses obtained from the respective algorithms in the learning step. And an input data designating step of selecting input data from the input candidate data based on a function value using a variance of predicted values of the plurality of learned hypotheses.
本発明による他の能動学習法は、 データのリサンプリングを繰り返すことによ つて受動的な学習アルゴリズムの精度を向上させるブースティング技術によって 前記学習段階の学習精度を向上させるブースティング段階をさらに有する。 Another active learning method according to the present invention is to repeat data resampling. The method further includes a boosting step of improving the learning accuracy of the learning step by a boosting technique for improving the accuracy of a passive learning algorithm.
本発明による能動学習法の一構成例は、 学習段階及び関数予測段階において、 ランダムに選ばれた複数の入力候補データに対して下位アルゴリズムがそれぞれ 学習を行う。 また、 入力データ指定段階において、 複数の仮説の予測値の分散が 最大であるデータを入力候補データとして選択する。 In one configuration example of the active learning method according to the present invention, in a learning stage and a function prediction stage, a lower algorithm performs learning on a plurality of input candidate data selected at random. In the input data designating step, the data having the largest variance of the predicted values of a plurality of hypotheses is selected as input candidate data.
本発明による能動学習法のプログラムを記録した記録媒体は、 選択した入出力 のデータを基に、 入出力データ間に少なくとも近似的に成り立つ関数関係の推定 を行う能動学習法を実行する能動学習システムとしてコンピュータを機能させる ためのプログラムを記録したコンピュータ読み取り可能な記録媒体であって、 コ ンピュータに、 過去の入出力データから一定の表現形を用いて関数関係を推定し かつ予め定められた単一及び複数種類のうちのいずれかの複数の学習アルゴリズ ムを下位アルゴリズムとして用いて前記下位アルゴリズムに次入力データを選択 する時点までの入出力データを訓練データとして学習させる学習処理と、 前記学 習処理各々のアルゴリズムから得られた複数の仮説から得られる最終仮説を用レ、 て複数の入力候補データに対する関数値の予測を行う関数予測処理と、 前記学習 された複数の仮説の予測値の分散を使った関数の値を基に前記入力候補データの 中から入力データを選択する入力データ指定処理とを実行させるためのプロダラ ムを記録している。 A recording medium on which a program for an active learning method according to the present invention is recorded is an active learning system for executing an active learning method for estimating a functional relationship that at least approximately holds between input and output data based on selected input and output data. A computer-readable recording medium on which a program for causing a computer to function as a computer is recorded. A learning process of learning input / output data as training data up to a point at which the next input data is selected by the lower algorithm using a plurality of learning algorithms of any of a plurality of types as lower algorithms; and the learning process. Using the final hypothesis obtained from the multiple hypotheses obtained from each algorithm, Function prediction processing for predicting a function value for force candidate data; and input data for selecting input data from the input candidate data based on a function value using a variance of predicted values of the plurality of learned hypotheses. The program for executing the specified process is recorded.
本発明による他の能動学習法のプログラムを記録した記録媒体は、 コンビユー タに、 データのリサンプリングを,操り返すことによつて受動的な学習アルゴリズ ムの精度を向上させるブースティング技術によつて学習処理の学習精度を向上さ せるブースティング処理をさらに実行させるためのプログラムを記録している。 本発明による能動学習法のプログラムを記録した記録媒体の一構成例は、 コン ピュータに、 学習処理及び関数予測処理において、 ランダムに選ばれた複数の入 力候補データに対して下位アルゴリズムがそれぞれ学習を行う処理を実行させる ためのプログラムを記録している。 また、 この記録媒体の別の構成例は、 コンビ ユータに、 入力データ指定処理において、 複数の仮説の予測値の分散が最大であ るデ一タを入力候補データとして選択する処理を実行させるためのプログラムを 記録している。 A recording medium on which another active learning method program according to the present invention is recorded is provided with a boosting technique for improving the accuracy of a passive learning algorithm by repeating the resampling of data by a computer. A program for further executing the boosting process for improving the learning accuracy of the learning process is recorded. One configuration example of a recording medium on which the program of the active learning method according to the present invention is recorded is a computer in which a lower algorithm learns a plurality of randomly selected input candidate data in a learning process and a function prediction process. It records a program to execute the process of performing Another example of the configuration of the recording medium is to allow a computer to execute a process of selecting data having a maximum variance of predicted values of a plurality of hypotheses as input candidate data in an input data designation process. Program Have recorded.
本発明による能動学習法のプログラムは、 選択した入出力のデータを基に、 入 出力データ間に少なくとも近似的に成り立つ関数関係の推定を行う能動学習法を 実行する能動学習システムとしてコンピュータを機能させるためのプログラムで あって、 コンピュータに、 過去の入出力データから一定の表現形を用いて関数関 係を推定しかつ予め定められた単一及ぴ複数種類のうちのいずれかの複数の学習 アルゴリズムを下位アルゴリズムとして用いて前記下位アルゴリズムに次入力デ ータを選択する時点までの入出力データを訓練データとして学習させる学習処理 と、 前記学習処理各々のアルゴリズムから得られた複数の仮説から得られる最終 仮説を用いて複数の入力候補データに対する関数値の予測を行う関数予測処理と、 前記学習された複数の仮説の予測値の分散を使った関数の値を基に前記入力候補 データの中から入力データを選択する入力データ指定処理とを実行させている。 本発明による他の能動学習法のプログラムは、 コンピュータに、 データのリサ ンプリングを繰り返すことによつて受動的な学習アルゴリズムの精度を向上させ るブースティング技術によって学習処理の学習精度を向上させるブースティング 処理をさらに実行させている。 A program for an active learning method according to the present invention causes a computer to function as an active learning system that executes an active learning method for estimating a functional relationship that at least approximately holds between input and output data based on selected input and output data. A program for estimating a functional relationship from a past input / output data using a certain expression form, and learning one or a plurality of learning algorithms of one or more of predetermined types. Is used as a lower algorithm, the lower algorithm learns input / output data up to the point at which the next input data is selected as training data, and a plurality of hypotheses obtained from the respective algorithms of the learning process. A function prediction process for predicting a function value for a plurality of input candidate data using a final hypothesis; More it is made to execute the input data designation processing for selecting input data from the input candidate data distribution based on the value of functions with the predicted value of the hypothesis. Another active learning program according to the present invention is a boosting technique for improving the accuracy of a learning process by boosting a computer to improve the accuracy of a passive learning algorithm by repeating data resampling. The process is being executed further.
本発明による能動学習法のプログラムの一構成例は、 コンピュータに、 学習処 理及び関数予測処理において、 ランダムに選ばれた複数の入力候補データに対し て下位アルゴリズムがそれぞれ学習を行う処理を実行させている。 また、 このプ ログラムの別の構成例は、 コンピュータに、 入力データ指定処理において、 複数 の仮説の予測値の分散が最大であるデータを入力候補データとして選択する処理 を実行させている。 One example of the configuration of the program of the active learning method according to the present invention causes a computer to execute a process in which a lower-level algorithm performs learning on a plurality of randomly selected input candidate data in a learning process and a function prediction process. ing. Another configuration example of the program causes a computer to execute a process of selecting, as input candidate data, data in which a variance of predicted values of a plurality of hypotheses is maximum in an input data designation process.
すなわち、 本発明の能動学習システムは、 上述した課題を解決するために、 連 続入出力データ用の能動学習を用いた実験計画法を提案している。 つまり、 本発 明の能動学習システムでは、 受動的な学習アルゴリズムの精度を、 データのリサ ンプリングを繰り返すことによって向上させる手法である 「ブースティング」 技 術を、 能動学習の有力手法の一つである集団質問学習と適切に組合せることで、 精度の高い能動学習を実現し、 実験計画の立案に適用する。 That is, the active learning system of the present invention proposes an experiment design method using active learning for continuous input / output data in order to solve the above-mentioned problems. In other words, in the active learning system of the present invention, the “boosting” technology, which is a method of improving the accuracy of the passive learning algorithm by repeating data sampling, is one of the leading methods of active learning. By appropriately combining with a certain group question learning, high-accuracy active learning is realized and applied to the planning of experiment.
上記のように、 本発明の能動学習システムでは、 入力と出力との間に正確にま たは近似的に成り立つ関数関係について、 実験によって得られる入出力データを 基に推定を行う実験計画法において、 過去の入出力データから一定の表現形を用 いて関数関係を推定する予め定められた学習アルゴリズムを下位アルゴリズムと して用い、 下位アルゴリズムに当該時点までの入出力データを訓練データとして 学習させる学習段階と、 学習段階の学習精度をブースティング技術によって向上 させるブースティング段階と、 ブースティング段階によって複数の仮説の重み付 き平均として得られる最終仮説を用いて複数の入力候補データに対する関数値の 予測を行う関数値予測段階と、 複数の仮説の予測値の分散を使った関数の値を基 に入力候補データを入力として選択する入力データ指定段階とを行うことを特徴 とする実験計画法が得られる。 As described above, in the active learning system of the present invention, the accuracy between the input and the output is In a design of experiment, which estimates functional relationships that approximately hold based on input / output data obtained from experiments, a predetermined expression that estimates functional relationships from past input / output data using a certain expression A learning step in which the learning algorithm is used as a lower-level algorithm, and the lower-level algorithm learns input / output data up to that point as training data; a boosting step in which the learning accuracy of the learning step is improved by boosting technology; and a boosting step. A function value prediction step of predicting function values for a plurality of input candidate data using a final hypothesis obtained as a weighted average of a plurality of hypotheses, and a function value using a variance of predicted values of a plurality of hypotheses. Input data selection step of selecting input candidate data as input based on the input data. Programming can be obtained.
言い換えると、 本発明の能動学習システムは、 実験によって得られる入出力デ ータを基に入力と出力との間に成り立つ関数関係を推定する実験計画法において、 受動学習によって過去のデータから仮説を推定する仮説推定段階と、 仮説に基づ いて次の入力候補データを指示する入力データ指定段階と、 入力候捕データに対 する関数値を実験によって求める実験段階と、 入力候補データ及び関数値を過去 のデータに追加するデータ更新段階とを含むことを特徴とする実験計画法が得ら れる。 この場合、 上記の各段階は次の実験の入力データを選択する各時点で、 つ まり関数関係の推定が終了するまで、 次の実験の入力データを選択する毎に実行 される。 In other words, in the active learning system of the present invention, in an experimental design method for estimating a functional relationship established between an input and an output based on input / output data obtained by experiments, a hypothesis is obtained from past data by passive learning. Estimating a hypothesis, estimating the next input candidate data based on the hypothesis, specifying input data based on the hypothesis, experimentally obtaining a function value for the input caught data by experiment, and calculating the input candidate data and the function value. An experimental design method is obtained, which includes a data updating step to be added to past data. In this case, the above steps are executed at each time of selecting the input data of the next experiment, that is, each time the input data of the next experiment is selected until the estimation of the functional relationship is completed.
ここで、 仮説推定段階においてはブースティングを用いることが好ましい。 ま た、 入力データ指定段階は過去のデータにない複数の入力候補データを生成する 入力候補データ生成段階と、 仮説に基づいて複数の入力候補データから次の入力 候補データを選択する入力候補デ一タ選択段階とからなる。 Here, it is preferable to use boosting in the hypothesis estimation stage. In the input data designating step, an input candidate data generating step of generating a plurality of input candidate data not present in the past data, and an input candidate data of selecting the next input candidate data from the plurality of input candidate data based on a hypothesis. Data selection stage.
また、 本発明の能動学習システムでは、 入力と出力との間に正確にまたは近似 的に成り立つ関数関係について、 実験によって得られる入出力データを基に推定 を行う実験計画プログラムにおいて、 過去の入出力データから一定の表現形を用 いて関数関係を推定する任意の学習アルゴリズムを下位アルゴリズムとして用い、 下位アルゴリズムに当該時点までの入出力データを訓練データとして学習させる 学習処理と、 学習処理の学習精度をブースティング技術によって向上させるブー スティング処理と、 ブースティング処理によって精度が高められた複数の仮説の 予測値の分散を使った関数の値を基に入力候補データを入力として選択する入力 データ指定処理とをコンピュータに実行させている。 Further, in the active learning system of the present invention, an experimental design program for estimating a functional relationship that is established accurately or approximately between an input and an output based on input / output data obtained by an experiment includes: An arbitrary learning algorithm that estimates a functional relationship from data using a certain expression is used as a lower algorithm, and the lower algorithm learns input / output data up to that point as training data. Boo boosted by boosting technology The computer performs a sting process and an input data designation process that selects input candidate data as input based on the value of a function that uses the variance of the predicted values of multiple hypotheses that have been improved by the boosting process. .
言い換えると、 本発明の能動学習システムでは、 実験によって得られる入出力 データを基に入力と出力との間に成り立つ関数関係を推定する実験計画プログラ ムにおいて、 受動学習によって過去のデータから仮説を推定する仮説推定処理と、 仮説に基づいて次の入力候補データを指示する入力候補データ指示処理と、 入力 候補データに対する関数値を実験によって求める実験処理と、 入力候補データ及 び関数値を過去のデータに追加するデータ更新処理とをコンピュータに実行させ ている。 この場合、 上記の各段階は次の実験の入力点を選択する各時点で、 つま り関数関係の推定が終了するまで、 次の実験の入力を選択する毎に実行される。 ここで、 仮説推定処理はブースティング処理を含むことが好ましい。 また、 入 力候補データ処理は、 過去のデータにない複数の入力候補データを生成する入力 候補データ生成処理と、 仮説に基づいて複数の入力候補データから次の入力候補 データを選択する入力候捕データ選択処理とからなる。 In other words, in the active learning system of the present invention, a hypothesis is estimated from past data by passive learning in an experiment design program for estimating a functional relationship between input and output based on input / output data obtained by experiments. Hypothesis estimation processing, input candidate data instruction processing to indicate the next input candidate data based on the hypothesis, experimental processing to experimentally determine the function value for the input candidate data, and past input data and function values to past data The data update process to be added to the computer is executed by the computer. In this case, each of the above steps is performed at each time of selecting the input point of the next experiment, that is, each time the input of the next experiment is selected until the estimation of the functional relationship is completed. Here, the hypothesis estimation processing preferably includes a boosting processing. The input candidate data processing includes input candidate data generation processing for generating a plurality of input candidate data that are not included in past data, and input candidate data selection for selecting the next input candidate data from the plurality of input candidate data based on a hypothesis. And data selection processing.
本発明にかかる 「実験計画法」 は、 上記の従来の 「実験計画法」 に対し、 従来 の 「実験計画法」 の表現を用いると、 次のようになる。 すなわち、 本発明にかか る 「実験計画法」 は、 因子を入力として特性を出力とするような関数 f を推定す る際に、 結果を知りたい実験の因子の水準を逐一指定することによって、 少ない 実験回数で関数 ίを推定する方法である。 さらに、 表現を一般化すると、 入力 X とそのラベル (出力) yとの関係を表す関数 (仮説) f (y= f (x) ) を効率 良く推定するために行う実験の順序を計画することを指す。 The “experimental design method” according to the present invention is as follows when the expression of the conventional “experimental design method” is used with respect to the above-mentioned conventional “experimental design method”. That is, the "experimental design method" according to the present invention is such that, when estimating a function f having factors as inputs and characteristics as outputs, the level of the factor of the experiment for which the result is desired is to be specified one by one. This is a method to estimate the function で with a small number of experiments. Furthermore, generalizing the expression, the order of experiments to be performed to efficiently estimate the function (hypothesis) f (y = f (x)) representing the relationship between the input X and its label (output) y Point to.
本発明では上記の実験計画法を能動学習によって実現する。 能動学習とは、 例 えば、 安倍と中村による解説 「能動学習概要」 ( 「情報処理」 38卷 7号、 55 8— 56 1頁、 1 997年発行) 等に記載されているように、 学習者が環境から 選択的に情報を獲得し、 学習する方法を指す。 In the present invention, the above-described experiment design method is realized by active learning. Active learning refers to learning as described in, for example, Abe and Nakamura's commentary “Active Learning Overview” (“Information Processing”, Vol. 38, No. 55, pp. 58-56, 1997). Is a method by which a person selectively obtains information from the environment and learns.
能動学習の手法としては、 多くの方法が提案されている。 例えば、 チャロナー (C h a 1 o n e r ) とベルデイネリイ (Ve r d i n e l l i ) との論文 「ベ イジアンェクスペリメンタノレデザイン : アレビユー」 (B a y e s i a n e x p e r i me n t a 1 d e s i g n : a R e v i ew). ( 「スタティスティカ /レサイエンス」 (S t a t i s t i c a l S c i e n c e) , 第 10卷, 27 3— 304頁, 1995年発行) では、 べィズ決定理論に基づく能動学習手法が 提案されている。 Many methods have been proposed for active learning. For example, the paper "Baijian esperimentanore design: Alebiyu" by Charoner (Cha 1 oner) and Verdenerelli (Vayesianex) peri me nta 1 design: a Revi ew). ("Statistical Science", Vol. 10, pp. 273-304, published in 1995) An active learning method based on it has been proposed.
また、 ノ ゥム (B a um) の論文 「ニューラルネットアルゴリズムズザットラ ーンインポリノミアルタイムフロムィグザンプルズアンドクエリーズ」 (Ne u r a l n e t a l g o r i t hms t h a t l e a r n i n p o 1 y n o m i a 1 t i me f r om e x amp l e s a n d q u e r i e s ) ( 「アイトリプルィートランザクションズオン-ユーラルネットワークス」 (I EEE Tr a n s a c t i o n s o n Ne u r a l Ne t wo r k s) , 第 2卷, 5— 1 9頁, 1 991年発行) では、 過去の学習データから境界 点と推定される点を新たな入力候補データとする学習方法が提案されている。 これらの手法と共に、 データ効率性を実現するための 「集団質問学習」 と呼ば れる学習方法が提案されており、 有力な能動学習手法の一^ 3とされている。 代表 的な集団質問学習手法としては、 セング (S e un g) らの論文 「クエリーパイ コミッティ」 (Qu e r y b y C o mm i t t e e) ( 「プロシーディング スォブザフィフスァニュアルエ一シーエムワークショップオンコンビュテーショ ナノレラーニングセォリー (P r o c e e d i n g s o f t h e f i f t h a nn u a 1 ACM wo r k s h o p o n c omp u t a t i o n a 1 l e a r n i n g t h o e r y) 」 の予稿集, 287— 294頁, 199 2年発行) に記載された方法がある。 この方法ではランダム化された複数の学習 者に予測をさせ、 その不一致度が最大の点を新たな入力候補データとし、 学習を 行っている。 Also, a paper by Baum, “Neuralnetalgorithms thatlearninpo 1 ynomia 1 time fr om ex amp lesandqueries” (“I EEE Transactions on Neural Networks”, Vol. 2, pp. 5-19, published in 991) is based on past learning data. A learning method has been proposed in which a point estimated as is used as new input candidate data. Along with these methods, a learning method called "group query learning" for realizing data efficiency has been proposed, and is considered as one of the leading active learning methods. A typical group question learning method is the query by the SENG et al., “Query pie committee” (“Proceeding sub-sub-fifth annual EM workshop on-committee”). The method is described in the Proceedings of the Shonan Reeling Series (Proceedings of fiftha nn ua 1 ACM wo rkshoponc omp utationa 1 learningthoery), pp. 287-294, 1992. In this method, a plurality of randomized learners make predictions, and learning is performed with the point having the highest degree of mismatch as new input candidate data.
一方、 能動学習をしない学習は受動学習と呼ばれる。'受動学習において、 学習 者は環境から情報を選択的に得ることはなく、 所与の大量のデータからデータに 内在する規則 (仮説) を推定する。 受動学習において、 近年、 与えられたデータ からそれまでの仮説が不得意とするデータ領域を重点的に学習させるようにリサ ンプリングを繰り返すことによって学習精度を向上させる 「ブースティング」 と 呼ばれる技術が注目されており、 実験的にも高い性能を有することが確かめられ てきている。 On the other hand, learning without active learning is called passive learning. 'In passive learning, the learner does not selectively obtain information from the environment, but estimates the rules (hypotheses) inherent in the data from a given large amount of data. In passive learning, in recent years, a technique called `` boosting '', which improves learning accuracy by repeating resampling so that data areas for which the hypothesis up to that point is not good at given data is trained, has been attracting attention in recent years. Has been confirmed to have high performance experimentally. Is coming.
「ブースティング」 の手法については、 フロインド (F r e un d) とシャピ レ (S h a p i r e) との論文 「ァデシジョンセォレティックジエネラライゼ一 ションォプオンラインラーニングアンドアンアプリケーショントウーブースティ ンク」 (A d e c i s i o n— t h e o r e t i c g e n e r a l i z a t i o n o f o n— l i n e l e a r n i n g a n a a n a p p i i c a t i o n t o b o o s t i n g) ( 「プロシーディングスォブザセカンド ョ一口ピアンコンファレンスオンコンビュテーショナノレラーニングセオリー (P r o c e e d i n g s o f t h e s e c o n d e u r o p e a n c o n f e r e n c e o n c omp u t a t i o n a l 1 e a r n i n g t h o e r y) 」 の予稿集, 23— 37頁, 1 995年発行) に記載されている。 さらに、 フロインド (F r e un d) とシャピレ (Sh a p i r e) との論文 「ェタスペリメンッウイズァニューブースティングアルゴリズム」 (E X p e r i m e n t s w i t h a n ew b o o s t i n g a l g o r ι t h m) ( 「プロシーディングスォブザサーティーンスインタ一ナショナルコンファ レンスオンマシンラーニング (P r o c e e d i n g s o f t h e t h i r t e e n t h i n t e r n a t i o n a l c o n i e r e n c e o n ma c h i n e l e a r n i n g) 」 の予稿集, 148— 156頁, 1 996 年発行) においては、 受動学習における 「ブースティング」 技術が、 そこそこの 学習性能を持つ学習アルゴリズムを使用し、 任意の学習精度を持つ学習アルゴリ ズムを構成するという、 優位性を実験的に示している。 The method of "boosting" is described in a paper by Freund and Shapire, "Advice Decision Theoretic Generalization-Online Learning and Application Application Boostink" (A decision—theoreticgeneralizationofon—linelearninganaanappi icationtoboosting) (Proceedings of the Proceedings of the second europeanconferenceoncomp utational 1 earningthoery, pp. 23, 23, 9-95) It is described in. In addition, a paper by Freund and Shapire, "EX perimentswithan ew boostingalgor ltm" ("Proceedings of the Third International") Proceedings of Conference on Machine Learning (Proceedingsofthethirte enthinternationalconierenceon machinelearning), pp. 148-156, published in 1996), the “boosting” technique in passive learning is a technique for learning algorithms with reasonable learning performance. It demonstrates experimentally the superiority of using it to construct a learning algorithm with arbitrary learning accuracy.
しかしながら、 ブースティング技術は、 受動学習でのみ使用されており、 能動 学習にはまだ適用されていない。 すなわち、 能動学習においては、 リサンプリン グを繰り返すといった 「ブースティング」 技術特有の考え方はこれまで無く、 こ の技術を能動学習の枠組で使用する方法は提案されていない。 However, boosting techniques have only been used in passive learning and have not yet been applied to active learning. In other words, in active learning, there has never been a concept unique to “boosting” technology, such as repeating resampling, and no method of using this technology in the framework of active learning has been proposed.
また、 能動学習としては、 特開平 1 1— 316754号公報に開示された技術 が既に知られている。 ところが、 この能動学習法では、 その特長として、 重みの 総和が最大の出力値の重み和と、 重みの総和とが次に大きい出力値の重みの総和 との差が最も小さい点を入力候補点とし、 それらの中から入力として選択するこ とをあげているが、 この手法は出力値がある特定の離散的な数値だけを対象とし ている。 本発明は連続的な値の場合の課題の解決手法に、 分散の関数を使ってい る点が最大の特徴で、 特開平 1 1一 3 1 6 7 5 4号公報に開示された技術とは異 なる。 図面の簡単な説明 As active learning, a technique disclosed in Japanese Patent Application Laid-Open No. 11-316754 is already known. However, the feature of this active learning method is that the point at which the difference between the sum of weights of the output values with the largest sum of the weights and the sum of the weights of the output values with the next largest sum is the smallest is the input candidate point. And select them as input from them. However, this method targets only certain discrete numerical values with output values. The most significant feature of the present invention is that a variance function is used as a solution to the problem in the case of continuous values, and the technique disclosed in Japanese Patent Application Laid-Open No. different. BRIEF DESCRIPTION OF THE FIGURES
図 1は、 本発明の実施例による能動学習のアルゴリズムの流れを示すための図 である。 FIG. 1 is a diagram showing a flow of an active learning algorithm according to an embodiment of the present invention.
図 2は、 本発明の実施例による能動学習機械の構成を示すブロック図である。 図 3は、 本発明の第 1の実施例による能動学習システムの構成を示すブロック 図である。 FIG. 2 is a block diagram showing a configuration of the active learning machine according to the embodiment of the present invention. FIG. 3 is a block diagram showing the configuration of the active learning system according to the first embodiment of the present invention.
図 4は、 本発明の第 2の実施例による能動学習システムの構成を示すブロック 図である。 FIG. 4 is a block diagram showing a configuration of an active learning system according to a second embodiment of the present invention.
図 5は、 本発明の第 3の実施例による能動学習システムの構成を示すプロック 図である。 FIG. 5 is a block diagram showing the configuration of the active learning system according to the third embodiment of the present invention.
図 6は、 本発明の第 4の実施例による能動学習システムの構成を示すブロック 図である。 FIG. 6 is a block diagram showing a configuration of an active learning system according to a fourth embodiment of the present invention.
図 7は、 本発明の第 5の実施例による能動学習システムの構成を示すプロック 図である。 FIG. 7 is a block diagram showing the configuration of the active learning system according to the fifth embodiment of the present invention.
図 8は、 本発明の第 6の実施例による能動学習システムの構成を示すブロック 図である。 FIG. 8 is a block diagram showing a configuration of an active learning system according to a sixth embodiment of the present invention.
図 9は、 本発明の第 7の実施例による能動学習システムの構成を示すプロック 図である。 FIG. 9 is a block diagram showing the configuration of the active learning system according to the seventh embodiment of the present invention.
図 1 0は、 本発明の第 8の実施例による能動学習機械の構成を示すブロック図 である。 実施例の詳細な説明 FIG. 10 is a block diagram showing a configuration of an active learning machine according to an eighth embodiment of the present invention. Detailed description of the embodiment
次に、 本発明の実施例について図面を参照して説明する。 図 1は本発明の実施 例による能動学習のアルゴリズムの流れを示すための図であり、 図 2は本発明の 実施例による能動学習機械の構成を示すブロック図である。 Next, embodiments of the present invention will be described with reference to the drawings. FIG. 1 is a diagram showing a flow of an active learning algorithm according to an embodiment of the present invention, and FIG. It is a block diagram showing composition of an active learning machine by an example.
図 1において、 ブースティング装置 1は学習予測誤差計算手段 1 1と、 データ 分布更新手段 1 2と、 判定手段 1 3とを備え、 関数計算手段 2 1と、 入力候補デ ータ選択手段 2 2と、 関数値入力手段 2 3と、 データ更新手段 2 4と、 判定手段 2 5と、 実験シミュレーション 3 0とを含む装置と、 入力手段 1 0 0とに接続さ れている。 ここで、 実験シミュレーション 3 0は、 必ずしも装置に含まなくとも よく、 外部シミュレータゃ人手による実験 'シミュレーション計算などで代替え してもよい。 尚、 上記の各手段をステップに置き換えることで、 プログラムによ つて実現することも可能である。 In FIG. 1, the boosting device 1 includes a learning prediction error calculation unit 11, a data distribution updating unit 12, and a judgment unit 13, a function calculation unit 21, and input candidate data selection unit 22. , A function value input means 23, a data update means 24, a determination means 25, an apparatus including an experiment simulation 30, and an input means 100. Here, the experiment simulation 30 need not necessarily be included in the apparatus, and may be replaced with an external simulator / manual experiment / simulation calculation. It should be noted that it is also possible to realize by a program by replacing each of the above means with a step.
図 2において、 能動学習機械 4は前処理装置 4 1と、 下位学習機械 4 2 1〜4 2 Nからなる下位学習機械装置 4 2と、 仮説 '予測値記憶領域 4 3 1〜4 3 Nを 持つ仮説 ·予測値記憶装置 4 3と、 後処理装置 4 4と、 上記の各装置をコンビュ 一タで実現する際にそのコンピュータで実行させるプログラムを記録する記録媒 体 4 5とから構成されている。 In FIG. 2, the active learning machine 4 includes a preprocessing device 41, a lower learning machine device 42 composed of lower learning machines 4 2 1 to 4 2 N, and a hypothesis' predicted value storage area 4 3 1 to 4 3 N. It is composed of a prediction value storage device 43, a post-processing device 44, and a recording medium 45 for recording a program to be executed by a computer when each of the above devices is realized by a computer. I have.
この能動学習機械 4は、 コンピュータと、 このコンピュータを能動学習機械 4 として機能させるソフトウェアとからなる。 この場合、 記録媒体 4 5に記録され たプログラムがコンピュータに読み込まれることにより、 このプログラムとコン ピュータのハードウェア資源とが協働し、 コンピュータに前処理装置 4 1と下位 学習機械装置 4 2と仮説 ·予測値記憶装置 4 3と後処理装置 4 4とを実現させる。 これら図 1及び図 2を用いて本発明の実施例による実験計画法について説明す る。 本実施例では、 まず、 入力手段 1 0 0において、 初期データ、 そのデータの 初期分布を入力する。 The active learning machine 4 includes a computer and software that causes the computer to function as the active learning machine 4. In this case, when the program recorded on the recording medium 45 is read into the computer, this program and the hardware resources of the computer cooperate with each other, and the preprocessing device 41 and the lower learning machine device 42 are connected to the computer. Hypothesis · The predicted value storage device 43 and the post-processing device 44 are realized. The experiment design method according to the embodiment of the present invention will be described with reference to FIGS. In this embodiment, first, initial data and initial distribution of the data are input to the input means 100.
次に、 図 1に示すように、 学習予測誤差計算手段 1 1から判定手段 2 5の間で 能動学習における質問の繰り返しに入る。 繰り返しの最初の部分は、 ブースティ ング装置 1の処理で実行される。 この処理は能動学習機械 4において前処理装置 4 1と下位学習機械 4 2とで行われる。 Next, as shown in FIG. 1, the questions in the active learning are repeated between the learning prediction error calculating means 11 and the judging means 25. The first part of the iteration is performed in the process of the boosting device 1. This processing is performed in the active learning machine 4 by the preprocessing device 41 and the lower learning machine 42.
ブースティング装置 1において、 学習予測誤差計算手段 1 1とデータ分布更新 手段 1 2と判定手段 1 3とによってデータのリサンプリングが繰り返される。 まず、 ブースティング装置 1では学習予測誤差計算手段 1 1において、 下位学 習アルゴリズムを使用して現在のデータからの学習が行われ、 学習結果として仮 説を得る。 下位学習アルゴリズムは回帰木等、 連続値を扱う学習アルゴリズムな らどんなものでもよい。 その結果は仮説 ·予測値記憶装置 4 3に出力される。 さらに、 学習予測誤差計算手段 1 1において、 仮説から現在のデータへの予測 値を計算し、 予測値から誤差を計算する。 次に、 データ分布更新手段 1 2におい て、 学習予測誤差計算手段手段 1 1で得られた誤差から、 データのリサンプリン グ分布を更新する。 ' In the boosting device 1, resampling of data is repeated by the learning prediction error calculating means 11, the data distribution updating means 12, and the judging means 13. First, in the boosting device 1, the learning prediction error calculation means 11 Learning from the current data is performed using a learning algorithm, and a hypothesis is obtained as the learning result. The lower learning algorithm may be any learning algorithm that handles continuous values, such as a regression tree. The result is output to the hypothesis / predicted value storage device 43. Further, the learning prediction error calculation means 11 calculates a prediction value from the hypothesis to the current data, and calculates an error from the prediction value. Next, the data distribution updating means 12 updates the data resampling distribution from the error obtained by the learning prediction error calculating means 11. '
最後に、 判定手段 1 3において、 リサンプリング回数が一定値 ( T) に達して いれば、 ブースティング装置 1による処理を終了し、 そうでなければ学習予測誤 差計算手段 1 1に戻り、 上述した処理を繰り返しを行う。 Finally, if the number of times of resampling has reached the fixed value (T) in the judging means 13, the processing by the boosting device 1 ends, otherwise, the process returns to the learning prediction error calculating means 11, and The above processing is repeated.
また、 ブースティング装置 1の処理の反復回数は 1回であってもよい。 その時、 データ分布更新手段 1 2のデータ分布の更新と判定手段 1 3の条件判断文、 判定 手段 1 3から学習予測誤差計算手段 1 1に戻るループは必要なくなる。 つまり、 ブースティング処理なしの下位学習処理となる。 Further, the number of repetitions of the processing of the boosting device 1 may be one. At that time, the data distribution updating means 12 updates the data distribution and the condition judgment statement of the judging means 13 and the loop returning from the judging means 13 to the learning prediction error calculating means 11 becomes unnecessary. In other words, lower learning processing without boosting processing is performed.
ブースティング装置 1による処理終了後、 後処理装置 4 4による処理を行う。 この場合、 後処理装置 4 4は、 関数計算手段 2 1と、 入力候補データ選択手段 2 2と、 関数値入力手段 2 3と、 データ更新手段 2 4と、 判定手段 2 5と、 実験シ ミュレーシヨン 3 0とを有している。 関数計算手段 2 1では学習された複数の仮 説の予測値の分散を使った関数の値と複数の仮説を使って最終仮説を計算し、 最 終仮説を出力する。 入力候補データ選択手段 2 2では関数計算手段 2 1による計 算結果から得られた分散を使った関数の値を基に入力候補データの選択を行い、 この入力候補データを実験者による実験 (実験シミュレーション 3 0 ) に対する 新しい実験として指示する。 After the processing by the boosting device 1 is completed, the processing by the post-processing device 4 is performed. In this case, the post-processing device 44 includes function calculation means 21, input candidate data selection means 22, function value input means 23, data update means 24, determination means 25, and an experimental simulation. 30. The function calculation means 21 calculates the final hypothesis using the function value using the variance of the predicted values of the plurality of learned hypotheses and the plurality of hypotheses, and outputs the final hypothesis. The input candidate data selection means 22 selects input candidate data based on the value of the function using the variance obtained from the calculation result by the function calculation means 21. Indicate as a new experiment for simulation 30).
関数値入力手段 2 3では実験者による実験 (実験シミュレーション 3 0 ) によ つて試された入力候補データの関数値を受け取る。 データ更新手段 2 4では得ら れた入力候補データとその関数値とを過去のデータに加える。 最後に、 判定手段 2 5において、 質問回数が一定値 (N) に達しているか否かを判定し、 一定値に 達していないときは、 ブースティング装置 1による処理に戻り能動学習を続ける。 質問回数が一定値に達していれば、 能動学習を終了する。 次に、 図 2で示した各装置と図 1で示した主要な各手段との対応関係について 説明する。 この実施例では、 下位学習機械装置 4 2と仮説 ·予測値記憶装置 4 3 とが組み合わされ、 学習予測誤差計算手段 1 1として動作することにより、 下位 アルゴリズムに次入力データを選択する時点までの入出力データを訓練データと して学習させる学習手段および複数の入力候補データに対する関数値の予測を行 う関数予測手段として作用する。 The function value input means 23 receives the function value of the input candidate data tested in the experiment (experimental simulation 30) by the experimenter. The data updating means 24 adds the obtained input candidate data and its function value to the past data. Finally, the determining means 25 determines whether or not the number of questions has reached a certain value (N). If not, the process returns to the boosting device 1 to continue active learning. If the number of questions reaches a certain value, the active learning ends. Next, the correspondence between each device shown in FIG. 2 and each main means shown in FIG. 1 will be described. In this embodiment, the lower-order learning machine device 42 and the hypothesis / prediction value storage device 4 3 are combined and operated as the learning prediction error calculation means 11, so that the time until the next input data is selected for the lower-order algorithm is obtained. It functions as learning means for learning input / output data as training data and function prediction means for predicting function values for a plurality of input candidate data.
さらに、 下位学習機械装置 4 2は、 データ分布更新手段 1 2および判定手段 1 3としても動作し、 学習手段の学習精度を向上させるブースティング手段として 作用する。 Further, the lower learning machine device 42 also operates as the data distribution updating device 12 and the judging device 13 and acts as a boosting device for improving the learning accuracy of the learning device.
また、 後処理装置 4 4が関数計算手段 2 1および入力候補データ選択手段 2 2 として動作することにより、 学習された複数の仮説の予測値の分散を使った関数 の値を基に入力候補データの中から入力データを選択する入力データ指定手段と して作用する。 さらに、 前処理装置 4 1と後処理装置 4 4とが連携して、 関数値 入力手段 2 3、 データ更新手段 2 4および実験シミュレーシヨン 3 0として動作 することにより、 入力候補データ及び関数値を過去のデータに追加するデータ更 新手段として作用する。 これにより、 人手を介することなく実験結果データべ一 ス 6の更新と、 能動学習機械 4への入力が可能となる。 このようにすることで、 ェつの学習システムと見なせ、 利用者の手間を大幅に減らすことができる。 以上、 本発明にかかる能動学習機械 4の実施例について説明したが、 本発明は これに限定されるものではなく、 当業者の通常の知識の範囲内でその変更や改良 が可能であることは明らかである。 In addition, the post-processing device 44 operates as the function calculating means 21 and the input candidate data selecting means 22, so that the input candidate data based on the variance of the predicted values of the plurality of learned hypotheses is used. Acts as input data designating means for selecting input data from Further, the pre-processing device 41 and the post-processing device 44 cooperate with each other to operate as the function value input means 23, the data updating means 24, and the experiment simulation 30, so that the input candidate data and the function value can be obtained. Acts as a data updating means to add to past data. As a result, it becomes possible to update the experimental result database 6 and input to the active learning machine 4 without manual intervention. In this way, it can be regarded as another learning system, and the user's labor can be greatly reduced. Although the embodiment of the active learning machine 4 according to the present invention has been described above, the present invention is not limited to this, and it is possible to make changes and improvements within the ordinary knowledge of those skilled in the art. it is obvious.
例えば、 学習手段及び関数予測手段として作用する学習予測誤差計算手段 1 1 において、 ランダムに選ばれた複数の入力候補データに対して下位アルゴリズム がそれぞれ学習を行うようにしてもよい。 ランダムに選ぶことによって下位学習 アルゴリズムはお互いに独立なものとなり、 並列化が容易になる。 さらに、 乱数 を発生させるのみであるため、 計算時間がほかの方法に比べ短くなる。 For example, in the learning prediction error calculation means 11 acting as the learning means and the function prediction means, the lower algorithm may perform learning on a plurality of input candidate data selected at random. By choosing at random, the lower learning algorithms become independent of each other and parallelization becomes easier. Furthermore, since only random numbers are generated, the computation time is shorter than other methods.
また、 入力データ指定手段の一部として作用する入力候補データ選択手段 2 2 において、 複数の仮説の予測値の分散が最大であるデータを入力候補データとし て選択するようにしてもよい。 分散が最大であるデータは、 得られる情報量が最 大であるため、 入力候補データとして選択することにより、 学習の効果を最大限 に高めることができる。 Further, the input candidate data selecting means 22 acting as a part of the input data designating means may select, as the input candidate data, data in which the variance of the predicted values of a plurality of hypotheses is maximum. Data with the largest variance has the least amount of information available. Since it is large, the effect of learning can be maximized by selecting it as input candidate data.
さらに、 学習手段として作用する学習予測誤差計算手段 1 1において、 下位ァ ルゴリズムの入力データとしてそれぞれ任意の規則性にしたがった結論及びシミ ユレーシヨン計算によって求めた結果のいずれかを用いるようにしてもよレ、。 こ のようなシステムにすることによって、 実験結果若しくはシミュレーション結果 をもとに、 計算若しくは変換された値を扱うことが可能になり、 さらに広範な範 囲のデータを能動学習することが可能になる。 Further, in the learning prediction error calculating means 11 acting as a learning means, any one of a conclusion according to an arbitrary regularity and a result obtained by simulation calculation may be used as input data of the lower algorithm. Les ,. By adopting such a system, it is possible to handle calculated or converted values based on experimental or simulation results, and to actively learn a wider range of data. .
図 3は本発明の第 1の実施例による能動学習システムの構成を示すプロック図 である。 図 3において、 能動学習システム 5は能動学習機械 4と、 過去の入出力 を記憶する記憶装置 [実験結果データベース (属性値と結果) ] 6とを備えてい る。 FIG. 3 is a block diagram showing the configuration of the active learning system according to the first embodiment of the present invention. In FIG. 3, the active learning system 5 includes an active learning machine 4 and a storage device [experiment result database (attribute values and results)] 6 for storing past inputs and outputs.
ここで、 能動学習機械 4は上述した本発明にかかる能動学習機械 4の実施例と 同様の動作を行う。 尚、 能動学習システム 5は図 2に示す能動学習機械 4と入出 力が同じであるが、 外部に記憶装置を必要としないシステムを構築することがで きる。 Here, the active learning machine 4 performs the same operation as the above-described embodiment of the active learning machine 4 according to the present invention. Although the active learning system 5 has the same input and output as the active learning machine 4 shown in FIG. 2, a system that does not require an external storage device can be constructed.
図 4は本発明の第 2の実施例による能動学習システムの構成を示すブロック図 である。 図 4において、 本発明の第 2の実施例による能動学習システムは能動学 習機械 4を組入れたシステムの一例である。 FIG. 4 is a block diagram showing a configuration of an active learning system according to a second embodiment of the present invention. In FIG. 4, the active learning system according to the second embodiment of the present invention is an example of a system incorporating an active learning machine 4.
能動学習機械 4のシステムによって出力された次の入力候補を、 人手による実 験 · シミュレーション計算 7で実験者が実際に実験を行い、 その入力候補と実験 結果との組を記憶装置 6に記憶する。 また、 能動学習機械 4はそれを訓練データ として学習を行い、 次の入力候補を出力する。 The next input candidate output by the system of the active learning machine 4 is manually experimented by the experimenter and simulation calculation 7 by the experimenter.The experimenter actually performs the experiment, and the set of the input candidate and the experiment result is stored in the storage device 6. . In addition, the active learning machine 4 performs learning by using it as training data, and outputs the next input candidate.
上記のような操作を繰返し行うことで、 実験者は能動学習機械 4の指示に従え ば、 無数にある実験対象の中から効率よく実験対象を選ぶことができるようにな り、 目的に添った実験を効率よく行うことができるようになる。 By repeating the above operations, the experimenter can efficiently select an infinite number of experimental objects from the innumerable experimental objects according to the instructions of the active learning machine 4. Experiments can be performed efficiently.
図 5は本発明の第 3の実施例による能勲学習システムの構成を示すプロック図 である。 図 5において、 能動学習システム 5は能動学習機械 4に記憶装置 6を組 み合わせたシステムである。 本実施例では本発明の第 2の実施例において必要であった記憶装置 6が能動学 習システム 5の外部に必要なくなり、 場所の低減化やデータベース ·能動学習機 械間のデータの転送時間を減らすことができる。 FIG. 5 is a block diagram showing a configuration of the Noh learning system according to the third embodiment of the present invention. In FIG. 5, an active learning system 5 is a system in which a storage device 6 is combined with an active learning machine 4. In this embodiment, the storage device 6, which was required in the second embodiment of the present invention, is no longer required outside the active learning system 5, so that the space can be reduced and the data transfer time between the database and the active learning machine can be reduced. Can be reduced.
図 6は本発明の第 4の実施例による能動学習システムの構成を示すプロック図 である。 図 6において、 本発明の第 4の実施例による能動学習システムは能動学 習機械 4を,組入れたシステムの一例である。 能動学習機械 4のシステムによって 出力された次の入力候補に対応したシミュレーシヨンをシミュレータ 8が自動的 に行い、 その入力候補とシミユレーション結果との組を記憶装置 6に記憶する。 さらに、 能動学習機械 4はそれを訓練データとして学習を行い、 次の入力候補 を出力する。 このような操作を繰返し行うことで、 無数にある実験対象の中から 効率よく実験対象を選ぶことができるようになる上、 利用者は入力候補データの 属性とその結果値との組合せ方だけを与えておけば、 目的に添った実験を自動的 に行うことができるようになる。 これにより、 実験者が能動学習エンジンの選ん だ入力候補データを何らか加工する必要がなくなるので、 実験者の負担を減らす ことができる。 伹し、 この能動学習システムではユーザがシミュレータ 8等に必 要なシステム (例えば、 コンピュータやシミュレーションを行うための装置) を 準備しておく必要がある。 FIG. 6 is a block diagram showing a configuration of an active learning system according to a fourth embodiment of the present invention. In FIG. 6, the active learning system according to the fourth embodiment of the present invention is an example of a system incorporating an active learning machine 4. The simulator 8 automatically performs a simulation corresponding to the next input candidate output by the system of the active learning machine 4, and stores a set of the input candidate and the simulation result in the storage device 6. Further, the active learning machine 4 performs learning using the training data as training data, and outputs the next input candidate. By repeating such an operation, the experiment target can be efficiently selected from the myriad of experiment targets, and the user can select only the combination of the attribute of the input candidate data and the result value. If you do, you will be able to automatically carry out experiments that meet your purpose. This eliminates the need for the experimenter to process the input candidate data selected by the active learning engine, thus reducing the burden on the experimenter. However, in this active learning system, it is necessary for the user to prepare a system (for example, a computer or a device for performing a simulation) necessary for the simulator 8 or the like.
図 7は本発明の第 5の実施例による能動学習システムの構成を示すプロック図 である。 図 7において、 能動学習システム 5は能動学習機械 4に記憶装置 6を組 込んで構成されている。 FIG. 7 is a block diagram showing a configuration of an active learning system according to a fifth embodiment of the present invention. In FIG. 7, an active learning system 5 is configured by incorporating a storage device 6 into an active learning machine 4.
この能動学習システム 5を用いることによって、 上述した本発明の第 4の実施 例において必要であった記憶装置 6がシステムの外部に必要なくなり、 場所の低 減化や記憶装置 6と能動学習機械 4との間のデータ転送時間を減らすことができ る。 By using this active learning system 5, the storage device 6, which was required in the above-described fourth embodiment of the present invention, is not required outside the system. The data transfer time between the two can be reduced.
図 8は本発明の第 6の実施例による能動学習システムの構成を示すプロック図 である。 図 8において、 能動学習システム 9は能動学習機械 4にシミュレータ 8 を組込んで構成されている。 このようなシステムにすることによって、 シミュレ —ションを含んだ能動学習システムを構築することができ、 人手を介在すること なく、 最適な予測シミュレーションを自動的に行うことが可能になる。 これによ り、 人は初期条件さえ入力すれば所望の結果を手間をかけることなく自動的に得 ることができるようになるので、 利用者の負担を減らすことができる。 FIG. 8 is a block diagram showing a configuration of an active learning system according to a sixth embodiment of the present invention. In FIG. 8, an active learning system 9 is configured by incorporating a simulator 8 into an active learning machine 4. By adopting such a system, an active learning system including simulation can be constructed, and optimal prediction simulation can be automatically performed without human intervention. This In other words, the user can obtain the desired result automatically without any trouble if only the initial conditions are input, so that the burden on the user can be reduced.
この能動学習システム 9ではシステムの目的が限定されてしまうものの、 シス テム利用者は記憶装置 6だけを準備しておけばよく、 上述した本発明の第 5の実 施例に比べてシミュレータ 8等の外部システムを準備する手間を減らすことがで きる。 Although the purpose of the active learning system 9 is limited, the system user only needs to prepare the storage device 6, and the simulator 8 and the like are different from those of the above-described fifth embodiment of the present invention. It is possible to reduce the trouble of preparing an external system.
図 9は本発明の第 7の実施例による能動学習システムの構成を示すブロック図 である。 図 9において、 能動学習システム 1 0は能動学習機械 4にシミュレータ FIG. 9 is a block diagram showing a configuration of an active learning system according to a seventh embodiment of the present invention. In FIG. 9, the active learning system 10
8と記憶装置 6とを組込んで構成されている。 このようなシステムにすることに よって、 利用者の負担は上述した本発明の第 1〜第 6の実施例に比べて激減し、 人は入力候補データの属性とその結果値との組合せ方だけを与えておけば、 目的 に添った実験を自動的に行うことができるようになる。 8 and a storage device 6. By adopting such a system, the burden on the user is drastically reduced as compared with the above-described first to sixth embodiments of the present invention, and the person is required only to combine the attribute of the input candidate data with the result value. If you give, you will be able to automatically perform the experiment that meets your purpose.
図 1 0は本発明の第 8の実施例による能動学習機械の構成を示すプロック図で ある。 図 1 0において、 本発明の第 8の実施例による能動学習機械 4 0は前処理 装置 4 1と下位学習機械装置 4 2との間に下位学習機械装置 4 2の入力データ生 成機械装置 4 6を設けた以外は図 2に示す能動学習機械 4と同様の構成となって おり、 同一構成要素には同一符号を付してある。 FIG. 10 is a block diagram showing a configuration of an active learning machine according to an eighth embodiment of the present invention. In FIG. 10, an active learning machine 40 according to an eighth embodiment of the present invention includes an input data generation machine 4 of a lower learning machine 42 between a preprocessing device 41 and a lower learning machine 42. The configuration is the same as that of the active learning machine 4 shown in FIG. 2 except that 6 is provided, and the same components are denoted by the same reference numerals.
入力データ生成機械装置 4 6は入力データ生成機械 4 6 1〜4 6 Nからなり、 記憶装置 6に蓄えられたデータに加工を加えて下位学習機械装置 4 2に供給して いるので、 下位学習機械装置 4 2が入力データ生成機械装置 4 6から与えられる データで学習することが可能となり、 より幅の広い問題に適用することが可能に なる。 The input data generation machine 46 includes an input data generation machine 46 1 to 46 N. The data stored in the storage device 6 is processed and supplied to the lower learning machine 42, so that the lower learning is performed. The mechanical device 42 can learn from data provided from the input data generating mechanical device 46, and can be applied to a wider range of problems.
この能動学習機械 4 0は、 コンピュータと、 このコンピュータを能動学習機械 4 0として機能させるソフトウェアとからなる。 この場合、 記録媒体 4 5に記録 されたプログラムがコンピュータに読み込まれることにより、 このプログラムと コンピュータのハードウェア資源とが協働し、 コンピュータに前処理装置 4 1と 入力データ生成機械装置 4 6と下位学習機械装置 4 2と仮説 '予測値記憶装置 4 3と後処理装置 4 4とを実現させる。 The active learning machine 40 includes a computer and software that causes the computer to function as the active learning machine 40. In this case, when the program recorded on the recording medium 45 is read into the computer, the program and the hardware resources of the computer cooperate with each other, and the preprocessing device 41 and the input data generation device 46 are connected to the computer. The lower learning machine device 42, the hypothesis 'predicted value storage device 43' and the post-processing device 44 are realized.
このように、 本発明では、 受動学習において、 ほどほどの精度を挙げることが 可能な学習アルゴリズムであれば任意の精度にまで予測精度を向上させることが 可能なブースティング技術を、 能動学習の手法である集団質問学習と適切な方法 とによつて組合せることによって、 精度の高い能動学習手法を実現することがで さる。 Thus, according to the present invention, in passive learning, moderate accuracy can be obtained. Boosting technology that can improve the prediction accuracy to an arbitrary accuracy if it is a possible learning algorithm is combined with group query learning, which is an active learning method, by using an appropriate method to improve accuracy. It is possible to realize a high active learning method.
また、 本発明では、 実験の数が多いあるいは一回の試行の負担が大きいために、 現実的にすべての試行を行うことが不可能な実験に対して、 上述した方法を実験 計画の立案に適用することによって、 効率的でかつ精度の高い実験計画の実現が 可能となる。 以上説明したように本発明の能動学習システムは、 選択した入出力のデータを 基に、 入出力データ間に少なくとも近似的に成り立つ関数関係の推定を行う能動 学習法を実行する能動学習システムにおいて、 過去の入出力データから一定の表 現形を用いて関数関係を推定しかつ予め定められた単一及び複数種類のうちのい ずれかの複数の学習アルゴリズムを下位アルゴリズムとして用いて下位アルゴリ ズムに次入力データを選択する時点までの入出力データを訓練データとして学習 させる学習手段と、 学習手段各々のアルゴリズムから得られた複数の仮説から得 られる最終仮説を用いて複数の入力候補データに対する関数値の予測を行う関数 予測手段と、 学習された複数の仮説の予測値の分散を使った関数の値を基に入力 候捕データの中から入力データを選択する入力データ指定手段とを設けることに よって、 入出力データが連続的な値の時に、 対象となる全試行の遂行が事実上不 可能な実験を行う際に、 少ない実験回数で目的関数の推定を行うことができると いう効果が得られる。 In addition, in the present invention, the above-described method is used to formulate an experiment plan for experiments in which it is impossible to perform all trials realistically because the number of experiments is large or the load of one trial is large. By applying this, it is possible to realize an efficient and accurate experimental design. As described above, the active learning system of the present invention is an active learning system that executes an active learning method for estimating a functional relationship that at least approximately holds between input and output data based on the selected input and output data. A functional relationship is estimated from past input / output data using a certain expression form, and a plurality of predetermined learning algorithms, one of a single type or a plurality of types, are used as lower-level algorithms, and the next lower-level algorithms are used. A learning means for learning input / output data up to the point of selecting input data as training data, and a final hypothesis obtained from a plurality of hypotheses obtained from each algorithm of the learning means, to obtain a function value for a plurality of input candidate data using a final hypothesis. Function to make predictions Predictive means and input based on function values using the variance of predicted values of multiple learned hypotheses By providing input data designating means for selecting input data from among them, when the input / output data is a continuous value, when performing an experiment in which it is virtually impossible to perform all the trials in question, a small number of experiments can be performed. The effect that the objective function can be estimated by the number of times is obtained.
また、 本発明の他の能動学習システムは、 データのリサンプリングを繰り返す ことによって受動的な学習アルゴリズムの精度を向上させるブースティング技術 によって学習手段の学習精度を向上させるブースティング手段をさらに有するの で、 受動学習において、 ほどほどの精度を挙げることが可能な学習アルゴリズム であれば任意の精度にまで予測精度を向上させることが可能なブースティング技 術を、 能動学習の手法である集団質問学習と適切な方法とによつて組合せること によって、 精度の高い能動学習を実現することができるという効果が得られる。 以上のように、 本発明にかかる能動学習システム及びそれに用いる能動学習法 は、 入出力データが連続的な値の時に、 対象となる全試行の遂行が事実上不可能 な実験を行う際に、 少ない実験回数で目的関数の推定を行うシステム及び方法と して有用であり、 特に分子生物学等の分野の実験や設計を効率的に遂行する計画 の立案に適している。 Another active learning system of the present invention further includes a boosting means for improving the learning accuracy of the learning means by a boosting technique for improving the accuracy of a passive learning algorithm by repeating resampling of data. In the case of passive learning, a boosting technique that can improve the prediction accuracy to an arbitrary level as long as it is a learning algorithm that can achieve a moderate level of accuracy is called a group learning method that is an active learning method. By combining these methods with other methods, it is possible to achieve an effect that highly accurate active learning can be realized. As described above, the active learning system according to the present invention and the active learning method used in the active learning system perform an experiment in which it is practically impossible to perform all the trials when the input and output data are continuous values. It is useful as a system and method for estimating an objective function with a small number of experiments, and is particularly suitable for making a plan for efficiently performing experiments and designs in fields such as molecular biology.
Claims
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2002042430A JP2005258480A (en) | 2002-02-20 | 2002-02-20 | Active learning system, active learning method used in the same and program for the same |
| JP2002-42430 | 2002-02-20 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2003071480A1 true WO2003071480A1 (en) | 2003-08-28 |
Family
ID=27750496
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/JP2003/001184 Ceased WO2003071480A1 (en) | 2002-02-20 | 2003-02-05 | Active learning system, active learning method used in the same, program for the same, and recording medium containing the program |
Country Status (2)
| Country | Link |
|---|---|
| JP (1) | JP2005258480A (en) |
| WO (1) | WO2003071480A1 (en) |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2008117343A (en) * | 2006-11-08 | 2008-05-22 | Nec Corp | Learning processor |
| US10387430B2 (en) | 2015-02-26 | 2019-08-20 | International Business Machines Corporation | Geometry-directed active question selection for question answering systems |
| US11741113B2 (en) | 2019-02-13 | 2023-08-29 | Hitachi, Ltd. | Measurement guide device and simulation computing device used therefor |
Families Citing this family (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP5321596B2 (en) * | 2008-10-21 | 2013-10-23 | 日本電気株式会社 | Statistical model learning apparatus, statistical model learning method, and program |
| US11537869B2 (en) * | 2017-02-17 | 2022-12-27 | Twitter, Inc. | Difference metric for machine learning-based processing systems |
-
2002
- 2002-02-20 JP JP2002042430A patent/JP2005258480A/en active Pending
-
2003
- 2003-02-05 WO PCT/JP2003/001184 patent/WO2003071480A1/en not_active Ceased
Non-Patent Citations (2)
| Title |
|---|
| NAOKI ABE, TAKU MAMIZUKA: "Bit Bessatsu.Hakken Kagaku to Data Mining", 5 May 2000, KYORITSU SHUPPAN CO., LTD., article "Nodo gakushu to hakken kagaku", pages: 64 - 72, XP002975372 * |
| PAASS HERHARD: "Yosoku to model sentaku no tame no shitsumon sentaku", JOHO SHORI, vol. 38, no. 7, 15 July 1997 (1997-07-15), pages 562 - 568, XP002975373 * |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2008117343A (en) * | 2006-11-08 | 2008-05-22 | Nec Corp | Learning processor |
| US10387430B2 (en) | 2015-02-26 | 2019-08-20 | International Business Machines Corporation | Geometry-directed active question selection for question answering systems |
| US11741113B2 (en) | 2019-02-13 | 2023-08-29 | Hitachi, Ltd. | Measurement guide device and simulation computing device used therefor |
Also Published As
| Publication number | Publication date |
|---|---|
| JP2005258480A (en) | 2005-09-22 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Hu et al. | Planning goals for exploration | |
| KR102355489B1 (en) | Method for predicting drug-target protein interactions and device thereof | |
| US6327582B1 (en) | Method and system for genetic programming | |
| Li et al. | The max-min high-order dynamic Bayesian network for learning gene regulatory networks with time-delayed regulations | |
| Dong et al. | Autohas: Differentiable hyper-parameter and architecture search | |
| CN115968480A (en) | Production process optimization method and production process optimization system | |
| Shao et al. | Towards hardware accelerated reinforcement learning for application-specific robotic control | |
| Singh et al. | Ask4help: Learning to leverage an expert for embodied tasks | |
| JP2014160457A (en) | Interactive variable selection device, interactive variable selection method and interactive variable selection program | |
| Knowles et al. | Noisy multiobjective optimization on a budget of 250 evaluations | |
| JPH11316754A (en) | Experimental design and recording medium recording experimental design program | |
| WO2003071480A1 (en) | Active learning system, active learning method used in the same, program for the same, and recording medium containing the program | |
| Seiler et al. | Using reinforcement learning for per-instance algorithm configuration on the TSP | |
| Zhang et al. | Stepwise model selection for sequence prediction via deep kernel learning | |
| JPWO2023133462A5 (en) | ||
| KR102559605B1 (en) | Method and apparatus for function optimization | |
| CN119439910A (en) | A flexible job shop scheduling method based on deep reinforcement learning | |
| Tithi et al. | A convolutional neural network for goal recognition | |
| CN117012303A (en) | A DTI prediction method, system, storage medium and device based on reinforcement learning | |
| Lan | Device Placement Optimization with Deep Reinforcement Learning | |
| CN108549979B (en) | Open-source software development team extension method based on precise embedded representation | |
| Saleh et al. | Should models be accurate? | |
| Potolea et al. | Meta-learning framework for prediction strategy evaluation | |
| JPH0713959A (en) | Method for predicting protein stereoscopic structure | |
| Furukawa et al. | Regularization for parameter identification using multi-objective optimization |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AK | Designated states |
Kind code of ref document: A1 Designated state(s): CN JP US |
|
| AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PT SE SI SK TR |
|
| DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
| 122 | Ep: pct application non-entry in european phase | ||
| NENP | Non-entry into the national phase |
Ref country code: JP |