TECHNOLOGY TREND PREDICTION METHOD AND SYSTEM
20230043735 · 2023-02-09
Assignee
Inventors
Cpc classification
G06N7/01
PHYSICS
Y02D10/00
GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
G06Q10/0637
PHYSICS
G06N3/0442
PHYSICS
International classification
Abstract
A technology trend prediction method and system are provided. The method comprises acquiring paper data, and further comprises following steps: processing the paper data to generate a candidate technology lexicon; screening the candidate technology lexicon based on mutual information; calculating an independent word forming probability of an OOV word; extracting missed words in a title using a bidirectional long short-term memory network and a conditional random field (BI-LSTM+CRF) model; predicting a technology trend. The technology trend prediction method and system provided analyzes relationship of technology changes in a high-dimensional space, and predicts a development of technology trend based on time by extracting technical features of papers through natural language processing and time sequence algorithms.
Claims
1. A technology trend prediction method, comprising: acquiring paper data, the technology trend prediction method further comprises following steps: step 1: processing the paper data to generate a candidate technology lexicon; step 2: screening the candidate technology lexicon based on mutual information; step 3: calculating an independent word forming probability of an out-of-vocabulary (OOV) word; step 4: extracting missed words in a title using a bidirectional long short-term memory network and a conditional random field (BI-LSTM+CRF) model; step 5: predicting a technology trend.
2. The technology trend prediction method according to claim 1, wherein the acquiring of the paper data further comprises constructing a set of paper data: wherein step 1 further comprises performing a part-of-speech filtering by using an existing part-of-speech tagging, obtaining a preliminary lexicon after the part-of-speech filtering is completed, and improving an OOV word discovery of the candidate technology lexicon by using a Hidden Markov Model (HMM) method; wherein a formula of the HMM method is:
3. (canceled)
4. (canceled)
5. (canceled)
6. The technology trend prediction method according to claim 2, wherein step 2 further comprises calculating the mutual information of the OOV words, selecting a first threshold, and removing the OOV words with the mutual information lower than the first threshold, and a calculation formula is:
7. (canceled)
8. The technology trend prediction method according to claim 6, wherein step 3 further comprises selecting a second threshold, and removing the OOV words with the independent word forming probability lower than the second suitable threshold, formulas are as follows:
9. The technology trend prediction method according to claim 8, wherein a training method of the BI-LSTM+CRF model comprises following sub-steps: step 41: constructing a labeled corpus according to a technology lexicon, taking words in a title and also in the technology lexicon obtained in step 3 as a training corpus of the BI-LSTM+CRF model, taking the other words in the title as a predicted corpus of the BI-LSTM+CRF model, and labeling the words in the title of the training corpus with B, I, and O, three types of tags, wherein B represents a beginning character of a new word, I represents an internal character of the new word, and O represents a non-technical noun word; step 42: converting the words into word vectors, and encoding the word vectors by using the BI-LSTM model; step 43: mapping an encoded result to a sequence vector with a dimension of a number of the tags through a fully connected layer; and step 44: decoding the sequence vector by the CRF model; wherein step 4 further comprises applying the trained BI-LSTM+CRF model to the predicted corpus, and extracting words labeled as B and I as new words discovered.
10. (canceled)
11. The technology trend prediction method according to claim 1, wherein step 5 further comprises following sub-steps: step 51: extracting keywords of the paper data using a technology lexicon and an existing word segmentation system; step 52: calculating word vectors of the extracted keywords in a high dimension to obtain X.sub.tϵ.sup.d, wherein d is a spatial dimension, and
is a set of word vector; step 53: matching a technology word group w={w1,w2,w3 . . . } corresponding to a specific technology through a technical map, calculating correlated words of a word in the technology word group w in the paper data to obtain wt={w1t, w2t, w3t . . . }, wherein t is the time when the word appears for the first time in a paper; step 54: performing K-means clustering for the correlated words generated after calculation to obtain a same or similar technology set; step 55: obtaining a corresponding technical representation of the same or similar technology set using a weighted reverse maximum matching method, wherein different technical keywords have different weights in the technical map; step 56: calculating a number of papers at different times for the specific technology by a Jaccard index to obtain a published time sequence of paper related with the specific technology, and a formula of the Jaccard index is:
12. The technology trend prediction method according to claim 11, wherein the keywords are extracted using a weighted term frequency-inverse document frequency (TF-IDF) formula:
13. (canceled)
14. (canceled)
15. A technology trend prediction system, comprises an acquisition module configured for acquiring paper data, wherein the technology trend prediction system further comprises: a processing module configured to process the paper data to generate a candidate technology lexicon; a screening module configured to screen the candidate technology lexicon based on mutual information; a calculation module configured to calculate an independent word forming probability of an OOV word; an extraction module configured to extract missed words in a title by using a bidirectional long short-term memory network and a conditional random field (BI-LSTM+CRF) model; and a prediction module configured to predict a technology trend.
16. The technology trend prediction system according to claim 15, wherein the acquisition module is further configured to construct a set of paper data; wherein the processing module is further configured to perform a part-of-speech filtering by using an existing part-of-speech tagging, obtaining a preliminary lexicon after the part-of-speech filtering is completed, and improving OOV word discovery of the candidate technology lexicon by using a Hidden Markov Model (HMM) method; and wherein a formula of the HMM method is:
17. (canceled)
18. (canceled)
19. (canceled)
20. The technology trend prediction system according to claim 16, wherein the screening module is further configured to calculate the mutual information of the OOV words, select a first threshold, and remove the OOV words with the mutual information lower than the first threshold, a calculation formula is:
21. (canceled)
22. The technology trend prediction system according to claim 20, wherein the calculation module is further configured to select a second threshold and remove the OOV words with the independent word forming probability lower than the second suitable threshold, formulas are:
23. The technology trend prediction system according to claim 22, wherein a training method of the BI-LSTM+CRF model comprises following sub-steps: step 41: constructing a labeled corpus according to a technology lexicon, taking words in a title and also in the technology lexicon obtained in step 1 to step 3 as a training corpus of the BI-LSTM+CRF model, taking the other words in the title as a predicted corpus of the BI-LSTM+CRF model, and labeling the words in the title of the training corpus with B, I, and O, three types of tags, wherein B represents a beginning character of a new word, I represents an internal character of the new word, and O represents a non-technical noun word; step 42: converting the words into word vectors, and encoding the word vectors by using the BI-LSTM model; step 43: mapping an encoded result to a sequence vector with a dimension of a number of the tags through a fully connected layer; and step 44: decoding the sequence vector by the CRF model.
24. The technology trend prediction system according to claim 23, wherein the extraction module is further configured to apply the trained BI-LSTM+CRF model to the predicted corpus and extract words labeled as B and I as new words discovered.
25. The technology trend prediction system according to claim 15, wherein an operation of the prediction module further comprises following sub-steps: step 51: extracting keywords of the paper data using a technology lexicon and an existing word segmentation system; step 52: calculating word vectors of the extracted keywords in a high dimension to obtain x.sub.t.sup.d, wherein d is a spatial dimension, and
is a set of word vectors; step 53: matching a technology word group w={w1,w2,w3 . . . } corresponding to a certain specific technology through a technical map, calculating correlated words of a word in the technology word group w in the paper data to obtain wt={w1t, w2t, w3t . . . }, wherein t is the time when the word appears for the first time same or similar in a paper; step 54: performing K-means clustering for the correlated words generated after calculation to obtain a same or similar technology set; step 55: obtaining a corresponding technical representation of the same or similar technology set using a weighted reverse maximum matching method, wherein different technology keywords have different weights in the technical map; step 56: calculating a number of papers at different times for the specific technology by a Jaccard index to obtain a published time sequence of papers related with the specific technology, and a formula of the Jaccard index is:
26. The technology trend prediction system according to claim 25, wherein the keywords are extracted using a weighted term frequency-inverse document frequency (TF-IDF) formula:
27. (canceled)
28. (canceled)
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0037]
[0038]
[0039]
[0040]
[0041]
DETAILED DESCRIPTION OF THE EMBODIMENTS
[0042] Further description of the present invention is provided below with reference to specific embodiments and drawings.
Embodiment 1
[0043] As shown in
[0044] Step 110 is performed to process the paper data to generate a candidate technology lexicon using a processing module 210. Part-of-speech filtering is performed by using an existing part-of-speech tagging, and a preliminary lexicon is obtained after the part-of-speech filtering is completed. OOV (out-of-vocabulary) word discovery of the technology lexicon is improved by using a Hidden Markov Model (HMM) method. A formula of the HMM method is:
wherein, x is an observation sequence, y is a state sequence, π(y.sub.1) represents a probability that the first state is y.sub.1, P represents a state transition probability, i represents the i-th state, and n represents the number of states.
[0045] Step 120 is performed to screen the technology lexicon based on mutual information using a screening module 220. The mutual information of the OOV words is calculated, a suitable threshold is selected, and the OOV words with the mutual information lower than this threshold is removed, and a calculation formula is:
wherein, t.sub.1 t.sub.2 . . . t.sub.i represents the OOV word, t.sub.i represents characters forming the OOV word, f(t.sub.1 t.sub.2 . . . t.sub.i) represents a frequency of the OOV word appearing in patent, L represents a total word frequency of all words in the patent, i represents the number of characters forming the OOV word, P(t.sub.1 t.sub.2 . . . t.sub.i) represents a probability that the t.sub.1 t.sub.2 . . . t.sub.i appears in the patent.
[0046] A result above is compensated with a word length when the frequency of long word is less than the that of short word appearing in text, and the compensated result is:
wherein, N.sub.i=i log.sub.2 i.
[0047] Step 130 is performed to calculate an independent word forming probability of the OOV words using a calculation module 230. Another suitable threshold is selected, and the OOV words with the independent word forming probability lower than this threshold are removed, formulas are as follows:
wherein, str represents a substring, pstr represents a parent string, Rpstr represents a right parent string, Lpstr represents a left parent string, p(⋅) represents the probability that a character string appears, f(⋅) represents the frequency of the character string, Ldp represents dependence of the substring on the left parent string, Rdp represents the dependence of the substring on the right parent string, Idp represents the independent word forming probability of the sub string, dp represents the dependence of the sub string on the parent string and is the maximum value of the Idp and the Rdp.
[0048] Step 140 is performed, and an extraction module 240 extracts missed words in a title by using a bidirectional long short-term memory network BI-LSTM and a conditional random field CRF (BI-LSTM+CRF) model.
[0049] As shown in
[0050] Step 150 is performed to predict a technology trend using a prediction module 250. As shown in
wherein, T.sub.ij is a feature word, tf.sub.ij is a feature word frequency, idf.sub.j is an inverse document frequency, n.sub.ij is the number of occurrences of the feature word in the paper d.sub.j, k is the number of words in one paper, n.sub.kj is the total number of words in the paper d.sub.j, D is the total number of all papers in the corpus, |{j: term.sub.i∈d.sub.1}| is the number of documents containing the feature word term.sub.i.
[0051] Step 152 is performed to calculate word vectors of the extracted keywords in a high dimension to obtain x.sub.tϵ.sup.d, wherein d is a spatial dimension, and
is a set of word vectors.
[0052] Step 153 is performed to match a technology word group w={w1,w2,w3 . . . } corresponding to a certain technology through a technical map, calculate correlated word of the word in the technology word group w in the paper data, and obtain wt={w1t, w2t, w3t . . . }, wherein t is the time when the word appears for the first time in the paper.
[0053] Step 154 is performed to perform K-means clustering for the correlated words generated after calculation to obtain the same or similar technology set. A formula of the clustering is:
wherein, x is a technology word group vector, μ is a technology core word vector.
[0054] Step 155 is performed to obtain a corresponding technical representation of the technology set using a weighted reverse maximum matching method, wherein different technology keywords have different weights in the technical map.
[0055] Step 156 is performed to calculate the number of papers at different times for the technology by a Jaccard index to obtain a published time sequence of papers related with the technology, and a Jaccard index formula is:
wherein, w1 is a keyword in the technology set, and w2 is a keyword in the paper.
[0056] Step 157 is performed to calculate the technology trend by an ARIMA (Autoregressive Integrated Moving Average) model. A formula for calculating the technology trend:
[0057] wherein, p is an autoregressive term, ϕ is a slope coefficient, L is a lag operator, d is a fractional order, X is a technical correlation, q is a corresponding number of moving average terms, θ is a moving average coefficient, ε is a technical coefficient.
[0058] Step 158 is performed to use an unweighted maximum matching and edge cutting algorithm, finally obtain technology relevance without communication to calculate a technology change trend between technology clusters.
Embodiment 2
[0059] The present invention comprises the following steps.
[0060] The first step: processing a technology lexicon.
[0061] a) Step 1: acquiring paper data to construct a set of paper data.
[0062] b) Step 2: generating a candidate technology lexicon. A specific realization method is: performing part-of-speech filtering by using an existing part-of-speech tagging, and obtaining a preliminary lexicon after the part-of-speech filtering is completed, a method of the part-of-speech filtering is as follow:
TABLE-US-00001 two-word terms three-word terms N + N N + N + N N + V V + N + N V + N N + V + N A + N V + V + N D + N B + V + N B + N N + M + N
wherein, N represents a noun, V represents a verb, B represents a distinguishing word, A represents an adjective, D represents a adverb, M represents a numeral, and a multi-word term is generated by different combinations of part-of-speech.
[0063] c) Step 3: improving OOV word discovery of the technology lexicon using a Hidden Markov Model (HMM) method, a formula of the HMM method is:
wherein, x is an observation sequence, y is a state sequence, π(y.sub.1) represents a probability that the first state is
[0064] d) Step 4: screening the lexicon generated above using a mutual information method. The mutual information of the OOV words is calculated, a suitable threshold is selected, and the OOV words with the mutual information lower than this threshold are removed, and a formula is:
wherein, t.sub.1 t.sub.2 . . . t.sub.i represents the OOV word, t.sub.i represents the characters forming the OOV word, f(t.sub.1 t.sub.2 . . . t.sub.i) represents a frequency of the OOV word appearing in patent, L represents a total word frequency of all words in the patent, i represents the number of characters forming the OOV word, P(t.sub.1 t.sub.2 . . . t.sub.i) represents a probability that the t.sub.1 t.sub.2 . . . t.sub.i appears in the patent.
[0065] According to a statistical result, the frequency of the long word is less than that of short word appearing in text, so a result above is compensated with a word length, and the compensated result is:
wherein, N.sub.i=i log.sub.2 i.
[0066] e) Step 5: reducing broken strings in the lexicon generated above. An independent word forming probability of an OOV word is calculated, another suitable threshold is selected, and the OOV words with independent word forming probability lower than this threshold are removed, a formula is:
wherein, str represents a substring, pstr represents a parent string, Rpstr represents a right parent string, Lpstr represents a left parent string, p(⋅) represents the probability that a character string appears, f(⋅) represents the frequency of the character string, Ldp represents dependence of the substring on the left parent string, Rdp represents the dependence of the substring on the right parent string, Idp represents the independent word forming probability of the sub string.
[0067] f) Step 6: extracting missed words in a title after the above steps to improve a recall rate using a BI-LSTM+CRF model:1. i. constructing a labeled corpus according to the technology lexicon obtained after the above steps, taking the words in the title which also in the lexicon as a training corpus of the model, taking the other words in the title as a predicted corpus of the model, and labeling the words in the title of the training corpus with B, I and O three types of tags, wherein B represents a beginning character of new word, I represents an internal character of the new word, and O represents a non-technical noun word;
ii. converting the words into word vectors, and then encoding them by using the BI-LSTM;
iii. mapping an encoded result to a sequence vector with the dimension of the number of the tags through a fully connected layer;
iv. decoding the sequence vector obtained above by the CRF;
v. training a model according to the above steps, and then applying the trained BI-LSTM+CRF model to the predicted corpus, and extracting words labeled as B and I e as new words discovered.
[0068] The second step: predicting a technology trend.
[0069] a). Keywords of the paper data are extracted using the technology lexicon generated in the first step and an existing word segmentation system; the keywords are extracted using a weighted (TF-IDF) method, and a formula is:
wherein, t(w)=weight(T.sub.ij) is a TF-IDF value of a feature T.sub.ij in a document d.sub.j; title(w) is the weight of the word w when the word w appears in the title, and tec(w) is the weight of the word w when the word w appears in the technology field.
[0070] b). Word vectors of the extracted keywords in a high dimension are calculated to obtain x.sub.tϵ.sup.d, wherein d is a spatial dimension.
[0071] c). A technology word group w={w1,w2,w3 . . . } corresponding to a certain technology is matched through a technical map, and then correlated word of the word in the technology word group w in the paper data is calculated to obtain wt={w1t, w2t, w3t . . . }, wherein t is the time when the word appears for the first time in the paper.
[0072] d). K-means clustering is performed for the correlated words generated after calculation to obtain the same or similar technology set:
[0073] e). A corresponding technical representation of the technology set is obtained by using a weighted reverse maximum matching method, wherein different technology keywords have different weights in the technical map.
[0074] f). The number of papers at different times for the technology is calculated to obtain a published time sequence of papers related with the technology.
[0075] g). the technology trend is calculated by a ARIMA model, and a formula is:
wherein, L is a lag operator, d is in Z and d>0.
[0076] h). An unweighted maximum matching and edge cutting algorithm is used to obtain technology relevance without communication, and a technology change trend between e technology clusters is calculated.
[0077] In order to better understand the present invention, the detailed description is made above in conjunction with the specific embodiments of the present invention, but it is not a limitation of the present invention. Any simple modification to the above embodiments based on the technical essence of the present invention still belongs to the scope of the technical solution of the present invention. Each embodiment in this specification focuses on differences from other embodiments, and the same or similar parts between the various embodiments can be referred to each other. As for the system embodiment, since it basically corresponds to the method embodiment, the description is relatively simple, and the relevant part can refer to the part of the description of the method embodiment.