CN103049526A - Cross-media retrieval method based on double space learning - Google Patents
Cross-media retrieval method based on double space learning Download PDFInfo
- Publication number
- CN103049526A CN103049526A CN201210559081XA CN201210559081A CN103049526A CN 103049526 A CN103049526 A CN 103049526A CN 201210559081X A CN201210559081X A CN 201210559081XA CN 201210559081 A CN201210559081 A CN 201210559081A CN 103049526 A CN103049526 A CN 103049526A
- Authority
- CN
- China
- Prior art keywords
- medium data
- matrix
- vector
- proper vector
- different modalities
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 46
- 238000013507 mapping Methods 0.000 claims abstract description 26
- 238000012360 testing method Methods 0.000 claims abstract description 15
- 239000013598 vector Substances 0.000 claims description 43
- 239000011159 matrix material Substances 0.000 claims description 40
- 238000012549 training Methods 0.000 claims description 15
- 239000000284 extract Substances 0.000 claims description 8
- 238000000354 decomposition reaction Methods 0.000 claims description 4
- 230000002902 bimodal effect Effects 0.000 claims 3
- 238000005259 measurement Methods 0.000 abstract description 2
- 230000006870 function Effects 0.000 description 5
- 238000010586 diagram Methods 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 241001269238 Data Species 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000013506 data mapping Methods 0.000 description 1
- 230000004069 differentiation Effects 0.000 description 1
- 239000002360 explosive Substances 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000012804 iterative process Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
Images
Abstract
The invention discloses a cross-media retrieval method based on double space learning. The method comprises the following steps: firstly extracting the characteristics of multimedia data in different modes, utilizing double space learning method to learn and obtain two mapping matrixes, and mapping the data in different modes to the same space; dividing a test sample set into two parts of a query data set and a target data set, using the mapping matrixes obtained by learning to map the data of the two data sets to a unified space, and then measuring the distance between query data and target data, and obtaining the target data mostly close to the query data in distance. According to the method, the multimedia data in different modes can be mapped into the unified space for measurement, and the characteristics are selected while mapping, so that the robustness and accuracy of retrieval are improved, and the method is good in application prospect.
Description
Technical field
The present invention relates to mode identification technology, particularly a kind ofly stride the media search method based on double space study.
Background technology
In the last few years, along with the rise of Web2.0 technology, the multi-medium data on the internet presented explosive growth, and the user need to retrieve the contents such as the article oneself wanted, picture, video from these mass datas.And present main flow search engine such as Baidu, Google etc., the search technique of employing still is based on key word, and this search technique can't solve the semantic gap that exists between each media.In order to realize striding the media retrieval, need to provide a kind of integrated retrieval method from the another kind of multi-medium data of a class multimedia data retrieval.
The research of striding the media search method has caused the attention of domestic and international academia.At present propose in the world certain methods and come relation between the multi-medium data of modeling different modalities, existing method can be divided into two classes: 1) multi-medium data with two mode is mapped to same subspace, then the similarity between the multimedia of tolerance different modalities in the subspace that this study is arrived realizes striding the media retrieval with this.2) multi-medium data of a class mode is transformed into the similarity measurement that carries out the different modalities data in the feature space of multi-medium data of another kind of mode, realizes striding the media retrieval with this.Yet existing method has all been ignored another major issue, namely how to select simultaneously the most relevant from the feature space of different modalities and the feature of differentiation power is arranged most, and this problem is called the double space feature selecting.For in Learning Subspaces, carry out the double space feature selecting, the present invention propose a kind of based on double space study stride the media search method, the method has realized the leap of the semantic gap between the different modalities media effectively, and then so that the result that search engine returns is more accurate.
Summary of the invention
The problem that exists in order to solve prior art the purpose of this invention is to provide and a kind ofly strides the media search method based on double space study.The method can be carried out sub-space learning and double space feature selecting simultaneously by minimizing an objective function, in the similarity of the multi-medium data of the subspace vacuum metrics different modalities that study is arrived, realizes striding the media retrieval with this.
A kind of media search method of striding based on double space study that the present invention proposes may further comprise the steps:
Step S1, the multi-medium data sample of collection different modalities is set up and is striden media retrieval knowledge database, and described database is divided into training set and test set;
Step S2 extracts the proper vector of different modalities multi-medium data sample in the described database;
Step S3 obtains and described different modalities corresponding mapping matrix respectively based on the proper vector of the different modalities data in the described training set;
Step S4, according to the mode classification of multi-medium data sample in the described test set, the mapping matrix that utilizes described step S3 to obtain is mapped to the same space with them;
Step S5, as query set, the multi-medium data of another mode classification is as object set with the multi-medium data of same mode classification in the test set that is mapped to behind the same space;
Step S6 for a multi-medium data in the described query set, according to the similarity between the multi-medium data, obtains described target tightening and its most similar multi-medium data, thereby obtains striding the media result for retrieval.
The method according to this invention can be mapped into the multi-medium data of different modalities unified space and measure, and has carried out feature selecting in mapping, has improved robustness and the accuracy of retrieval, has good utilization prospect.The inventive method can have been crossed over the semantic gap between the different modalities media effectively, and then to stride the result that the media research engine returns more accurate.
Description of drawings
Fig. 1 be the present invention is based on double space study stride media search method process flow diagram;
Fig. 2 strides the retrieval effectiveness synoptic diagram of media search method from the text to the image according to the present invention.
Embodiment
For making the purpose, technical solutions and advantages of the present invention clearer, below in conjunction with specific embodiment, and with reference to accompanying drawing, the present invention is described in more detail.
The present invention learns two mapping matrixes by double space, and the data-mapping of different modalities in same subspace, and has been carried out feature selecting in mapping; In the similarity of the data of the subspace vacuum metrics different modalities that study is arrived, realize striding the purpose of media retrieval with this.
Fig. 1 be the present invention is based on double space study stride media search method process flow diagram, as shown in Figure 1, a kind of cross-module attitude search method based on double space study that the present invention proposes comprises following step:
Step S1, the multi-medium data sample of collection different modalities is set up and is striden media retrieval knowledge database, and described database is divided into training set and test set;
Described different modalities is such as being the mode such as text, image, and for the ease of the present invention will be described, hereinafter the data instance with text and these two mode of image makes an explanation.
When described database is divided into training set and test set, can divide according to actual needs, such as the data of 80% in the described database can being divided into training set, remaining 20% data are divided into test set.
Step S2 extracts the proper vector of different modalities multi-medium data sample in the described database;
Among the present invention, use respectively hidden Di Lei Cray distribution (Latent Dirichlet Allocation for text and image, LDA) algorithm and the conversion of yardstick invariant features (Scale-Invariant Feature Transform, SIFT) algorithm carries out feature extraction.
Step S3 obtains and described different modalities corresponding mapping matrix respectively based on the proper vector of the different modalities data in the described training set;
Before this step is described in detail, introduce first some matrix operation symbols:
For matrix
Be the space of matrices of n * m, the i row element of matrix M is expressed as m
i, the j column element of matrix M is expressed as m
j, and the Frobenius norm of matrix M is called for short the F-norm and is defined as:
Two norms of the row vector of matrix M with ‖ M ‖
21Be defined as:
The trace norm of matrix M is defined as:
Wherein, σ
iI the singular value of representing matrix M.
For vector
Represent that a diagonal entry is u
iDiagonal matrix, wherein, u
iI the element that represents vectorial u.
If the matrix that the proper vector of described training set Chinese version multi-medium data forms is
D1 is the dimension of proper vector, and n is the number of proper vector, and a represents the mode classification; The matrix that the proper vector of image multi-medium data forms is
D2 is the dimension of proper vector, and n is the number of proper vector, and b represents the mode classification; The class label matrix is
N is the number of proper vector, and c is the number of mode classification.
Described step S3 further may further comprise the steps:
Step S31, set up objective function based on the proper vector of different modalities data in the described training set:
Wherein, U
aAnd U
bBe two mapping matrixes corresponding with text and image difference, λ
1And λ
2Be undetermined parameter, span is 0.0001 to 0.1;
Step S32, find the solution described objective function (such as passing through iterative algorithm) and obtain the mapping matrix corresponding with described different modalities difference:
Described step S32 further may further comprise the steps:
Step S321 arranges mapping matrix U
aAnd U
bInitial value be null matrix, maximum iteration time k is set, and the initial value that iterations is set is 1;
Step S322 is to matrix
Carry out Eigenvalues Decomposition, obtain Eigenvalues Decomposition result: VDiag (d) V
T, wherein, V is with matrix
Feature value vector be the matrix that column vector forms; D is the vector that feature value vector characteristic of correspondence value forms.
Step S323 sets intermediate variable S
-1=VDiag (c) V
T, wherein, c is a vector, its i element is
Wherein, d
iBe i the element of vectorial d, μ is a minimum number.
Step S324 is according to i the element p of following formula calculating intermediate vector p and q
iAnd q
i:
And make P=Diag (p), and Q=Diag (q), wherein, ε is a minimum number.
Step S325 tries to achieve described mapping matrix U by two linear problems of finding the solution in the following formula
aAnd U
b:
Whether step S326 judges the iterations of this moment less than maximum iteration time k, if then turn to described step S322 to continue iteration; If not, then stop iteration, obtain described mapping matrix U
aAnd U
b
Step S4, according to the mode classification of multi-medium data sample in the described test set, the mapping matrix that utilizes described step S3 to obtain is mapped to the same space with them;
Step S5, as query set, the multi-medium data of another mode classification is as object set with the multi-medium data of same mode classification in the test set that is mapped to behind the same space;
Step S6 for a multi-medium data in the described query set, according to the similarity between the multi-medium data, obtains described target tightening and its most similar multi-medium data, thereby obtains striding the media result for retrieval.
Similarity between the described multi-medium data can be measured with Euclidean distance, and for Euclidean distance, distance is nearer just to illustrate that these two multi-medium datas are more similar.
In order to verify the validity of the inventive method, next method proposed by the invention is applied to a multimedia database.The multi-medium data that comprises 10 semantic domains in this database, the text and the image that comprise 2866 pairs of identical semantic informations, choose at random wherein the 2173 pairs of texts and image pattern as training set, choose wherein the 693 pairs of texts and image pattern as test set, the concrete steps of using the inventive method are as follows:
1) extracts different proper vectors for the difference of the data based mode classification in the described database: for text modality, extract the LDA semantic feature of 10 dimensions, for image modalities, extract the SIFT feature of 128 dimensions;
2) with the proper vector composition characteristic matrix of proper vector and the view data of described training set Chinese version data, the structure objective function, learn to obtain two mapping matrixes corresponding with text and image by minimizing objective function, the learning process of described mapping matrix realizes by an iterative process;
3) according to the difference of described mode classification, the mapping matrix that the proper vector of the proper vector of the text data in the test set and view data is obtained by study is mapped in the same space;
4) with the text data in the test set as query set, view data is as object set, a given text document can access the image of target tightening the most similar with it (being that Euclidean distance is nearest).Fig. 2 strides the retrieval effectiveness synoptic diagram of media search method from the text to the image according to the present invention, Fig. 2 left side is two text documents, the right side be obtain according to the inventive method to these two images that text document is the most similar.
To sum up, the present invention proposes and a kind ofly effectively stride the media search method based on double space study.The present invention is easy to realize, stable performance, and recognition accuracy is high.The present invention proposes effectively to have crossed over the semantic gap between the different modalities media, strides media method relatively with traditional, and the present invention shows and uses widely prospect and larger marketable value.
Above-described specific embodiment; purpose of the present invention, technical scheme and beneficial effect are further described; institute is understood that; the above only is specific embodiments of the invention; be not limited to the present invention; within the spirit and principles in the present invention all, any modification of making, be equal to replacement, improvement etc., all should be included within protection scope of the present invention.
Claims (9)
- One kind based on double space study stride the media search method, it is characterized in that the method may further comprise the steps:Step S1, the multi-medium data sample of collection different modalities is set up and is striden media retrieval knowledge database, and described database is divided into training set and test set;Step S2 extracts the proper vector of different modalities multi-medium data sample in the described database;Step S3 obtains and described different modalities corresponding mapping matrix respectively based on the proper vector of the different modalities data in the described training set;Step S4, according to the mode classification of multi-medium data sample in the described test set, the mapping matrix that utilizes described step S3 to obtain is mapped to the same space with them;Step S5, as query set, the multi-medium data of another mode classification is as object set with the multi-medium data of same mode classification in the test set that is mapped to behind the same space;Step S6 for a multi-medium data in the described query set, according to the similarity between the multi-medium data, obtains described target tightening and its most similar multi-medium data, thereby obtains striding the media result for retrieval.
- 2. method according to claim 1 is characterized in that, described different modalities is bimodal.
- 3. method according to claim 2 is characterized in that, described bimodal is text and image.
- 4. method according to claim 3 is characterized in that, for text multi-medium data sample, extracts its hidden Di Lei Cray distribution characteristics vector; For image multi-medium data sample, extract its yardstick invariant features transform characteristics vector.
- 5. method according to claim 1 is characterized in that, described step S3 further may further comprise the steps:Step S31 sets up objective function based on the proper vector of different modalities data in the described training set;Step S32 finds the solution described objective function and obtains the mapping matrix corresponding with described different modalities difference.
- 6. method according to claim 5 is characterized in that, described objective function is:Wherein, U aAnd U bBe two mapping matrixes corresponding with the bimodal difference, λ 1And λ 2Be undetermined parameter, ‖ ‖ FFor asking for the operation of F-norm, ‖ ‖ 21For asking for vectorial two norms of row and operation, ‖ ‖ *For asking for the trace norm operation.
- 7. method according to claim 5 is characterized in that, carries out for finding the solution with iterative algorithm of described objective function.
- 8. method according to claim 7 is characterized in that, described step S32 further may further comprise the steps:Step S321 arranges mapping matrix U aAnd U bInitial value be null matrix, maximum iteration time k is set, the initial value of iterations is 1;Step S322 is to matrix Carry out Eigenvalues Decomposition, obtain Eigenvalues Decomposition result: VDiag (d) V T, wherein, Be the matrix that the proper vector of described training set Chinese version multi-medium data forms, d1 is the dimension of proper vector, and n is the number of proper vector, and a represents the mode classification; Be the matrix that the proper vector of image multi-medium data in the described training set forms, d2 is the dimension of proper vector, and n is the number of proper vector, and b represents the mode classification; V is with matrix Feature value vector be the matrix that column vector forms; D is the vector that feature value vector characteristic of correspondence value forms;Step S323 sets intermediate variable S -1=VDiag (c) V T, wherein, c is a vector, its i element is Wherein, d iBe i the element of vectorial d, μ is a minimum number;Step S324 is according to i the element p of following formula calculating intermediate vector p and q iAnd q i:And make P=Diag (p), and Q=Diag (q), wherein, ε is a minimum number;Step S325 tries to achieve described mapping matrix U by two linear problems of finding the solution in the following formula aAnd U b:Wherein, Be the class label matrix, n is the number of proper vector, and c is the number of mode classification;Whether step S326 judges the iterations of this moment less than maximum iteration time k, if then turn to described step S322 to continue iteration; If not, then stop iteration, obtain described mapping matrix U aAnd U b
- 9. method according to claim 1 is characterized in that, the similarity between the described multi-medium data is measured with Euclidean distance.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201210559081.XA CN103049526B (en) | 2012-12-20 | 2012-12-20 | Based on the cross-media retrieval method of double space study |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201210559081.XA CN103049526B (en) | 2012-12-20 | 2012-12-20 | Based on the cross-media retrieval method of double space study |
Publications (2)
Publication Number | Publication Date |
---|---|
CN103049526A true CN103049526A (en) | 2013-04-17 |
CN103049526B CN103049526B (en) | 2015-08-05 |
Family
ID=48062167
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201210559081.XA Active CN103049526B (en) | 2012-12-20 | 2012-12-20 | Based on the cross-media retrieval method of double space study |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103049526B (en) |
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103995903A (en) * | 2014-06-12 | 2014-08-20 | 武汉科技大学 | Cross-media search method based on isomorphic subspace mapping and optimization |
CN105574133A (en) * | 2015-12-15 | 2016-05-11 | 苏州贝多环保技术有限公司 | Multi-mode intelligent question answering system and method |
CN105701227A (en) * | 2016-01-15 | 2016-06-22 | 北京大学 | Cross-media similarity measure method and search method based on local association graph |
CN105701225A (en) * | 2016-01-15 | 2016-06-22 | 北京大学 | Cross-media search method based on unification association supergraph protocol |
CN106056043A (en) * | 2016-05-19 | 2016-10-26 | 中国科学院自动化研究所 | Animal behavior identification method and apparatus based on transfer learning |
CN107169061A (en) * | 2017-05-02 | 2017-09-15 | 广东工业大学 | A kind of text multi-tag sorting technique for merging double information sources |
CN107273517A (en) * | 2017-06-21 | 2017-10-20 | 复旦大学 | Picture and text cross-module state search method based on the embedded study of figure |
WO2017210949A1 (en) * | 2016-06-06 | 2017-12-14 | 北京大学深圳研究生院 | Cross-media retrieval method |
CN108399414A (en) * | 2017-02-08 | 2018-08-14 | 南京航空航天大学 | Method of Sample Selection and device |
CN109284414A (en) * | 2018-09-30 | 2019-01-29 | 中国科学院计算技术研究所 | The cross-module state content search method and system kept based on semanteme |
CN109992676A (en) * | 2019-04-01 | 2019-07-09 | 中国传媒大学 | Across the media resource search method of one kind and searching system |
CN110851641A (en) * | 2018-08-01 | 2020-02-28 | 杭州海康威视数字技术股份有限公司 | Cross-modal retrieval method and device and readable storage medium |
CN111639197A (en) * | 2020-05-28 | 2020-09-08 | 山东大学 | Cross-modal multimedia data retrieval method and system with label embedded online hash |
CN116383342A (en) * | 2023-04-07 | 2023-07-04 | 四川大学 | Robust cross-domain text retrieval method under noise label |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107220337B (en) * | 2017-05-25 | 2020-12-22 | 北京大学 | Cross-media retrieval method based on hybrid migration network |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040267774A1 (en) * | 2003-06-30 | 2004-12-30 | Ibm Corporation | Multi-modal fusion in content-based retrieval |
CN1920818A (en) * | 2006-09-14 | 2007-02-28 | 浙江大学 | Transmedia search method based on multi-mode information convergence analysis |
CN101021849A (en) * | 2006-09-14 | 2007-08-22 | 浙江大学 | Transmedia searching method based on content correlation |
CN102521368A (en) * | 2011-12-16 | 2012-06-27 | 武汉科技大学 | Similarity matrix iteration based cross-media semantic digesting and optimizing method |
-
2012
- 2012-12-20 CN CN201210559081.XA patent/CN103049526B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040267774A1 (en) * | 2003-06-30 | 2004-12-30 | Ibm Corporation | Multi-modal fusion in content-based retrieval |
CN1920818A (en) * | 2006-09-14 | 2007-02-28 | 浙江大学 | Transmedia search method based on multi-mode information convergence analysis |
CN101021849A (en) * | 2006-09-14 | 2007-08-22 | 浙江大学 | Transmedia searching method based on content correlation |
CN102521368A (en) * | 2011-12-16 | 2012-06-27 | 武汉科技大学 | Similarity matrix iteration based cross-media semantic digesting and optimizing method |
Non-Patent Citations (1)
Title |
---|
KAIYE WANG 等: "Learning Coupled Feature Spaces for Cross-modal Matching", 《2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION》 * |
Cited By (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103995903B (en) * | 2014-06-12 | 2017-04-12 | 武汉科技大学 | Cross-media search method based on isomorphic subspace mapping and optimization |
CN103995903A (en) * | 2014-06-12 | 2014-08-20 | 武汉科技大学 | Cross-media search method based on isomorphic subspace mapping and optimization |
CN105574133A (en) * | 2015-12-15 | 2016-05-11 | 苏州贝多环保技术有限公司 | Multi-mode intelligent question answering system and method |
CN105701225A (en) * | 2016-01-15 | 2016-06-22 | 北京大学 | Cross-media search method based on unification association supergraph protocol |
CN105701227A (en) * | 2016-01-15 | 2016-06-22 | 北京大学 | Cross-media similarity measure method and search method based on local association graph |
CN105701227B (en) * | 2016-01-15 | 2019-02-01 | 北京大学 | A kind of across media method for measuring similarity and search method based on local association figure |
CN105701225B (en) * | 2016-01-15 | 2019-02-01 | 北京大学 | A kind of cross-media retrieval method based on unified association hypergraph specification |
CN106056043A (en) * | 2016-05-19 | 2016-10-26 | 中国科学院自动化研究所 | Animal behavior identification method and apparatus based on transfer learning |
CN106056043B (en) * | 2016-05-19 | 2019-07-30 | 中国科学院自动化研究所 | Animal behavior recognition methods and device based on transfer learning |
WO2017210949A1 (en) * | 2016-06-06 | 2017-12-14 | 北京大学深圳研究生院 | Cross-media retrieval method |
CN108399414B (en) * | 2017-02-08 | 2021-06-01 | 南京航空航天大学 | Sample selection method and device applied to cross-modal data retrieval field |
CN108399414A (en) * | 2017-02-08 | 2018-08-14 | 南京航空航天大学 | Method of Sample Selection and device |
WO2018145604A1 (en) * | 2017-02-08 | 2018-08-16 | 南京航空航天大学 | Sample selection method, apparatus and server |
US10885390B2 (en) | 2017-02-08 | 2021-01-05 | Nanjing University Of Aeronautics And Astronautics | Sample selection method and apparatus and server |
CN107169061A (en) * | 2017-05-02 | 2017-09-15 | 广东工业大学 | A kind of text multi-tag sorting technique for merging double information sources |
CN107273517A (en) * | 2017-06-21 | 2017-10-20 | 复旦大学 | Picture and text cross-module state search method based on the embedded study of figure |
CN107273517B (en) * | 2017-06-21 | 2021-07-23 | 复旦大学 | Graph-text cross-modal retrieval method based on graph embedding learning |
CN110851641A (en) * | 2018-08-01 | 2020-02-28 | 杭州海康威视数字技术股份有限公司 | Cross-modal retrieval method and device and readable storage medium |
CN109284414B (en) * | 2018-09-30 | 2020-12-04 | 中国科学院计算技术研究所 | Cross-modal content retrieval method and system based on semantic preservation |
CN109284414A (en) * | 2018-09-30 | 2019-01-29 | 中国科学院计算技术研究所 | The cross-module state content search method and system kept based on semanteme |
CN109992676A (en) * | 2019-04-01 | 2019-07-09 | 中国传媒大学 | Across the media resource search method of one kind and searching system |
CN109992676B (en) * | 2019-04-01 | 2020-12-25 | 中国传媒大学 | Cross-media resource retrieval method and retrieval system |
CN111639197A (en) * | 2020-05-28 | 2020-09-08 | 山东大学 | Cross-modal multimedia data retrieval method and system with label embedded online hash |
CN116383342A (en) * | 2023-04-07 | 2023-07-04 | 四川大学 | Robust cross-domain text retrieval method under noise label |
CN116383342B (en) * | 2023-04-07 | 2023-11-14 | 四川大学 | Robust cross-domain text retrieval method under noise label |
Also Published As
Publication number | Publication date |
---|---|
CN103049526B (en) | 2015-08-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN103049526B (en) | Based on the cross-media retrieval method of double space study | |
Liu et al. | Cross-modality binary code learning via fusion similarity hashing | |
Zhu et al. | Exploring auxiliary context: discrete semantic transfer hashing for scalable image retrieval | |
Saleh et al. | Toward automated discovery of artistic influence | |
Gordo | Supervised mid-level features for word image representation | |
Wang et al. | Learning compact hash codes for multimodal representations using orthogonal deep structure | |
Qian et al. | Unsupervised feature selection for multi-view clustering on text-image web news data | |
Lu et al. | Image classification by visual bag-of-words refinement and reduction | |
CN104834693A (en) | Depth-search-based visual image searching method and system thereof | |
CN104463247A (en) | Extracting method of optical spectrum vector cross-correlation features in hyper-spectral image classification | |
CN102663447A (en) | Cross-media searching method based on discrimination correlation analysis | |
CN105930873A (en) | Self-paced cross-modal matching method based on subspace | |
CN105808752A (en) | CCA and 2PKNN based automatic image annotation method | |
CN103995903B (en) | Cross-media search method based on isomorphic subspace mapping and optimization | |
Afzalan et al. | An automated spectral clustering for multi-scale data | |
Zheng et al. | MMDF-LDA: An improved Multi-Modal Latent Dirichlet Allocation model for social image annotation | |
Xu et al. | A lightweight intrinsic mean for remote sensing classification with lie group kernel function | |
Jiang et al. | A hierarchal BoW for image retrieval by enhancing feature salience | |
Jiang et al. | Co-regularized PLSA for multi-view clustering | |
CN106844481A (en) | Font similarity and font replacement method | |
CN103177121A (en) | Locality preserving projection method for adding pearson relevant coefficient | |
Ahsan et al. | Clustering social event images using kernel canonical correlation analysis | |
CN103886072A (en) | Retrieved result clustering system in coal mine search engine | |
Han et al. | Dayside aurora classification via BIFs-based sparse representation using manifold learning | |
Zhang et al. | Large-scale clustering with structured optimal bipartite graph |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant |