CN103049526A - Cross-media retrieval method based on double space learning - Google Patents

Cross-media retrieval method based on double space learning Download PDF

Info

Publication number
CN103049526A
CN103049526A CN201210559081XA CN201210559081A CN103049526A CN 103049526 A CN103049526 A CN 103049526A CN 201210559081X A CN201210559081X A CN 201210559081XA CN 201210559081 A CN201210559081 A CN 201210559081A CN 103049526 A CN103049526 A CN 103049526A
Authority
CN
China
Prior art keywords
medium data
matrix
vector
proper vector
different modalities
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201210559081XA
Other languages
Chinese (zh)
Other versions
CN103049526B (en
Inventor
王亮
谭铁牛
赫然
王开业
王威
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Automation of Chinese Academy of Science
Original Assignee
Institute of Automation of Chinese Academy of Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Automation of Chinese Academy of Science filed Critical Institute of Automation of Chinese Academy of Science
Priority to CN201210559081.XA priority Critical patent/CN103049526B/en
Publication of CN103049526A publication Critical patent/CN103049526A/en
Application granted granted Critical
Publication of CN103049526B publication Critical patent/CN103049526B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Abstract

The invention discloses a cross-media retrieval method based on double space learning. The method comprises the following steps: firstly extracting the characteristics of multimedia data in different modes, utilizing double space learning method to learn and obtain two mapping matrixes, and mapping the data in different modes to the same space; dividing a test sample set into two parts of a query data set and a target data set, using the mapping matrixes obtained by learning to map the data of the two data sets to a unified space, and then measuring the distance between query data and target data, and obtaining the target data mostly close to the query data in distance. According to the method, the multimedia data in different modes can be mapped into the unified space for measurement, and the characteristics are selected while mapping, so that the robustness and accuracy of retrieval are improved, and the method is good in application prospect.

Description

Stride the media search method based on double space study
Technical field
The present invention relates to mode identification technology, particularly a kind ofly stride the media search method based on double space study.
Background technology
In the last few years, along with the rise of Web2.0 technology, the multi-medium data on the internet presented explosive growth, and the user need to retrieve the contents such as the article oneself wanted, picture, video from these mass datas.And present main flow search engine such as Baidu, Google etc., the search technique of employing still is based on key word, and this search technique can't solve the semantic gap that exists between each media.In order to realize striding the media retrieval, need to provide a kind of integrated retrieval method from the another kind of multi-medium data of a class multimedia data retrieval.
The research of striding the media search method has caused the attention of domestic and international academia.At present propose in the world certain methods and come relation between the multi-medium data of modeling different modalities, existing method can be divided into two classes: 1) multi-medium data with two mode is mapped to same subspace, then the similarity between the multimedia of tolerance different modalities in the subspace that this study is arrived realizes striding the media retrieval with this.2) multi-medium data of a class mode is transformed into the similarity measurement that carries out the different modalities data in the feature space of multi-medium data of another kind of mode, realizes striding the media retrieval with this.Yet existing method has all been ignored another major issue, namely how to select simultaneously the most relevant from the feature space of different modalities and the feature of differentiation power is arranged most, and this problem is called the double space feature selecting.For in Learning Subspaces, carry out the double space feature selecting, the present invention propose a kind of based on double space study stride the media search method, the method has realized the leap of the semantic gap between the different modalities media effectively, and then so that the result that search engine returns is more accurate.
Summary of the invention
The problem that exists in order to solve prior art the purpose of this invention is to provide and a kind ofly strides the media search method based on double space study.The method can be carried out sub-space learning and double space feature selecting simultaneously by minimizing an objective function, in the similarity of the multi-medium data of the subspace vacuum metrics different modalities that study is arrived, realizes striding the media retrieval with this.
A kind of media search method of striding based on double space study that the present invention proposes may further comprise the steps:
Step S1, the multi-medium data sample of collection different modalities is set up and is striden media retrieval knowledge database, and described database is divided into training set and test set;
Step S2 extracts the proper vector of different modalities multi-medium data sample in the described database;
Step S3 obtains and described different modalities corresponding mapping matrix respectively based on the proper vector of the different modalities data in the described training set;
Step S4, according to the mode classification of multi-medium data sample in the described test set, the mapping matrix that utilizes described step S3 to obtain is mapped to the same space with them;
Step S5, as query set, the multi-medium data of another mode classification is as object set with the multi-medium data of same mode classification in the test set that is mapped to behind the same space;
Step S6 for a multi-medium data in the described query set, according to the similarity between the multi-medium data, obtains described target tightening and its most similar multi-medium data, thereby obtains striding the media result for retrieval.
The method according to this invention can be mapped into the multi-medium data of different modalities unified space and measure, and has carried out feature selecting in mapping, has improved robustness and the accuracy of retrieval, has good utilization prospect.The inventive method can have been crossed over the semantic gap between the different modalities media effectively, and then to stride the result that the media research engine returns more accurate.
Description of drawings
Fig. 1 be the present invention is based on double space study stride media search method process flow diagram;
Fig. 2 strides the retrieval effectiveness synoptic diagram of media search method from the text to the image according to the present invention.
Embodiment
For making the purpose, technical solutions and advantages of the present invention clearer, below in conjunction with specific embodiment, and with reference to accompanying drawing, the present invention is described in more detail.
The present invention learns two mapping matrixes by double space, and the data-mapping of different modalities in same subspace, and has been carried out feature selecting in mapping; In the similarity of the data of the subspace vacuum metrics different modalities that study is arrived, realize striding the purpose of media retrieval with this.
Fig. 1 be the present invention is based on double space study stride media search method process flow diagram, as shown in Figure 1, a kind of cross-module attitude search method based on double space study that the present invention proposes comprises following step:
Step S1, the multi-medium data sample of collection different modalities is set up and is striden media retrieval knowledge database, and described database is divided into training set and test set;
Described different modalities is such as being the mode such as text, image, and for the ease of the present invention will be described, hereinafter the data instance with text and these two mode of image makes an explanation.
When described database is divided into training set and test set, can divide according to actual needs, such as the data of 80% in the described database can being divided into training set, remaining 20% data are divided into test set.
Step S2 extracts the proper vector of different modalities multi-medium data sample in the described database;
Among the present invention, use respectively hidden Di Lei Cray distribution (Latent Dirichlet Allocation for text and image, LDA) algorithm and the conversion of yardstick invariant features (Scale-Invariant Feature Transform, SIFT) algorithm carries out feature extraction.
Step S3 obtains and described different modalities corresponding mapping matrix respectively based on the proper vector of the different modalities data in the described training set;
Before this step is described in detail, introduce first some matrix operation symbols:
For matrix
Figure BDA00002621093900031
Figure BDA00002621093900032
Be the space of matrices of n * m, the i row element of matrix M is expressed as m i, the j column element of matrix M is expressed as m j, and the Frobenius norm of matrix M is called for short the F-norm and is defined as:
| | M | | F = Σ i = 1 n | | m i | | 2 2 - - - ( 1 )
Two norms of the row vector of matrix M with ‖ M ‖ 21Be defined as:
| | M | | 21 = Σ i = 1 n | | m i | | 2 - - - ( 2 )
The trace norm of matrix M is defined as:
| | M | | * = Σ i = 1 min ( m , n ) σ i - - - ( 3 )
Wherein, σ iI the singular value of representing matrix M.
For vector
Figure BDA00002621093900042
Represent that a diagonal entry is u iDiagonal matrix, wherein, u iI the element that represents vectorial u.
If the matrix that the proper vector of described training set Chinese version multi-medium data forms is
Figure BDA00002621093900043
D1 is the dimension of proper vector, and n is the number of proper vector, and a represents the mode classification; The matrix that the proper vector of image multi-medium data forms is
Figure BDA00002621093900044
D2 is the dimension of proper vector, and n is the number of proper vector, and b represents the mode classification; The class label matrix is
Figure BDA00002621093900045
N is the number of proper vector, and c is the number of mode classification.
Described step S3 further may further comprise the steps:
Step S31, set up objective function based on the proper vector of different modalities data in the described training set:
min U a , U b 1 2 ( | | X a T U a - Y | | F 2 + | | X b T U b - Y | | F 2 ) + λ 1 ( | | U a | | 21 + | | U b | | 21 ) + λ 2 | | [ X a T U a X b T U b ] | | * - - - ( 4 )
Wherein, U aAnd U bBe two mapping matrixes corresponding with text and image difference, λ 1And λ 2Be undetermined parameter, span is 0.0001 to 0.1;
Step S32, find the solution described objective function (such as passing through iterative algorithm) and obtain the mapping matrix corresponding with described different modalities difference:
Described step S32 further may further comprise the steps:
Step S321 arranges mapping matrix U aAnd U bInitial value be null matrix, maximum iteration time k is set, and the initial value that iterations is set is 1;
Step S322 is to matrix
Figure BDA00002621093900047
Carry out Eigenvalues Decomposition, obtain Eigenvalues Decomposition result: VDiag (d) V T, wherein, V is with matrix
Figure BDA00002621093900048
Feature value vector be the matrix that column vector forms; D is the vector that feature value vector characteristic of correspondence value forms.
Step S323 sets intermediate variable S -1=VDiag (c) V T, wherein, c is a vector, its i element is
Figure BDA00002621093900049
Wherein, d iBe i the element of vectorial d, μ is a minimum number.
Step S324 is according to i the element p of following formula calculating intermediate vector p and q iAnd q i:
p i = 1 2 | | u a i | | 2 2 + ϵ q i = 1 2 | | u b i | | 2 2 + ϵ - - - ( 5 )
And make P=Diag (p), and Q=Diag (q), wherein, ε is a minimum number.
Step S325 tries to achieve described mapping matrix U by two linear problems of finding the solution in the following formula aAnd U b:
( X a X a T + λ 1 P a + λ 2 X a S - 1 X a T ) U a = X a Y ( X b X b T + λ 1 P b + λ 2 X b S - 1 X b T ) U b = X b Y - - - ( 6 )
Whether step S326 judges the iterations of this moment less than maximum iteration time k, if then turn to described step S322 to continue iteration; If not, then stop iteration, obtain described mapping matrix U aAnd U b
Step S4, according to the mode classification of multi-medium data sample in the described test set, the mapping matrix that utilizes described step S3 to obtain is mapped to the same space with them;
Step S5, as query set, the multi-medium data of another mode classification is as object set with the multi-medium data of same mode classification in the test set that is mapped to behind the same space;
Step S6 for a multi-medium data in the described query set, according to the similarity between the multi-medium data, obtains described target tightening and its most similar multi-medium data, thereby obtains striding the media result for retrieval.
Similarity between the described multi-medium data can be measured with Euclidean distance, and for Euclidean distance, distance is nearer just to illustrate that these two multi-medium datas are more similar.
In order to verify the validity of the inventive method, next method proposed by the invention is applied to a multimedia database.The multi-medium data that comprises 10 semantic domains in this database, the text and the image that comprise 2866 pairs of identical semantic informations, choose at random wherein the 2173 pairs of texts and image pattern as training set, choose wherein the 693 pairs of texts and image pattern as test set, the concrete steps of using the inventive method are as follows:
1) extracts different proper vectors for the difference of the data based mode classification in the described database: for text modality, extract the LDA semantic feature of 10 dimensions, for image modalities, extract the SIFT feature of 128 dimensions;
2) with the proper vector composition characteristic matrix of proper vector and the view data of described training set Chinese version data, the structure objective function, learn to obtain two mapping matrixes corresponding with text and image by minimizing objective function, the learning process of described mapping matrix realizes by an iterative process;
3) according to the difference of described mode classification, the mapping matrix that the proper vector of the proper vector of the text data in the test set and view data is obtained by study is mapped in the same space;
4) with the text data in the test set as query set, view data is as object set, a given text document can access the image of target tightening the most similar with it (being that Euclidean distance is nearest).Fig. 2 strides the retrieval effectiveness synoptic diagram of media search method from the text to the image according to the present invention, Fig. 2 left side is two text documents, the right side be obtain according to the inventive method to these two images that text document is the most similar.
To sum up, the present invention proposes and a kind ofly effectively stride the media search method based on double space study.The present invention is easy to realize, stable performance, and recognition accuracy is high.The present invention proposes effectively to have crossed over the semantic gap between the different modalities media, strides media method relatively with traditional, and the present invention shows and uses widely prospect and larger marketable value.
Above-described specific embodiment; purpose of the present invention, technical scheme and beneficial effect are further described; institute is understood that; the above only is specific embodiments of the invention; be not limited to the present invention; within the spirit and principles in the present invention all, any modification of making, be equal to replacement, improvement etc., all should be included within protection scope of the present invention.

Claims (9)

  1. One kind based on double space study stride the media search method, it is characterized in that the method may further comprise the steps:
    Step S1, the multi-medium data sample of collection different modalities is set up and is striden media retrieval knowledge database, and described database is divided into training set and test set;
    Step S2 extracts the proper vector of different modalities multi-medium data sample in the described database;
    Step S3 obtains and described different modalities corresponding mapping matrix respectively based on the proper vector of the different modalities data in the described training set;
    Step S4, according to the mode classification of multi-medium data sample in the described test set, the mapping matrix that utilizes described step S3 to obtain is mapped to the same space with them;
    Step S5, as query set, the multi-medium data of another mode classification is as object set with the multi-medium data of same mode classification in the test set that is mapped to behind the same space;
    Step S6 for a multi-medium data in the described query set, according to the similarity between the multi-medium data, obtains described target tightening and its most similar multi-medium data, thereby obtains striding the media result for retrieval.
  2. 2. method according to claim 1 is characterized in that, described different modalities is bimodal.
  3. 3. method according to claim 2 is characterized in that, described bimodal is text and image.
  4. 4. method according to claim 3 is characterized in that, for text multi-medium data sample, extracts its hidden Di Lei Cray distribution characteristics vector; For image multi-medium data sample, extract its yardstick invariant features transform characteristics vector.
  5. 5. method according to claim 1 is characterized in that, described step S3 further may further comprise the steps:
    Step S31 sets up objective function based on the proper vector of different modalities data in the described training set;
    Step S32 finds the solution described objective function and obtains the mapping matrix corresponding with described different modalities difference.
  6. 6. method according to claim 5 is characterized in that, described objective function is:
    min U a , U b 1 2 ( | | X a T U a - Y | | F 2 + | | X b T U b - Y | | F 2 ) + λ 1 ( | | U a | | 21 + | | U b | | 21 ) + λ 2 | | [ X a T U a X b T U b ] | | * ,
    Wherein, U aAnd U bBe two mapping matrixes corresponding with the bimodal difference, λ 1And λ 2Be undetermined parameter, ‖ ‖ FFor asking for the operation of F-norm, ‖ ‖ 21For asking for vectorial two norms of row and operation, ‖ ‖ *For asking for the trace norm operation.
  7. 7. method according to claim 5 is characterized in that, carries out for finding the solution with iterative algorithm of described objective function.
  8. 8. method according to claim 7 is characterized in that, described step S32 further may further comprise the steps:
    Step S321 arranges mapping matrix U aAnd U bInitial value be null matrix, maximum iteration time k is set, the initial value of iterations is 1;
    Step S322 is to matrix
    Figure FDA00002621093800022
    Carry out Eigenvalues Decomposition, obtain Eigenvalues Decomposition result: VDiag (d) V T, wherein,
    Figure FDA00002621093800023
    Be the matrix that the proper vector of described training set Chinese version multi-medium data forms, d1 is the dimension of proper vector, and n is the number of proper vector, and a represents the mode classification; Be the matrix that the proper vector of image multi-medium data in the described training set forms, d2 is the dimension of proper vector, and n is the number of proper vector, and b represents the mode classification; V is with matrix
    Figure FDA00002621093800025
    Feature value vector be the matrix that column vector forms; D is the vector that feature value vector characteristic of correspondence value forms;
    Step S323 sets intermediate variable S -1=VDiag (c) V T, wherein, c is a vector, its i element is
    Figure FDA00002621093800026
    Wherein, d iBe i the element of vectorial d, μ is a minimum number;
    Step S324 is according to i the element p of following formula calculating intermediate vector p and q iAnd q i:
    p i = 1 2 | | u a i | | 2 2 + ϵ q i = 1 2 | | u b i | | 2 2 + ϵ ,
    And make P=Diag (p), and Q=Diag (q), wherein, ε is a minimum number;
    Step S325 tries to achieve described mapping matrix U by two linear problems of finding the solution in the following formula aAnd U b:
    ( X a X a T + λ 1 P a + λ 2 X a S - 1 X a T ) U a = X a Y ( X b X b T + λ 1 P b + λ 2 X b S - 1 X b T ) U b = X b Y ,
    Wherein,
    Figure FDA00002621093800031
    Be the class label matrix, n is the number of proper vector, and c is the number of mode classification;
    Whether step S326 judges the iterations of this moment less than maximum iteration time k, if then turn to described step S322 to continue iteration; If not, then stop iteration, obtain described mapping matrix U aAnd U b
  9. 9. method according to claim 1 is characterized in that, the similarity between the described multi-medium data is measured with Euclidean distance.
CN201210559081.XA 2012-12-20 2012-12-20 Based on the cross-media retrieval method of double space study Active CN103049526B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210559081.XA CN103049526B (en) 2012-12-20 2012-12-20 Based on the cross-media retrieval method of double space study

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210559081.XA CN103049526B (en) 2012-12-20 2012-12-20 Based on the cross-media retrieval method of double space study

Publications (2)

Publication Number Publication Date
CN103049526A true CN103049526A (en) 2013-04-17
CN103049526B CN103049526B (en) 2015-08-05

Family

ID=48062167

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210559081.XA Active CN103049526B (en) 2012-12-20 2012-12-20 Based on the cross-media retrieval method of double space study

Country Status (1)

Country Link
CN (1) CN103049526B (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103995903A (en) * 2014-06-12 2014-08-20 武汉科技大学 Cross-media search method based on isomorphic subspace mapping and optimization
CN105574133A (en) * 2015-12-15 2016-05-11 苏州贝多环保技术有限公司 Multi-mode intelligent question answering system and method
CN105701227A (en) * 2016-01-15 2016-06-22 北京大学 Cross-media similarity measure method and search method based on local association graph
CN105701225A (en) * 2016-01-15 2016-06-22 北京大学 Cross-media search method based on unification association supergraph protocol
CN106056043A (en) * 2016-05-19 2016-10-26 中国科学院自动化研究所 Animal behavior identification method and apparatus based on transfer learning
CN107169061A (en) * 2017-05-02 2017-09-15 广东工业大学 A kind of text multi-tag sorting technique for merging double information sources
CN107273517A (en) * 2017-06-21 2017-10-20 复旦大学 Picture and text cross-module state search method based on the embedded study of figure
WO2017210949A1 (en) * 2016-06-06 2017-12-14 北京大学深圳研究生院 Cross-media retrieval method
CN108399414A (en) * 2017-02-08 2018-08-14 南京航空航天大学 Method of Sample Selection and device
CN109284414A (en) * 2018-09-30 2019-01-29 中国科学院计算技术研究所 The cross-module state content search method and system kept based on semanteme
CN109992676A (en) * 2019-04-01 2019-07-09 中国传媒大学 Across the media resource search method of one kind and searching system
CN110851641A (en) * 2018-08-01 2020-02-28 杭州海康威视数字技术股份有限公司 Cross-modal retrieval method and device and readable storage medium
CN111639197A (en) * 2020-05-28 2020-09-08 山东大学 Cross-modal multimedia data retrieval method and system with label embedded online hash
CN116383342A (en) * 2023-04-07 2023-07-04 四川大学 Robust cross-domain text retrieval method under noise label

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107220337B (en) * 2017-05-25 2020-12-22 北京大学 Cross-media retrieval method based on hybrid migration network

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040267774A1 (en) * 2003-06-30 2004-12-30 Ibm Corporation Multi-modal fusion in content-based retrieval
CN1920818A (en) * 2006-09-14 2007-02-28 浙江大学 Transmedia search method based on multi-mode information convergence analysis
CN101021849A (en) * 2006-09-14 2007-08-22 浙江大学 Transmedia searching method based on content correlation
CN102521368A (en) * 2011-12-16 2012-06-27 武汉科技大学 Similarity matrix iteration based cross-media semantic digesting and optimizing method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040267774A1 (en) * 2003-06-30 2004-12-30 Ibm Corporation Multi-modal fusion in content-based retrieval
CN1920818A (en) * 2006-09-14 2007-02-28 浙江大学 Transmedia search method based on multi-mode information convergence analysis
CN101021849A (en) * 2006-09-14 2007-08-22 浙江大学 Transmedia searching method based on content correlation
CN102521368A (en) * 2011-12-16 2012-06-27 武汉科技大学 Similarity matrix iteration based cross-media semantic digesting and optimizing method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
KAIYE WANG 等: "Learning Coupled Feature Spaces for Cross-modal Matching", 《2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION》 *

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103995903B (en) * 2014-06-12 2017-04-12 武汉科技大学 Cross-media search method based on isomorphic subspace mapping and optimization
CN103995903A (en) * 2014-06-12 2014-08-20 武汉科技大学 Cross-media search method based on isomorphic subspace mapping and optimization
CN105574133A (en) * 2015-12-15 2016-05-11 苏州贝多环保技术有限公司 Multi-mode intelligent question answering system and method
CN105701225A (en) * 2016-01-15 2016-06-22 北京大学 Cross-media search method based on unification association supergraph protocol
CN105701227A (en) * 2016-01-15 2016-06-22 北京大学 Cross-media similarity measure method and search method based on local association graph
CN105701227B (en) * 2016-01-15 2019-02-01 北京大学 A kind of across media method for measuring similarity and search method based on local association figure
CN105701225B (en) * 2016-01-15 2019-02-01 北京大学 A kind of cross-media retrieval method based on unified association hypergraph specification
CN106056043A (en) * 2016-05-19 2016-10-26 中国科学院自动化研究所 Animal behavior identification method and apparatus based on transfer learning
CN106056043B (en) * 2016-05-19 2019-07-30 中国科学院自动化研究所 Animal behavior recognition methods and device based on transfer learning
WO2017210949A1 (en) * 2016-06-06 2017-12-14 北京大学深圳研究生院 Cross-media retrieval method
CN108399414B (en) * 2017-02-08 2021-06-01 南京航空航天大学 Sample selection method and device applied to cross-modal data retrieval field
CN108399414A (en) * 2017-02-08 2018-08-14 南京航空航天大学 Method of Sample Selection and device
WO2018145604A1 (en) * 2017-02-08 2018-08-16 南京航空航天大学 Sample selection method, apparatus and server
US10885390B2 (en) 2017-02-08 2021-01-05 Nanjing University Of Aeronautics And Astronautics Sample selection method and apparatus and server
CN107169061A (en) * 2017-05-02 2017-09-15 广东工业大学 A kind of text multi-tag sorting technique for merging double information sources
CN107273517A (en) * 2017-06-21 2017-10-20 复旦大学 Picture and text cross-module state search method based on the embedded study of figure
CN107273517B (en) * 2017-06-21 2021-07-23 复旦大学 Graph-text cross-modal retrieval method based on graph embedding learning
CN110851641A (en) * 2018-08-01 2020-02-28 杭州海康威视数字技术股份有限公司 Cross-modal retrieval method and device and readable storage medium
CN109284414B (en) * 2018-09-30 2020-12-04 中国科学院计算技术研究所 Cross-modal content retrieval method and system based on semantic preservation
CN109284414A (en) * 2018-09-30 2019-01-29 中国科学院计算技术研究所 The cross-module state content search method and system kept based on semanteme
CN109992676A (en) * 2019-04-01 2019-07-09 中国传媒大学 Across the media resource search method of one kind and searching system
CN109992676B (en) * 2019-04-01 2020-12-25 中国传媒大学 Cross-media resource retrieval method and retrieval system
CN111639197A (en) * 2020-05-28 2020-09-08 山东大学 Cross-modal multimedia data retrieval method and system with label embedded online hash
CN116383342A (en) * 2023-04-07 2023-07-04 四川大学 Robust cross-domain text retrieval method under noise label
CN116383342B (en) * 2023-04-07 2023-11-14 四川大学 Robust cross-domain text retrieval method under noise label

Also Published As

Publication number Publication date
CN103049526B (en) 2015-08-05

Similar Documents

Publication Publication Date Title
CN103049526B (en) Based on the cross-media retrieval method of double space study
Liu et al. Cross-modality binary code learning via fusion similarity hashing
Zhu et al. Exploring auxiliary context: discrete semantic transfer hashing for scalable image retrieval
Saleh et al. Toward automated discovery of artistic influence
Gordo Supervised mid-level features for word image representation
Wang et al. Learning compact hash codes for multimodal representations using orthogonal deep structure
Qian et al. Unsupervised feature selection for multi-view clustering on text-image web news data
Lu et al. Image classification by visual bag-of-words refinement and reduction
CN104834693A (en) Depth-search-based visual image searching method and system thereof
CN104463247A (en) Extracting method of optical spectrum vector cross-correlation features in hyper-spectral image classification
CN102663447A (en) Cross-media searching method based on discrimination correlation analysis
CN105930873A (en) Self-paced cross-modal matching method based on subspace
CN105808752A (en) CCA and 2PKNN based automatic image annotation method
CN103995903B (en) Cross-media search method based on isomorphic subspace mapping and optimization
Afzalan et al. An automated spectral clustering for multi-scale data
Zheng et al. MMDF-LDA: An improved Multi-Modal Latent Dirichlet Allocation model for social image annotation
Xu et al. A lightweight intrinsic mean for remote sensing classification with lie group kernel function
Jiang et al. A hierarchal BoW for image retrieval by enhancing feature salience
Jiang et al. Co-regularized PLSA for multi-view clustering
CN106844481A (en) Font similarity and font replacement method
CN103177121A (en) Locality preserving projection method for adding pearson relevant coefficient
Ahsan et al. Clustering social event images using kernel canonical correlation analysis
CN103886072A (en) Retrieved result clustering system in coal mine search engine
Han et al. Dayside aurora classification via BIFs-based sparse representation using manifold learning
Zhang et al. Large-scale clustering with structured optimal bipartite graph

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant