+

CN104077419A - Long inquiring image searching reordering algorithm based on semantic and visual information - Google Patents

Long inquiring image searching reordering algorithm based on semantic and visual information Download PDF

Info

Publication number
CN104077419A
CN104077419A CN201410346066.6A CN201410346066A CN104077419A CN 104077419 A CN104077419 A CN 104077419A CN 201410346066 A CN201410346066 A CN 201410346066A CN 104077419 A CN104077419 A CN 104077419A
Authority
CN
China
Prior art keywords
visual
formula
concept
image
visual concept
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201410346066.6A
Other languages
Chinese (zh)
Other versions
CN104077419B (en
Inventor
洪日昌
高鹏飞
汪萌
刘学亮
郝世杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hefei University of Technology
Original Assignee
Hefei University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hefei University of Technology filed Critical Hefei University of Technology
Priority to CN201410346066.6A priority Critical patent/CN104077419B/en
Publication of CN104077419A publication Critical patent/CN104077419A/en
Application granted granted Critical
Publication of CN104077419B publication Critical patent/CN104077419B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Library & Information Science (AREA)
  • Computational Linguistics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

本发明公开了一种结合语义与视觉信息的长查询图像检索重排序方法,其特征是按如下步骤进行:1.输入长查询语句获得初始返回列表;2.构建视觉词典;3.将长查询语句进行分割,提取视觉概念;4.由视觉概念分别获得各自的初始返回列表;5.提取文本特征和视觉特征;6.建立概率模型;7.语义相关性估计;8.视觉相关性估计;9.结合语义与视觉的相关性估计;10.重排序获得重排序结果。本发明能够充分利用图像特征信息,从而有效提高图像检索重排序的准确性。The invention discloses a long-query image retrieval reordering method combining semantic and visual information, which is characterized in that the steps are as follows: 1. Input a long query sentence to obtain an initial return list; 2. Build a visual dictionary; 3. Convert the long query Segment the sentence and extract the visual concept; 4. Obtain the respective initial return list from the visual concept; 5. Extract text features and visual features; 6. Establish a probability model; 7. Estimation of semantic correlation; 8. Estimation of visual correlation; 9. Combining semantic and visual correlation estimation; 10. Reordering to obtain reordering results. The invention can make full use of image feature information, thereby effectively improving the accuracy of image retrieval and reordering.

Description

Retrieve with the long query image of visual information the algorithm that reorders in conjunction with semantic
Technical field
The invention belongs to technical field of information retrieval, specifically a kind of combination is semantic retrieves method for reordering with the long query image of visual information.
Background technology
21 century is the information age, is accompanied by the development of Internet technology and network share service, and network epigraph data increase by geometric progression, and the retrieval of image has become an activity requisite in people's daily life.Along with the network user's retrieval behavior is more and more accurate, what query word became becomes increasingly complex, and complicated long inquiry can be expressed more specific and accurate information than simple queries.But the result for retrieval that existing network search engines returns for long inquiry has wrong sequence conventionally.Trace it to its cause, be mainly because: first, long inquiry is made up of multiple concepts, and this has just further expanded the semantic gap between text query word and vision content.Secondly,, because length is inquired about the rare of positive sample, cause the results of learning based on model poor.For the performance of improving retrieval is experienced and satisfaction to improve user, on initial search result, combining image characteristic information carries out retrieving result reordering and has become a popular research point.
Generally speaking, the characteristic information of image comprises the visual information of text message and the image of image.Existing web image search engine, depends on text matches definite between query statement and textual description, and the result that search is returned is easy to make user dissatisfied.At present, most of images algorithm that reorders adopts visual signature to reorder, and summary is got up, and can be divided into two class algorithms below: based on spurious correlation feedback and reordering based on figure.This two classes method for reordering all relies on visual signature and reorders.But much research points out, only using image vision information to reorder can not achieve satisfactory results.Meanwhile, in the time using long inquiry to retrieve, initial retrieval result is normally insecure, comes image and query word correlativity before initial retrieval result very low.
Summary of the invention
In order to overcome the deficiencies in the prior art, the present invention proposes a kind of combination long query image semantic and visual information and retrieves the algorithm that reorders, and can make full use of image feature information, thereby effectively improve the accuracy that image retrieval reorders.
The present invention is that technical solution problem adopts following technical scheme:
The feature that a kind of combination of the present invention long query image semantic and visual information is retrieved the algorithm that reorders is to carry out as follows:
Step 1, on search engine, input long query statement Q and carry out image retrieval, return to several long query image, choose the long query image that in described long query image, sequence is front N, grow query image by described top n and form initial return-list X={x 1, x 2..., x u..., x n, x ube illustrated in u long query image in described initial return-list, u represents described long query image x uposition in initial return-list is u, u=0, and 1 ..., N;
Step 2, utilize reptile instrument to obtain unique question and answer pair, and utilize part-of-speech tagging device to collect verb and the noun of described unique question and answer centering, and remove the stop words in described verb and noun, thereby build visual dictionary;
Step 3, utilize partition tools to cut apart described long query statement Q, obtain some statement blocks, and each statement block and described visual dictionary are compared, choose the statement block of the verb that includes in described visual dictionary or noun as visual concept; And form visual concept set C={q by τ visual concept 0, q 1..., q c..., q τ-1; q cbe illustrated in c visual concept in described visual concept set C, c=0,1 ..., τ-1;
Step 4, on search engine, respectively the each visual concept in described visual concept set C is carried out to image retrieval, return to several visual concept images corresponding with each visual concept, choose the visual concept image that in described visual concept image, sequence is front L, by described front L visual concept image construction sample set D={ (X 0; q 0), (X 1; q 1) ..., (X c; q c) ..., (X τ-1; q τ-1); And X 0=(x n+1, x n+2..., x n+L), X 1=(x n+L+1, x n+L+2..., x n+2L), X c=(x n+cL+1, x n+cL+2..., x n+cL+ ζ..., x n+ (c+1) L), X τ-1=(x n+ (τ-1) L+1, x n+ (τ-1) L+2..., x n+ τ L), X crepresent and described c visual concept q ccorresponding visual concept image collection; x n+cL+ ζrepresent with described c visual concept q cζ the visual concept image returning while carrying out image retrieval;
Step 5, described N long query image is extracted respectively to text feature and visual signature, obtain long query text characteristic set with long inquiry visual signature set F={f 1, f 2..., f u..., f n; represent u long query image x ulist of labels, and formed t by n label μrepresent μ label; f urepresent u long query image x uvisual signature;
Described sample set D is extracted to visual signature, obtain and a described front L Image Visual Feature that visual concept image is corresponding respectively; By the set of described Image Visual Feature constitutive characteristic represent and described c visual concept q ccorresponding visual concept image collection X cthe visual signature extracting; f n+cL+ ζrepresent with described c visual concept q cζ the visual concept image x returning while carrying out image retrieval n+cL+ ζcorresponding Image Visual Feature;
Step 6, utilize formula (1) to set up probability model Score (Q, x u):
Score ( Q , x u ) = Σ q c ∈ C P ( q c | Q ) log P ( q c | x u ) - - - ( 1 )
In formula (1), P (q c| Q) c visual concept q of expression cfor the significance level of described long query statement Q, P (q c| x u) c visual concept q of expression cwith described u long query image x urelevance;
Step 7, semantic dependency are estimated:
Step 7.1, utilize formula (2) to estimate the semantic dependency between any two visual concepts:
Sim(q i,q j)=Sim co(q i,q j)×Sim wd(q i,q j)×Sim wiki(q i,q j) (2)
In formula (2), Sim co(q i, q j) represent any two visual concept q iand q jbetween send out altogether frequency similarity, i, j ∈ 0,1 ..., τ-1, and have:
Sim co ( q i , q j ) = exp ( - max ( log f ( q i ) , log f ( q j ) ) - log f ( q i , q j ) log I - min ( log f ( q i ) , log f ( q j ) ) ) - - - ( 3 )
In formula (3), I represents total number of images all on described search engine; F (q i) and f (q j) be illustrated respectively in and on described search engine, input visual concept q iand q jthe rear visual concept total number of images of returning respectively; F (q i, q j) be illustrated in and on described search engine, input visual concept q simultaneously iand q jafter the total number of images returned;
In formula (2), Sim wd(q i, q j) represent any two visual concept q of obtaining by WordNet thesaurus tools iand q jbetween similarity, and have:
Sim wd ( q i , q j ) = # ( q i ) + # ( q j ) # ( words q j ) wd + # ( words q i ) wd - - - ( 4 )
In formula (4), # (q i) expression use visual concept q jafter inquiring about in described WordNet dictionary, visual concept q in the Query Result returning ithe number of times occurring; # (q j) expression use visual concept q iafter inquiring about in described WordNet dictionary, visual concept q in the Query Result returning jthe number of times occurring; represent to use visual concept q jafter inquiring about in described WordNet dictionary, the total number of word of the Query Result returning; represent to use visual concept q iafter inquiring about in described WordNet dictionary, the total number of word of the Query Result returning;
In formula (2), Sim wiki(q i, q j) represent any two visual concept q of obtaining by wikipedia iand q jbetween similarity, and have:
Sim wiki ( q i , q j ) = # ( q i ) + # ( q j ) # ( words q j ) wiki + # ( words q i ) wiki - - - ( 5 )
In formula (5), # (q i) expression use visual concept q jafter inquiring about in described wikipedia, visual concept q in the Query Result returning ithe number of times occurring; # (q j) expression use visual concept q iafter inquiring about in described wikipedia, visual concept q in the Query Result returning jthe number of times occurring; represent to use visual concept q jafter inquiring about in described wikipedia, the total number of word of the Query Result returning; represent to use visual concept q iafter inquiring about in described wikipedia, the total number of word of the Query Result returning;
Step 7.2, utilize formula (6) to obtain described long query statement Q and c visual concept q cbetween semantic dependency G (q c, Q):
G ( q c , Q ) = 1 τ Σ q j ∈ C Sim ( q c , q j ) - - - ( 6 )
Step 7.3, utilize formula (7) obtain c visual concept q cwith u long query image x ubetween correlativity G (q c, x u):
G ( q c , x u ) = Σ t μ ∈ T x u Sim ( q c , t μ ) | T x u | - - - ( 7 )
In formula (7), represent described u long query image x ulist of labels radix;
Step 8, visual correlation are estimated:
Step 8.1, utilize formula (8) to obtain described long query statement Q and c visual concept q cbetween visual correlation V (q c, Q):
V ( q c , Q ) = 1 | X c | × | X | Σ f N + cL + ζ ∈ F Xc , f u ∈ F K ( f N + cL + ζ , f u ) - - - ( 8 )
In formula (8), | X| represents the radix of described initial return-list X; | X c| represent described and described c visual concept q ccorresponding visual concept image collection X cradix; K (f n+cL+ ζ, f u) represent Gauss's similar function, and have:
K(f N+cL+ζ,f u)=exp(-||f N+cL+ζ-f u|| 22) (9)
In formula (9), δ is scale parameter;
Step 8.2, utilize formula (10) by described c visual concept q cwith u long query image x ubetween visual correlation V (q c, x u) further decompose:
V ( q c , x u ) = Σ ω = N + 1 N + τL P ( q c | x ω ) P ( x ω | x u ) - - - ( 10 )
In formula (10): x ωrepresent any one visual concept image in sample set D;
Step 8.3, based on markov Random Walk Algorithm, regard described N long query image and τ L visual concept image as node, set up symmetrical κ neighbour and scheme; Through type (11) obtains the connection weight W between φ node and ψ node φ ψ:
In formula (11), N κ (φ) represents the indexed set of the symmetrical κ neighbour figure of ψ the node calculating by Euclidean distance; N κ (ψ) represents the indexed set of the symmetrical κ neighbour figure of φ the node calculating by Euclidean distance; φ, ψ ∈ (0,1 ..., N+ τ L);
Represent a step transition probability matrix, the elements A in a described step transition probability matrix A with A ω urepresent to transfer to from ω node the probability of u node, A ω u=W ω u/ Σ ψw ω ψ; Utilize formula (12) to obtain from ω node and shift the probability P at u Nodes through s step s|0(x u| x ω):
P s|0(x u|x ω)=[A s] ωu (12)
Utilize formula (13) to obtain with described any one visual concept image x ωfor starting point stops at u long query image x through s step uthe conditional probability P at place 0|s(x ω| x u):
P 0 | s ( x ω | x u ) = P s | 0 ( x u | x ω ) P 0 ( x ω ) Σ ψ P s | 0 ( x u | x ψ ) P 0 ( x ψ ) - - - ( 13 )
Utilize P 0(x ω)=P 0(x ψ), formula (13) is rewritten as:
P 0 | s ( x ω | x u ) = = [ A s ] ωu Σ ψ [ A s ] ψu - - - ( 14 )
Step 8.4, travel through each the visual concept image in described sample set D, obtain any one visual concept image x ωwith c visual concept q cbetween relevance scores P (q c| x ω):
P ( q c | x ω ) = 1 Z Σ x N + cL + ζ ∈ X c P 0 | s ( x N + cL + ζ | x ω ) - - - ( 15 )
In formula (11), Z = Σ q c ∈ C Σ x N + cL + ζ ∈ X c P 0 | s ( x N + cL + ζ | x ω ) ;
Step 9: the correlation estimation in conjunction with semantic and vision:
Step 9.1, utilize formula (6) and formula (8), obtain c visual concept q cand final associated score P (q between long query statement Q c| Q):
P(q c|Q)=αV(q c,Q)+(1-α)G(q c,Q) (15)
In formula (12), α represents that balance semanteme and vision are to described final associated score P (q c| Q) parameter of significance level, α ∈ (0,1);
Step 9.2, utilize formula (7) and formula (10), obtain c visual concept q cwith u long query image x ubetween final associated score P (q c| x u):
P(q c|x u)=βV(q c,x u)+(1-β)G(q c,x u) (16)
In formula (13), β represents that balance semanteme and vision are to described final associated score P (q c| x u) parameter of significance level, β ∈ (0,1);
Step 10: the probability model Score (Q, the x that obtain according to formula (1) u) N long query image set X is reordered, thereby obtain the result that reorders of described N long query image.
Compared with the prior art, beneficial effect of the present invention is embodied in:
1, long query statement is considered as multiple visual concepts by the present invention, the result for retrieval of visual concept not only can be expressed the Partial Feature of the result for retrieval of the long inquiry of its composition, the accuracy rate of its retrieval is simultaneously high, and then promotes the correlation estimation accuracy between long inquiry and image;
2, the present invention is by building a probability model, analyze the correlativity of visual concept and long inquiry and initial retrieval result thereof, this method can not be subject to the impact of image sequence in initial retrieval result, overcome the defect that traditional images method for reordering relies on initial ranking results, can effectively promote the performance that reorders;
3, the present invention is in calculating relevance scores process, by text feature and visual signature in conjunction with utilization, adopt the method for linear combination, the semanteme calculating and visual correlation mark are combined, to overcome in image retrieval reorders characteristics of image utilization problem not fully;
4, the present invention, in the correlation process of calculating between concept, combines send out the altogether frequency and WordNet and three kinds of different resources of wikipedia of concept, thereby can estimate more accurately semantic dependency between concept.
Embodiment
In the present embodiment, a kind of combination is semantic retrieves with the long query image of visual information the algorithm that reorders, and is that the Search Results returning for image search engine is resequenced, and carries out as follows:
Step 1, on search engine, input long query statement Q and carry out image retrieval, long query statement has by several a natural language querying statement that the concept that is closely connected forms, for example, long inquiry " people dancing on the wedding " comprises three concepts, is respectively " people ", " dancing ", " wedding ", and be closely connected between these three concepts.From returning several long query image, choosing sequence in long query image is the long query image of front N, forms initial return-list X={x by the long query image of top n 1, x 2..., x u..., x n, x ube illustrated in u long query image in initial return-list, u represents long query image x uposition in initial return-list is u, u=0, and 1 ..., N;
Step 2, utilize reptile instrument to obtain the unique question and answer pair of dimension on base question and answer website, and utilize part-of-speech tagging device, as Part-Of-Speech Tagger collects verb and the noun of unique question and answer centering, and remove the stop words in verb and noun, thereby build visual dictionary;
Step 3, utilize partition tools, as openNLP instrument, long query statement Q is cut apart, obtain some statement blocks, and each statement block and visual dictionary are compared, choose the statement block of the verb that includes in visual dictionary or noun as visual concept; And form visual concept set C={q by τ visual concept 0, q 1..., q c..., q τ-1; q cbe illustrated in c visual concept in visual concept set C, c=0,1 ..., τ-1;
Step 4, on search engine, respectively the each visual concept in visual concept set C is carried out to image retrieval, return to several visual concept images corresponding with each visual concept, choose the visual concept image that in visual concept image, sequence is front L, by front L visual concept image construction sample set D={ (X 0; q 0), (X 1; q 1) ..., (X c; q c) ..., (X τ-1; q τ-1); And X 0=(x n+1, x n+2..., x n+L), X 1=(x n+L+1, x n+L+2..., x n+2L), X c=(x n+cL+1, x n+cL+2..., x n+cL+ ζ..., x n+ (c+1) L), X τ-1=(x n+ (τ-1) L+1, x n+ (τ-1) L+2..., x n+ τ L), X crepresent and c visual concept q ccorresponding visual concept image collection; x n+cL+ ζrepresent with c visual concept q cζ the visual concept image returning while carrying out image retrieval;
Step 5, N long query image is extracted respectively to text feature and visual signature, text feature extracts from image tag text, and visual signature uses the overall situation and local two kinds of features: the global characteristics of 428 dimensions comprises: the Wavelet Texture of the color moment of 225 dimensions, the marginal distribution histogram of 75 dimensions, 128 dimensions; Local feature refers to: use the key point of the every piece image of DOG function check, then extract SIFT feature at the regional area of these key points.Based on K means clustering method, the proper vector of extracting is carried out to cluster, obtain the code book of one 1000 dimension, the word bag histogram that finally obtains 1000 dimensions of every piece image represents.Thereby obtain long query text characteristic set with long inquiry visual signature set F={f 1, f 2..., f u..., f n; represent u long query image x ulist of labels, and formed t by n label μrepresent μ label; f urepresent u long query image x uvisual signature;
Sample set D is extracted to visual signature, obtain and front L the Image Visual Feature that visual concept image is corresponding respectively; By the set of Image Visual Feature constitutive characteristic represent and c visual concept q ccorresponding visual concept image collection X cthe visual signature extracting; f n+cL+ ζrepresent with c visual concept q cζ the visual concept image x returning while carrying out image retrieval n+cL+ ζcorresponding Image Visual Feature;
Step 6, employing associated score method, utilize formula (1) to set up probability model Score (Q, x u):
Score ( Q , x u ) = Σ q c ∈ C P ( q c | Q ) log P ( q c | x u ) - - - ( 1 )
In formula (1), P (q c| Q) c visual concept q of expression cfor the significance level of long query statement Q, P (q c| x u) c visual concept q of expression cwith u long query image x urelevance;
Step 7, semantic dependency are estimated:
Step 7.1, utilize formula (2) to estimate the semantic dependency between any two visual concepts:
Sim(q i,q j)=Sim co(q i,q j)×Sim wd(q i,q j)×Sim wiki(q i,q j) (2)
In formula (2), Sim co(q i, q j) represent any two visual concept q iand q jbetween send out altogether frequency similarity, i, j ∈ 0,1 ..., τ-1, and have:
Sim co ( q i , q j ) = exp ( - max ( log f ( q i ) , log f ( q j ) ) - log f ( q i , q j ) log I - min ( log f ( q i ) , log f ( q j ) ) ) - - - ( 3 )
In formula (3), I represents total number of images all on search engine; F (q i) and f (q j) be illustrated respectively in and on search engine, input visual concept q iand q jthe rear visual concept total number of images of returning respectively; F (q i, q j) be illustrated in and on search engine, input visual concept q simultaneously iand q jafter the total number of images returned;
In formula (2), Sim wd(q i, q j) represent any two visual concept q of obtaining by WordNet thesaurus tools iand q jbetween similarity, and have:
Sim wd ( q i , q j ) = # ( q i ) + # ( q j ) # ( words q j ) wd + # ( words q i ) wd - - - ( 4 )
In formula (4), # (q i) expression use visual concept q jafter inquiring about in WordNet dictionary, visual concept q in the Query Result returning ithe number of times occurring; # (q j) expression use visual concept q iafter inquiring about in WordNet dictionary, visual concept q in the Query Result returning jthe number of times occurring; represent to use visual concept q jafter inquiring about in WordNet dictionary, the total number of word of the Query Result returning; represent to use visual concept q iafter inquiring about in WordNet dictionary, the total number of word of the Query Result returning;
In formula (2), Sim wiki(q i, q j) represent any two visual concept q of obtaining by wikipedia iand q jbetween similarity, and have:
Sim wiki ( q i , q j ) = # ( q i ) + # ( q j ) # ( words q j ) wiki + # ( words q i ) wiki - - - ( 5 )
In formula (5), # (q i) expression use visual concept q jafter inquiring about in wikipedia, visual concept q in the Query Result returning ithe number of times occurring; # (q j) expression use visual concept q iafter inquiring about in wikipedia, visual concept q in the Query Result returning jthe number of times occurring; represent to use visual concept q jafter inquiring about in wikipedia, the total number of word of the Query Result returning; represent to use visual concept q iafter inquiring about in wikipedia, the total number of word of the Query Result returning;
Step 7.2, utilize formula (6) to obtain long query statement Q and c visual concept q cbetween semantic dependency G (q c, Q):
G ( q c , Q ) = 1 τ Σ q j ∈ C Sim ( q c , q j ) - - - ( 6 )
Step 7.3, adopt simple linear fusion method, so-called linear fusion, by different data values, is merged arrangement, thereby is obtained a new data value, and the new data value obtaining so more definitely also has more generality; Utilize formula (7) to obtain c visual concept q cwith u long query image x ubetween correlativity G (q c, x u):
G ( q c , x u ) = Σ t μ ∈ T x u Sim ( q c , t μ ) | T x u | - - - ( 7 )
In formula (7), represent u long query image x ulist of labels radix;
Step 8, visual correlation are estimated:
Step 8.1, utilize formula (8) to obtain long query statement Q and c visual concept q cbetween visual correlation V (q c, Q):
V ( q c , Q ) = 1 | X c | × | X | Σ f N + cL + ζ ∈ F Xc , f u ∈ F K ( f N + cL + ζ , f u ) - - - ( 8 )
In formula (8), | X| represents the radix of initial return-list X; | X c| represent and c visual concept q ccorresponding visual concept image collection X cradix; K (f n+cL+ ζ, f u) represent Gauss's similar function, and have:
K(f N+cL+ζ,f u)=exp(-||f N+cL+ζ-f u|| 22) (9)
In formula (9), δ is scale parameter, is made as the right Euclidean distance median of all images;
Step 8.2, utilize formula (10) by c visual concept q cwith u long query image x ubetween visual correlation V (q c, x u) further decompose:
V ( q c , x u ) = Σ ω = N + 1 N + τL P ( q c | x ω ) P ( x ω | x u ) - - - ( 10 )
In formula (10): x ωrepresent any one visual concept image in sample set D;
Step 8.3, based on markov Random Walk Algorithm, regard N long query image and τ L visual concept image as node, set up symmetrical κ neighbour and scheme; Through type (11) obtains the connection weight W between φ node and ψ node φ ψ:
In formula (11), N κ (φ) represents the indexed set of the symmetrical κ neighbour figure of ψ the node calculating by Euclidean distance; N κ (ψ) represents the indexed set of the symmetrical κ neighbour figure of φ the node calculating by Euclidean distance; φ, ψ ∈ (0,1 ..., N+ τ L);
Represent a step transition probability matrix, the elements A in a step transition probability matrix A with A ω urepresent to transfer to from ω node the probability of u node, A ω u=W ω u/ Σ ψw ω ψ; Utilize formula (12) to obtain from ω node and shift the probability P at u Nodes through s step s|0(x u| x ω):
P s|0(x u|x ω)=[A s] ωu (12)
Utilize formula (13) to obtain with any one visual concept image x ωfor starting point stops at u long query image x through s step uthe conditional probability P at place 0|s(x ω| x u):
P 0 | s ( x ω | x u ) = P s | 0 ( x u | x ω ) P 0 ( x ω ) Σ ψ P s | 0 ( x u | x ψ ) P 0 ( x ψ ) - - - ( 13 )
Markov random walk starting point is evenly random, utilizes P 0(x ω)=P 0(x ψ), formula (13) is rewritten as:
P 0 | s ( x ω | x u ) = = [ A s ] ωu Σ ψ [ A s ] ψu - - - ( 14 )
Step 8.4, based on Normalizing Relatedness Cross Concepts method, propose at " Partially labeled classification with Markov random walks " article in 2002, the method can be considered the contact between visual concept in same long inquiry simultaneously; Each visual concept image in traversal sample set D, obtains any one visual concept image x ωwith c visual concept q cbetween relevance scores P (q c| x ω):
P ( q c | x ω ) = 1 Z Σ x N + cL + ζ ∈ X c P 0 | s ( x N + cL + ζ | x ω ) - - - ( 15 )
In formula (11), Z is normalized factor, and has: Z = Σ q c ∈ C Σ x N + cL + ζ ∈ X c P 0 | s ( x N + cL + ζ | x ω ) ;
Step 9: the correlation estimation in conjunction with semantic and vision:
Step 9.1, adopt the method for linear combination, utilize formula (6) and formula (8), obtain c visual concept q cand final associated score P (q between long query statement Q c| Q):
P(q c|Q)=αV(q c,Q)+(1-α)G(q c,Q) (15)
In formula (12), α represents that balance semanteme and vision are to final associated score P (q c| Q) parameter of significance level, α ∈ (0,1), considers that visual correlation mark plays a major role, in the present embodiment, α=0.8;
Step 9.2, utilize formula (7) and formula (10), obtain c visual concept q cwith u long query image x ubetween final associated score P (q c| x u):
P(q c|x u)=βV(q c,x u)+(1-β)G(q c,x u) (16)
In formula (13), β represents that balance semanteme and vision are to final associated score P (q c| x u) parameter of significance level, β ∈ (0,1), considers that visual correlation mark plays a major role, in the present embodiment, and β=0.8;
Step 10: the probability model Score (Q, the x that obtain according to formula (1) u) N long query image set X reordered, mark, by descending sort, is generated to new sorted lists, thereby obtain the result that reorders of N long query image.

Claims (1)

1.一种结合语义与视觉信息的长查询图像检索重排序算法,其特征是按如下步骤进行:1. A long query image retrieval reordering algorithm combining semantics and visual information, characterized in that it proceeds as follows: 步骤1、在搜索引擎上,输入长查询语句Q进行图像检索,返回若干个长查询图像,选取所述长查询图像中排序为前N的长查询图像,由所述前N个长查询图像构成初始返回列表X={x1,x2,…,xu,…,xN},xu表示在所述初始返回列表中第u个长查询图像,u表示所述长查询图像xu在初始返回列表中的位置为第u个,u=0,1,…,N;Step 1. On the search engine, input a long query sentence Q to perform image retrieval, return several long query images, and select the long query images sorted as the top N in the long query images, which are composed of the first N long query images The initial return list X={x 1 , x 2 ,...,x u ,...,x N }, x u represents the uth long query image in the initial return list, u represents the long query image x u in The position in the initial return list is the uth, u=0,1,...,N; 步骤2、利用爬虫工具获得唯一问答对,并利用词性标注器收集所述唯一问答对中的动词和名词,并去除所述动词和名词里的停用词,从而构建视觉词典;Step 2, utilize crawler tool to obtain unique question and answer pair, and utilize part-of-speech tagger to collect the verb and noun in described unique question and answer pair, and remove the stop words in described verb and noun, thereby construct visual dictionary; 步骤3、利用分割工具对所述长查询语句Q进行分割,获得若干语句块,并将每个语句块与所述视觉词典进行比较,选取包含有所述视觉词典中的动词或名词的语句块作为视觉概念;并由τ个视觉概念构成视觉概念集合C={q0,q1,…,qc,…,qτ-1};qc表示在所述视觉概念集合C中第c个视觉概念,c=0,1,…,τ-1;Step 3, using a segmentation tool to segment the long query statement Q to obtain several statement blocks, and compare each statement block with the visual dictionary, and select a statement block containing verbs or nouns in the visual dictionary As a visual concept; and a visual concept set C={q 0 , q 1 ,...,q c ,...,q τ-1 } is composed of τ visual concepts; q c represents the cth in the visual concept set C visual concept, c = 0,1,...,τ-1; 步骤4、在搜索引擎上,分别对所述视觉概念集合C中的每个视觉概念进行图像检索,返回与每个视觉概念相对应的若干个视觉概念图像,选取所述视觉概念图像中排序为前L的视觉概念图像,由所述前L个视觉概念图像构成样本集合D={(X0;q0),(X1;q1),…,(Xc;qc),…,(Xτ-1;qτ-1)};且X0=(xN+1,xN+2,…,xN+L),X1=(xN+L+1,xN+L+2,…,xN+2L),Xc=(xN+cL+1,xN+cL+2,…,xN+cL+ζ,…,xN+(c+1)L),Xτ-1=(xN+(τ-1)L+1,xN+(τ-1)L+2,…,xN+τL),Xc表示与所述第c个视觉概念qc相对应的视觉概念图像集合;xN+cL+ζ表示以所述第c个视觉概念qc进行图像检索时所返回的第ζ个视觉概念图像;Step 4. On the search engine, perform image retrieval for each visual concept in the visual concept set C, return several visual concept images corresponding to each visual concept, and select the visual concept images sorted as For the previous L visual concept images, a sample set D={(X 0 ; q 0 ),(X 1 ;q 1 ),...,(X c ;q c ),..., (X τ-1 ; q τ-1 )}; and X 0 =(x N+1 ,x N+2 ,…,x N+L ), X 1 =(x N+L+1 ,x N+ L+2 ,...,x N+2L ), X c =(x N+cL+1 ,x N+cL+2 ,...,x N+cL+ζ ,...,x N+(c+1)L ) , X τ-1 =(x N+(τ-1)L+1 ,x N+(τ-1)L+2 ,…,x N+τL ), X c represents the c-th visual concept q c Corresponding set of visual concept images; xN +cL+ζ represents the ζth visual concept image returned when performing image retrieval with the c-th visual concept qc; 步骤5、对所述N个长查询图像分别提取文本特征和视觉特征,获得长查询文本特征集合和长查询视觉特征集合F={f1,f2,…,fu,…,fN};表示第u个长查询图像xu的标签列表,并由n个标签构成,tμ表示第μ个标签;fu表示第u个长查询图像xu的视觉特征;Step 5. Extracting text features and visual features from the N long query images respectively to obtain a set of long query text features And long query visual feature set F={f 1 ,f 2 ,…,f u ,…,f N }; Indicates the tag list of the u-th long query image x u , and consists of n tags, t μ denotes the μ-th tag; f u denotes the visual features of the u-th long query image x u ; 对所述样本集合D提取视觉特征,分别获得与所述前L个视觉概念图像相对应的图像视觉特征;由所述图像视觉特征构成特征集合 表示与所述第c个视觉概念qc相对应的视觉概念图像集合Xc所提取的视觉特征;fN+cL+ζ表示以所述第c个视觉概念qc进行图像检索时所返回的第ζ个视觉概念图像xN+cL+ζ相对应的图像视觉特征;Extracting visual features to the sample set D, respectively obtaining image visual features corresponding to the first L visual concept images; forming a feature set by the image visual features Indicates the visual features extracted from the visual concept image set Xc corresponding to the c-th visual concept qc ; The image visual features corresponding to the ζth visual concept image x N+cL+ζ ; 步骤6、利用式(1)建立概率模型Score(Q,xu):Step 6. Establish a probability model Score(Q,x u ) using formula (1): ScoreScore (( QQ ,, xx uu )) == ΣΣ qq cc ∈∈ CC PP (( qq cc || QQ )) loglog PP (( qq cc || xx uu )) -- -- -- (( 11 )) 式(1)中,P(q|cQ)表示第c个视觉概念qc对于所述长查询语句Q的重要程度,P(qc|xu)表示第c个视觉概念qc与所述第u个长查询图像xu的关联性;In formula (1), P(q| c Q) represents the importance of the c-th visual concept q c to the long query statement Q, and P(q c |x u ) represents the relationship between the c-th visual concept q c and all Describe the relevance of the uth long query image x u ; 步骤7、语义相关性估计:Step 7. Semantic relevance estimation: 步骤7.1、利用式(2)估计任意两个视觉概念之间的语义相关性:Step 7.1, use formula (2) to estimate the semantic correlation between any two visual concepts: Sim(qi,qj)=Simco(qi,qj)×Simwd(qi,qj)×Simwiki(qi,qj)   (2)Sim(q i ,q j )=Sim co (q i ,q j )×Sim wd (q i ,q j )×Sim wiki (q i ,q j ) (2) 式(2)中,Simco(qi,qj)表示任意两个视觉概念qi和qj之间的共发频率相似度,i,j∈0,1,…,τ-1,并有:In formula (2), Sim co (q i ,q j ) represents the co-occurrence frequency similarity between any two visual concepts q i and q j , i,j∈0,1,…,τ-1, and have: SimSim coco (( qq ii ,, qq jj )) == expexp (( -- maxmax (( loglog ff (( qq ii )) ,, loglog ff (( qq jj )) )) -- loglog ff (( qq ii ,, qq jj )) loglog II -- minmin (( loglog ff (( qq ii )) ,, loglog ff (( qq jj )) )) )) -- -- -- (( 33 )) 式(3)中,I表示所述搜索引擎上所有的图像总数;f(qi)和f(qj)分别表示在所述搜索引擎上输入视觉概念qi和qj后分别返回的视觉概念图像总数;f(qi,qj)表示在所述搜索引擎上同时输入视觉概念qi和qj后返回的图像总数;In formula (3), I represents the total number of images on the search engine; f(q i ) and f(q j ) represent the visual images returned after inputting visual concepts q i and q j respectively on the search engine The total number of concept images; f(q i , q j ) represents the total number of images returned after inputting visual concepts q i and q j simultaneously on the search engine; 式(2)中,Simwd(qi,qj)表示通过WordNet词典工具获得的任意两个视觉概念qi和qj之间的相似度,并有:In formula (2), Sim wd (q i ,q j ) represents the similarity between any two visual concepts q i and q j obtained through the WordNet dictionary tool, and has: SimSim wdwd (( qq ii ,, qq jj )) == ## (( qq ii )) ++ ## (( qq jj )) ## (( wordswords qq jj )) wdwd ++ ## (( wordswords qq ii )) wdwd -- -- -- (( 44 )) 式(4)中,#(qi)表示使用视觉概念qj在所述WordNet词典中进行查询后,返回的查询结果中视觉概念qi出现的次数;#(qj)表示使用视觉概念qi在所述WordNet词典中进行查询后,返回的查询结果中视觉概念qj出现的次数;表示使用视觉概念qj在所述WordNet词典中进行查询后,返回的查询结果的总字数;表示使用视觉概念qi在所述WordNet词典中进行查询后,返回的查询结果的总字数;In formula (4), #(q i ) represents the number of occurrences of visual concept q i in the returned query results after using visual concept q j to query in the WordNet dictionary; #(q j ) represents the use of visual concept q After i inquires in described WordNet dictionary, the number of times that visual concept q j occurs in the query result returned; After expressing to use visual concept qj to inquire in described WordNet dictionary, the total number of words of the query result that returns; Represents the total number of words of the query result returned after using the visual concept q i to inquire in the WordNet dictionary; 式(2)中,Simwiki(qi,qj)表示通过维基百科获得的任意两个视觉概念qi和qj之间的相似度,并有:In formula (2), Sim wiki (q i ,q j ) represents the similarity between any two visual concepts q i and q j obtained through Wikipedia, and has: SimSim wikiwiki (( qq ii ,, qq jj )) == ## (( qq ii )) ++ ## (( qq jj )) ## (( wordswords qq jj )) wikiwiki ++ ## (( wordswords qq ii )) wikiwiki -- -- -- (( 55 )) 式(5)中,#(qi)表示使用视觉概念qj在所述维基百科中进行查询后,返回的查询结果中视觉概念qi出现的次数;#(qj)表示使用视觉概念qi在所述维基百科中进行查询后,返回的查询结果中视觉概念qj出现的次数;表示使用视觉概念qj在所述维基百科中进行查询后,返回的查询结果的总字数;表示使用视觉概念qi在所述维基百科中进行查询后,返回的查询结果的总字数;In formula (5), #(q i ) represents the number of occurrences of visual concept q i in the returned query results after using visual concept q j to query in the Wikipedia; #(q j ) represents the use of visual concept q After i performs a query in said Wikipedia, the number of occurrences of visual concept qj in the returned query result; Indicates the total number of words in the query result returned after using the visual concept q j to query in the Wikipedia; Indicates the total number of words in the query result returned after using the visual concept q i to query in the Wikipedia; 步骤7.2、利用式(6)获得所述长查询语句Q与第c个视觉概念qc之间的语义相关性G(qc,Q):Step 7.2, using formula (6) to obtain the semantic correlation G(q c , Q) between the long query statement Q and the c-th visual concept q c : GG (( qq cc ,, QQ )) == 11 ττ ΣΣ qq jj ∈∈ CC SimSim (( qq cc ,, qq jj )) -- -- -- (( 66 )) 步骤7.3、利用式(7)获得第c个视觉概念qc与第u个长查询图像xu之间的相关性G(qc,xu):Step 7.3. Use formula (7) to obtain the correlation G(q c ,x u ) between the c-th visual concept q c and the u-th long query image x u : GG (( qq cc ,, xx uu )) == ΣΣ tt μμ ∈∈ TT xx uu SimSim (( qq cc ,, tt μμ )) || TT xx uu || -- -- -- (( 77 )) 式(7)中,表示所述第u个长查询图像xu的标签列表的基数;In formula (7), A list of labels representing the u-th long query image x u base of 步骤8、视觉相关性估计:Step 8. Visual correlation estimation: 步骤8.1、利用式(8)获得所述长查询语句Q与第c个视觉概念qc之间的视觉相关性V(qc,Q):Step 8.1, using formula (8) to obtain the visual correlation V(q c , Q) between the long query statement Q and the c-th visual concept q c : VV (( qq cc ,, QQ )) == 11 || Xx cc || ×× || Xx || ΣΣ ff NN ++ cLc ++ ζζ ∈∈ Ff XcXc ,, ff uu ∈∈ Ff KK (( ff NN ++ cLc ++ ζζ ,, ff uu )) -- -- -- (( 88 )) 式(8)中,|X|表示所述初始返回列表X的基数;|Xc|表示所述与所述第c个视觉概念qc相对应的视觉概念图像集合Xc的基数;K(fN+cL+ζ,fu)表示高斯相似函数,并有:In formula (8), |X| represents the cardinality of the initial return list X; |X c | represents the cardinality of the visual concept image set X c corresponding to the c-th visual concept q c ; K( f N+cL+ζ ,f u ) represents the Gaussian similarity function, and has: K(fN+cL+ζ,fu)=exp(-||fN+cL+ζ-fu||22)   (9)K(f N+cL+ζ ,f u )=exp(-||f N+cL+ζ -f u || 22 ) (9) 式(9)中,δ为尺度参数;In formula (9), δ is a scale parameter; 步骤8.2、利用式(10)将所述第c个视觉概念qc与第u个长查询图像xu之间的视觉相关性V(qc,xu)进一步分解:Step 8.2, use formula (10) to further decompose the visual correlation V(q c , x u ) between the c-th visual concept q c and the u-th long query image x u : VV (( qq cc ,, xx uu )) == ΣΣ ωω == NN ++ 11 NN ++ τLτL PP (( qq cc || xx ωω )) PP (( xx ωω || xx uu )) -- -- -- (( 1010 )) 式(10)中:xω表示样本集合D中任意一个视觉概念图像;In formula (10): x ω represents any visual concept image in the sample set D; 步骤8.3、基于马尔可夫随机游走算法,将所述N个长查询图像和τL个视觉概念图像看做节点,建立对称κ近邻图;则通过式(11)获得第φ个节点和第ψ个节点之间的连接权重WφψStep 8.3. Based on the Markov random walk algorithm, the N long query images and τL visual concept images are regarded as nodes, and a symmetrical κ neighbor graph is established; then the φth node and the ψth node are obtained through formula (11). The connection weight W φψ between nodes: 式(11)中,Nκ(φ)表示通过欧式距离计算的第ψ个节点的对称κ近邻图的索引集;Nκ(ψ)表示通过欧式距离计算的第φ个节点的对称κ近邻图的索引集;φ、ψ∈(0,1,…,N+τL);In formula (11), Nκ(φ) represents the index set of the symmetric κ neighbor graph of the ψth node calculated by the Euclidean distance; Nκ(ψ) represents the index of the symmetric κ neighbor graph of the φth node calculated by the Euclidean distance set; φ, ψ∈(0,1,...,N+τL); 用A表示一步转移概率矩阵,所述一步转移概率矩阵A中的元素Aωu表示从第ω个节点转移到第u个节点的概率,Aωu=WωuψWωψ;则利用式(12)获得从第ω个节点出发经过s步转移在第u个节点处的概率Ps|0(xu|xω):Use A to represent a one-step transition probability matrix, and the element A ωu in the one-step transition probability matrix A represents the probability of transferring from the ω-th node to the u-th node, A ωu =W ωu / Σψ W ωψ ; then use the formula ( 12) Obtain the probability P s|0 (x u |x ω ) of moving from the ω-th node to the u-th node after s steps: Ps|0(xu|xω)=[As]ωu   (12)P s|0 (x u |x ω )=[A s ] ωu (12) 利用式(13)获得以所述任意一个视觉概念图像xω为起点经过s步停止在第u个长查询图像xu处的条件概率P0|s(xω|xu):Use formula (13) to obtain the conditional probability P 0|s (x ω |x u ) of the u-th long query image x u after s steps starting from any one of the visual concept images x ω : PP 00 || sthe s (( xx ωω || xx uu )) == PP sthe s || 00 (( xx uu || xx ωω )) PP 00 (( xx ωω )) ΣΣ ψψ PP sthe s || 00 (( xx uu || xx ψψ )) PP 00 (( xx ψψ )) -- -- -- (( 1313 )) 利用P0(xω)=P0(xψ),将式(13)改写为:Using P 0 (x ω )=P 0 (x ψ ), rewrite formula (13) as: PP 00 || sthe s (( xx ωω || xx uu )) == == [[ AA sthe s ]] ωuωu ΣΣ ψψ [[ AA sthe s ]] ψuψu -- -- -- (( 1414 )) 步骤8.4、遍历所述样本集合D中的每一个视觉概念图像,获得任意一个视觉概念图像xω与第c个视觉概念qc之间的相关性分数P(qc|xω):Step 8.4, traverse each visual concept image in the sample set D, and obtain the correlation score P(q c |x ω ) between any visual concept image x ω and the c-th visual concept q c : PP (( qq cc || xx ωω )) == 11 ZZ ΣΣ xx NN ++ cLc ++ ζζ ∈∈ Xx cc PP 00 || sthe s (( xx NN ++ cLc ++ ζζ || xx ωω )) -- -- -- (( 1515 )) 式(11)中, Z = Σ q c ∈ C Σ x N + cL + ζ ∈ X c P 0 | s ( x N + cL + ζ | x ω ) ; In formula (11), Z = Σ q c ∈ C Σ x N + c + ζ ∈ x c P 0 | the s ( x N + c + ζ | x ω ) ; 步骤9:结合语义与视觉的相关性估计:Step 9: Combine semantic and visual correlation estimation: 步骤9.1、利用式(6)和式(8),获得第c个视觉概念qc和长查询语句Q之间的最终相关分数P(qc|Q):Step 9.1. Using formula (6) and formula (8), obtain the final correlation score P(q c |Q) between the c-th visual concept q c and the long query sentence Q: P(qc|Q)=αV(qc,Q)+(1-α)G(qc,Q)   (15)P(q c |Q)=αV(q c ,Q)+(1-α)G(q c ,Q) (15) 式(12)中,α表示权衡语义与视觉对所述最终相关分数P(qc|Q)重要程度的参数,α∈(0,1);In formula (12), α represents a parameter that weighs the importance of semantics and vision to the final correlation score P(q c |Q), α∈(0,1); 步骤9.2、利用式(7)和式(10),获得第c个视觉概念qc和第u个长查询图像xu之间的最终相关分数P(qc|xu):Step 9.2. Using formula (7) and formula (10), obtain the final correlation score P(q c | x u ) between the c-th visual concept q c and the u-th long query image x u : P(qc|xu)=βV(qc,xu)+(1-β)G(qc,xu)   (16)P(q c |x u )=βV(q c ,x u )+(1-β)G(q c ,x u ) (16) 式(13)中,β表示权衡语义与视觉对所述最终相关分数P(qc|xu)重要程度的参数,β∈(0,1);In formula (13), β represents a parameter weighing the importance of semantics and vision to the final correlation score P(q c |x u ), β∈(0,1); 步骤10:根据式(1)所获得的概率模型Score(Q,xu)对N个长查询图像集合X进行重排序,从而获得所述N个长查询图像的重排序结果。Step 10: Reorder the N long query image sets X according to the probability model Score(Q,x u ) obtained by formula (1), so as to obtain the reordering results of the N long query images.
CN201410346066.6A 2014-07-18 2014-07-18 With reference to semantic method for reordering is retrieved with the long query image of visual information Active CN104077419B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410346066.6A CN104077419B (en) 2014-07-18 2014-07-18 With reference to semantic method for reordering is retrieved with the long query image of visual information

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410346066.6A CN104077419B (en) 2014-07-18 2014-07-18 With reference to semantic method for reordering is retrieved with the long query image of visual information

Publications (2)

Publication Number Publication Date
CN104077419A true CN104077419A (en) 2014-10-01
CN104077419B CN104077419B (en) 2018-05-01

Family

ID=51598673

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410346066.6A Active CN104077419B (en) 2014-07-18 2014-07-18 With reference to semantic method for reordering is retrieved with the long query image of visual information

Country Status (1)

Country Link
CN (1) CN104077419B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105653701A (en) * 2015-12-31 2016-06-08 百度在线网络技术(北京)有限公司 Model generating method and device as well as word weighting method and device
CN109740599A (en) * 2019-01-03 2019-05-10 华东理工大学 An image description generation method based on visually enhanced gLSTM
CN110647612A (en) * 2019-09-18 2020-01-03 合肥工业大学 Visual conversation generation method based on double-visual attention network
CN114168839A (en) * 2020-09-11 2022-03-11 北京搜狗科技发展有限公司 Picture information processing method and device and electronic equipment
CN118364127A (en) * 2024-06-20 2024-07-19 中南大学 Home textile image retrieval and re-ranking method and device based on feature fusion

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120221572A1 (en) * 2011-02-24 2012-08-30 Nec Laboratories America, Inc. Contextual weighting and efficient re-ranking for vocabulary tree based image retrieval
CN103399951A (en) * 2013-08-19 2013-11-20 山东大学 Semi-supervised image reordering method with self-feedback characteristic based on heterogeneous diagram

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120221572A1 (en) * 2011-02-24 2012-08-30 Nec Laboratories America, Inc. Contextual weighting and efficient re-ranking for vocabulary tree based image retrieval
CN103399951A (en) * 2013-08-19 2013-11-20 山东大学 Semi-supervised image reordering method with self-feedback characteristic based on heterogeneous diagram

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
LIQIANG NIE等: "Harvesting Visual Concepts for Image Search with Complex Queries", 《2012 ACM》 *
RICHANG HONG等: "Multimedia encyclopedia construction by mining", 《SIGNAL PROCESSING》 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105653701A (en) * 2015-12-31 2016-06-08 百度在线网络技术(北京)有限公司 Model generating method and device as well as word weighting method and device
WO2017113592A1 (en) * 2015-12-31 2017-07-06 百度在线网络技术(北京)有限公司 Model generation method, word weighting method, apparatus, device and computer storage medium
CN105653701B (en) * 2015-12-31 2019-01-15 百度在线网络技术(北京)有限公司 Model generating method and device, word assign power method and device
US10565253B2 (en) 2015-12-31 2020-02-18 Baidu Online Network Technology (Beijing) Co., Ltd. Model generation method, word weighting method, device, apparatus, and computer storage medium
CN109740599A (en) * 2019-01-03 2019-05-10 华东理工大学 An image description generation method based on visually enhanced gLSTM
CN109740599B (en) * 2019-01-03 2024-05-14 华东理工大学 Image description generation method based on visual enhancement gLSTM
CN110647612A (en) * 2019-09-18 2020-01-03 合肥工业大学 Visual conversation generation method based on double-visual attention network
CN114168839A (en) * 2020-09-11 2022-03-11 北京搜狗科技发展有限公司 Picture information processing method and device and electronic equipment
CN118364127A (en) * 2024-06-20 2024-07-19 中南大学 Home textile image retrieval and re-ranking method and device based on feature fusion
CN118364127B (en) * 2024-06-20 2024-09-06 中南大学 Home textile image retrieval and re-ranking method and device based on feature fusion

Also Published As

Publication number Publication date
CN104077419B (en) 2018-05-01

Similar Documents

Publication Publication Date Title
Liu et al. Mining quality phrases from massive text corpora
CN105893523B (en) It is measured using the assessment of answer relevance ranking come the method for computational problem similarity
Li et al. Twiner: named entity recognition in targeted twitter stream
CN103778227B (en) A Method of Filtering Useful Images from Retrieval Images
CN105045875B (en) Personalized search and device
CN106649597B (en) Method for auto constructing is indexed after a kind of books book based on book content
CN103617157B (en) Based on semantic Text similarity computing method
CN108255813B (en) Text matching method based on word frequency-inverse document and CRF
CN108052593A (en) A kind of subject key words extracting method based on descriptor vector sum network structure
CN105653706A (en) Multilayer quotation recommendation method based on literature content mapping knowledge domain
CN105718579A (en) Information push method based on internet-surfing log mining and user activity recognition
CN101751455B (en) A Method of Automatically Generating Headlines Using Artificial Intelligence Technology
CN104834735A (en) A method for automatic extraction of document summaries based on word vectors
CN110633365A (en) A hierarchical multi-label text classification method and system based on word vectors
CN112148885A (en) Intelligent searching method and system based on knowledge graph
CN101944099A (en) Method for automatically classifying text documents by utilizing body
CN103886020B (en) A kind of real estate information method for fast searching
CN104199965A (en) Semantic information retrieval method
CN111061939B (en) Scientific research academic news keyword matching recommendation method based on deep learning
CN112989208B (en) Information recommendation method and device, electronic equipment and storage medium
CN103678412A (en) Document retrieval method and device
CN104077419A (en) Long inquiring image searching reordering algorithm based on semantic and visual information
CN114090861A (en) Education field search engine construction method based on knowledge graph
CN106708929A (en) Video program searching method and device
CN107239564A (en) A kind of text label based on supervision topic model recommends method

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
CB03 Change of inventor or designer information

Inventor after: Hong Richang

Inventor after: Gao Pengfei

Inventor after: Wang Meng

Inventor after: Liu Yiqun

Inventor after: Sun Maosong

Inventor after: Liu Xueliang

Inventor after: Hao Shijie

Inventor before: Hong Richang

Inventor before: Gao Pengfei

Inventor before: Wang Meng

Inventor before: Liu Xueliang

Inventor before: Hao Shijie

COR Change of bibliographic data
GR01 Patent grant
GR01 Patent grant
点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载