Class Similarity

本文详细介绍了Lucene搜索引擎中的打分机制,包括查询规范化、文档规范化、逆文档频率等核心概念,解释了如何通过这些因素计算文档的相关性得分。

[转载] http://lucene.apache.org/java/2_4_1/api/org/apache/lucene/search/Similarity.html


public abstract class Similarity

extends Object implements Serializable

 

Expert: Scoring API.

Subclasses implement search scoring.

The score of query q for document d correlates to the cosine-distance or dot-product between document and query vectors in a Vector Space Model (VSM) of Information Retrieval. A document whose vector is closer to the query vector in that model is scored higher. The score is computed as follows:

 

score(q,d)   =   coord(q,d)  ·  queryNorm(q)  · 

( tf(t in d)  ·  idf(t)2  ·  t.getBoost() ·  norm(t,d) )

 

t in q

 

where

  1. tf(t in d) correlates to the term's frequency , defined as the number of times term t appears in the currently scored document d . Documents that have more occurrences of a given term receive a higher score. The default computation for tf(t in d) in DefaultSimilarity is:

     

    tf(t in d)   =  

    frequency½

     

  2. idf(t) stands for Inverse Document Frequency. This value correlates to the inverse of docFreq (the number of documents in which the term t appears). This means rarer terms give higher contribution to the total score. The default computation for idf(t) in DefaultSimilarity is:

     

    idf(t)   =  

    1 + log (

    numDocs

    –––––––––

    docFreq+1

    )

     

  3. coord(q,d) is a score factor based on how many of the query terms are found in the specified document. Typically, a document that contains more of the query's terms will receive a higher score than another document with fewer query terms. This is a search time factor computed in coord(q,d) by the Similarity in effect at search time.  

  4. queryNorm(q) is a normalizing factor used to make scores between queries comparable. This factor does not affect document ranking (since all ranked documents are multiplied by the same factor), but rather just attempts to make scores from different queries (or even different indexes) comparable. This is a search time factor computed by the Similarity in effect at search time. The default computation in DefaultSimilarity is:

     

    queryNorm(q)   =   queryNorm(sumOfSquaredWeights)   =  

    1

    ––––––––––––––

    sumOfSquaredWeights½

     

    The sum of squared weights (of the query terms) is computed by the query Weight object. For example, a boolean query computes this value as:

     

    sumOfSquaredWeights   =   q.getBoost() 2  · 

    ( idf(t)  ·  t.getBoost() ) 2

     

    t in q

     

     

  5. t.getBoost() is a search time boost of term t in the query q as specified in the query text (see query syntax), or as set by application calls to setBoost() . Notice that there is really no direct API for accessing a boost of one term in a multi term query, but rather multi terms are represented in a query as multi TermQuery objects, and so the boost of a term in the query is accessible by calling the sub-query getBoost() .  

  6. norm(t,d) encapsulates a few (indexing time) boost and length factors:

    • Document boost - set by calling doc.setBoost() before adding the document to the index.

    • Field boost - set by calling field.setBoost() before adding the field to a document.

    • lengthNorm (field) - computed when the document is added to the index in accordance with the number of tokens of this field in the document, so that shorter fields contribute more to the score. LengthNorm is computed by the Similarity class in effect at indexing.

    When a document is added to the index, all the above factors are multiplied. If the document has multiple fields with the same name, all their boosts are multiplied together:

    norm(t,d)   =   doc.getBoost()  ·  lengthNorm(field)  · 

    f.getBoost ()

     

    field f in d named as t

     

     

    However the resulted norm value is encoded as a single byte before being stored. At search time, the norm byte value is read from the index directory and decoded back to a float norm value. This encoding/decoding, while reducing index size, comes with the price of precision loss - it is not guaranteed that decode(encode(x)) = x. For instance, decode(encode(0.89)) = 0.75. Also notice that search time is too late to modify this norm part of scoring, e.g. by using a different Similarity for search.  

 

 

See Also:
setDefault(Similarity) , IndexWriter.setSimilarity(Similarity) , Searcher.setSimilarity(Similarity) , Serialized Form

Constructor Summary

Similarity ()

 

 

Method Summary

abstract  float

coord (int overlap, int maxOverlap)

          Computes a score factor based on the fraction of all query terms that a document contains.

static float

decodeNorm (byte b)

          Decodes a normalization factor stored in an index.

static byte

encodeNorm (float f)

          Encodes a normalization factor for storage in an index.

static Similarity

getDefault ()

          Return the default Similarity implementation used by indexing and search code.

static float[]

getNormDecoder ()

          Returns a table for decoding normalization bytes.

 float

idf (Collection terms, Searcher searcher)

          Computes a score factor for a phrase.

abstract  float

idf (int docFreq, int numDocs)

          Computes a score factor based on a term's document frequency (the number of documents which contain the term).

 float

idf (Term term, Searcher searcher)

          Computes a score factor for a simple term.

abstract  float

lengthNorm (String fieldName, int numTokens)

          Computes the normalization value for a field given the total number of terms contained in a field.

abstract  float

queryNorm (float sumOfSquaredWeights)

          Computes the normalization value for a query given the sum of the squared weights of each of the query terms.

 float

scorePayload (String fieldName, byte[] payload, int offset, int length)

          Calculate a scoring factor based on the data in the payload.

static void

setDefault (Similarity similarity)

          Set the default Similarity implementation used by indexing and search code.

abstract  float

sloppyFreq (int distance)

          Computes the amount of a sloppy phrase match, based on an edit distance.

abstract  float

tf (float freq)

          Computes a score factor based on a term or phrase's frequency in a document.

 float

tf (int freq)

          Computes a score factor based on a term or phrase's frequency in a document.

 

Methods inherited from class java.lang.Object

clone, equals, finalize, getClass, hashCode, notify, notifyAll, toString, wait, wait, wait

 

Constructor Detail

Similarity

public Similarity ()

Method Detail

setDefault

public static void setDefault (Similarity similarity)

Set the default Similarity implementation used by indexing and search code.

 

See Also:
Searcher.setSimilarity(Similarity) , IndexWriter.setSimilarity(Similarity)

getDefault

public static Similarity getDefault ()

Return the default Similarity implementation used by indexing and search code.

This is initially an instance of DefaultSimilarity .

 

See Also:
Searcher.setSimilarity(Similarity) , IndexWriter.setSimilarity(Similarity)

decodeNorm

public static float decodeNorm (byte b)

Decodes a normalization factor stored in an index.

 

See Also:
encodeNorm(float)

getNormDecoder

public static float[] getNormDecoder ()

Returns a table for decoding normalization bytes.

 

See Also:
encodeNorm(float)

lengthNorm

public abstract float lengthNorm (String fieldName, int numTokens)

Computes the normalization value for a field given the total number of terms contained in a field. These values, together with field boosts, are stored in an index and multipled into scores for hits on each field by the search code.

Matches in longer fields are less precise, so implementations of this method usually return smaller values when numTokens is large, and larger values when numTokens is small.

That these values are computed under IndexWriter.addDocument(org.apache.lucene.document.Document) and stored then using encodeNorm(float) . Thus they have limited precision, and documents must be re-indexed if this method is altered.

 

Parameters:
fieldName - the name of the field
numTokens - the total number of tokens contained in fields named fieldName of doc .
Returns:
a normalization factor for hits on this field of this document
See Also:
AbstractField.setBoost(float)

queryNorm

public abstract float queryNorm (float sumOfSquaredWeights)

Computes the normalization value for a query given the sum of the squared weights of each of the query terms. This value is then multipled into the weight of each query term.

This does not affect ranking, but rather just attempts to make scores from different queries comparable.

 

Parameters:
sumOfSquaredWeights - the sum of the squares of query term weights
Returns:
a normalization factor for query weights

encodeNorm

public static byte encodeNorm (float f)

Encodes a normalization factor for storage in an index.

The encoding uses a three-bit mantissa, a five-bit exponent, and the zero-exponent point at 15, thus representing values from around 7x10^9 to 2x10^-9 with about one significant decimal digit of accuracy. Zero is also represented. Negative numbers are rounded up to zero. Values too large to represent are rounded down to the largest representable value. Positive values too small to represent are rounded up to the smallest positive representable value.

 

See Also:
AbstractField.setBoost(float) , SmallFloat

tf

public float tf (int freq)

Computes a score factor based on a term or phrase's frequency in a document. This value is multiplied by the idf(Term, Searcher) factor for each term in the query and these products are then summed to form the initial score for a document.

Terms and phrases repeated in a document indicate the topic of the document, so implementations of this method usually return larger values when freq is large, and smaller values when freq is small.

The default implementation calls tf(float) .

 

Parameters:
freq - the frequency of a term within a document
Returns:
a score factor based on a term's within-document frequency

sloppyFreq

public abstract float sloppyFreq (int distance)

Computes the amount of a sloppy phrase match, based on an edit distance. This value is summed for each sloppy phrase match in a document to form the frequency that is passed to tf(float) .

A phrase match with a small edit distance to a document passage more closely matches the document, so implementations of this method usually return larger values when the edit distance is small and smaller values when it is large.

 

Parameters:
distance - the edit distance of this sloppy phrase match
Returns:
the frequency increment for this match
See Also:
PhraseQuery.setSlop(int)

tf

public abstract float tf (float freq)

Computes a score factor based on a term or phrase's frequency in a document. This value is multiplied by the idf(Term, Searcher) factor for each term in the query and these products are then summed to form the initial score for a document.

Terms and phrases repeated in a document indicate the topic of the document, so implementations of this method usually return larger values when freq is large, and smaller values when freq is small.

 

Parameters:
freq - the frequency of a term within a document
Returns:
a score factor based on a term's within-document frequency

idf

public float idf (Term term, Searcher searcher)
throws IOException

Computes a score factor for a simple term.

The default implementation is:

   return idf(searcher.docFreq(term), searcher.maxDoc());
Note that Searcher.maxDoc() is used instead of IndexReader.numDocs() because it is proportional to Searcher.docFreq(Term) , i.e., when one is inaccurate, so is the other, and in the same direction.

 

Parameters:
term - the term in question
searcher - the document collection being searched
Returns:
a score factor for the term
Throws:
IOException

idf

public float idf (Collection terms, Searcher searcher)
throws IOException

Computes a score factor for a phrase.

The default implementation sums the idf(Term,Searcher) factor for each term in the phrase.

 

Parameters:
terms - the terms in the phrase
searcher - the document collection being searched
Returns:
a score factor for the phrase
Throws:
IOException

idf

public abstract float idf (int docFreq, int numDocs)

Computes a score factor based on a term's document frequency (the number of documents which contain the term). This value is multiplied by the tf(int) factor for each term in the query and these products are then summed to form the initial score for a document.

Terms that occur in fewer documents are better indicators of topic, so implementations of this method usually return larger values for rare terms, and smaller values for common terms.

 

Parameters:
docFreq - the number of documents which contain the term
numDocs - the total number of documents in the collection
Returns:
a score factor based on the term's document frequency

coord

public abstract float coord (int overlap, int maxOverlap)

Computes a score factor based on the fraction of all query terms that a document contains. This value is multiplied into scores.

The presence of a large portion of the query terms indicates a better match with the query, so implementations of this method usually return larger values when the ratio between these parameters is large and smaller values when the ratio between them is small.

 

Parameters:
overlap - the number of query terms matched in the document
maxOverlap - the total number of terms in the query
Returns:
a score factor based on term overlap with the query

scorePayload

public float scorePayload (String fieldName, byte[] payload, int offset, int length)

Calculate a scoring factor based on the data in the payload. Overriding implementations are responsible for interpreting what is in the payload. Lucene makes no assumptions about what is in the byte array.

The default implementation returns 1.

 

Parameters:
fieldName - The fieldName of the term this payload belongs to
payload - The payload byte array to be scored
offset - The offset into the payload array
length - The length in the array
Returns:
An implementation dependent float to be used as a scoring factor
采用PyQt5框架与Python编程语言构建图书信息管理平台 本项目基于Python编程环境,结合PyQt5图形界面开发库,设计实现了一套完整的图书信息管理解决方案。该系统主要面向图书馆、书店等机构的日常运营需求,通过模块化设计实现了图书信息的标准化管理流程。 系统架构采用典型的三层设计模式,包含数据存储层、业务逻辑层和用户界面层。数据持久化方案支持SQLite轻量级数据库与MySQL企业级数据库的双重配置选项,通过统一的数据库操作接口实现数据存取隔离。在数据建模方面,设计了包含图书基本信息、读者档案、借阅记录等核心数据实体,各实体间通过主外键约束建立关联关系。 核心功能模块包含六大子系统: 1. 图书编目管理:支持国际标准书号、中国图书馆分类法等专业元数据的规范化著录,提供批量导入与单条录入两种数据采集方式 2. 库存动态监控:实时追踪在架数量、借出状态、预约队列等流通指标,设置库存预警阈值自动提醒补货 3. 读者服务管理:建立完整的读者信用评价体系,记录借阅历史与违规行为,实施差异化借阅权限管理 4. 流通业务处理:涵盖借书登记、归还处理、续借申请、逾期计算等标准业务流程,支持射频识别技术设备集成 5. 统计报表生成:按日/月/年周期自动生成流通统计、热门图书排行、读者活跃度等多维度分析图表 6. 系统维护配置:提供用户权限分级管理、数据备份恢复、操作日志审计等管理功能 在技术实现层面,界面设计遵循Material Design设计规范,采用QSS样式表实现视觉定制化。通过信号槽机制实现前后端数据双向绑定,运用多线程处理技术保障界面响应流畅度。数据验证机制包含前端格式校验与后端业务规则双重保障,关键操作均设有二次确认流程。 该系统适用于中小型图书管理场景,通过可扩展的插件架构支持功能模块的灵活组合。开发过程中特别注重代码的可维护性,采用面向对象编程范式实现高内聚低耦合的组件设计,为后续功能迭代奠定技术基础。 资源来源于网络分享,仅用于学习交流使用,请勿用于商业,如有侵权请联系我删除!
### 相似度计算的概念与方法 相似度计算是一种衡量两个对象之间相似程度的技术,在自然语言处理、数据挖掘以及机器学习等领域具有广泛应用。以下是几种常见的相似度计算方法及其特点。 #### 1. 基于词频的相似度计算 通过比较两段文本中词语出现频率的不同,可以评估其内容上的相似性。这种方法的核心理念在于:如果两段文字使用的词汇越接近,则它们表达的意思也可能更相近[^1]。具体实现可以通过统计每一段文本中的单词数量并构建向量表示,随后利用距离度量(如欧几里得距离)或者角度差(如余弦相似度)来进行量化分析。 #### 2. Cell Similarity (CSIM) 算法 这是一种专门用于路径规划领域内的轨迹匹配技术,能够有效判断两条或多条移动物体行走线路之间的近似关系。下面给出了一段简化版 Python 实现代码作为参考: ```python import math def calculate_csim(traj1, traj2): """Calculate the CSIM between two trajectories.""" min_length = min(len(traj1), len(traj2)) similarity_sum = sum( math.exp(-abs(p1 - p2)) for p1, p2 in zip(traj1[:min_length], traj2[:min_length]) ) return similarity_sum / min_length # Example usage: trajectory_a = [0.1, 0.5, 0.9] trajectory_b = [0.2, 0.4, 0.8] csim_value = calculate_csim(trajectory_a, trajectory_b) print(f"The CSIM value is {csim_value:.4f}") ``` 此函数接受两个列表形式输入参数 `traj1` 和 `traj2` ,分别代表待对比的两条运动记录序列;返回值则介于 `[0,1]` 范围内数值越高表明两者越趋同[^2]。 #### 3. 余弦相似度 (Cosine Similarity) 作为一种经典的矢量空间模型下的测距方式,它主要关注的是方向而非大小差异。给定任意两个非零n维实数列A=(a₁,...an),B=(b₁,...bn),定义如下公式求解夹角θ余弦值得到最终结果S(A,B): \[ S(A,B)=\frac{\sum_{i=1}^{n}{a_i b_i}} {\sqrt{(\sum_{i=1}^n a_i ^2)} \cdot \sqrt{( \sum _{j=1 }^m b_j ^2 ) }} \] 其中分子部分为点乘运算结果而分母则是各自模长平方根相乘所得产物。当且仅当 A 平行 B 即完全一致时取最大可能值即等于正一(+1); 反之若反向排列则最小可达负一 (-1)[^3]. #### 4. Solr 中自定义 Similarity 类型配置 Apache Solr 提供灵活机制允许开发者替换默认 BM25 排序逻辑换成其他更适合特定应用场景需求的新方案比如 TF-IDF 。只需修改 schema.xml 文件相应节点即可完成切换操作: ```xml <fieldType name="text_general" class="solr.TextField"> <analyzer> <!-- Other analyzer settings --> </analyzer> </fieldType> <!-- Define custom similarity factory --> <similarity class="org.apache.solr.search.similarities.ClassicSimilarityFactory"/> ``` 上述 XML 片段展示了如何指定经典相似度工厂替代原有设置过程[^4]。 ---
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值