sphinx全文检索引擎

官网和文档:http://sphinxsearch.com/docs/

中文参考:http://www.sphinxsearch.org/

python sdk:http://pypi.python.org/pypi/sphinxapi

增量索引参考:http://blog.csdn.net/jianglei421/article/details/5431946

校友张晏的blog有写他对sphinx的使用:http://blog.s135.com/post/360/

理论:

1.sphinx支持Mysql协议,所以除了使用普通Sphinx SDK封装的API以外,还可以使用mysql客户端对searchd发起连接和查询(使用SphinxQL语言)

2.重新编译MYSQL可以把sphinx的Mysql存储引擎编译进去,这样可以创建sphinxSE数据表,用于查询的时候跟searchd通信,直接获得查询结果

3.命令行工具:indexer/searchd/search/...

4.增量索引需要注意的是:1>下次建增量索引的时候老的增量索引文件会被覆盖,所以应该在下次建增量索引之前把老的增量索引合并进全量索引中或者合并到某个指定索引中(然后定时合并到全量索引)  2>每次建增量索引时需要把该次增量的标记记下,下次增量从该标记开始即可,如此反复...不记录标记的话增量需要处理的数据会原来越大越来越慢,最终失去增量在时效上的意义

5.中文分词需要注意的问题:做好配置的情况下(配置如下所示,参考自http://www.sphinxsearch.org/sphinx-tutorial),sphinx是可以对中文做单个汉字的切词和索引的;如果需要根据语义做的词语分词就需要安装一些中文分词的插件了(比如sfc神马的)

ngram_len = 1 # 对于非字母型数据的长度切割
ngram_chars = U+4E00..U+9FBF, U+3400..U+4DBF, U+20000..U+2A6DF, U+F900..U+FAFF,\
U+2F800..U+2FA1F, U+2E80..U+2EFF, U+2F00..U+2FDF, U+3100..U+312F, U+31A0..U+31BF,\
U+3040..U+309F, U+30A0..U+30FF, U+31F0..U+31FF, U+AC00..U+D7AF, U+1100..U+11FF,\
U+3130..U+318F, U+A000..U+A48F, U+A490..U+A4CF

6.对于json字符串建立索引:今天(2013.1.18)业务上有个需求是对数据表中存储json串建全文索引,那么就有两种情况了,一种是内部utf8编码的utf8 json串、另一种是对内部unicode编码直接加\转义的utf8 json串。前者和我们一般的处理方式没什么区别;后者需要把搜索词也做unicode转义然后进行搜索

下面是菜鸟的第一次实战:

1.安装(sphinx的rpm包不支持自定义安装路径--prefix,error: package sphinx is not relocatable)

[dongsong@bogon ~]$ sudo rpm -i sphinx-2.0.4-1.rhel6.x86_64.rpm

Sphinx installed!
Now create a full-text index, start the search daemon, and you're all set.

To manage indexes:
    editor /etc/sphinx/sphinx.conf

To rebuild all disk indexes:
    sudo -u sphinx indexer --all --rotate

To start/stop search daemon:
    service searchd start/stop

To query search daemon using MySQL client:
    mysql -h 0 -P 9306
    mysql> SELECT * FROM test1 WHERE MATCH('test');

See the manual at /usr/share/doc/sphinx-2.0.4 for details.

For commercial support please contact Sphinx Technologies Inc at
http://sphinxsearch.com/contacts.html

可以做一个符号链接把本地html手册链接到apache目录下,方便本地查看帮助

[dongsong@bogon python_study]$ sudo ln -sf /usr/share/doc/sphinx-2.0.4/sphinx.html ./sphinx.html
http://172.26.16.100/sphinx.html

2.使用

[root@bogon sphinx]# indexer --config /etc/sphinx/sphinx.conf spider
Sphinx 2.0.4-id64-release (r3135)
Copyright (c) 2001-2012, Andrew Aksyonoff
Copyright (c) 2008-2012, Sphinx Technologies Inc (http://sphinxsearch.com)

using config file '/etc/sphinx/sphinx.conf'...
indexing index 'spider'...
WARNING: attribute 'id' not found - IGNORING
WARNING: Attribute count is 0: switching to none docinfo
collected 20011 docs, 115.0 MB
sorted 5.4 Mhits, 100.0% done
total 20011 docs, 115049820 bytes
total 33.003 sec, 3486001 bytes/sec, 606.33 docs/sec
total 2 reads, 0.123 sec, 25973.1 kb/call avg, 61.9 msec/call avg
total 188 writes, 2.964 sec, 585.8 kb/call avg, 15.7 msec/call avg

[root@bogon sphinx]# search -c /etc/sphinx/sphinx.conf 中国
Sphinx 2.0.4-id64-release (r3135)
Copyright (c) 2001-2012, Andrew Aksyonoff
Copyright (c) 2008-2012, Sphinx Technologies Inc (http://sphinxsearch.com)

using config file '/etc/sphinx/sphinx.conf'...
index 'spider': query '中国 ': returned 150 matches of 150 total in 0.000 sec

displaying matches:
1. document=719806, weight=2654
2. document=1397236, weight=2654
3. document=3733569, weight=1729
4. document=13384, weight=1722
5. document=3563788, weight=1705
6. document=3742995, weight=1705
7. document=17777, weight=1698
8. document=3741757, weight=1698
9. document=3888109, weight=1698
10. document=2472909, weight=1689
11. document=3741705, weight=1689
12. document=2145250, weight=1676
13. document=2600863, weight=1676
14. document=3561074, weight=1676
15. document=3737639, weight=1676
16. document=3746591, weight=1676
17. document=3805049, weight=1676
18. document=1822, weight=1654
19. document=7755, weight=1654
20. document=13399, weight=1654

words:
1. '中国': 150 documents, 237 hits
3.search可以搜到数据,api搜索报错(api搜索需要启动searchd进程,searchd启动过程没有报错,但是并没有在指定端口上监听,也没有实际的searchd进程存在)

[dongsong@bogon api]$ vpython test.py -h localhost -p 9312 -i spider 中国
query failed: connection to localhost;9312 failed ([Errno 111] Connection refused)
在/etc/sphinx/sphinx.conf中查看searchd的日志文件位置

searchd
{
        listen                  = 9312
        listen                  = 9306:mysql41
        log                     = /var/log/sphinx/searchd.log
        query_log               = /var/log/sphinx/query.log
        read_timeout            = 5
        max_children            = 30
        pid_file                = /var/run/sphinx/searchd.pid
        max_matches             = 1000
        seamless_rotate         = 1
        preopen_indexes         = 1
        unlink_old              = 1
        workers                 = threads # for RT to work
        binlog_path            = /var/data
}
打开日志文件/var/log/sphinx/search.log找到问题的根源
[Fri Jun 15 10:28:44.583 2012] [ 7889] listening on all interfaces, port=9312
[Fri Jun 15 10:28:44.583 2012] [ 7889] listening on all interfaces, port=9306
[Fri Jun 15 10:28:44.585 2012] [ 7889] FATAL: failed to open '/var/data/binlog.lock': 2 'No such file or directory'
[Fri Jun 15 10:28:44.585 2012] [ 7888] Child process 7889 has been forked
[Fri Jun 15 10:28:44.585 2012] [ 7888] Child process 7889 has been finished, exit code 1. Watchdog finishes also. Good bye!
[Fri Jun 15 10:29:09.968 2012] [ 7905] Child process 7906 has been forked
[Fri Jun 15 10:29:09.970 2012] [ 7906] listening on all interfaces, port=9312
[Fri Jun 15 10:29:09.970 2012] [ 7906] listening on all interfaces, port=9306
[Fri Jun 15 10:29:09.987 2012] [ 7906] FATAL: failed to open '/var/data/binlog.lock': 2 'No such file or directory'
[Fri Jun 15 10:29:09.993 2012] [ 7905] Child process 7906 has been finished, exit code 1. Watchdog finishes also. Good bye!

把binlog_path的配置注释掉就ok了

[dongsong@bogon api]$ vpython test.py -h localhost -p 9312 -i spider 中国      
Query '中国 ' retrieved 3 of 3 matches in 0.005 sec
Query stats:
        '中国' found 4 times in 3 documents

Matches:
1. doc_id=5, weight=100
2. doc_id=80, weight=100
3. doc_id=2012, weight=100

4.对于中文数据的检索,不在conf的index里面设置下面这项就搜不到中文

charset_table = U+FF10..U+FF19->0..9, 0..9, U+FF41..U+FF5A->a..z, U+FF21..U+FF3A->a..z,\
A..Z->a..z, a..z, U+0149, U+017F, U+0138, U+00DF, U+00FF, U+00C0..U+00D6->U+00E0..U+00F6,\
U+00E0..U+00F6, U+00D8..U+00DE->U+00F8..U+00FE, U+00F8..U+00FE, U+0100->U+0101, U+0101,\
U+0102->U+0103, U+0103, U+0104->U+0105, U+0105, U+0106->U+0107, U+0107, U+0108->U+0109,\
U+0109, U+010A->U+010B, U+010B, U+010C->U+010D, U+010D, U+010E->U+010F, U+010F,\
U+0110->U+0111, U+0111, U+0112->U+0113, U+0113, U+0114->U+0115, U+0115, \
U+0116->U+0117,U+0117, U+0118->U+0119, U+0119, U+011A->U+011B, U+011B, U+011C->U+011D,\
 U+011D,U+011E->U+011F, U+011F, U+0130->U+0131, U+0131, U+0132->U+0133, U+0133, \
U+0134->U+0135,U+0135, U+0136->U+0137, U+0137, U+0139->U+013A, U+013A, U+013B->U+013C, \
U+013C,U+013D->U+013E, U+013E, U+013F->U+0140, U+0140, U+0141->U+0142, U+0142, \
U+0143->U+0144,U+0144, U+0145->U+0146, U+0146, U+0147->U+0148, U+0148, U+014A->U+014B, \
U+014B,U+014C->U+014D, U+014D, U+014E->U+014F, U+014F, U+0150->U+0151, U+0151, \
U+0152->U+0153,U+0153, U+0154->U+0155, U+0155, U+0156->U+0157, U+0157, U+0158->U+0159,\
 U+0159,U+015A->U+015B, U+015B, U+015C->U+015D, U+015D, U+015E->U+015F, U+015F, \
U+0160->U+0161,U+0161, U+0162->U+0163, U+0163, U+0164->U+0165, U+0165, U+0166->U+0167, \
U+0167,U+0168->U+0169, U+0169, U+016A->U+016B, U+016B, U+016C->U+016D, U+016D, \
U+016E->U+016F,U+016F, U+0170->U+0171, U+0171, U+0172->U+0173, U+0173, U+0174->U+0175,\
 U+0175,U+0176->U+0177, U+0177, U+0178->U+00FF, U+00FF, U+0179->U+017A, U+017A, \
U+017B->U+017C,U+017C, U+017D->U+017E, U+017E, U+0410..U+042F->U+0430..U+044F, \
U+0430..U+044F,U+05D0..U+05EA, U+0531..U+0556->U+0561..U+0586, U+0561..U+0587, \
U+0621..U+063A, U+01B9,U+01BF, U+0640..U+064A, U+0660..U+0669, U+066E, U+066F, \
U+0671..U+06D3, U+06F0..U+06FF,U+0904..U+0939, U+0958..U+095F, U+0960..U+0963, \
U+0966..U+096F, U+097B..U+097F,U+0985..U+09B9, U+09CE, U+09DC..U+09E3, U+09E6..U+09EF, \
U+0A05..U+0A39, U+0A59..U+0A5E,U+0A66..U+0A6F, U+0A85..U+0AB9, U+0AE0..U+0AE3, \
U+0AE6..U+0AEF, U+0B05..U+0B39,U+0B5C..U+0B61, U+0B66..U+0B6F, U+0B71, U+0B85..U+0BB9, \
U+0BE6..U+0BF2, U+0C05..U+0C39,U+0C66..U+0C6F, U+0C85..U+0CB9, U+0CDE..U+0CE3, \
U+0CE6..U+0CEF, U+0D05..U+0D39, U+0D60,U+0D61, U+0D66..U+0D6F, U+0D85..U+0DC6, \
U+1900..U+1938, U+1946..U+194F, U+A800..U+A805,U+A807..U+A822, U+0386->U+03B1, \
U+03AC->U+03B1, U+0388->U+03B5, U+03AD->U+03B5,U+0389->U+03B7, U+03AE->U+03B7, \
U+038A->U+03B9, U+0390->U+03B9, U+03AA->U+03B9,U+03AF->U+03B9, U+03CA->U+03B9, \
U+038C->U+03BF, U+03CC->U+03BF, U+038E->U+03C5,U+03AB->U+03C5, U+03B0->U+03C5, \
U+03CB->U+03C5, U+03CD->U+03C5, U+038F->U+03C9,U+03CE->U+03C9, U+03C2->U+03C3, \
U+0391..U+03A1->U+03B1..U+03C1,U+03A3..U+03A9->U+03C3..U+03C9, U+03B1..U+03C1, \
U+03C3..U+03C9, U+0E01..U+0E2E,U+0E30..U+0E3A, U+0E40..U+0E45, U+0E47, U+0E50..U+0E59, \
U+A000..U+A48F, U+4E00..U+9FBF,U+3400..U+4DBF, U+20000..U+2A6DF, U+F900..U+FAFF, \
U+2F800..U+2FA1F, U+2E80..U+2EFF,U+2F00..U+2FDF, U+3100..U+312F, U+31A0..U+31BF, \
U+3040..U+309F, U+30A0..U+30FF,U+31F0..U+31FF, U+AC00..U+D7AF, U+1100..U+11FF, \
U+3130..U+318F, U+A000..U+A48F,U+A490..U+A4CF
charset就不说了吧....
charset_type            = utf-8

5.处理增量

[root@bogon sphinx]# indexer --config /etc/sphinx/sphinx.conf spiderinc --rotate
Sphinx 2.0.4-id64-release (r3135)
Copyright (c) 2001-2012, Andrew Aksyonoff
Copyright (c) 2008-2012, Sphinx Technologies Inc (http://sphinxsearch.com)

using config file '/etc/sphinx/sphinx.conf'...
indexing index 'spiderinc'...
WARNING: attribute 'id' not found - IGNORING
WARNING: Attribute count is 0: switching to none docinfo
collected 17 docs, 0.1 MB
sorted 0.0 Mhits, 100.0% done
total 17 docs, 87216 bytes
total 0.060 sec, 1444643 bytes/sec, 281.58 docs/sec
total 2 reads, 0.000 sec, 23.4 kb/call avg, 0.0 msec/call avg
total 6 writes, 0.008 sec, 16.9 kb/call avg, 1.4 msec/call avg
rotating indices: succesfully sent SIGHUP to searchd (pid=10459).
6.对于正在对外提供服务的索引(searchd已经征用的索引),调用indexer创建索引的时候加上--rotate可以不打断其服务(建立新的索引然后覆盖旧的,如不加会创建索引失败)

   如果第一次建立索引就用indexer --rotate则会失败(因为没有老版本索引可以覆盖)

7.示例代码

#查询
	cl = SphinxClient()
	cl.SetServer ( host, port )
	cl.SetWeights ( [100, 1] )
	cl.SetMatchMode ( mode )
	if filtervals:
		cl.SetFilter ( filtercol, filtervals )
	if groupby:
		cl.SetGroupBy ( groupby, SPH_GROUPBY_ATTR, groupsort )
	if sortby:
		cl.SetSortMode ( SPH_SORT_ATTR_DESC, sortby )
	#cl.SetLimits ( offset, limit, limit+offset ) #如果过于追求效率可以把启用该行 命中必须的数据后立马返回 markbyxds 
	cl.SetLimits ( offset, limit)
	cl.SetConnectTimeout(60.0)
	res = cl.Query ( query, index )
	if not res:
		return HttpResponse(json.dumps({'page':page, 'count':count, 'total':0, 'datas':[]}))

	#去数据库取实际数据
	ids = [match['id'] for match in res['matches']]
	rawDatas = RawData.objects.filter(id__in = ids).order_by('-create_time')
	response = {'page':page, 'count':count, 'total':res['total'], 'datas':[]}
	response['datas'] = construct_response_data(rawDatas)
	
	#对内容生成高亮摘要
	'''
	cl.SetConnectTimeout(5.0)
	bodyDatas = [tmpData['data'] for tmpData in response['datas']]
	try:
		excerpts = cl.BuildExcerpts(bodyDatas,
									'spider',
									query,
									{'before_match':'<span "style":"color:red;">',
									'after_match':'</span>',
									'query_mode':mode}) 
	except Exception,e:
		import pdb
		pdb.set_trace()
		pass
	listIndex = 0
	for excerpt in excerpts:
		response['datas'][listIndex]['data'] = excerpt.decode('utf-8')
		listIndex += 1
	'''
	cl.SetConnectTimeout(0.1)
	for listIndex in range(len(response['datas'])):
		tmpData = response['datas'][listIndex]['data']
		for retry in range(3):
			try:
				excerpt = cl.BuildExcerpts([tmpData],
											'spider', #用多个索引返回空list spider;spiderinc markbyxds 
											query,
											{'before_match':'<span style="color:red;">',
											'after_match':'</span>',
											'query_mode':mode})
			except Exception,e:
				logging.error("%s:%s" % (type(e),str(e)))
				excerpt = None
				break
			else:
				if excerpt != None:
					response['datas'][listIndex]['data'] = excerpt[0].decode('utf-8')
					break
				else:
					logging.warning('return none . (timeout is too small), to retrying。。。')
					continue
		if excerpt == None:
			snippetLen = 1024
			response['datas'][listIndex]['data'] = tmpData[0:snippetLen]
			if len(tmpData) > snippetLen:
				 response['datas'][listIndex]['data'] += '...'
		
	#json串
	jsonStr = json.dumps(response, ensure_ascii = False)
	if isinstance(jsonStr, unicode):
		jsonStr = jsonStr.encode('utf-8')
	
	return HttpResponse(jsonStr)

8.今天(2014.2.20)找到对微博内容建sphinx索引后搜索出来的id到mysql差不到数据的原因:

以微博id微博sphinx文档id、以微博内容为文本建sphinx索引时因为微博id是bigint大整数(8bytes),到sphinx这边被截断了(默认安装的sphinx只支持4byte的id)

详见:http://sphinxsearch.com/forum/view.html?id=2064

你可能感兴趣的:(c,search,query,全文检索,引擎)