mongodb-river-elasticsearch源码解析
来源:互联网 发布:sql万能钥匙 编辑:程序博客网 时间:2024/05/16 10:37
MongoDBRiverPlugin
MongoDBRiverPlugin类是插件注册类,它继承自AbstractPlugin,其功能是
1. 在RiverModule中注册一个MongoDBRiver
2. 在RestModule中注册一个RestMongoDBRiverAction
package org.elasticsearch.plugin.river.mongodb;import org.elasticsearch.plugins.AbstractPlugin;import org.elasticsearch.rest.RestModule;import org.elasticsearch.rest.action.mongodb.RestMongoDBRiverAction;import org.elasticsearch.river.RiversModule;import org.elasticsearch.river.mongodb.MongoDBRiver;import org.elasticsearch.river.mongodb.MongoDBRiverModule;/*** @author flaper87 (Flavio Percoco Premoli)* @author aparo (Alberto Paro)* @author kryptt (Rodolfo Hansen)*/public class MongoDBRiverPlugin extends AbstractPlugin {@Overridepublic String name() {return MongoDBRiver.NAME;}@Overridepublic String description() {return MongoDBRiver.DESCRIPTION;}/*** Register the MongoDB river to Elasticsearch node** @param module*/public void onModule(RiversModule module) {module.registerRiver(MongoDBRiver.TYPE, MongoDBRiverModule.class);}/*** Register the REST move to Elasticsearch node** @param module*/public void onModule(RestModule module) {module.addRestAction(RestMongoDBRiverAction.class);}}
MongoDBRiver
首先看river部分 org.elasticsearch.river.mongodb.MongoDBRiver是核心类,构造函数中都是都是elasticsearch 的配置信息和服务
参数类型
参数名称含义取值RiverNameriverName名称 RiverSettingssettings设置信息 StringriverIndexName索引名 Clientclient客户端 ScriptServicescriptService 脚本服务 MongoDBRiverDefinitiondefinition解析后的定义MongoDBRiverDefinition.parseSettings(riverName.name(),riverIndexName, settings, scriptService);
还有一个参数stream表示操作流,用来存储需要放在mongo oplog中的数据队列
BlockingQueue<QueueEntry> stream = definition.getThrottleSize() == -1 ? new LinkedTransferQueue<QueueEntry>() : new ArrayBlockingQueue<QueueEntry>(definition.getThrottleSize());
可以看到,如果definition中设定的阈值大小没有设定的话,使用一个链表数据结构作为队列,否则使用一个数组队列。不过两种情况使用的数据结构都是多线程使用的数据结构BlockingQueue阻塞队列。阻塞队列是用在“生产者-消费者”模式的主要数据结构,其作用是如果队列空,则消费者阻塞;如果队列满,则生产者阻塞。而且队列支持多个生产者和消费者线程。其中QueueEntry定义如下,其中Operation是一个枚举,包含了各种mongodb操作:INSERT,UPDATE, DELETE, DROP_COLLECTION, DROP_DATABASE, COMMAND, UNKNOWN;
protected static class QueueEntry {private final DBObject data;private final Operation operation;private final Timestamp<?> oplogTimestamp;private final String collection;public QueueEntry(DBObject data, String collection) {this(null, Operation.INSERT, data, collection);}public QueueEntry(Timestamp<?> oplogTimestamp, Operation oplogOperation, DBObject data, String collection) {this.data = data;this.operation = oplogOperation;this.oplogTimestamp = oplogTimestamp;this.collection = collection;}public boolean isOplogEntry() {return oplogTimestamp != null;}public boolean isAttachment() {return (data instanceof GridFSDBFile);}public DBObject getData() {return data;}public Operation getOperation() {return operation;}public Timestamp<?> getOplogTimestamp() {return oplogTimestamp;}public String getCollection() {return collection;}}}
最后MongoDBRiver构造函数里面还有一个全局参数SharedContext context,这个参数包含了这个队列的引用,并且包含了整体运行状态的一个上下文状态:UNKNOWN, START_FAILED, RUNNING, STOPPED, IMPORT_FAILED,INITIAL_IMPORT_FAILED, SCRIPT_IMPORT_FAILED, RIVER_STALE;
this.context = new SharedContext(stream, Status.STOPPED);初始化之后就可以,elasticsearch将通过start方法启动这个插件,启动逻辑如下:
* 首先是各种状态的检查:
1、 用client获取elastic的状态,转成Status
client.prepareGet("_river", "mongodb-river", "_riverstatus").get()XContentMapValues.extractValue("mongodb.status")2、如果状态是IMPORT_FAILED、INITIAL_IMPORT_FAILED、SCRIPT_IMPORT_FAILED Status.START_FAILED 或者 STOPPED;表示有问题,直接打印日志并返回
3、 如果没有问题,则使用方法设定river为启动状态:
MongoDBRiverHelper.setRiverStatus(client, riverName.getName(), Status.RUNNING);context.setStatus(Status.RUNNING);
4、如果不存在索引则建立之
// Create the index if it does not existclient.admin().indices().prepareCreate(definition.getIndexName()).get();
5、 如果是GridFS要做一些额外的索引工作
client.admin().indices().preparePutMapping(definition.getIndexName()).setType(definition.getTypeName()).setSource(getGridFSMapping()).get();6、 然后我们开始启动相关的线程:
如果是mongos,就启动多个OpLog处理线程,否则使用一个线程,创建方式如下:
EsExecutors.daemonThreadFactory(settings.globalSettings(), "mongodb_river_slurper").newThread( new Slurper(definition.getMongoServers(), definition, context, client));7、启动之后再启动Indexer进程
EsExecutors.daemonThreadFactory(settings.globalSettings(),"mongodb_river_indexer").newThread(new Indexer(this, definition, context, client, scriptService));8、最后再启动一个状态监测进程:
EsExecutors.daemonThreadFactory(settings.globalSettings(), "mongodb_river_status").newThread(new StatusChecker(this, definition, context));
* 所以代码的核心就是三个线程:
收割 new Slurper(definition.getMongoServers(), definition, context,client)
索引处理 new Indexer(this, definition, context, client, scriptService)
状态检查 new StatusChecker(this, definition, context)
可以看到共同的参数都是:一个definition包含所有的配置,context包含了操作队列和状态
Slurper收割线程
其逻辑是:
1、 如果driver的状态是Running,则查找OpLog的信息并放入stream队列中
2、 如果无法获取oplogCollection队列,则退出线程failed to assign oplogCollection orslurpedCollection
3、 增量处理是按照上次注入时间点为查询条件的
cursor = oplogCursor(startTimestamp);if (cursor == null) { cursor = processFullOplog();}
查询条件是
filter.put(MongoDBRiver.OPLOG_TIMESTAMP,new BasicDBObject(QueryOperators.GTE, time));ts > time
4、获得数据库指针之后,处理每一个OpLog的数据
while (cursor.hasNext()) { DBObject item = cursor.next(); startTimestamp = processOplogEntry(item, startTimestamp);}
处理这些数据最后就是调用 addToStream 或 addInsertToStream 加入stream中
初始化导入
上面的过程只适合于从当前时间开始的数据,如果需要把原来的数据导入的话,还需要做一个initialimport
当程序配置满足一下条件的时候,才会在第一次运行该线程的时候进行初始化导入:
SkipInitialImport == falseInitialTimestamp == null // initial timestamp 为空MongoDBRiver.getIndexCount(client, definition) == 0 // 没有index过MongoDBRiver.getLastTimestamp(client, definition) == null;Get the latest timestamp for a given namespace.
满足这些条件之后才会进行数据的初始化导入:初始化导入会查看一下设置,如果是ImportAllCollections,则检查每一个collection并注入否则,找出设定的collection并注入
核心代码是这样的:if (!definition.isSkipInitialImport()) { if (!riverHasIndexedFromOplog() && definition.getInitialTimestamp() == null) { if (!isIndexEmpty()) { MongoDBRiverHelper.setRiverStatus(client, definition.getRiverName(), Status.INITIAL_IMPORT_FAILED); break; } if (definition.isImportAllCollections()) { for (String name : slurpedDb.getCollectionNames()) { DBCollection collection = slurpedDb.getCollection(name); startTimestamp = doInitialImport(collection); } } else { DBCollection collection = slurpedDb.getCollection(definition.getMongoCollection()); startTimestamp = doInitialImport(collection); } } } else { logger.info("Skip initial import from collection {}", definition.getMongoCollection()); }
/*** Does an initial sync the same way MongoDB does.* https://groups.google.com/* forum/?fromgroups=#!topic/mongodb-user/sOKlhD_E2ns** @return the last oplog timestamp before the import began* @throws InterruptedException* if the blocking queue stream is interrupted while waiting*/protected Timestamp<?> doInitialImport(DBCollection collection) throws InterruptedException {// TODO: ensure the index type is empty// DBCollection slurpedCollection =// slurpedDb.getCollection(definition.getMongoCollection());logger.info("MongoDBRiver is beginning initial import of " + collection.getFullName());Timestamp<?> startTimestamp = getCurrentOplogTimestamp();boolean inProgress = true;String lastId = null;while (inProgress) {DBCursor cursor = null;try {if (definition.isDisableIndexRefresh()) {updateIndexRefresh(definition.getIndexName(), -1L);}if (!definition.isMongoGridFS()) {logger.info("Collection {} - count: {}", collection.getName(), collection.count());long count = 0;cursor = collection.find(getFilterForInitialImport(definition.getMongoCollectionFilter(), lastId));while (cursor.hasNext()) {DBObject object = cursor.next();count++;if (cursor.hasNext()) {lastId = addInsertToStream(null, applyFieldFilter(object), collection.getName());} else {logger.debug("Last entry for initial import - add timestamp: {}", startTimestamp);lastId = addInsertToStream(startTimestamp, applyFieldFilter(object), collection.getName());}}inProgress = false;logger.info("Number documents indexed: {}", count);} else {// TODO: To be optimized.// https://github.com/mongodb/mongo-java-driver/pull/48#issuecomment-25241988// possible option: Get the object id list from .fs// collection// then call GriDFS.findOneGridFS grid = new GridFS(mongo.getDB(definition.getMongoDb()), definition.getMongoCollection());cursor = grid.getFileList();while (cursor.hasNext()) {DBObject object = cursor.next();if (object instanceof GridFSDBFile) {GridFSDBFile file = grid.findOne(new ObjectId(object.get(MongoDBRiver.MONGODB_ID_FIELD).toString()));if (cursor.hasNext()) {lastId = addInsertToStream(null, file);} else {logger.debug("Last entry for initial import - add timestamp: {}", startTimestamp);lastId = addInsertToStream(startTimestamp, file);}}}inProgress = false;}} catch (MongoException.CursorNotFound e) {logger.info("Initial import - Cursor {} has been closed. About to open a new cusor.", cursor.getCursorId());logger.debug("Total document inserted [{}]", totalDocuments.get());} finally {if (cursor != null) {logger.trace("Closing initial import cursor");cursor.close();}if (definition.isDisableIndexRefresh()) {updateIndexRefresh(definition.getIndexName(), TimeValue.timeValueSeconds(1));}}}return startTimestamp;}private BasicDBObject getFilterForInitialImport(BasicDBObject filter, String id) {if (id == null) {return filter;} else {BasicDBObject filterId = new BasicDBObject(MongoDBRiver.MONGODB_ID_FIELD, new BasicBSONObject(QueryOperators.GT, id));if (filter == null) {return filterId;} else {List<BasicDBObject> values = ImmutableList.of(filter, filterId);return new BasicDBObject(QueryOperators.AND, values);}}
Indexer线程
其逻辑是:
1、如果driver的状态是Running,则从stream队列中获取信息并放入Index中
在构造函数初始化的时候会做一些MongoDBRiverBulkProcessor的创建 build:
SimpleEntry<String, String> entry = new SimpleEntry<String, String>(index, type); if (!processors.containsKey(entry)) { processors.put(new SimpleEntry<String, String>(index, type), new MongoDBRiverBulkProcessor.Builder(river, definition, client, index, type).build()); } return processors.get(entry);
然后在业务逻辑中读取entry,并processBlockingQueue processBlockingQueue就是根据不同的业务的内容做不同的处理,就是对不同的操作用相关的es api加以处理。
// 1. Attempt to fill as much of the bulk request as possible QueueEntry entry = context.getStream().take(); lastTimestamp = processBlockingQueue(entry); while ((entry = context.getStream().poll(definition.getBulk().getFlushInterval().millis(), MILLISECONDS)) != null) { lastTimestamp = processBlockingQueue(entry); } // 2. Update the timestamp if (lastTimestamp != null) { MongoDBRiver.setLastTimestamp(definition, lastTimestamp, getBulkProcessor(definition.getIndexName(), definition.getTypeName()).getBulkProcessor()); }
StatusChecker
状态检查就是更具用户的命令进行开/关
就是检查elastic中的最新状态【用户设定的状态】:MongoDBRiverHelper.getRiverStatus(client, riverName);
如果状态和当前状态不一致,就进行driver的start或stop
用一个流程图来解释这几个线程之间的关系就是这样的:
RestModule
注册这个模块的作用是在原来es支持的rest api基础上,增加针对mongodb的新的api类型,具体实现可以参考一下这篇文章,这里不再赘述了:
http://elasticsearchserverbook.com/creating-custom-elasticsearch-rest-action/
参考文档:
https://github.com/richardwilly98/elasticsearch-river-mongodb
http://elasticsearchserverbook.com/creating-custom-elasticsearch-rest-action/
http://blog.csdn.net/vernonzheng/article/details/8247564
http://www.cnblogs.com/jackyuj/archive/2010/11/24/1886553.html
- mongodb-river-elasticsearch源码解析
- mongodb-river-elasticsearch源码解析
- ES-MongoDB学习2_mongodb-river-elasticsearch源码解析
- mongodb-elasticsearch-rive源码解析
- 单机搭建elasticsearch和mongodb river的数据同步
- 在windows上配置elasticsearch和river-mongodb插件
- elasticsearch源码解析---AllocationDecider
- ElasticSearch Bulk 源码解析
- elasticsearch-river-jdbc
- HBase-River-to-Elasticsearch
- Elasticsearch分词器源码解析
- 用ElasticSearch和mongodb River搭建一个简单地search工程。
- Dianping River Plugin for Elasticsearch
- mongodb 源码解析内存管理
- ElasticSearch源码解析(三):索引创建
- elasticsearch使用river同步mysql数据
- elasticsearch使用river同步mysql数据
- 用elasticsearch-river-jdbc同步数据到elasticsearch
- Linux如何及时响应外部中断
- freopen()函数文件流重定向和文件流的清除
- Android签名与认证详细分析之二(CERT.RSA剖析)
- CSDN采访陶辉的关于开发Nginx模块的建议
- MFC中三种DLL区别
- mongodb-river-elasticsearch源码解析
- Python print 不换行打印
- projecteuler---->problem=31----Coin sums 无限背包计算可能存在的次数
- Objective-C学习笔记(1)
- Segmentation fault (core dumped)
- ip地址获取
- 轻量级IOC框架Beancontext发布
- 解决Retrying connect to server: 0.0.0.0/0.0.0.0:8032. Already tried 0 time(s); retry policy is...
- Oracle数据库视图与权限问题