diff options
author | Stephane Gamard <stephane.gamard@searchbox.com> | 2014-08-22 10:00:59 +0200 |
---|---|---|
committer | Stephane Gamard <stephane.gamard@searchbox.com> | 2014-08-22 10:01:35 +0200 |
commit | 9369c4fc6230b858596594473513f9b350d8e284 (patch) | |
tree | e9921d9f21d71b49c4edf2b24a4a5019ca527caf /server/sonar-server | |
parent | 90864dacb4972d0251c5659d4cb962fbd919b032 (diff) | |
download | sonarqube-9369c4fc6230b858596594473513f9b350d8e284.tar.gz sonarqube-9369c4fc6230b858596594473513f9b350d8e284.zip |
fix quality flaws
Diffstat (limited to 'server/sonar-server')
3 files changed, 60 insertions, 33 deletions
diff --git a/server/sonar-server/src/main/java/org/sonar/server/db/BaseDao.java b/server/sonar-server/src/main/java/org/sonar/server/db/BaseDao.java index cc5ff34ea8e..9724eda67f3 100644 --- a/server/sonar-server/src/main/java/org/sonar/server/db/BaseDao.java +++ b/server/sonar-server/src/main/java/org/sonar/server/db/BaseDao.java @@ -219,8 +219,7 @@ public abstract class BaseDao<M, E extends Dto<K>, K extends Serializable> imple session.enqueue(new UpsertDto<E>(getIndexType(), item)); } } catch (Exception e) { - e.printStackTrace(); - throw new IllegalStateException("Fail to insert item in db: " + item, e.getCause()); + throw new IllegalStateException("Fail to insert item in db: " + item, e); } } diff --git a/server/sonar-server/src/main/java/org/sonar/server/db/DbClient.java b/server/sonar-server/src/main/java/org/sonar/server/db/DbClient.java index 2b7b1351f24..846357d0bc4 100644 --- a/server/sonar-server/src/main/java/org/sonar/server/db/DbClient.java +++ b/server/sonar-server/src/main/java/org/sonar/server/db/DbClient.java @@ -34,6 +34,7 @@ import org.sonar.core.user.AuthorizationDao; import org.sonar.core.user.UserDao; import org.sonar.server.activity.db.ActivityDao; import org.sonar.server.component.persistence.ComponentDao; +import org.sonar.server.issue.db.IssueDao; import org.sonar.server.measure.persistence.MeasureDao; import org.sonar.server.measure.persistence.MetricDao; import org.sonar.server.qualityprofile.db.ActiveRuleDao; @@ -61,6 +62,7 @@ public class DbClient implements ServerComponent { private final ActivityDao activityDao; private final AuthorizationDao authorizationDao; private final UserDao userDao; + private final IssueDao issueDao; public DbClient(Database db, MyBatis myBatis, DaoComponent... daoComponents) { this.db = db; @@ -83,6 +85,7 @@ public class DbClient implements ServerComponent { activityDao = getDao(map, ActivityDao.class); authorizationDao = getDao(map, AuthorizationDao.class); userDao = getDao(map, UserDao.class); + issueDao = getDao(map, IssueDao.class); } public Database database() { @@ -101,6 +104,10 @@ public class DbClient implements ServerComponent { return activeRuleDao; } + public IssueDao issueDao() { + return issueDao; + } + public QualityProfileDao qualityProfileDao() { return qualityProfileDao; } diff --git a/server/sonar-server/src/main/java/org/sonar/server/search/IndexQueue.java b/server/sonar-server/src/main/java/org/sonar/server/search/IndexQueue.java index a8003978d94..b159aae781a 100644 --- a/server/sonar-server/src/main/java/org/sonar/server/search/IndexQueue.java +++ b/server/sonar-server/src/main/java/org/sonar/server/search/IndexQueue.java @@ -20,6 +20,8 @@ package org.sonar.server.search; import org.elasticsearch.action.ActionRequest; +import org.elasticsearch.action.admin.indices.refresh.RefreshRequestBuilder; +import org.elasticsearch.action.admin.indices.refresh.RefreshResponse; import org.elasticsearch.action.bulk.BulkRequestBuilder; import org.elasticsearch.action.bulk.BulkResponse; import org.elasticsearch.action.delete.DeleteRequest; @@ -54,7 +56,7 @@ public class IndexQueue extends LinkedBlockingQueue<Runnable> private static final Logger LOGGER = LoggerFactory.getLogger(IndexQueue.class); private static final Integer DEFAULT_QUEUE_SIZE = 200; - private static final int TIMEOUT = 30000; + private static final Integer CONCURRENT_NORMALIZATION_FACTOR = 5; public IndexQueue(Settings settings, SearchClient searchClient, ComponentContainer container) { super(DEFAULT_QUEUE_SIZE); @@ -71,28 +73,62 @@ public class IndexQueue extends LinkedBlockingQueue<Runnable> } try { - long normTime = System.currentTimeMillis(); - BulkRequestBuilder bulkRequestBuilder = new BulkRequestBuilder(searchClient); Map<String, Index> indexes = getIndexMap(); Set<String> indices = new HashSet<String>(); for (IndexActionRequest action : actions) { Index index = indexes.get(action.getIndexType()); action.setIndex(index); - indices.add(index.getIndexName()); + if (action.needsRefresh()) { + indices.add(index.getIndexName()); + } } - ExecutorService executorService = Executors.newFixedThreadPool(10); + BulkRequestBuilder bulkRequestBuilder = new BulkRequestBuilder(searchClient); - // Do we need to refresh - boolean requiresRefresh = false; - for (IndexActionRequest action : actions) { - if (action.needsRefresh()) { - requiresRefresh = true; - break; - } + long normTime = executeNormalization(bulkRequestBuilder, actions); + + //execute the request + long indexTime = System.currentTimeMillis(); + BulkResponse response = searchClient.execute(bulkRequestBuilder.setRefresh(false)); + indexTime = System.currentTimeMillis() - indexTime; + + long refreshTime = this.refreshRequiredIndex(indices); + + LOGGER.debug("-- submitted {} items with {}ms in normalization, {}ms indexing and {}ms refresh({}). Total: {}ms", + bulkRequestBuilder.numberOfActions(), normTime, indexTime, refreshTime, indices, (normTime + indexTime + refreshTime)); + + if (response.hasFailures()) { + throw new IllegalStateException("Errors while indexing stack: " + response.buildFailureMessage()); } - //invokeAll() blocks until ALL tasks submitted to executor complete + } catch (Exception e) { + LOGGER.error("Could not commit to ElasticSearch", e); + } + } + + + private long refreshRequiredIndex(Set<String> indices) { + + long refreshTime = System.currentTimeMillis(); + if (!indices.isEmpty()) { + RefreshRequestBuilder refreshRequest = searchClient.admin().indices() + .prepareRefresh(indices.toArray(new String[indices.size()])) + .setForce(false); + + RefreshResponse refreshResponse = searchClient.execute(refreshRequest); + + if (refreshResponse.getFailedShards() > 0) { + LOGGER.warn("{} Shard(s) did not refresh", refreshResponse.getFailedShards()); + } + } + return System.currentTimeMillis() - refreshTime; + } + + private long executeNormalization(BulkRequestBuilder bulkRequestBuilder, List<IndexActionRequest> actions) { + long normTime = System.currentTimeMillis(); + ExecutorService executorService = Executors.newFixedThreadPool(CONCURRENT_NORMALIZATION_FACTOR); + //invokeAll() blocks until ALL tasks submitted to executor complete + try { for (Future<List<ActionRequest>> updateRequests : executorService.invokeAll(actions)) { for (ActionRequest update : updateRequests.get()) { if (UpdateRequest.class.isAssignableFrom(update.getClass())) { @@ -104,26 +140,11 @@ public class IndexQueue extends LinkedBlockingQueue<Runnable> } } } - executorService.shutdown(); - normTime = System.currentTimeMillis() - normTime; - - //execute the request - long indexTime = System.currentTimeMillis(); - BulkResponse response = searchClient.execute(bulkRequestBuilder.setRefresh(false)); - indexTime = System.currentTimeMillis() - indexTime; - - long refreshTime = System.currentTimeMillis(); - if (requiresRefresh) { - searchClient.admin().indices().prepareRefresh(indices.toArray(new String[indices.size()])).setForce(false).get(); - } - refreshTime = System.currentTimeMillis() - refreshTime; - - LOGGER.debug("-- submitted {} items with {}ms in normalization, {}ms indexing and {}ms refresh({}). Total: {}ms", - bulkRequestBuilder.numberOfActions(), normTime, indexTime, refreshTime, indices, (normTime + indexTime + refreshTime)); - } catch (Exception e) { - e.printStackTrace(); + throw new IllegalStateException("Could not execute normalization for stack", e); } + executorService.shutdown(); + return System.currentTimeMillis() - normTime; } private Map<String, Index> getIndexMap() { |