Java Code Examples for org.apache.solr.common.cloud.Slice#getState()
The following examples show how to use
org.apache.solr.common.cloud.Slice#getState() .
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example. You may check out the related API usage on the sidebar.
Example 1
Source File: AbstractDistribZkTestBase.java From lucene-solr with Apache License 2.0 | 6 votes |
protected static void assertAllActive(String collection, ZkStateReader zkStateReader) throws KeeperException, InterruptedException { zkStateReader.forceUpdateCollection(collection); ClusterState clusterState = zkStateReader.getClusterState(); final DocCollection docCollection = clusterState.getCollectionOrNull(collection); if (docCollection == null || docCollection.getSlices() == null) { throw new IllegalArgumentException("Cannot find collection:" + collection); } Map<String,Slice> slices = docCollection.getSlicesMap(); for (Map.Entry<String,Slice> entry : slices.entrySet()) { Slice slice = entry.getValue(); if (slice.getState() != Slice.State.ACTIVE) { fail("Not all shards are ACTIVE - found a shard " + slice.getName() + " that is: " + slice.getState()); } Map<String,Replica> shards = slice.getReplicasMap(); for (Map.Entry<String,Replica> shard : shards.entrySet()) { Replica replica = shard.getValue(); if (replica.getState() != Replica.State.ACTIVE) { fail("Not all replicas are ACTIVE - found a replica " + replica.getName() + " that is: " + replica.getState()); } } } }
Example 2
Source File: DistributedZkUpdateProcessor.java From lucene-solr with Apache License 2.0 | 6 votes |
/** For {@link org.apache.solr.common.params.CollectionParams.CollectionAction#SPLITSHARD} */ protected boolean amISubShardLeader(DocCollection coll, Slice parentSlice, String id, SolrInputDocument doc) throws InterruptedException { // Am I the leader of a shard in "construction/recovery" state? String myShardId = cloudDesc.getShardId(); Slice mySlice = coll.getSlice(myShardId); final Slice.State state = mySlice.getState(); if (state == Slice.State.CONSTRUCTION || state == Slice.State.RECOVERY) { Replica myLeader = zkController.getZkStateReader().getLeaderRetry(collection, myShardId); boolean amILeader = myLeader.getName().equals(cloudDesc.getCoreNodeName()); if (amILeader) { // Does the document belong to my hash range as well? DocRouter.Range myRange = mySlice.getRange(); if (myRange == null) myRange = new DocRouter.Range(Integer.MIN_VALUE, Integer.MAX_VALUE); if (parentSlice != null) { boolean isSubset = parentSlice.getRange() != null && myRange.isSubsetOf(parentSlice.getRange()); return isSubset && coll.getRouter().isTargetSlice(id, doc, req.getParams(), myShardId, coll); } else { // delete by query case -- as long as I am a sub shard leader we're fine return true; } } } return false; }
Example 3
Source File: DistributedZkUpdateProcessor.java From lucene-solr with Apache License 2.0 | 6 votes |
/** For {@link org.apache.solr.common.params.CollectionParams.CollectionAction#SPLITSHARD} */ protected List<SolrCmdDistributor.Node> getSubShardLeaders(DocCollection coll, String shardId, String docId, SolrInputDocument doc) { Collection<Slice> allSlices = coll.getSlices(); List<SolrCmdDistributor.Node> nodes = null; for (Slice aslice : allSlices) { final Slice.State state = aslice.getState(); if (state == Slice.State.CONSTRUCTION || state == Slice.State.RECOVERY) { DocRouter.Range myRange = coll.getSlice(shardId).getRange(); if (myRange == null) myRange = new DocRouter.Range(Integer.MIN_VALUE, Integer.MAX_VALUE); boolean isSubset = aslice.getRange() != null && aslice.getRange().isSubsetOf(myRange); if (isSubset && (docId == null // in case of deletes || coll.getRouter().isTargetSlice(docId, doc, req.getParams(), aslice.getName(), coll))) { Replica sliceLeader = aslice.getLeader(); // slice leader can be null because node/shard is created zk before leader election if (sliceLeader != null && clusterState.liveNodesContain(sliceLeader.getNodeName())) { if (nodes == null) nodes = new ArrayList<>(); ZkCoreNodeProps nodeProps = new ZkCoreNodeProps(sliceLeader); nodes.add(new SolrCmdDistributor.StdNode(nodeProps, coll.getName(), aslice.getName())); } } } } return nodes; }
Example 4
Source File: SolrCore.java From lucene-solr with Apache License 2.0 | 6 votes |
/** * Set UpdateLog to buffer updates if the slice is in construction. */ private void bufferUpdatesIfConstructing(CoreDescriptor coreDescriptor) { if (coreContainer != null && coreContainer.isZooKeeperAware()) { if (reqHandlers.get("/get") == null) { log.warn("WARNING: RealTimeGetHandler is not registered at /get. SolrCloud will always use full index replication instead of the more efficient PeerSync method."); } // ZK pre-register would have already happened so we read slice properties now final ClusterState clusterState = coreContainer.getZkController().getClusterState(); final DocCollection collection = clusterState.getCollection(coreDescriptor.getCloudDescriptor().getCollectionName()); final Slice slice = collection.getSlice(coreDescriptor.getCloudDescriptor().getShardId()); if (slice.getState() == Slice.State.CONSTRUCTION) { // set update log to buffer before publishing the core getUpdateHandler().getUpdateLog().bufferUpdates(); } } }
Example 5
Source File: DistributedZkUpdateProcessor.java From lucene-solr with Apache License 2.0 | 5 votes |
/** For {@link org.apache.solr.common.params.CollectionParams.CollectionAction#SPLITSHARD} */ private boolean couldIbeSubShardLeader(DocCollection coll) { // Could I be the leader of a shard in "construction/recovery" state? String myShardId = cloudDesc.getShardId(); Slice mySlice = coll.getSlice(myShardId); Slice.State state = mySlice.getState(); return state == Slice.State.CONSTRUCTION || state == Slice.State.RECOVERY; }
Example 6
Source File: RecoveryStrategy.java From lucene-solr with Apache License 2.0 | 5 votes |
final private void sendPrepRecoveryCmd(String leaderBaseUrl, String leaderCoreName, Slice slice) throws SolrServerException, IOException, InterruptedException, ExecutionException { WaitForState prepCmd = new WaitForState(); prepCmd.setCoreName(leaderCoreName); prepCmd.setNodeName(zkController.getNodeName()); prepCmd.setCoreNodeName(coreZkNodeName); prepCmd.setState(Replica.State.RECOVERING); prepCmd.setCheckLive(true); prepCmd.setOnlyIfLeader(true); final Slice.State state = slice.getState(); if (state != Slice.State.CONSTRUCTION && state != Slice.State.RECOVERY && state != Slice.State.RECOVERY_FAILED) { prepCmd.setOnlyIfLeaderActive(true); } int conflictWaitMs = zkController.getLeaderConflictResolveWait(); // timeout after 5 seconds more than the max timeout (conflictWait + 3 seconds) on the server side int readTimeout = conflictWaitMs + Integer.parseInt(System.getProperty("prepRecoveryReadTimeoutExtraWait", "8000")); try (HttpSolrClient client = buildRecoverySolrClient(leaderBaseUrl)) { client.setSoTimeout(readTimeout); HttpUriRequestResponse mrr = client.httpUriRequest(prepCmd); prevSendPreRecoveryHttpUriRequest = mrr.httpUriRequest; log.info("Sending prep recovery command to [{}]; [{}]", leaderBaseUrl, prepCmd); mrr.future.get(); } }
Example 7
Source File: DistributedZkUpdateProcessor.java From lucene-solr with Apache License 2.0 | 4 votes |
private void doDefensiveChecks(DistribPhase phase) { boolean isReplayOrPeersync = (updateCommand.getFlags() & (UpdateCommand.REPLAY | UpdateCommand.PEER_SYNC)) != 0; if (isReplayOrPeersync) return; String from = req.getParams().get(DISTRIB_FROM); DocCollection docCollection = clusterState.getCollection(collection); Slice mySlice = docCollection.getSlice(cloudDesc.getShardId()); boolean localIsLeader = cloudDesc.isLeader(); if (DistribPhase.FROMLEADER == phase && localIsLeader && from != null) { // from will be null on log replay String fromShard = req.getParams().get(DISTRIB_FROM_PARENT); if (fromShard != null) { if (mySlice.getState() == Slice.State.ACTIVE) { throw new SolrException(SolrException.ErrorCode.SERVICE_UNAVAILABLE, "Request says it is coming from parent shard leader but we are in active state"); } // shard splitting case -- check ranges to see if we are a sub-shard Slice fromSlice = docCollection.getSlice(fromShard); DocRouter.Range parentRange = fromSlice.getRange(); if (parentRange == null) parentRange = new DocRouter.Range(Integer.MIN_VALUE, Integer.MAX_VALUE); if (mySlice.getRange() != null && !mySlice.getRange().isSubsetOf(parentRange)) { throw new SolrException(SolrException.ErrorCode.SERVICE_UNAVAILABLE, "Request says it is coming from parent shard leader but parent hash range is not superset of my range"); } } else { String fromCollection = req.getParams().get(DISTRIB_FROM_COLLECTION); // is it because of a routing rule? if (fromCollection == null) { log.error("Request says it is coming from leader, but we are the leader: {}", req.getParamString()); SolrException solrExc = new SolrException(SolrException.ErrorCode.SERVICE_UNAVAILABLE, "Request says it is coming from leader, but we are the leader"); solrExc.setMetadata("cause", "LeaderChanged"); throw solrExc; } } } int count = 0; while (((isLeader && !localIsLeader) || (isSubShardLeader && !localIsLeader)) && count < 5) { count++; // re-getting localIsLeader since we published to ZK first before setting localIsLeader value localIsLeader = cloudDesc.isLeader(); try { Thread.sleep(500); } catch (InterruptedException e) { Thread.currentThread().interrupt(); } } if ((isLeader && !localIsLeader) || (isSubShardLeader && !localIsLeader)) { log.error("ClusterState says we are the leader, but locally we don't think so"); throw new SolrException(SolrException.ErrorCode.SERVICE_UNAVAILABLE, "ClusterState says we are the leader (" + zkController.getBaseUrl() + "/" + req.getCore().getName() + "), but locally we don't think so. Request came from " + from); } }
Example 8
Source File: CloudUtil.java From lucene-solr with Apache License 2.0 | 4 votes |
/** * Return a {@link CollectionStatePredicate} that returns true if a collection has the expected * number of shards and replicas. * <p>Note: for shards marked as inactive the current Solr behavior is that replicas remain active. * {@link org.apache.solr.cloud.autoscaling.sim.SimCloudManager} follows this behavior.</p> * @param expectedShards expected number of shards * @param expectedReplicas expected number of active replicas per shard * @param withInactive if true then count also inactive shards * @param requireLeaders if true then require that each shard has a leader */ public static CollectionStatePredicate clusterShape(int expectedShards, int expectedReplicas, boolean withInactive, boolean requireLeaders) { return (liveNodes, collectionState) -> { if (collectionState == null) { log.debug("-- null collection"); return false; } Collection<Slice> slices = withInactive ? collectionState.getSlices() : collectionState.getActiveSlices(); if (slices.size() != expectedShards) { if (log.isDebugEnabled()) { log.debug("-- wrong number of slices for collection {}, expected={}, found={}: {}", collectionState.getName(), expectedShards, collectionState.getSlices().size(), collectionState.getSlices()); } return false; } Set<String> leaderless = new HashSet<>(); for (Slice slice : slices) { int activeReplicas = 0; if (requireLeaders && slice.getState() != Slice.State.INACTIVE && slice.getLeader() == null) { leaderless.add(slice.getName()); continue; } // skip other checks, we're going to fail anyway if (!leaderless.isEmpty()) { continue; } for (Replica replica : slice) { if (replica.isActive(liveNodes)) activeReplicas++; } if (activeReplicas != expectedReplicas) { if (log.isDebugEnabled()) { log.debug("-- wrong number of active replicas for collection {} in slice {}, expected={}, found={}", collectionState.getName(), slice.getName(), expectedReplicas, activeReplicas); } return false; } } if (leaderless.isEmpty()) { return true; } else { log.info("-- shards without leaders: {}", leaderless); return false; } }; }
Example 9
Source File: SharedFSAutoReplicaFailoverTest.java From lucene-solr with Apache License 2.0 | 4 votes |
private boolean waitingForReplicasNotLive(ZkStateReader zkStateReader, int timeoutInMs, List<JettySolrRunner> jetties) { Set<String> nodeNames = jetties.stream() .filter(jetty -> jetty.getCoreContainer() != null) .map(JettySolrRunner::getNodeName) .collect(Collectors.toSet()); long timeout = System.nanoTime() + TimeUnit.NANOSECONDS.convert(timeoutInMs, TimeUnit.MILLISECONDS); boolean success = false; while (!success && System.nanoTime() < timeout) { success = true; ClusterState clusterState = zkStateReader.getClusterState(); if (clusterState != null) { Map<String, DocCollection> collections = clusterState.getCollectionsMap(); for (Map.Entry<String, DocCollection> entry : collections.entrySet()) { DocCollection docCollection = entry.getValue(); Collection<Slice> slices = docCollection.getSlices(); for (Slice slice : slices) { // only look at active shards if (slice.getState() == Slice.State.ACTIVE) { Collection<Replica> replicas = slice.getReplicas(); for (Replica replica : replicas) { if (nodeNames.contains(replica.getNodeName())) { boolean live = clusterState.liveNodesContain(replica .getNodeName()); if (live) { success = false; } } } } } } if (!success) { try { Thread.sleep(500); } catch (InterruptedException e) { Thread.currentThread().interrupt(); throw new SolrException(SolrException.ErrorCode.SERVER_ERROR, "Interrupted"); } } } } return success; }