code
stringlengths
73
34.1k
label
stringclasses
1 value
protected void populateTasksByStealer(List<StealerBasedRebalanceTask> sbTaskList) { // Setup mapping of stealers to work for this run. for(StealerBasedRebalanceTask task: sbTaskList) { if(task.getStealInfos().size() != 1) { throw new VoldemortException("StealerBasedRebalanceTasks should have a list of RebalancePartitionsInfo of length 1."); } RebalanceTaskInfo stealInfo = task.getStealInfos().get(0); int stealerId = stealInfo.getStealerId(); if(!this.tasksByStealer.containsKey(stealerId)) { this.tasksByStealer.put(stealerId, new ArrayList<StealerBasedRebalanceTask>()); } this.tasksByStealer.get(stealerId).add(task); } if(tasksByStealer.isEmpty()) { return; } // Shuffle order of each stealer's work list. This randomization // helps to get rid of any "patterns" in how rebalancing tasks were // added to the task list passed in. for(List<StealerBasedRebalanceTask> taskList: tasksByStealer.values()) { Collections.shuffle(taskList); } }
java
protected synchronized StealerBasedRebalanceTask scheduleNextTask(boolean executeService) { // Make sure there is work left to do. if(doneSignal.getCount() == 0) { logger.info("All tasks completion signaled... returning"); return null; } // Limit number of tasks outstanding. if(this.numTasksExecuting >= maxParallelRebalancing) { logger.info("Executing more tasks than [" + this.numTasksExecuting + "] the parallel allowed " + maxParallelRebalancing); return null; } // Shuffle list of stealer IDs each time a new task to schedule needs to // be found. Randomizing the order should avoid prioritizing one // specific stealer's work ahead of all others. List<Integer> stealerIds = new ArrayList<Integer>(tasksByStealer.keySet()); Collections.shuffle(stealerIds); for(int stealerId: stealerIds) { if(nodeIdsWithWork.contains(stealerId)) { logger.info("Stealer " + stealerId + " is already working... continuing"); continue; } for(StealerBasedRebalanceTask sbTask: tasksByStealer.get(stealerId)) { int donorId = sbTask.getStealInfos().get(0).getDonorId(); if(nodeIdsWithWork.contains(donorId)) { logger.info("Stealer " + stealerId + " Donor " + donorId + " is already working... continuing"); continue; } // Book keeping addNodesToWorkerList(Arrays.asList(stealerId, donorId)); numTasksExecuting++; // Remove this task from list thus destroying list being // iterated over. This is safe because returning directly out of // this branch. tasksByStealer.get(stealerId).remove(sbTask); try { if(executeService) { logger.info("Stealer " + stealerId + " Donor " + donorId + " going to schedule work"); service.execute(sbTask); } } catch(RejectedExecutionException ree) { logger.error("Stealer " + stealerId + "Rebalancing task rejected by executor service.", ree); throw new VoldemortRebalancingException("Stealer " + stealerId + "Rebalancing task rejected by executor service."); } return sbTask; } } printRemainingTasks(stealerIds); return null; }
java
public synchronized void addNodesToWorkerList(List<Integer> nodeIds) { // Bookkeeping for nodes that will be involved in the next task nodeIdsWithWork.addAll(nodeIds); logger.info("Node IDs with work: " + nodeIdsWithWork + " Newly added nodes " + nodeIds); }
java
public synchronized void doneTask(int stealerId, int donorId) { removeNodesFromWorkerList(Arrays.asList(stealerId, donorId)); numTasksExecuting--; doneSignal.countDown(); // Try and schedule more tasks now that resources may be available to do // so. scheduleMoreTasks(); }
java
private List<Long> collectLongMetric(String metricGetterName) { List<Long> vals = new ArrayList<Long>(); for(BdbEnvironmentStats envStats: environmentStatsTracked) { vals.add((Long) ReflectUtils.callMethod(envStats, BdbEnvironmentStats.class, metricGetterName, new Class<?>[0], new Object[0])); } return vals; }
java
public static Iterable<String> toHexStrings(Iterable<ByteArray> arrays) { ArrayList<String> ret = new ArrayList<String>(); for(ByteArray array: arrays) ret.add(ByteUtils.toHexString(array.get())); return ret; }
java
@Override public void sendResponse(StoreStats performanceStats, boolean isFromLocalZone, long startTimeInMs) throws Exception { /* * Pay attention to the code below. Note that in this method we wrap a multiPart object with a mimeMessage. * However when writing to the outputStream we only send the multiPart object and not the entire * mimeMessage. This is intentional. * * In the earlier version of this code we used to create a multiPart object and just send that multiPart * across the wire. * * However, we later discovered that upon setting the content of a MimeBodyPart, JavaMail internally creates * a DataHandler object wrapping the object you passed in. The part's Content-Type header is not updated * immediately. In order to get the headers updated, one needs to to call MimeMessage.saveChanges() on the * enclosing message, which cascades down the MIME structure into a call to MimeBodyPart.updateHeaders() * on the body part. It's this updateHeaders call that transfers the content type from the * DataHandler to the part's MIME Content-Type header. * * To make sure that the Content-Type headers are being updated (without changing too much code), we decided * to wrap the multiPart in a mimeMessage, call mimeMessage.saveChanges() and then just send the multiPart. * This is to make sure multiPart's headers are updated accurately. */ MimeMessage message = new MimeMessage(Session.getDefaultInstance(new Properties())); MimeMultipart multiPart = new MimeMultipart(); ByteArrayOutputStream outputStream = new ByteArrayOutputStream(); String base64Key = RestUtils.encodeVoldemortKey(key.get()); String contentLocationKey = "/" + this.storeName + "/" + base64Key; for(Versioned<byte[]> versionedValue: versionedValues) { byte[] responseValue = versionedValue.getValue(); VectorClock vectorClock = (VectorClock) versionedValue.getVersion(); String eTag = RestUtils.getSerializedVectorClock(vectorClock); numVectorClockEntries += vectorClock.getVersionMap().size(); // Create the individual body part for each versioned value of the // requested key MimeBodyPart body = new MimeBodyPart(); try { // Add the right headers body.addHeader(CONTENT_TYPE, "application/octet-stream"); body.addHeader(CONTENT_TRANSFER_ENCODING, "binary"); body.addHeader(RestMessageHeaders.X_VOLD_VECTOR_CLOCK, eTag); body.setContent(responseValue, "application/octet-stream"); body.addHeader(RestMessageHeaders.CONTENT_LENGTH, Integer.toString(responseValue.length)); multiPart.addBodyPart(body); } catch(MessagingException me) { logger.error("Exception while constructing body part", me); outputStream.close(); throw me; } } message.setContent(multiPart); message.saveChanges(); try { multiPart.writeTo(outputStream); } catch(Exception e) { logger.error("Exception while writing multipart to output stream", e); outputStream.close(); throw e; } ChannelBuffer responseContent = ChannelBuffers.dynamicBuffer(); responseContent.writeBytes(outputStream.toByteArray()); // Create the Response object HttpResponse response = new DefaultHttpResponse(HTTP_1_1, OK); // Set the right headers response.setHeader(CONTENT_TYPE, "multipart/binary"); response.setHeader(CONTENT_TRANSFER_ENCODING, "binary"); response.setHeader(CONTENT_LOCATION, contentLocationKey); // Copy the data into the payload response.setContent(responseContent); response.setHeader(CONTENT_LENGTH, response.getContent().readableBytes()); // Write the response to the Netty Channel if(logger.isDebugEnabled()) { String keyStr = RestUtils.getKeyHexString(this.key); debugLog("GET", this.storeName, keyStr, startTimeInMs, System.currentTimeMillis(), numVectorClockEntries); } this.messageEvent.getChannel().write(response); if(performanceStats != null && isFromLocalZone) { recordStats(performanceStats, startTimeInMs, Tracked.GET); } outputStream.close(); }
java
public String getPublicConfigValue(String key) throws ConfigurationException { if (!allProps.containsKey(key)) { throw new UndefinedPropertyException("The requested config key does not exist."); } if (restrictedConfigs.contains(key)) { throw new ConfigurationException("The requested config key is not publicly available!"); } return allProps.get(key); }
java
private void checkRateLimit(String quotaKey, Tracked trackedOp) { String quotaValue = null; try { if(!metadataStore.getQuotaEnforcingEnabledUnlocked()) { return; } quotaValue = quotaStore.cacheGet(quotaKey); // Store may not have any quotas if(quotaValue == null) { return; } // But, if it does float currentRate = getThroughput(trackedOp); float allowedRate = Float.parseFloat(quotaValue); // TODO the histogram should be reasonably accurate to do all // these things.. (ghost qps and all) // Report the current quota usage level quotaStats.reportQuotaUsed(trackedOp, Utils.safeGetPercentage(currentRate, allowedRate)); // check if we have exceeded rate. if(currentRate > allowedRate) { quotaStats.reportRateLimitedOp(trackedOp); throw new QuotaExceededException("Exceeded rate limit for " + quotaKey + ". Maximum allowed : " + allowedRate + " Current: " + currentRate); } } catch(NumberFormatException nfe) { // move on, if we cannot parse quota value properly logger.debug("Invalid formatting of quota value for key " + quotaKey + " : " + quotaValue); } }
java
public synchronized void submitOperation(int requestId, AsyncOperation operation) { if(this.operations.containsKey(requestId)) throw new VoldemortException("Request " + requestId + " already submitted to the system"); this.operations.put(requestId, operation); scheduler.scheduleNow(operation); logger.debug("Handling async operation " + requestId); }
java
public synchronized boolean isComplete(int requestId, boolean remove) { if (!operations.containsKey(requestId)) throw new VoldemortException("No operation with id " + requestId + " found"); if (operations.get(requestId).getStatus().isComplete()) { if (logger.isDebugEnabled()) logger.debug("Operation complete " + requestId); if (remove) operations.remove(requestId); return true; } return false; }
java
@JmxOperation(description = "Retrieve operation status") public String getStatus(int id) { try { return getOperationStatus(id).toString(); } catch(VoldemortException e) { return "No operation with id " + id + " found"; } }
java
public List<Integer> getAsyncOperationList(boolean showCompleted) { /** * Create a copy using an immutable set to avoid a * {@link java.util.ConcurrentModificationException} */ Set<Integer> keySet = ImmutableSet.copyOf(operations.keySet()); if(showCompleted) return new ArrayList<Integer>(keySet); List<Integer> keyList = new ArrayList<Integer>(); for(int key: keySet) { AsyncOperation operation = operations.get(key); if(operation != null && !operation.getStatus().isComplete()) keyList.add(key); } return keyList; }
java
@JmxOperation public String stopAsyncOperation(int requestId) { try { stopOperation(requestId); } catch(VoldemortException e) { return e.getMessage(); } return "Stopping operation " + requestId; }
java
@Override public void updateStoreDefinition(StoreDefinition storeDef) { this.storeDef = storeDef; if(storeDef.hasRetentionPeriod()) this.retentionTimeMs = storeDef.getRetentionDays() * Time.MS_PER_DAY; }
java
private List<Versioned<byte[]>> filterExpiredEntries(ByteArray key, List<Versioned<byte[]>> vals) { Iterator<Versioned<byte[]>> valsIterator = vals.iterator(); while(valsIterator.hasNext()) { Versioned<byte[]> val = valsIterator.next(); VectorClock clock = (VectorClock) val.getVersion(); // omit if expired if(clock.getTimestamp() < (time.getMilliseconds() - this.retentionTimeMs)) { valsIterator.remove(); // delete stale value if configured if(deleteExpiredEntries) { getInnerStore().delete(key, clock); } } } return vals; }
java
private synchronized void flushData() { BufferedWriter writer = null; try { writer = new BufferedWriter(new FileWriter(new File(this.inputPath))); for(String key: this.metadataMap.keySet()) { writer.write(NEW_PROPERTY_SEPARATOR + key.toString() + "]" + NEW_LINE); writer.write(this.metadataMap.get(key).toString()); writer.write("" + NEW_LINE + "" + NEW_LINE); } writer.flush(); } catch(IOException e) { logger.error("IO exception while flushing data to file backed storage: " + e.getMessage()); } try { if(writer != null) writer.close(); } catch(Exception e) { logger.error("Error while flushing data to file backed storage: " + e.getMessage()); } }
java
public static String getSerializedVectorClock(VectorClock vc) { VectorClockWrapper vcWrapper = new VectorClockWrapper(vc); String serializedVC = ""; try { serializedVC = mapper.writeValueAsString(vcWrapper); } catch(Exception e) { e.printStackTrace(); } return serializedVC; }
java
public static String getSerializedVectorClocks(List<VectorClock> vectorClocks) { List<VectorClockWrapper> vectorClockWrappers = new ArrayList<VectorClockWrapper>(); for(VectorClock vc: vectorClocks) { vectorClockWrappers.add(new VectorClockWrapper(vc)); } String serializedVC = ""; try { serializedVC = mapper.writeValueAsString(vectorClockWrappers); } catch(Exception e) { e.printStackTrace(); } return serializedVC; }
java
public static String constructSerializerInfoXml(StoreDefinition storeDefinition) { Element store = new Element(StoreDefinitionsMapper.STORE_ELMT); store.addContent(new Element(StoreDefinitionsMapper.STORE_NAME_ELMT).setText(storeDefinition.getName())); Element keySerializer = new Element(StoreDefinitionsMapper.STORE_KEY_SERIALIZER_ELMT); StoreDefinitionsMapper.addSerializer(keySerializer, storeDefinition.getKeySerializer()); store.addContent(keySerializer); Element valueSerializer = new Element(StoreDefinitionsMapper.STORE_VALUE_SERIALIZER_ELMT); StoreDefinitionsMapper.addSerializer(valueSerializer, storeDefinition.getValueSerializer()); store.addContent(valueSerializer); XMLOutputter serializer = new XMLOutputter(Format.getPrettyFormat()); return serializer.outputString(store); }
java
public void updateMetadataVersions() { Properties versionProps = MetadataVersionStoreUtils.getProperties(this.systemStoreRepository.getMetadataVersionStore()); Long newVersion = fetchNewVersion(SystemStoreConstants.CLUSTER_VERSION_KEY, null, versionProps); if(newVersion != null) { this.currentClusterVersion = newVersion; } }
java
public static void validateClusterStores(final Cluster cluster, final List<StoreDefinition> storeDefs) { // Constructing a StoreRoutingPlan has the (desirable in this // case) side-effect of verifying that the store definition is congruent // with the cluster definition. If there are issues, exceptions are // thrown. for(StoreDefinition storeDefinition: storeDefs) { new StoreRoutingPlan(cluster, storeDefinition); } return; }
java
public static void validateCurrentFinalCluster(final Cluster currentCluster, final Cluster finalCluster) { validateClusterPartitionCounts(currentCluster, finalCluster); validateClusterNodeState(currentCluster, finalCluster); return; }
java
public static void validateInterimFinalCluster(final Cluster interimCluster, final Cluster finalCluster) { validateClusterPartitionCounts(interimCluster, finalCluster); validateClusterZonesSame(interimCluster, finalCluster); validateClusterNodeCounts(interimCluster, finalCluster); validateClusterNodeState(interimCluster, finalCluster); return; }
java
public static void validateClusterPartitionCounts(final Cluster lhs, final Cluster rhs) { if(lhs.getNumberOfPartitions() != rhs.getNumberOfPartitions()) throw new VoldemortException("Total number of partitions should be equal [ lhs cluster (" + lhs.getNumberOfPartitions() + ") not equal to rhs cluster (" + rhs.getNumberOfPartitions() + ") ]"); }
java
public static void validateClusterPartitionState(final Cluster subsetCluster, final Cluster supersetCluster) { if(!supersetCluster.getNodeIds().containsAll(subsetCluster.getNodeIds())) { throw new VoldemortException("Superset cluster does not contain all nodes from subset cluster[ subset cluster node ids (" + subsetCluster.getNodeIds() + ") are not a subset of superset cluster node ids (" + supersetCluster.getNodeIds() + ") ]"); } for(int nodeId: subsetCluster.getNodeIds()) { Node supersetNode = supersetCluster.getNodeById(nodeId); Node subsetNode = subsetCluster.getNodeById(nodeId); if(!supersetNode.getPartitionIds().equals(subsetNode.getPartitionIds())) { throw new VoldemortRebalancingException("Partition IDs do not match between clusters for nodes with id " + nodeId + " : subset cluster has " + subsetNode.getPartitionIds() + " and superset cluster has " + supersetNode.getPartitionIds()); } } Set<Integer> nodeIds = supersetCluster.getNodeIds(); nodeIds.removeAll(subsetCluster.getNodeIds()); for(int nodeId: nodeIds) { Node supersetNode = supersetCluster.getNodeById(nodeId); if(!supersetNode.getPartitionIds().isEmpty()) { throw new VoldemortRebalancingException("New node " + nodeId + " in superset cluster already has partitions: " + supersetNode.getPartitionIds()); } } }
java
public static void validateClusterZonesSame(final Cluster lhs, final Cluster rhs) { Set<Zone> lhsSet = new HashSet<Zone>(lhs.getZones()); Set<Zone> rhsSet = new HashSet<Zone>(rhs.getZones()); if(!lhsSet.equals(rhsSet)) throw new VoldemortException("Zones are not the same [ lhs cluster zones (" + lhs.getZones() + ") not equal to rhs cluster zones (" + rhs.getZones() + ") ]"); }
java
public static void validateClusterNodeCounts(final Cluster lhs, final Cluster rhs) { if(!lhs.getNodeIds().equals(rhs.getNodeIds())) { throw new VoldemortException("Node ids are not the same [ lhs cluster node ids (" + lhs.getNodeIds() + ") not equal to rhs cluster node ids (" + rhs.getNodeIds() + ") ]"); } }
java
public static Cluster vacateZone(Cluster currentCluster, int dropZoneId) { Cluster returnCluster = Cluster.cloneCluster(currentCluster); // Go over each node in the zone being dropped for(Integer nodeId: currentCluster.getNodeIdsInZone(dropZoneId)) { // For each node grab all the partitions it hosts for(Integer partitionId: currentCluster.getNodeById(nodeId).getPartitionIds()) { // Now for each partition find a new home..which would be a node // in one of the existing zones int finalZoneId = -1; int finalNodeId = -1; int adjacentPartitionId = partitionId; do { adjacentPartitionId = (adjacentPartitionId + 1) % currentCluster.getNumberOfPartitions(); finalNodeId = currentCluster.getNodeForPartitionId(adjacentPartitionId).getId(); finalZoneId = currentCluster.getZoneForPartitionId(adjacentPartitionId).getId(); if(adjacentPartitionId == partitionId) { logger.error("PartitionId " + partitionId + "stays unchanged \n"); } else { logger.info("PartitionId " + partitionId + " goes together with partition " + adjacentPartitionId + " on node " + finalNodeId + " in zone " + finalZoneId); returnCluster = UpdateClusterUtils.createUpdatedCluster(returnCluster, finalNodeId, Lists.newArrayList(partitionId)); } } while(finalZoneId == dropZoneId); } } return returnCluster; }
java
public static Cluster dropZone(Cluster intermediateCluster, int dropZoneId) { // Filter out nodes that don't belong to the zone being dropped Set<Node> survivingNodes = new HashSet<Node>(); for(int nodeId: intermediateCluster.getNodeIds()) { if(intermediateCluster.getNodeById(nodeId).getZoneId() != dropZoneId) { survivingNodes.add(intermediateCluster.getNodeById(nodeId)); } } // Filter out dropZoneId from all zones Set<Zone> zones = new HashSet<Zone>(); for(int zoneId: intermediateCluster.getZoneIds()) { if(zoneId == dropZoneId) { continue; } List<Integer> proximityList = intermediateCluster.getZoneById(zoneId) .getProximityList(); proximityList.remove(new Integer(dropZoneId)); zones.add(new Zone(zoneId, proximityList)); } return new Cluster(intermediateCluster.getName(), Utils.asSortedList(survivingNodes), Utils.asSortedList(zones)); }
java
public static List<Integer> getStolenPrimaryPartitions(final Cluster currentCluster, final Cluster finalCluster, final int stealNodeId) { List<Integer> finalList = new ArrayList<Integer>(finalCluster.getNodeById(stealNodeId) .getPartitionIds()); List<Integer> currentList = new ArrayList<Integer>(); if(currentCluster.hasNodeWithId(stealNodeId)) { currentList = currentCluster.getNodeById(stealNodeId).getPartitionIds(); } else { if(logger.isDebugEnabled()) { logger.debug("Current cluster does not contain stealer node (cluster : [[[" + currentCluster + "]]], node id " + stealNodeId + ")"); } } finalList.removeAll(currentList); return finalList; }
java
public static List<StoreDefinition> validateRebalanceStore(List<StoreDefinition> storeDefList) { List<StoreDefinition> returnList = new ArrayList<StoreDefinition>(storeDefList.size()); for(StoreDefinition def: storeDefList) { if(!def.isView() && !canRebalanceList.contains(def.getType())) { throw new VoldemortException("Rebalance does not support rebalancing of stores of type " + def.getType() + " - " + def.getName()); } else if(!def.isView()) { returnList.add(def); } else { logger.debug("Ignoring view " + def.getName() + " for rebalancing"); } } return returnList; }
java
public static void dumpClusters(Cluster currentCluster, Cluster finalCluster, String outputDirName, String filePrefix) { dumpClusterToFile(outputDirName, filePrefix + currentClusterFileName, currentCluster); dumpClusterToFile(outputDirName, filePrefix + finalClusterFileName, finalCluster); }
java
public static void dumpClusters(Cluster currentCluster, Cluster finalCluster, String outputDirName) { dumpClusters(currentCluster, finalCluster, outputDirName, ""); }
java
public static void dumpClusterToFile(String outputDirName, String fileName, Cluster cluster) { if(outputDirName != null) { File outputDir = new File(outputDirName); if(!outputDir.exists()) { Utils.mkdirs(outputDir); } try { FileUtils.writeStringToFile(new File(outputDirName, fileName), new ClusterMapper().writeCluster(cluster)); } catch(IOException e) { logger.error("IOException during dumpClusterToFile: " + e); } } }
java
public static void dumpStoreDefsToFile(String outputDirName, String fileName, List<StoreDefinition> storeDefs) { if(outputDirName != null) { File outputDir = new File(outputDirName); if(!outputDir.exists()) { Utils.mkdirs(outputDir); } try { FileUtils.writeStringToFile(new File(outputDirName, fileName), new StoreDefinitionsMapper().writeStoreList(storeDefs)); } catch(IOException e) { logger.error("IOException during dumpStoreDefsToFile: " + e); } } }
java
public static void dumpAnalysisToFile(String outputDirName, String baseFileName, PartitionBalance partitionBalance) { if(outputDirName != null) { File outputDir = new File(outputDirName); if(!outputDir.exists()) { Utils.mkdirs(outputDir); } try { FileUtils.writeStringToFile(new File(outputDirName, baseFileName + ".analysis"), partitionBalance.toString()); } catch(IOException e) { logger.error("IOException during dumpAnalysisToFile: " + e); } } }
java
public static void dumpPlanToFile(String outputDirName, RebalancePlan plan) { if(outputDirName != null) { File outputDir = new File(outputDirName); if(!outputDir.exists()) { Utils.mkdirs(outputDir); } try { FileUtils.writeStringToFile(new File(outputDirName, "plan.out"), plan.toString()); } catch(IOException e) { logger.error("IOException during dumpPlanToFile: " + e); } } }
java
public static List<RebalanceTaskInfo> filterTaskPlanWithStores(List<RebalanceTaskInfo> existingPlanList, List<StoreDefinition> storeDefs) { List<RebalanceTaskInfo> plans = Lists.newArrayList(); List<String> storeNames = StoreDefinitionUtils.getStoreNames(storeDefs); for(RebalanceTaskInfo existingPlan: existingPlanList) { RebalanceTaskInfo info = RebalanceTaskInfo.create(existingPlan.toJsonString()); // Filter the plans only for stores given HashMap<String, List<Integer>> storeToPartitions = info.getStoreToPartitionIds(); HashMap<String, List<Integer>> newStoreToPartitions = Maps.newHashMap(); for(String storeName: storeNames) { if(storeToPartitions.containsKey(storeName)) newStoreToPartitions.put(storeName, storeToPartitions.get(storeName)); } info.setStoreToPartitionList(newStoreToPartitions); plans.add(info); } return plans; }
java
public static void executorShutDown(ExecutorService executorService, long timeOutSec) { try { executorService.shutdown(); executorService.awaitTermination(timeOutSec, TimeUnit.SECONDS); } catch(Exception e) { logger.warn("Error while stoping executor service.", e); } }
java
public List<T> resolveConflicts(List<T> values) { if(values.size() > 1) return values; else return Collections.singletonList(values.get(0)); }
java
public static <K, V, T> List<Versioned<V>> get(Store<K, V, T> storageEngine, K key, T transform) { Map<K, List<Versioned<V>>> result = storageEngine.getAll(Collections.singleton(key), Collections.singletonMap(key, transform)); if(result.size() > 0) return result.get(key); else return Collections.emptyList(); }
java
public static <K, V, T> Map<K, List<Versioned<V>>> getAll(Store<K, V, T> storageEngine, Iterable<K> keys, Map<K, T> transforms) { Map<K, List<Versioned<V>>> result = newEmptyHashMap(keys); for(K key: keys) { List<Versioned<V>> value = storageEngine.get(key, transforms != null ? transforms.get(key) : null); if(!value.isEmpty()) result.put(key, value); } return result; }
java
public static <K, V> HashMap<K, V> newEmptyHashMap(Iterable<?> iterable) { if(iterable instanceof Collection<?>) return Maps.newHashMapWithExpectedSize(((Collection<?>) iterable).size()); return Maps.newHashMap(); }
java
public static void assertValidMetadata(ByteArray key, RoutingStrategy routingStrategy, Node currentNode) { List<Node> nodes = routingStrategy.routeRequest(key.get()); for(Node node: nodes) { if(node.getId() == currentNode.getId()) { return; } } throw new InvalidMetadataException("Client accessing key belonging to partitions " + routingStrategy.getPartitionList(key.get()) + " not present at " + currentNode); }
java
public static void assertValidNode(MetadataStore metadataStore, Integer nodeId) { if(!metadataStore.getCluster().hasNodeWithId(nodeId)) { throw new InvalidMetadataException("NodeId " + nodeId + " is not or no longer in this cluster"); } }
java
@SuppressWarnings("unchecked") public static <T> Serializer<T> unsafeGetSerializer(SerializerFactory serializerFactory, SerializerDefinition serializerDefinition) { return (Serializer<T>) serializerFactory.getSerializer(serializerDefinition); }
java
public static StoreDefinition getStoreDef(List<StoreDefinition> list, String name) { for(StoreDefinition def: list) if(def.getName().equals(name)) return def; return null; }
java
public static List<String> getStoreNames(List<StoreDefinition> list, boolean ignoreViews) { List<String> storeNameSet = new ArrayList<String>(); for(StoreDefinition def: list) if(!def.isView() || !ignoreViews) storeNameSet.add(def.getName()); return storeNameSet; }
java
private void plan() { // Mapping of stealer node to list of primary partitions being moved final TreeMultimap<Integer, Integer> stealerToStolenPrimaryPartitions = TreeMultimap.create(); // Output initial and final cluster if(outputDir != null) RebalanceUtils.dumpClusters(currentCluster, finalCluster, outputDir); // Determine which partitions must be stolen for(Node stealerNode: finalCluster.getNodes()) { List<Integer> stolenPrimaryPartitions = RebalanceUtils.getStolenPrimaryPartitions(currentCluster, finalCluster, stealerNode.getId()); if(stolenPrimaryPartitions.size() > 0) { numPrimaryPartitionMoves += stolenPrimaryPartitions.size(); stealerToStolenPrimaryPartitions.putAll(stealerNode.getId(), stolenPrimaryPartitions); } } // Determine plan batch-by-batch int batches = 0; Cluster batchCurrentCluster = Cluster.cloneCluster(currentCluster); List<StoreDefinition> batchCurrentStoreDefs = this.currentStoreDefs; List<StoreDefinition> batchFinalStoreDefs = this.finalStoreDefs; Cluster batchFinalCluster = RebalanceUtils.getInterimCluster(this.currentCluster, this.finalCluster); while(!stealerToStolenPrimaryPartitions.isEmpty()) { int partitions = 0; List<Entry<Integer, Integer>> partitionsMoved = Lists.newArrayList(); for(Entry<Integer, Integer> stealerToPartition: stealerToStolenPrimaryPartitions.entries()) { partitionsMoved.add(stealerToPartition); batchFinalCluster = UpdateClusterUtils.createUpdatedCluster(batchFinalCluster, stealerToPartition.getKey(), Lists.newArrayList(stealerToPartition.getValue())); partitions++; if(partitions == batchSize) break; } // Remove the partitions moved for(Iterator<Entry<Integer, Integer>> partitionMoved = partitionsMoved.iterator(); partitionMoved.hasNext();) { Entry<Integer, Integer> entry = partitionMoved.next(); stealerToStolenPrimaryPartitions.remove(entry.getKey(), entry.getValue()); } if(outputDir != null) RebalanceUtils.dumpClusters(batchCurrentCluster, batchFinalCluster, outputDir, "batch-" + Integer.toString(batches) + "."); // Generate a plan to compute the tasks final RebalanceBatchPlan RebalanceBatchPlan = new RebalanceBatchPlan(batchCurrentCluster, batchCurrentStoreDefs, batchFinalCluster, batchFinalStoreDefs); batchPlans.add(RebalanceBatchPlan); numXZonePartitionStoreMoves += RebalanceBatchPlan.getCrossZonePartitionStoreMoves(); numPartitionStoreMoves += RebalanceBatchPlan.getPartitionStoreMoves(); nodeMoveMap.add(RebalanceBatchPlan.getNodeMoveMap()); zoneMoveMap.add(RebalanceBatchPlan.getZoneMoveMap()); batches++; batchCurrentCluster = Cluster.cloneCluster(batchFinalCluster); // batchCurrentStoreDefs can only be different from // batchFinalStoreDefs for the initial batch. batchCurrentStoreDefs = batchFinalStoreDefs; } logger.info(this); }
java
private String storageOverhead(Map<Integer, Integer> finalNodeToOverhead) { double maxOverhead = Double.MIN_VALUE; PartitionBalance pb = new PartitionBalance(currentCluster, currentStoreDefs); StringBuilder sb = new StringBuilder(); sb.append("Per-node store-overhead:").append(Utils.NEWLINE); DecimalFormat doubleDf = new DecimalFormat("####.##"); for(int nodeId: finalCluster.getNodeIds()) { Node node = finalCluster.getNodeById(nodeId); String nodeTag = "Node " + String.format("%4d", nodeId) + " (" + node.getHost() + ")"; int initialLoad = 0; if(currentCluster.getNodeIds().contains(nodeId)) { initialLoad = pb.getNaryPartitionCount(nodeId); } int toLoad = 0; if(finalNodeToOverhead.containsKey(nodeId)) { toLoad = finalNodeToOverhead.get(nodeId); } double overhead = (initialLoad + toLoad) / (double) initialLoad; if(initialLoad > 0 && maxOverhead < overhead) { maxOverhead = overhead; } String loadTag = String.format("%6d", initialLoad) + " + " + String.format("%6d", toLoad) + " -> " + String.format("%6d", initialLoad + toLoad) + " (" + doubleDf.format(overhead) + " X)"; sb.append(nodeTag + " : " + loadTag).append(Utils.NEWLINE); } sb.append(Utils.NEWLINE) .append("**** Max per-node storage overhead: " + doubleDf.format(maxOverhead) + " X.") .append(Utils.NEWLINE); return (sb.toString()); }
java
public static Boolean askConfirm(Boolean confirm, String opDesc) throws IOException { if(confirm) { System.out.println("Confirmed " + opDesc + " in command-line."); return true; } else { System.out.println("Are you sure you want to " + opDesc + "? (yes/no)"); BufferedReader buffer = new BufferedReader(new InputStreamReader(System.in)); String text = buffer.readLine().toLowerCase(Locale.ENGLISH); boolean go = text.equals("yes") || text.equals("y"); if (!go) { System.out.println("Did not confirm; " + opDesc + " aborted."); } return go; } }
java
public static List<String> getValueList(List<String> valuePairs, String delim) { List<String> valueList = Lists.newArrayList(); for(String valuePair: valuePairs) { String[] value = valuePair.split(delim, 2); if(value.length != 2) throw new VoldemortException("Invalid argument pair: " + valuePair); valueList.add(value[0]); valueList.add(value[1]); } return valueList; }
java
public static <V> Map<V, V> convertListToMap(List<V> list) { Map<V, V> map = new HashMap<V, V>(); if(list.size() % 2 != 0) throw new VoldemortException("Failed to convert list to map."); for(int i = 0; i < list.size(); i += 2) { map.put(list.get(i), list.get(i + 1)); } return map; }
java
public static AdminClient getAdminClient(String url) { ClientConfig config = new ClientConfig().setBootstrapUrls(url) .setConnectionTimeout(5, TimeUnit.SECONDS); AdminClientConfig adminConfig = new AdminClientConfig().setAdminSocketTimeoutSec(5); return new AdminClient(adminConfig, config); }
java
public static List<Integer> getAllNodeIds(AdminClient adminClient) { List<Integer> nodeIds = Lists.newArrayList(); for(Integer nodeId: adminClient.getAdminClientCluster().getNodeIds()) { nodeIds.add(nodeId); } return nodeIds; }
java
public static List<String> getAllUserStoreNamesOnNode(AdminClient adminClient, Integer nodeId) { List<String> storeNames = Lists.newArrayList(); List<StoreDefinition> storeDefinitionList = adminClient.metadataMgmtOps.getRemoteStoreDefList(nodeId) .getValue(); for(StoreDefinition storeDefinition: storeDefinitionList) { storeNames.add(storeDefinition.getName()); } return storeNames; }
java
public static void validateUserStoreNamesOnNode(AdminClient adminClient, Integer nodeId, List<String> storeNames) { List<StoreDefinition> storeDefList = adminClient.metadataMgmtOps.getRemoteStoreDefList(nodeId) .getValue(); Map<String, Boolean> existingStoreNames = new HashMap<String, Boolean>(); for(StoreDefinition storeDef: storeDefList) { existingStoreNames.put(storeDef.getName(), true); } for(String storeName: storeNames) { if(!Boolean.TRUE.equals(existingStoreNames.get(storeName))) { Utils.croak("Store " + storeName + " does not exist!"); } } }
java
public static List<Integer> getAllPartitions(AdminClient adminClient) { List<Integer> partIds = Lists.newArrayList(); partIds = Lists.newArrayList(); for(Node node: adminClient.getAdminClientCluster().getNodes()) { partIds.addAll(node.getPartitionIds()); } return partIds; }
java
public static List<QuotaType> getQuotaTypes(List<String> strQuotaTypes) { if(strQuotaTypes.size() < 1) { throw new VoldemortException("Quota type not specified."); } List<QuotaType> quotaTypes; if(strQuotaTypes.size() == 1 && strQuotaTypes.get(0).equals(AdminToolUtils.QUOTATYPE_ALL)) { quotaTypes = Arrays.asList(QuotaType.values()); } else { quotaTypes = new ArrayList<QuotaType>(); for(String strQuotaType: strQuotaTypes) { QuotaType type = QuotaType.valueOf(strQuotaType); quotaTypes.add(type); } } return quotaTypes; }
java
public static File createDir(String dir) { // create outdir File directory = null; if(dir != null) { directory = new File(dir); if(!(directory.exists() || directory.mkdir())) { Utils.croak("Can't find or create directory " + dir); } } return directory; }
java
public static Map<String, StoreDefinition> getSystemStoreDefMap() { Map<String, StoreDefinition> sysStoreDefMap = Maps.newHashMap(); List<StoreDefinition> storesDefs = SystemStoreConstants.getAllSystemStoreDefs(); for(StoreDefinition def: storesDefs) { sysStoreDefMap.put(def.getName(), def); } return sysStoreDefMap; }
java
public static Map<String, StoreDefinition> getUserStoreDefMapOnNode(AdminClient adminClient, Integer nodeId) { List<StoreDefinition> storeDefinitionList = adminClient.metadataMgmtOps.getRemoteStoreDefList(nodeId) .getValue(); Map<String, StoreDefinition> storeDefinitionMap = Maps.newHashMap(); for(StoreDefinition storeDefinition: storeDefinitionList) { storeDefinitionMap.put(storeDefinition.getName(), storeDefinition); } return storeDefinitionMap; }
java
public static RebalanceTaskInfo decodeRebalanceTaskInfoMap(VAdminProto.RebalanceTaskInfoMap rebalanceTaskInfoMap) { RebalanceTaskInfo rebalanceTaskInfo = new RebalanceTaskInfo( rebalanceTaskInfoMap.getStealerId(), rebalanceTaskInfoMap.getDonorId(), decodeStoreToPartitionIds(rebalanceTaskInfoMap.getPerStorePartitionIdsList()), new ClusterMapper().readCluster(new StringReader(rebalanceTaskInfoMap.getInitialCluster()))); return rebalanceTaskInfo; }
java
public static RebalanceTaskInfoMap encodeRebalanceTaskInfoMap(RebalanceTaskInfo stealInfo) { return RebalanceTaskInfoMap.newBuilder() .setStealerId(stealInfo.getStealerId()) .setDonorId(stealInfo.getDonorId()) .addAllPerStorePartitionIds(ProtoUtils.encodeStoreToPartitionsTuple(stealInfo.getStoreToPartitionIds())) .setInitialCluster(new ClusterMapper().writeCluster(stealInfo.getInitialCluster())) .build(); }
java
public Versioned<E> getVersionedById(int id) { Versioned<VListNode<E>> listNode = getListNode(id); if(listNode == null) throw new IndexOutOfBoundsException(); return new Versioned<E>(listNode.getValue().getValue(), listNode.getVersion()); }
java
public E setById(final int id, final E element) { VListKey<K> key = new VListKey<K>(_key, id); UpdateElementById<K, E> updateElementAction = new UpdateElementById<K, E>(key, element); if(!_storeClient.applyUpdate(updateElementAction)) throw new ObsoleteVersionException("update failed"); return updateElementAction.getResult(); }
java
private void allClustersEqual(final List<String> clusterUrls) { Validate.notEmpty(clusterUrls, "clusterUrls cannot be null"); // If only one clusterUrl return immediately if (clusterUrls.size() == 1) return; AdminClient adminClientLhs = adminClientPerCluster.get(clusterUrls.get(0)); Cluster clusterLhs = adminClientLhs.getAdminClientCluster(); for (int index = 1; index < clusterUrls.size(); index++) { AdminClient adminClientRhs = adminClientPerCluster.get(clusterUrls.get(index)); Cluster clusterRhs = adminClientRhs.getAdminClientCluster(); if (!areTwoClustersEqual(clusterLhs, clusterRhs)) throw new VoldemortException("Cluster " + clusterLhs.getName() + " is not the same as " + clusterRhs.getName()); } }
java
private synchronized JsonSchema getInputPathJsonSchema() throws IOException { if (inputPathJsonSchema == null) { // No need to query Hadoop more than once as this shouldn't change mid-run, // thus, we can lazily initialize and cache the result. inputPathJsonSchema = HadoopUtils.getSchemaFromPath(getInputPath()); } return inputPathJsonSchema; }
java
private synchronized Schema getInputPathAvroSchema() throws IOException { if (inputPathAvroSchema == null) { // No need to query Hadoop more than once as this shouldn't change mid-run, // thus, we can lazily initialize and cache the result. inputPathAvroSchema = AvroUtils.getAvroSchemaFromPath(getInputPath()); } return inputPathAvroSchema; }
java
public String getRecordSchema() throws IOException { Schema schema = getInputPathAvroSchema(); String recSchema = schema.toString(); return recSchema; }
java
public String getKeySchema() throws IOException { Schema schema = getInputPathAvroSchema(); String keySchema = schema.getField(keyFieldName).schema().toString(); return keySchema; }
java
public String getValueSchema() throws IOException { Schema schema = getInputPathAvroSchema(); String valueSchema = schema.getField(valueFieldName).schema().toString(); return valueSchema; }
java
private void verifyOrAddStore(String clusterURL, String keySchema, String valueSchema) { String newStoreDefXml = VoldemortUtils.getStoreDefXml( storeName, props.getInt(BUILD_REPLICATION_FACTOR, 2), props.getInt(BUILD_REQUIRED_READS, 1), props.getInt(BUILD_REQUIRED_WRITES, 1), props.getNullableInt(BUILD_PREFERRED_READS), props.getNullableInt(BUILD_PREFERRED_WRITES), props.getString(PUSH_FORCE_SCHEMA_KEY, keySchema), props.getString(PUSH_FORCE_SCHEMA_VALUE, valueSchema), description, owners); log.info("Verifying store against cluster URL: " + clusterURL + "\n" + newStoreDefXml.toString()); StoreDefinition newStoreDef = VoldemortUtils.getStoreDef(newStoreDefXml); try { adminClientPerCluster.get(clusterURL).storeMgmtOps.verifyOrAddStore(newStoreDef, "BnP config/data", enableStoreCreation, this.storeVerificationExecutorService); } catch (UnreachableStoreException e) { log.info("verifyOrAddStore() failed on some nodes for clusterURL: " + clusterURL + " (this is harmless).", e); // When we can't reach some node, we just skip it and won't create the store on it. // Next time BnP is run while the node is up, it will get the store created. } // Other exceptions need to bubble up! storeDef = newStoreDef; }
java
public void syncInternalStateFromFileSystem(boolean alsoSyncRemoteState) { // Make sure versions missing from the file-system are cleaned up from the internal state for (Long version: versionToEnabledMap.keySet()) { File[] existingVersionDirs = ReadOnlyUtils.getVersionDirs(rootDir, version, version); if (existingVersionDirs.length == 0) { removeVersion(version, alsoSyncRemoteState); } } // Make sure we have all versions on the file-system in the internal state File[] versionDirs = ReadOnlyUtils.getVersionDirs(rootDir); if (versionDirs != null) { for (File versionDir: versionDirs) { long versionNumber = ReadOnlyUtils.getVersionId(versionDir); boolean versionEnabled = isVersionEnabled(versionDir); versionToEnabledMap.put(versionNumber, versionEnabled); } } // Identify the current version (based on a symlink in the file-system) File currentVersionDir = ReadOnlyUtils.getCurrentVersion(rootDir); if (currentVersionDir != null) { currentVersion = ReadOnlyUtils.getVersionId(currentVersionDir); } else { currentVersion = -1; // Should we throw instead? } logger.info("Successfully synced internal state from local file-system: " + this.toString()); }
java
private void persistDisabledVersion(long version) throws PersistenceFailureException { File disabledMarker = getDisabledMarkerFile(version); try { disabledMarker.createNewFile(); } catch (IOException e) { throw new PersistenceFailureException("Failed to create the disabled marker at path: " + disabledMarker.getAbsolutePath() + "\nThe store/version " + "will remain disabled only until the next restart.", e); } }
java
private void persistEnabledVersion(long version) throws PersistenceFailureException { File disabledMarker = getDisabledMarkerFile(version); if (disabledMarker.exists()) { if (!disabledMarker.delete()) { throw new PersistenceFailureException("Failed to create the disabled marker at path: " + disabledMarker.getAbsolutePath() + "\nThe store/version " + "will remain enabled only until the next restart."); } } }
java
private File getDisabledMarkerFile(long version) throws PersistenceFailureException { File[] versionDirArray = ReadOnlyUtils.getVersionDirs(rootDir, version, version); if (versionDirArray.length == 0) { throw new PersistenceFailureException("getDisabledMarkerFile did not find the requested version directory" + " on disk. Version: " + version + ", rootDir: " + rootDir); } File disabledMarkerFile = new File(versionDirArray[0], DISABLED_MARKER_NAME); return disabledMarkerFile; }
java
public Double getAvgEventValue() { resetIfNeeded(); synchronized(this) { long eventsLastInterval = numEventsLastInterval - numEventsLastLastInterval; if(eventsLastInterval > 0) return ((totalEventValueLastInterval - totalEventValueLastLastInterval) * 1.0) / eventsLastInterval; else return 0.0; } }
java
@SuppressWarnings("unchecked") public static void executeCommand(String[] args) throws IOException { OptionParser parser = getParser(); // declare parameters List<String> metaKeys = null; String url = null; // parse command-line input args = AdminToolUtils.copyArrayAddFirst(args, "--" + OPT_HEAD_META_CHECK); OptionSet options = parser.parse(args); if(options.has(AdminParserUtils.OPT_HELP)) { printHelp(System.out); return; } // check required options and/or conflicting options AdminParserUtils.checkRequired(options, OPT_HEAD_META_CHECK); AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_URL); // load parameters metaKeys = (List<String>) options.valuesOf(OPT_HEAD_META_CHECK); url = (String) options.valueOf(AdminParserUtils.OPT_URL); // execute command if(metaKeys.size() == 0 || (metaKeys.size() == 1 && metaKeys.get(0).equals(METAKEY_ALL))) { metaKeys = Lists.newArrayList(); metaKeys.add(MetadataStore.CLUSTER_KEY); metaKeys.add(MetadataStore.STORES_KEY); metaKeys.add(MetadataStore.SERVER_STATE_KEY); } AdminClient adminClient = AdminToolUtils.getAdminClient(url); doMetaCheck(adminClient, metaKeys); }
java
@SuppressWarnings("unchecked") public static void executeCommand(String[] args) throws IOException { OptionParser parser = getParser(); // declare parameters String url = null; List<Integer> nodeIds = null; Boolean allNodes = true; Boolean confirm = false; // parse command-line input OptionSet options = parser.parse(args); if(options.has(AdminParserUtils.OPT_HELP)) { printHelp(System.out); return; } // check required options and/or conflicting options AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_URL); AdminParserUtils.checkOptional(options, AdminParserUtils.OPT_NODE, AdminParserUtils.OPT_ALL_NODES); // load parameters url = (String) options.valueOf(AdminParserUtils.OPT_URL); if(options.has(AdminParserUtils.OPT_NODE)) { nodeIds = (List<Integer>) options.valuesOf(AdminParserUtils.OPT_NODE); allNodes = false; } if(options.has(AdminParserUtils.OPT_CONFIRM)) { confirm = true; } // print summary System.out.println("Remove metadata related to rebalancing"); System.out.println("Location:"); System.out.println(" bootstrap url = " + url); if(allNodes) { System.out.println(" node = all nodes"); } else { System.out.println(" node = " + Joiner.on(", ").join(nodeIds)); } // execute command if(!AdminToolUtils.askConfirm(confirm, "remove metadata related to rebalancing")) { return; } AdminClient adminClient = AdminToolUtils.getAdminClient(url); if(allNodes) { nodeIds = AdminToolUtils.getAllNodeIds(adminClient); } AdminToolUtils.assertServerNotInRebalancingState(adminClient, nodeIds); doMetaClearRebalance(adminClient, nodeIds); }
java
public static void doMetaClearRebalance(AdminClient adminClient, List<Integer> nodeIds) { AdminToolUtils.assertServerNotInOfflineState(adminClient, nodeIds); System.out.println("Setting " + MetadataStore.SERVER_STATE_KEY + " to " + MetadataStore.VoldemortState.NORMAL_SERVER); doMetaSet(adminClient, nodeIds, MetadataStore.SERVER_STATE_KEY, MetadataStore.VoldemortState.NORMAL_SERVER.toString()); RebalancerState state = RebalancerState.create("[]"); System.out.println("Cleaning up " + MetadataStore.REBALANCING_STEAL_INFO + " to " + state.toJsonString()); doMetaSet(adminClient, nodeIds, MetadataStore.REBALANCING_STEAL_INFO, state.toJsonString()); System.out.println("Cleaning up " + MetadataStore.REBALANCING_SOURCE_CLUSTER_XML + " to empty string"); doMetaSet(adminClient, nodeIds, MetadataStore.REBALANCING_SOURCE_CLUSTER_XML, ""); }
java
@SuppressWarnings("unchecked") public static void executeCommand(String[] args) throws IOException { OptionParser parser = getParser(); // declare parameters List<String> metaKeys = null; String url = null; String dir = null; List<Integer> nodeIds = null; Boolean allNodes = true; Boolean verbose = false; // parse command-line input args = AdminToolUtils.copyArrayAddFirst(args, "--" + OPT_HEAD_META_GET); OptionSet options = parser.parse(args); if(options.has(AdminParserUtils.OPT_HELP)) { printHelp(System.out); return; } // check required options and/or conflicting options AdminParserUtils.checkRequired(options, OPT_HEAD_META_GET); AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_URL); AdminParserUtils.checkOptional(options, AdminParserUtils.OPT_NODE, AdminParserUtils.OPT_ALL_NODES); // load parameters metaKeys = (List<String>) options.valuesOf(OPT_HEAD_META_GET); url = (String) options.valueOf(AdminParserUtils.OPT_URL); if(options.has(AdminParserUtils.OPT_DIR)) { dir = (String) options.valueOf(AdminParserUtils.OPT_DIR); } if(options.has(AdminParserUtils.OPT_NODE)) { nodeIds = (List<Integer>) options.valuesOf(AdminParserUtils.OPT_NODE); allNodes = false; } if(options.has(OPT_VERBOSE)) { verbose = true; } // execute command File directory = AdminToolUtils.createDir(dir); AdminClient adminClient = AdminToolUtils.getAdminClient(url); if(allNodes) { nodeIds = AdminToolUtils.getAllNodeIds(adminClient); } if(metaKeys.size() == 1 && metaKeys.get(0).equals(METAKEY_ALL)) { metaKeys = Lists.newArrayList(); for(Object key: MetadataStore.METADATA_KEYS) { metaKeys.add((String) key); } } doMetaGet(adminClient, nodeIds, metaKeys, directory, verbose); }
java
@SuppressWarnings("unchecked") public static void executeCommand(String[] args) throws IOException { OptionParser parser = getParser(); // declare parameters List<String> metaKeys = null; String url = null; List<Integer> nodeIds = null; Boolean allNodes = true; List<String> storeNames = null; // parse command-line input args = AdminToolUtils.copyArrayAddFirst(args, "--" + OPT_HEAD_META_GET_RO); OptionSet options = parser.parse(args); if(options.has(AdminParserUtils.OPT_HELP)) { printHelp(System.out); return; } // check required options and/or conflicting options AdminParserUtils.checkRequired(options, OPT_HEAD_META_GET_RO); AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_URL); AdminParserUtils.checkOptional(options, AdminParserUtils.OPT_NODE, AdminParserUtils.OPT_ALL_NODES); AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_STORE); // load parameters metaKeys = (List<String>) options.valuesOf(OPT_HEAD_META_GET_RO); url = (String) options.valueOf(AdminParserUtils.OPT_URL); if(options.has(AdminParserUtils.OPT_NODE)) { nodeIds = (List<Integer>) options.valuesOf(AdminParserUtils.OPT_NODE); allNodes = false; } storeNames = (List<String>) options.valuesOf(AdminParserUtils.OPT_STORE); // execute command AdminClient adminClient = AdminToolUtils.getAdminClient(url); if(allNodes) { nodeIds = AdminToolUtils.getAllNodeIds(adminClient); } if(metaKeys.size() == 1 && metaKeys.get(0).equals(METAKEY_ALL)) { metaKeys = Lists.newArrayList(); metaKeys.add(KEY_MAX_VERSION); metaKeys.add(KEY_CURRENT_VERSION); metaKeys.add(KEY_STORAGE_FORMAT); } doMetaGetRO(adminClient, nodeIds, storeNames, metaKeys); }
java
public static void doMetaGetRO(AdminClient adminClient, Collection<Integer> nodeIds, List<String> storeNames, List<String> metaKeys) throws IOException { for(String key: metaKeys) { System.out.println("Metadata: " + key); if(!key.equals(KEY_MAX_VERSION) && !key.equals(KEY_CURRENT_VERSION) && !key.equals(KEY_STORAGE_FORMAT)) { System.out.println(" Invalid read-only metadata key: " + key); } else { for(Integer nodeId: nodeIds) { String hostName = adminClient.getAdminClientCluster() .getNodeById(nodeId) .getHost(); System.out.println(" Node: " + hostName + ":" + nodeId); if(key.equals(KEY_MAX_VERSION)) { Map<String, Long> mapStoreToROVersion = adminClient.readonlyOps.getROMaxVersion(nodeId, storeNames); for(String storeName: mapStoreToROVersion.keySet()) { System.out.println(" " + storeName + ":" + mapStoreToROVersion.get(storeName)); } } else if(key.equals(KEY_CURRENT_VERSION)) { Map<String, Long> mapStoreToROVersion = adminClient.readonlyOps.getROCurrentVersion(nodeId, storeNames); for(String storeName: mapStoreToROVersion.keySet()) { System.out.println(" " + storeName + ":" + mapStoreToROVersion.get(storeName)); } } else if(key.equals(KEY_STORAGE_FORMAT)) { Map<String, String> mapStoreToROFormat = adminClient.readonlyOps.getROStorageFormat(nodeId, storeNames); for(String storeName: mapStoreToROFormat.keySet()) { System.out.println(" " + storeName + ":" + mapStoreToROFormat.get(storeName)); } } } } System.out.println(); } }
java
public static void doMetaUpdateVersionsOnStores(AdminClient adminClient, List<StoreDefinition> oldStoreDefs, List<StoreDefinition> newStoreDefs) { Set<String> storeNamesUnion = new HashSet<String>(); Map<String, StoreDefinition> oldStoreDefinitionMap = new HashMap<String, StoreDefinition>(); Map<String, StoreDefinition> newStoreDefinitionMap = new HashMap<String, StoreDefinition>(); List<String> storesChanged = new ArrayList<String>(); for(StoreDefinition storeDef: oldStoreDefs) { String storeName = storeDef.getName(); storeNamesUnion.add(storeName); oldStoreDefinitionMap.put(storeName, storeDef); } for(StoreDefinition storeDef: newStoreDefs) { String storeName = storeDef.getName(); storeNamesUnion.add(storeName); newStoreDefinitionMap.put(storeName, storeDef); } for(String storeName: storeNamesUnion) { StoreDefinition oldStoreDef = oldStoreDefinitionMap.get(storeName); StoreDefinition newStoreDef = newStoreDefinitionMap.get(storeName); if(oldStoreDef == null && newStoreDef != null || oldStoreDef != null && newStoreDef == null || oldStoreDef != null && newStoreDef != null && !oldStoreDef.equals(newStoreDef)) { storesChanged.add(storeName); } } System.out.println("Updating metadata version for the following stores: " + storesChanged); try { adminClient.metadataMgmtOps.updateMetadataversion(adminClient.getAdminClientCluster() .getNodeIds(), storesChanged); } catch(Exception e) { System.err.println("Error while updating metadata version for the specified store."); } }
java
public static void executeCommand(String[] args) throws IOException { OptionParser parser = getParser(); String url = null; Boolean confirm = false; // parse command-line input OptionSet options = parser.parse(args); if(options.has(AdminParserUtils.OPT_HELP)) { printHelp(System.out); return; } AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_URL); url = (String) options.valueOf(AdminParserUtils.OPT_URL); if(options.has(AdminParserUtils.OPT_CONFIRM)) { confirm = true; } // print summary System.out.println("Synchronize metadata versions across all nodes"); System.out.println("Location:"); System.out.println(" bootstrap url = " + url); System.out.println(" node = all nodes"); AdminClient adminClient = AdminToolUtils.getAdminClient(url); AdminToolUtils.assertServerNotInRebalancingState(adminClient); Versioned<Properties> versionedProps = mergeAllVersions(adminClient); printVersions(versionedProps); // execute command if(!AdminToolUtils.askConfirm(confirm, "do you want to synchronize metadata versions to all node")) return; adminClient.metadataMgmtOps.setMetadataVersion(versionedProps); }
java
public static void executeCommand(String[] args) throws IOException { OptionParser parser = getParser(); // declare parameters String url = null; // parse command-line input OptionSet options = parser.parse(args); if(options.has(AdminParserUtils.OPT_HELP)) { printHelp(System.out); return; } // check required options and/or conflicting options AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_URL); // load parameters url = (String) options.valueOf(AdminParserUtils.OPT_URL); // execute command AdminClient adminClient = AdminToolUtils.getAdminClient(url); doMetaCheckVersion(adminClient); }
java
private Integer getKeyPartitionId(byte[] key) { Integer keyPartitionId = storeInstance.getNodesPartitionIdForKey(nodeId, key); Utils.notNull(keyPartitionId); return keyPartitionId; }
java
protected boolean isItemAccepted(byte[] key) { boolean entryAccepted = false; if (!fetchOrphaned) { if (isKeyNeeded(key)) { entryAccepted = true; } } else { if (!StoreRoutingPlan.checkKeyBelongsToNode(key, nodeId, initialCluster, storeDef)) { entryAccepted = true; } } return entryAccepted; }
java
protected void accountForFetchedKey(byte[] key) { fetched++; if (streamStats != null) { streamStats.reportStreamingFetch(operation); } if (recordsPerPartition <= 0) { return; } Integer keyPartitionId = getKeyPartitionId(key); Long partitionFetch = partitionFetches.get(keyPartitionId); Utils.notNull(partitionFetch); partitionFetch++; partitionFetches.put(keyPartitionId, partitionFetch); if (partitionFetch == recordsPerPartition) { if (partitionsToFetch.contains(keyPartitionId)) { partitionsToFetch.remove(keyPartitionId); } else { logger.warn("Partitions to fetch did not contain expected partition ID: " + keyPartitionId); } } else if (partitionFetch > recordsPerPartition) { logger.warn("Partition fetch count larger than expected for partition ID " + keyPartitionId + " : " + partitionFetch); } }
java
protected StreamRequestHandlerState determineRequestHandlerState(String itemTag) { if (keyIterator.hasNext() && !fetchedEnoughForAllPartitions()) { return StreamRequestHandlerState.WRITING; } else { logger.info("Finished fetch " + itemTag + " for store '" + storageEngine.getName() + "' with partitions " + partitionIds); progressInfoMessage("Fetch " + itemTag + " (end of scan)"); return StreamRequestHandlerState.COMPLETE; } }
java
protected List<Versioned<V>> resolveAndConstructVersionsToPersist(List<Versioned<V>> valuesInStorage, List<Versioned<V>> multiPutValues) { List<Versioned<V>> obsoleteVals = new ArrayList<Versioned<V>>(multiPutValues.size()); // Go over all the values and determine whether the version is // acceptable for(Versioned<V> value: multiPutValues) { Iterator<Versioned<V>> iter = valuesInStorage.iterator(); boolean obsolete = false; // Compare the current version with a set of accepted versions while(iter.hasNext()) { Versioned<V> curr = iter.next(); Occurred occurred = value.getVersion().compare(curr.getVersion()); if(occurred == Occurred.BEFORE) { obsolete = true; break; } else if(occurred == Occurred.AFTER) { iter.remove(); } } if(obsolete) { // add to return value if obsolete obsoleteVals.add(value); } else { // else update the set of accepted versions valuesInStorage.add(value); } } return obsoleteVals; }
java
private <T> InternalProviderImpl installInternalProvider(Class<T> clazz, String bindingName, InternalProviderImpl<? extends T> internalProvider, boolean isBound, boolean isTestProvider) { if (bindingName == null) { if (isBound) { return installUnNamedProvider(mapClassesToUnNamedBoundProviders, clazz, internalProvider, isTestProvider); } else { return installUnNamedProvider(mapClassesToUnNamedUnBoundProviders, clazz, internalProvider, isTestProvider); } } else { return installNamedProvider(mapClassesToNamedBoundProviders, clazz, bindingName, internalProvider, isTestProvider); } }
java
@Override protected void reset() { super.reset(); mapClassesToNamedBoundProviders.clear(); mapClassesToUnNamedBoundProviders.clear(); hasTestModules = false; installBindingForScope(); }
java
public synchronized T get(Scope scope) { if (instance != null) { return instance; } if (providerInstance != null) { if (isProvidingSingletonInScope) { instance = providerInstance.get(); //gc providerInstance = null; return instance; } return providerInstance.get(); } if (factoryClass != null && factory == null) { factory = FactoryLocator.getFactory(factoryClass); //gc factoryClass = null; } if (factory != null) { if (!factory.hasScopeAnnotation() && !isCreatingSingletonInScope) { return factory.createInstance(scope); } instance = factory.createInstance(scope); //gc factory = null; return instance; } if (providerFactoryClass != null && providerFactory == null) { providerFactory = FactoryLocator.getFactory(providerFactoryClass); //gc providerFactoryClass = null; } if (providerFactory != null) { if (providerFactory.hasProvidesSingletonInScopeAnnotation() || isProvidingSingletonInScope) { instance = providerFactory.createInstance(scope).get(); //gc providerFactory = null; return instance; } if (providerFactory.hasScopeAnnotation() || isCreatingSingletonInScope) { providerInstance = providerFactory.createInstance(scope); //gc providerFactory = null; return providerInstance.get(); } return providerFactory.createInstance(scope).get(); } throw new IllegalStateException("A provider can only be used with an instance, a provider, a factory or a provider factory. Should not happen."); }
java
public static void closeScope(Object name) { //we remove the scope first, so that other threads don't see it, and see the next snapshot of the tree ScopeNode scope = (ScopeNode) MAP_KEY_TO_SCOPE.remove(name); if (scope != null) { ScopeNode parentScope = scope.getParentScope(); if (parentScope != null) { parentScope.removeChild(scope); } else { ConfigurationHolder.configuration.onScopeForestReset(); } removeScopeAndChildrenFromMap(scope); } }
java
public static void reset() { for (Object name : Collections.list(MAP_KEY_TO_SCOPE.keys())) { closeScope(name); } ConfigurationHolder.configuration.onScopeForestReset(); ScopeImpl.resetUnBoundProviders(); }
java
void notifyMwRevisionProcessors(MwRevision mwRevision, boolean isCurrent) { if (mwRevision == null || mwRevision.getPageId() <= 0) { return; } for (MwRevisionProcessorBroker.RevisionSubscription rs : this.revisionSubscriptions) { if (rs.onlyCurrentRevisions == isCurrent && (rs.model == null || mwRevision.getModel().equals( rs.model))) { rs.mwRevisionProcessor.processRevision(mwRevision); } } }
java
static ItemIdValueImpl fromIri(String iri) { int separator = iri.lastIndexOf('/') + 1; try { return new ItemIdValueImpl(iri.substring(separator), iri.substring(0, separator)); } catch (IllegalArgumentException e) { throw new IllegalArgumentException("Invalid Wikibase entity IRI: " + iri, e); } }
java