code stringlengths 73 34.1k | label stringclasses 1 value |
|---|---|
protected void populateTasksByStealer(List<StealerBasedRebalanceTask> sbTaskList) {
// Setup mapping of stealers to work for this run.
for(StealerBasedRebalanceTask task: sbTaskList) {
if(task.getStealInfos().size() != 1) {
throw new VoldemortException("StealerBasedRebalanceTasks should have a list of RebalancePartitionsInfo of length 1.");
}
RebalanceTaskInfo stealInfo = task.getStealInfos().get(0);
int stealerId = stealInfo.getStealerId();
if(!this.tasksByStealer.containsKey(stealerId)) {
this.tasksByStealer.put(stealerId, new ArrayList<StealerBasedRebalanceTask>());
}
this.tasksByStealer.get(stealerId).add(task);
}
if(tasksByStealer.isEmpty()) {
return;
}
// Shuffle order of each stealer's work list. This randomization
// helps to get rid of any "patterns" in how rebalancing tasks were
// added to the task list passed in.
for(List<StealerBasedRebalanceTask> taskList: tasksByStealer.values()) {
Collections.shuffle(taskList);
}
} | java |
protected synchronized StealerBasedRebalanceTask scheduleNextTask(boolean executeService) {
// Make sure there is work left to do.
if(doneSignal.getCount() == 0) {
logger.info("All tasks completion signaled... returning");
return null;
}
// Limit number of tasks outstanding.
if(this.numTasksExecuting >= maxParallelRebalancing) {
logger.info("Executing more tasks than [" + this.numTasksExecuting
+ "] the parallel allowed " + maxParallelRebalancing);
return null;
}
// Shuffle list of stealer IDs each time a new task to schedule needs to
// be found. Randomizing the order should avoid prioritizing one
// specific stealer's work ahead of all others.
List<Integer> stealerIds = new ArrayList<Integer>(tasksByStealer.keySet());
Collections.shuffle(stealerIds);
for(int stealerId: stealerIds) {
if(nodeIdsWithWork.contains(stealerId)) {
logger.info("Stealer " + stealerId + " is already working... continuing");
continue;
}
for(StealerBasedRebalanceTask sbTask: tasksByStealer.get(stealerId)) {
int donorId = sbTask.getStealInfos().get(0).getDonorId();
if(nodeIdsWithWork.contains(donorId)) {
logger.info("Stealer " + stealerId + " Donor " + donorId
+ " is already working... continuing");
continue;
}
// Book keeping
addNodesToWorkerList(Arrays.asList(stealerId, donorId));
numTasksExecuting++;
// Remove this task from list thus destroying list being
// iterated over. This is safe because returning directly out of
// this branch.
tasksByStealer.get(stealerId).remove(sbTask);
try {
if(executeService) {
logger.info("Stealer " + stealerId + " Donor " + donorId
+ " going to schedule work");
service.execute(sbTask);
}
} catch(RejectedExecutionException ree) {
logger.error("Stealer " + stealerId
+ "Rebalancing task rejected by executor service.", ree);
throw new VoldemortRebalancingException("Stealer "
+ stealerId
+ "Rebalancing task rejected by executor service.");
}
return sbTask;
}
}
printRemainingTasks(stealerIds);
return null;
} | java |
public synchronized void addNodesToWorkerList(List<Integer> nodeIds) {
// Bookkeeping for nodes that will be involved in the next task
nodeIdsWithWork.addAll(nodeIds);
logger.info("Node IDs with work: " + nodeIdsWithWork + " Newly added nodes " + nodeIds);
} | java |
public synchronized void doneTask(int stealerId, int donorId) {
removeNodesFromWorkerList(Arrays.asList(stealerId, donorId));
numTasksExecuting--;
doneSignal.countDown();
// Try and schedule more tasks now that resources may be available to do
// so.
scheduleMoreTasks();
} | java |
private List<Long> collectLongMetric(String metricGetterName) {
List<Long> vals = new ArrayList<Long>();
for(BdbEnvironmentStats envStats: environmentStatsTracked) {
vals.add((Long) ReflectUtils.callMethod(envStats,
BdbEnvironmentStats.class,
metricGetterName,
new Class<?>[0],
new Object[0]));
}
return vals;
} | java |
public static Iterable<String> toHexStrings(Iterable<ByteArray> arrays) {
ArrayList<String> ret = new ArrayList<String>();
for(ByteArray array: arrays)
ret.add(ByteUtils.toHexString(array.get()));
return ret;
} | java |
@Override
public void sendResponse(StoreStats performanceStats,
boolean isFromLocalZone,
long startTimeInMs) throws Exception {
/*
* Pay attention to the code below. Note that in this method we wrap a multiPart object with a mimeMessage.
* However when writing to the outputStream we only send the multiPart object and not the entire
* mimeMessage. This is intentional.
*
* In the earlier version of this code we used to create a multiPart object and just send that multiPart
* across the wire.
*
* However, we later discovered that upon setting the content of a MimeBodyPart, JavaMail internally creates
* a DataHandler object wrapping the object you passed in. The part's Content-Type header is not updated
* immediately. In order to get the headers updated, one needs to to call MimeMessage.saveChanges() on the
* enclosing message, which cascades down the MIME structure into a call to MimeBodyPart.updateHeaders()
* on the body part. It's this updateHeaders call that transfers the content type from the
* DataHandler to the part's MIME Content-Type header.
*
* To make sure that the Content-Type headers are being updated (without changing too much code), we decided
* to wrap the multiPart in a mimeMessage, call mimeMessage.saveChanges() and then just send the multiPart.
* This is to make sure multiPart's headers are updated accurately.
*/
MimeMessage message = new MimeMessage(Session.getDefaultInstance(new Properties()));
MimeMultipart multiPart = new MimeMultipart();
ByteArrayOutputStream outputStream = new ByteArrayOutputStream();
String base64Key = RestUtils.encodeVoldemortKey(key.get());
String contentLocationKey = "/" + this.storeName + "/" + base64Key;
for(Versioned<byte[]> versionedValue: versionedValues) {
byte[] responseValue = versionedValue.getValue();
VectorClock vectorClock = (VectorClock) versionedValue.getVersion();
String eTag = RestUtils.getSerializedVectorClock(vectorClock);
numVectorClockEntries += vectorClock.getVersionMap().size();
// Create the individual body part for each versioned value of the
// requested key
MimeBodyPart body = new MimeBodyPart();
try {
// Add the right headers
body.addHeader(CONTENT_TYPE, "application/octet-stream");
body.addHeader(CONTENT_TRANSFER_ENCODING, "binary");
body.addHeader(RestMessageHeaders.X_VOLD_VECTOR_CLOCK, eTag);
body.setContent(responseValue, "application/octet-stream");
body.addHeader(RestMessageHeaders.CONTENT_LENGTH,
Integer.toString(responseValue.length));
multiPart.addBodyPart(body);
} catch(MessagingException me) {
logger.error("Exception while constructing body part", me);
outputStream.close();
throw me;
}
}
message.setContent(multiPart);
message.saveChanges();
try {
multiPart.writeTo(outputStream);
} catch(Exception e) {
logger.error("Exception while writing multipart to output stream", e);
outputStream.close();
throw e;
}
ChannelBuffer responseContent = ChannelBuffers.dynamicBuffer();
responseContent.writeBytes(outputStream.toByteArray());
// Create the Response object
HttpResponse response = new DefaultHttpResponse(HTTP_1_1, OK);
// Set the right headers
response.setHeader(CONTENT_TYPE, "multipart/binary");
response.setHeader(CONTENT_TRANSFER_ENCODING, "binary");
response.setHeader(CONTENT_LOCATION, contentLocationKey);
// Copy the data into the payload
response.setContent(responseContent);
response.setHeader(CONTENT_LENGTH, response.getContent().readableBytes());
// Write the response to the Netty Channel
if(logger.isDebugEnabled()) {
String keyStr = RestUtils.getKeyHexString(this.key);
debugLog("GET",
this.storeName,
keyStr,
startTimeInMs,
System.currentTimeMillis(),
numVectorClockEntries);
}
this.messageEvent.getChannel().write(response);
if(performanceStats != null && isFromLocalZone) {
recordStats(performanceStats, startTimeInMs, Tracked.GET);
}
outputStream.close();
} | java |
public String getPublicConfigValue(String key) throws ConfigurationException {
if (!allProps.containsKey(key)) {
throw new UndefinedPropertyException("The requested config key does not exist.");
}
if (restrictedConfigs.contains(key)) {
throw new ConfigurationException("The requested config key is not publicly available!");
}
return allProps.get(key);
} | java |
private void checkRateLimit(String quotaKey, Tracked trackedOp) {
String quotaValue = null;
try {
if(!metadataStore.getQuotaEnforcingEnabledUnlocked()) {
return;
}
quotaValue = quotaStore.cacheGet(quotaKey);
// Store may not have any quotas
if(quotaValue == null) {
return;
}
// But, if it does
float currentRate = getThroughput(trackedOp);
float allowedRate = Float.parseFloat(quotaValue);
// TODO the histogram should be reasonably accurate to do all
// these things.. (ghost qps and all)
// Report the current quota usage level
quotaStats.reportQuotaUsed(trackedOp, Utils.safeGetPercentage(currentRate, allowedRate));
// check if we have exceeded rate.
if(currentRate > allowedRate) {
quotaStats.reportRateLimitedOp(trackedOp);
throw new QuotaExceededException("Exceeded rate limit for " + quotaKey
+ ". Maximum allowed : " + allowedRate
+ " Current: " + currentRate);
}
} catch(NumberFormatException nfe) {
// move on, if we cannot parse quota value properly
logger.debug("Invalid formatting of quota value for key " + quotaKey + " : "
+ quotaValue);
}
} | java |
public synchronized void submitOperation(int requestId, AsyncOperation operation) {
if(this.operations.containsKey(requestId))
throw new VoldemortException("Request " + requestId
+ " already submitted to the system");
this.operations.put(requestId, operation);
scheduler.scheduleNow(operation);
logger.debug("Handling async operation " + requestId);
} | java |
public synchronized boolean isComplete(int requestId, boolean remove) {
if (!operations.containsKey(requestId))
throw new VoldemortException("No operation with id " + requestId + " found");
if (operations.get(requestId).getStatus().isComplete()) {
if (logger.isDebugEnabled())
logger.debug("Operation complete " + requestId);
if (remove)
operations.remove(requestId);
return true;
}
return false;
} | java |
@JmxOperation(description = "Retrieve operation status")
public String getStatus(int id) {
try {
return getOperationStatus(id).toString();
} catch(VoldemortException e) {
return "No operation with id " + id + " found";
}
} | java |
public List<Integer> getAsyncOperationList(boolean showCompleted) {
/**
* Create a copy using an immutable set to avoid a
* {@link java.util.ConcurrentModificationException}
*/
Set<Integer> keySet = ImmutableSet.copyOf(operations.keySet());
if(showCompleted)
return new ArrayList<Integer>(keySet);
List<Integer> keyList = new ArrayList<Integer>();
for(int key: keySet) {
AsyncOperation operation = operations.get(key);
if(operation != null && !operation.getStatus().isComplete())
keyList.add(key);
}
return keyList;
} | java |
@JmxOperation
public String stopAsyncOperation(int requestId) {
try {
stopOperation(requestId);
} catch(VoldemortException e) {
return e.getMessage();
}
return "Stopping operation " + requestId;
} | java |
@Override
public void updateStoreDefinition(StoreDefinition storeDef) {
this.storeDef = storeDef;
if(storeDef.hasRetentionPeriod())
this.retentionTimeMs = storeDef.getRetentionDays() * Time.MS_PER_DAY;
} | java |
private List<Versioned<byte[]>> filterExpiredEntries(ByteArray key, List<Versioned<byte[]>> vals) {
Iterator<Versioned<byte[]>> valsIterator = vals.iterator();
while(valsIterator.hasNext()) {
Versioned<byte[]> val = valsIterator.next();
VectorClock clock = (VectorClock) val.getVersion();
// omit if expired
if(clock.getTimestamp() < (time.getMilliseconds() - this.retentionTimeMs)) {
valsIterator.remove();
// delete stale value if configured
if(deleteExpiredEntries) {
getInnerStore().delete(key, clock);
}
}
}
return vals;
} | java |
private synchronized void flushData() {
BufferedWriter writer = null;
try {
writer = new BufferedWriter(new FileWriter(new File(this.inputPath)));
for(String key: this.metadataMap.keySet()) {
writer.write(NEW_PROPERTY_SEPARATOR + key.toString() + "]" + NEW_LINE);
writer.write(this.metadataMap.get(key).toString());
writer.write("" + NEW_LINE + "" + NEW_LINE);
}
writer.flush();
} catch(IOException e) {
logger.error("IO exception while flushing data to file backed storage: "
+ e.getMessage());
}
try {
if(writer != null)
writer.close();
} catch(Exception e) {
logger.error("Error while flushing data to file backed storage: " + e.getMessage());
}
} | java |
public static String getSerializedVectorClock(VectorClock vc) {
VectorClockWrapper vcWrapper = new VectorClockWrapper(vc);
String serializedVC = "";
try {
serializedVC = mapper.writeValueAsString(vcWrapper);
} catch(Exception e) {
e.printStackTrace();
}
return serializedVC;
} | java |
public static String getSerializedVectorClocks(List<VectorClock> vectorClocks) {
List<VectorClockWrapper> vectorClockWrappers = new ArrayList<VectorClockWrapper>();
for(VectorClock vc: vectorClocks) {
vectorClockWrappers.add(new VectorClockWrapper(vc));
}
String serializedVC = "";
try {
serializedVC = mapper.writeValueAsString(vectorClockWrappers);
} catch(Exception e) {
e.printStackTrace();
}
return serializedVC;
} | java |
public static String constructSerializerInfoXml(StoreDefinition storeDefinition) {
Element store = new Element(StoreDefinitionsMapper.STORE_ELMT);
store.addContent(new Element(StoreDefinitionsMapper.STORE_NAME_ELMT).setText(storeDefinition.getName()));
Element keySerializer = new Element(StoreDefinitionsMapper.STORE_KEY_SERIALIZER_ELMT);
StoreDefinitionsMapper.addSerializer(keySerializer, storeDefinition.getKeySerializer());
store.addContent(keySerializer);
Element valueSerializer = new Element(StoreDefinitionsMapper.STORE_VALUE_SERIALIZER_ELMT);
StoreDefinitionsMapper.addSerializer(valueSerializer, storeDefinition.getValueSerializer());
store.addContent(valueSerializer);
XMLOutputter serializer = new XMLOutputter(Format.getPrettyFormat());
return serializer.outputString(store);
} | java |
public void updateMetadataVersions() {
Properties versionProps = MetadataVersionStoreUtils.getProperties(this.systemStoreRepository.getMetadataVersionStore());
Long newVersion = fetchNewVersion(SystemStoreConstants.CLUSTER_VERSION_KEY,
null,
versionProps);
if(newVersion != null) {
this.currentClusterVersion = newVersion;
}
} | java |
public static void validateClusterStores(final Cluster cluster,
final List<StoreDefinition> storeDefs) {
// Constructing a StoreRoutingPlan has the (desirable in this
// case) side-effect of verifying that the store definition is congruent
// with the cluster definition. If there are issues, exceptions are
// thrown.
for(StoreDefinition storeDefinition: storeDefs) {
new StoreRoutingPlan(cluster, storeDefinition);
}
return;
} | java |
public static void validateCurrentFinalCluster(final Cluster currentCluster,
final Cluster finalCluster) {
validateClusterPartitionCounts(currentCluster, finalCluster);
validateClusterNodeState(currentCluster, finalCluster);
return;
} | java |
public static void validateInterimFinalCluster(final Cluster interimCluster,
final Cluster finalCluster) {
validateClusterPartitionCounts(interimCluster, finalCluster);
validateClusterZonesSame(interimCluster, finalCluster);
validateClusterNodeCounts(interimCluster, finalCluster);
validateClusterNodeState(interimCluster, finalCluster);
return;
} | java |
public static void validateClusterPartitionCounts(final Cluster lhs, final Cluster rhs) {
if(lhs.getNumberOfPartitions() != rhs.getNumberOfPartitions())
throw new VoldemortException("Total number of partitions should be equal [ lhs cluster ("
+ lhs.getNumberOfPartitions()
+ ") not equal to rhs cluster ("
+ rhs.getNumberOfPartitions() + ") ]");
} | java |
public static void validateClusterPartitionState(final Cluster subsetCluster,
final Cluster supersetCluster) {
if(!supersetCluster.getNodeIds().containsAll(subsetCluster.getNodeIds())) {
throw new VoldemortException("Superset cluster does not contain all nodes from subset cluster[ subset cluster node ids ("
+ subsetCluster.getNodeIds()
+ ") are not a subset of superset cluster node ids ("
+ supersetCluster.getNodeIds() + ") ]");
}
for(int nodeId: subsetCluster.getNodeIds()) {
Node supersetNode = supersetCluster.getNodeById(nodeId);
Node subsetNode = subsetCluster.getNodeById(nodeId);
if(!supersetNode.getPartitionIds().equals(subsetNode.getPartitionIds())) {
throw new VoldemortRebalancingException("Partition IDs do not match between clusters for nodes with id "
+ nodeId
+ " : subset cluster has "
+ subsetNode.getPartitionIds()
+ " and superset cluster has "
+ supersetNode.getPartitionIds());
}
}
Set<Integer> nodeIds = supersetCluster.getNodeIds();
nodeIds.removeAll(subsetCluster.getNodeIds());
for(int nodeId: nodeIds) {
Node supersetNode = supersetCluster.getNodeById(nodeId);
if(!supersetNode.getPartitionIds().isEmpty()) {
throw new VoldemortRebalancingException("New node "
+ nodeId
+ " in superset cluster already has partitions: "
+ supersetNode.getPartitionIds());
}
}
} | java |
public static void validateClusterZonesSame(final Cluster lhs, final Cluster rhs) {
Set<Zone> lhsSet = new HashSet<Zone>(lhs.getZones());
Set<Zone> rhsSet = new HashSet<Zone>(rhs.getZones());
if(!lhsSet.equals(rhsSet))
throw new VoldemortException("Zones are not the same [ lhs cluster zones ("
+ lhs.getZones() + ") not equal to rhs cluster zones ("
+ rhs.getZones() + ") ]");
} | java |
public static void validateClusterNodeCounts(final Cluster lhs, final Cluster rhs) {
if(!lhs.getNodeIds().equals(rhs.getNodeIds())) {
throw new VoldemortException("Node ids are not the same [ lhs cluster node ids ("
+ lhs.getNodeIds()
+ ") not equal to rhs cluster node ids ("
+ rhs.getNodeIds() + ") ]");
}
} | java |
public static Cluster vacateZone(Cluster currentCluster, int dropZoneId) {
Cluster returnCluster = Cluster.cloneCluster(currentCluster);
// Go over each node in the zone being dropped
for(Integer nodeId: currentCluster.getNodeIdsInZone(dropZoneId)) {
// For each node grab all the partitions it hosts
for(Integer partitionId: currentCluster.getNodeById(nodeId).getPartitionIds()) {
// Now for each partition find a new home..which would be a node
// in one of the existing zones
int finalZoneId = -1;
int finalNodeId = -1;
int adjacentPartitionId = partitionId;
do {
adjacentPartitionId = (adjacentPartitionId + 1)
% currentCluster.getNumberOfPartitions();
finalNodeId = currentCluster.getNodeForPartitionId(adjacentPartitionId).getId();
finalZoneId = currentCluster.getZoneForPartitionId(adjacentPartitionId).getId();
if(adjacentPartitionId == partitionId) {
logger.error("PartitionId " + partitionId + "stays unchanged \n");
} else {
logger.info("PartitionId " + partitionId
+ " goes together with partition " + adjacentPartitionId
+ " on node " + finalNodeId + " in zone " + finalZoneId);
returnCluster = UpdateClusterUtils.createUpdatedCluster(returnCluster,
finalNodeId,
Lists.newArrayList(partitionId));
}
} while(finalZoneId == dropZoneId);
}
}
return returnCluster;
} | java |
public static Cluster dropZone(Cluster intermediateCluster, int dropZoneId) {
// Filter out nodes that don't belong to the zone being dropped
Set<Node> survivingNodes = new HashSet<Node>();
for(int nodeId: intermediateCluster.getNodeIds()) {
if(intermediateCluster.getNodeById(nodeId).getZoneId() != dropZoneId) {
survivingNodes.add(intermediateCluster.getNodeById(nodeId));
}
}
// Filter out dropZoneId from all zones
Set<Zone> zones = new HashSet<Zone>();
for(int zoneId: intermediateCluster.getZoneIds()) {
if(zoneId == dropZoneId) {
continue;
}
List<Integer> proximityList = intermediateCluster.getZoneById(zoneId)
.getProximityList();
proximityList.remove(new Integer(dropZoneId));
zones.add(new Zone(zoneId, proximityList));
}
return new Cluster(intermediateCluster.getName(),
Utils.asSortedList(survivingNodes),
Utils.asSortedList(zones));
} | java |
public static List<Integer> getStolenPrimaryPartitions(final Cluster currentCluster,
final Cluster finalCluster,
final int stealNodeId) {
List<Integer> finalList = new ArrayList<Integer>(finalCluster.getNodeById(stealNodeId)
.getPartitionIds());
List<Integer> currentList = new ArrayList<Integer>();
if(currentCluster.hasNodeWithId(stealNodeId)) {
currentList = currentCluster.getNodeById(stealNodeId).getPartitionIds();
} else {
if(logger.isDebugEnabled()) {
logger.debug("Current cluster does not contain stealer node (cluster : [[["
+ currentCluster + "]]], node id " + stealNodeId + ")");
}
}
finalList.removeAll(currentList);
return finalList;
} | java |
public static List<StoreDefinition> validateRebalanceStore(List<StoreDefinition> storeDefList) {
List<StoreDefinition> returnList = new ArrayList<StoreDefinition>(storeDefList.size());
for(StoreDefinition def: storeDefList) {
if(!def.isView() && !canRebalanceList.contains(def.getType())) {
throw new VoldemortException("Rebalance does not support rebalancing of stores of type "
+ def.getType() + " - " + def.getName());
} else if(!def.isView()) {
returnList.add(def);
} else {
logger.debug("Ignoring view " + def.getName() + " for rebalancing");
}
}
return returnList;
} | java |
public static void dumpClusters(Cluster currentCluster,
Cluster finalCluster,
String outputDirName,
String filePrefix) {
dumpClusterToFile(outputDirName, filePrefix + currentClusterFileName, currentCluster);
dumpClusterToFile(outputDirName, filePrefix + finalClusterFileName, finalCluster);
} | java |
public static void dumpClusters(Cluster currentCluster,
Cluster finalCluster,
String outputDirName) {
dumpClusters(currentCluster, finalCluster, outputDirName, "");
} | java |
public static void dumpClusterToFile(String outputDirName, String fileName, Cluster cluster) {
if(outputDirName != null) {
File outputDir = new File(outputDirName);
if(!outputDir.exists()) {
Utils.mkdirs(outputDir);
}
try {
FileUtils.writeStringToFile(new File(outputDirName, fileName),
new ClusterMapper().writeCluster(cluster));
} catch(IOException e) {
logger.error("IOException during dumpClusterToFile: " + e);
}
}
} | java |
public static void dumpStoreDefsToFile(String outputDirName,
String fileName,
List<StoreDefinition> storeDefs) {
if(outputDirName != null) {
File outputDir = new File(outputDirName);
if(!outputDir.exists()) {
Utils.mkdirs(outputDir);
}
try {
FileUtils.writeStringToFile(new File(outputDirName, fileName),
new StoreDefinitionsMapper().writeStoreList(storeDefs));
} catch(IOException e) {
logger.error("IOException during dumpStoreDefsToFile: " + e);
}
}
} | java |
public static void dumpAnalysisToFile(String outputDirName,
String baseFileName,
PartitionBalance partitionBalance) {
if(outputDirName != null) {
File outputDir = new File(outputDirName);
if(!outputDir.exists()) {
Utils.mkdirs(outputDir);
}
try {
FileUtils.writeStringToFile(new File(outputDirName, baseFileName + ".analysis"),
partitionBalance.toString());
} catch(IOException e) {
logger.error("IOException during dumpAnalysisToFile: " + e);
}
}
} | java |
public static void dumpPlanToFile(String outputDirName, RebalancePlan plan) {
if(outputDirName != null) {
File outputDir = new File(outputDirName);
if(!outputDir.exists()) {
Utils.mkdirs(outputDir);
}
try {
FileUtils.writeStringToFile(new File(outputDirName, "plan.out"), plan.toString());
} catch(IOException e) {
logger.error("IOException during dumpPlanToFile: " + e);
}
}
} | java |
public static List<RebalanceTaskInfo> filterTaskPlanWithStores(List<RebalanceTaskInfo> existingPlanList,
List<StoreDefinition> storeDefs) {
List<RebalanceTaskInfo> plans = Lists.newArrayList();
List<String> storeNames = StoreDefinitionUtils.getStoreNames(storeDefs);
for(RebalanceTaskInfo existingPlan: existingPlanList) {
RebalanceTaskInfo info = RebalanceTaskInfo.create(existingPlan.toJsonString());
// Filter the plans only for stores given
HashMap<String, List<Integer>> storeToPartitions = info.getStoreToPartitionIds();
HashMap<String, List<Integer>> newStoreToPartitions = Maps.newHashMap();
for(String storeName: storeNames) {
if(storeToPartitions.containsKey(storeName))
newStoreToPartitions.put(storeName, storeToPartitions.get(storeName));
}
info.setStoreToPartitionList(newStoreToPartitions);
plans.add(info);
}
return plans;
} | java |
public static void executorShutDown(ExecutorService executorService, long timeOutSec) {
try {
executorService.shutdown();
executorService.awaitTermination(timeOutSec, TimeUnit.SECONDS);
} catch(Exception e) {
logger.warn("Error while stoping executor service.", e);
}
} | java |
public List<T> resolveConflicts(List<T> values) {
if(values.size() > 1)
return values;
else
return Collections.singletonList(values.get(0));
} | java |
public static <K, V, T> List<Versioned<V>> get(Store<K, V, T> storageEngine, K key, T transform) {
Map<K, List<Versioned<V>>> result = storageEngine.getAll(Collections.singleton(key),
Collections.singletonMap(key,
transform));
if(result.size() > 0)
return result.get(key);
else
return Collections.emptyList();
} | java |
public static <K, V, T> Map<K, List<Versioned<V>>> getAll(Store<K, V, T> storageEngine,
Iterable<K> keys,
Map<K, T> transforms) {
Map<K, List<Versioned<V>>> result = newEmptyHashMap(keys);
for(K key: keys) {
List<Versioned<V>> value = storageEngine.get(key,
transforms != null ? transforms.get(key)
: null);
if(!value.isEmpty())
result.put(key, value);
}
return result;
} | java |
public static <K, V> HashMap<K, V> newEmptyHashMap(Iterable<?> iterable) {
if(iterable instanceof Collection<?>)
return Maps.newHashMapWithExpectedSize(((Collection<?>) iterable).size());
return Maps.newHashMap();
} | java |
public static void assertValidMetadata(ByteArray key,
RoutingStrategy routingStrategy,
Node currentNode) {
List<Node> nodes = routingStrategy.routeRequest(key.get());
for(Node node: nodes) {
if(node.getId() == currentNode.getId()) {
return;
}
}
throw new InvalidMetadataException("Client accessing key belonging to partitions "
+ routingStrategy.getPartitionList(key.get())
+ " not present at " + currentNode);
} | java |
public static void assertValidNode(MetadataStore metadataStore, Integer nodeId) {
if(!metadataStore.getCluster().hasNodeWithId(nodeId)) {
throw new InvalidMetadataException("NodeId " + nodeId + " is not or no longer in this cluster");
}
} | java |
@SuppressWarnings("unchecked")
public static <T> Serializer<T> unsafeGetSerializer(SerializerFactory serializerFactory,
SerializerDefinition serializerDefinition) {
return (Serializer<T>) serializerFactory.getSerializer(serializerDefinition);
} | java |
public static StoreDefinition getStoreDef(List<StoreDefinition> list, String name) {
for(StoreDefinition def: list)
if(def.getName().equals(name))
return def;
return null;
} | java |
public static List<String> getStoreNames(List<StoreDefinition> list, boolean ignoreViews) {
List<String> storeNameSet = new ArrayList<String>();
for(StoreDefinition def: list)
if(!def.isView() || !ignoreViews)
storeNameSet.add(def.getName());
return storeNameSet;
} | java |
private void plan() {
// Mapping of stealer node to list of primary partitions being moved
final TreeMultimap<Integer, Integer> stealerToStolenPrimaryPartitions = TreeMultimap.create();
// Output initial and final cluster
if(outputDir != null)
RebalanceUtils.dumpClusters(currentCluster, finalCluster, outputDir);
// Determine which partitions must be stolen
for(Node stealerNode: finalCluster.getNodes()) {
List<Integer> stolenPrimaryPartitions = RebalanceUtils.getStolenPrimaryPartitions(currentCluster,
finalCluster,
stealerNode.getId());
if(stolenPrimaryPartitions.size() > 0) {
numPrimaryPartitionMoves += stolenPrimaryPartitions.size();
stealerToStolenPrimaryPartitions.putAll(stealerNode.getId(),
stolenPrimaryPartitions);
}
}
// Determine plan batch-by-batch
int batches = 0;
Cluster batchCurrentCluster = Cluster.cloneCluster(currentCluster);
List<StoreDefinition> batchCurrentStoreDefs = this.currentStoreDefs;
List<StoreDefinition> batchFinalStoreDefs = this.finalStoreDefs;
Cluster batchFinalCluster = RebalanceUtils.getInterimCluster(this.currentCluster,
this.finalCluster);
while(!stealerToStolenPrimaryPartitions.isEmpty()) {
int partitions = 0;
List<Entry<Integer, Integer>> partitionsMoved = Lists.newArrayList();
for(Entry<Integer, Integer> stealerToPartition: stealerToStolenPrimaryPartitions.entries()) {
partitionsMoved.add(stealerToPartition);
batchFinalCluster = UpdateClusterUtils.createUpdatedCluster(batchFinalCluster,
stealerToPartition.getKey(),
Lists.newArrayList(stealerToPartition.getValue()));
partitions++;
if(partitions == batchSize)
break;
}
// Remove the partitions moved
for(Iterator<Entry<Integer, Integer>> partitionMoved = partitionsMoved.iterator(); partitionMoved.hasNext();) {
Entry<Integer, Integer> entry = partitionMoved.next();
stealerToStolenPrimaryPartitions.remove(entry.getKey(), entry.getValue());
}
if(outputDir != null)
RebalanceUtils.dumpClusters(batchCurrentCluster,
batchFinalCluster,
outputDir,
"batch-" + Integer.toString(batches) + ".");
// Generate a plan to compute the tasks
final RebalanceBatchPlan RebalanceBatchPlan = new RebalanceBatchPlan(batchCurrentCluster,
batchCurrentStoreDefs,
batchFinalCluster,
batchFinalStoreDefs);
batchPlans.add(RebalanceBatchPlan);
numXZonePartitionStoreMoves += RebalanceBatchPlan.getCrossZonePartitionStoreMoves();
numPartitionStoreMoves += RebalanceBatchPlan.getPartitionStoreMoves();
nodeMoveMap.add(RebalanceBatchPlan.getNodeMoveMap());
zoneMoveMap.add(RebalanceBatchPlan.getZoneMoveMap());
batches++;
batchCurrentCluster = Cluster.cloneCluster(batchFinalCluster);
// batchCurrentStoreDefs can only be different from
// batchFinalStoreDefs for the initial batch.
batchCurrentStoreDefs = batchFinalStoreDefs;
}
logger.info(this);
} | java |
private String storageOverhead(Map<Integer, Integer> finalNodeToOverhead) {
double maxOverhead = Double.MIN_VALUE;
PartitionBalance pb = new PartitionBalance(currentCluster, currentStoreDefs);
StringBuilder sb = new StringBuilder();
sb.append("Per-node store-overhead:").append(Utils.NEWLINE);
DecimalFormat doubleDf = new DecimalFormat("####.##");
for(int nodeId: finalCluster.getNodeIds()) {
Node node = finalCluster.getNodeById(nodeId);
String nodeTag = "Node " + String.format("%4d", nodeId) + " (" + node.getHost() + ")";
int initialLoad = 0;
if(currentCluster.getNodeIds().contains(nodeId)) {
initialLoad = pb.getNaryPartitionCount(nodeId);
}
int toLoad = 0;
if(finalNodeToOverhead.containsKey(nodeId)) {
toLoad = finalNodeToOverhead.get(nodeId);
}
double overhead = (initialLoad + toLoad) / (double) initialLoad;
if(initialLoad > 0 && maxOverhead < overhead) {
maxOverhead = overhead;
}
String loadTag = String.format("%6d", initialLoad) + " + "
+ String.format("%6d", toLoad) + " -> "
+ String.format("%6d", initialLoad + toLoad) + " ("
+ doubleDf.format(overhead) + " X)";
sb.append(nodeTag + " : " + loadTag).append(Utils.NEWLINE);
}
sb.append(Utils.NEWLINE)
.append("**** Max per-node storage overhead: " + doubleDf.format(maxOverhead) + " X.")
.append(Utils.NEWLINE);
return (sb.toString());
} | java |
public static Boolean askConfirm(Boolean confirm, String opDesc) throws IOException {
if(confirm) {
System.out.println("Confirmed " + opDesc + " in command-line.");
return true;
} else {
System.out.println("Are you sure you want to " + opDesc + "? (yes/no)");
BufferedReader buffer = new BufferedReader(new InputStreamReader(System.in));
String text = buffer.readLine().toLowerCase(Locale.ENGLISH);
boolean go = text.equals("yes") || text.equals("y");
if (!go) {
System.out.println("Did not confirm; " + opDesc + " aborted.");
}
return go;
}
} | java |
public static List<String> getValueList(List<String> valuePairs, String delim) {
List<String> valueList = Lists.newArrayList();
for(String valuePair: valuePairs) {
String[] value = valuePair.split(delim, 2);
if(value.length != 2)
throw new VoldemortException("Invalid argument pair: " + valuePair);
valueList.add(value[0]);
valueList.add(value[1]);
}
return valueList;
} | java |
public static <V> Map<V, V> convertListToMap(List<V> list) {
Map<V, V> map = new HashMap<V, V>();
if(list.size() % 2 != 0)
throw new VoldemortException("Failed to convert list to map.");
for(int i = 0; i < list.size(); i += 2) {
map.put(list.get(i), list.get(i + 1));
}
return map;
} | java |
public static AdminClient getAdminClient(String url) {
ClientConfig config = new ClientConfig().setBootstrapUrls(url)
.setConnectionTimeout(5, TimeUnit.SECONDS);
AdminClientConfig adminConfig = new AdminClientConfig().setAdminSocketTimeoutSec(5);
return new AdminClient(adminConfig, config);
} | java |
public static List<Integer> getAllNodeIds(AdminClient adminClient) {
List<Integer> nodeIds = Lists.newArrayList();
for(Integer nodeId: adminClient.getAdminClientCluster().getNodeIds()) {
nodeIds.add(nodeId);
}
return nodeIds;
} | java |
public static List<String> getAllUserStoreNamesOnNode(AdminClient adminClient, Integer nodeId) {
List<String> storeNames = Lists.newArrayList();
List<StoreDefinition> storeDefinitionList = adminClient.metadataMgmtOps.getRemoteStoreDefList(nodeId)
.getValue();
for(StoreDefinition storeDefinition: storeDefinitionList) {
storeNames.add(storeDefinition.getName());
}
return storeNames;
} | java |
public static void validateUserStoreNamesOnNode(AdminClient adminClient,
Integer nodeId,
List<String> storeNames) {
List<StoreDefinition> storeDefList = adminClient.metadataMgmtOps.getRemoteStoreDefList(nodeId)
.getValue();
Map<String, Boolean> existingStoreNames = new HashMap<String, Boolean>();
for(StoreDefinition storeDef: storeDefList) {
existingStoreNames.put(storeDef.getName(), true);
}
for(String storeName: storeNames) {
if(!Boolean.TRUE.equals(existingStoreNames.get(storeName))) {
Utils.croak("Store " + storeName + " does not exist!");
}
}
} | java |
public static List<Integer> getAllPartitions(AdminClient adminClient) {
List<Integer> partIds = Lists.newArrayList();
partIds = Lists.newArrayList();
for(Node node: adminClient.getAdminClientCluster().getNodes()) {
partIds.addAll(node.getPartitionIds());
}
return partIds;
} | java |
public static List<QuotaType> getQuotaTypes(List<String> strQuotaTypes) {
if(strQuotaTypes.size() < 1) {
throw new VoldemortException("Quota type not specified.");
}
List<QuotaType> quotaTypes;
if(strQuotaTypes.size() == 1 && strQuotaTypes.get(0).equals(AdminToolUtils.QUOTATYPE_ALL)) {
quotaTypes = Arrays.asList(QuotaType.values());
} else {
quotaTypes = new ArrayList<QuotaType>();
for(String strQuotaType: strQuotaTypes) {
QuotaType type = QuotaType.valueOf(strQuotaType);
quotaTypes.add(type);
}
}
return quotaTypes;
} | java |
public static File createDir(String dir) {
// create outdir
File directory = null;
if(dir != null) {
directory = new File(dir);
if(!(directory.exists() || directory.mkdir())) {
Utils.croak("Can't find or create directory " + dir);
}
}
return directory;
} | java |
public static Map<String, StoreDefinition> getSystemStoreDefMap() {
Map<String, StoreDefinition> sysStoreDefMap = Maps.newHashMap();
List<StoreDefinition> storesDefs = SystemStoreConstants.getAllSystemStoreDefs();
for(StoreDefinition def: storesDefs) {
sysStoreDefMap.put(def.getName(), def);
}
return sysStoreDefMap;
} | java |
public static Map<String, StoreDefinition> getUserStoreDefMapOnNode(AdminClient adminClient,
Integer nodeId) {
List<StoreDefinition> storeDefinitionList = adminClient.metadataMgmtOps.getRemoteStoreDefList(nodeId)
.getValue();
Map<String, StoreDefinition> storeDefinitionMap = Maps.newHashMap();
for(StoreDefinition storeDefinition: storeDefinitionList) {
storeDefinitionMap.put(storeDefinition.getName(), storeDefinition);
}
return storeDefinitionMap;
} | java |
public static RebalanceTaskInfo decodeRebalanceTaskInfoMap(VAdminProto.RebalanceTaskInfoMap rebalanceTaskInfoMap) {
RebalanceTaskInfo rebalanceTaskInfo = new RebalanceTaskInfo(
rebalanceTaskInfoMap.getStealerId(),
rebalanceTaskInfoMap.getDonorId(),
decodeStoreToPartitionIds(rebalanceTaskInfoMap.getPerStorePartitionIdsList()),
new ClusterMapper().readCluster(new StringReader(rebalanceTaskInfoMap.getInitialCluster())));
return rebalanceTaskInfo;
} | java |
public static RebalanceTaskInfoMap encodeRebalanceTaskInfoMap(RebalanceTaskInfo stealInfo) {
return RebalanceTaskInfoMap.newBuilder()
.setStealerId(stealInfo.getStealerId())
.setDonorId(stealInfo.getDonorId())
.addAllPerStorePartitionIds(ProtoUtils.encodeStoreToPartitionsTuple(stealInfo.getStoreToPartitionIds()))
.setInitialCluster(new ClusterMapper().writeCluster(stealInfo.getInitialCluster()))
.build();
} | java |
public Versioned<E> getVersionedById(int id) {
Versioned<VListNode<E>> listNode = getListNode(id);
if(listNode == null)
throw new IndexOutOfBoundsException();
return new Versioned<E>(listNode.getValue().getValue(), listNode.getVersion());
} | java |
public E setById(final int id, final E element) {
VListKey<K> key = new VListKey<K>(_key, id);
UpdateElementById<K, E> updateElementAction = new UpdateElementById<K, E>(key, element);
if(!_storeClient.applyUpdate(updateElementAction))
throw new ObsoleteVersionException("update failed");
return updateElementAction.getResult();
} | java |
private void allClustersEqual(final List<String> clusterUrls) {
Validate.notEmpty(clusterUrls, "clusterUrls cannot be null");
// If only one clusterUrl return immediately
if (clusterUrls.size() == 1)
return;
AdminClient adminClientLhs = adminClientPerCluster.get(clusterUrls.get(0));
Cluster clusterLhs = adminClientLhs.getAdminClientCluster();
for (int index = 1; index < clusterUrls.size(); index++) {
AdminClient adminClientRhs = adminClientPerCluster.get(clusterUrls.get(index));
Cluster clusterRhs = adminClientRhs.getAdminClientCluster();
if (!areTwoClustersEqual(clusterLhs, clusterRhs))
throw new VoldemortException("Cluster " + clusterLhs.getName()
+ " is not the same as " + clusterRhs.getName());
}
} | java |
private synchronized JsonSchema getInputPathJsonSchema() throws IOException {
if (inputPathJsonSchema == null) {
// No need to query Hadoop more than once as this shouldn't change mid-run,
// thus, we can lazily initialize and cache the result.
inputPathJsonSchema = HadoopUtils.getSchemaFromPath(getInputPath());
}
return inputPathJsonSchema;
} | java |
private synchronized Schema getInputPathAvroSchema() throws IOException {
if (inputPathAvroSchema == null) {
// No need to query Hadoop more than once as this shouldn't change mid-run,
// thus, we can lazily initialize and cache the result.
inputPathAvroSchema = AvroUtils.getAvroSchemaFromPath(getInputPath());
}
return inputPathAvroSchema;
} | java |
public String getRecordSchema() throws IOException {
Schema schema = getInputPathAvroSchema();
String recSchema = schema.toString();
return recSchema;
} | java |
public String getKeySchema() throws IOException {
Schema schema = getInputPathAvroSchema();
String keySchema = schema.getField(keyFieldName).schema().toString();
return keySchema;
} | java |
public String getValueSchema() throws IOException {
Schema schema = getInputPathAvroSchema();
String valueSchema = schema.getField(valueFieldName).schema().toString();
return valueSchema;
} | java |
private void verifyOrAddStore(String clusterURL,
String keySchema,
String valueSchema) {
String newStoreDefXml = VoldemortUtils.getStoreDefXml(
storeName,
props.getInt(BUILD_REPLICATION_FACTOR, 2),
props.getInt(BUILD_REQUIRED_READS, 1),
props.getInt(BUILD_REQUIRED_WRITES, 1),
props.getNullableInt(BUILD_PREFERRED_READS),
props.getNullableInt(BUILD_PREFERRED_WRITES),
props.getString(PUSH_FORCE_SCHEMA_KEY, keySchema),
props.getString(PUSH_FORCE_SCHEMA_VALUE, valueSchema),
description,
owners);
log.info("Verifying store against cluster URL: " + clusterURL + "\n" + newStoreDefXml.toString());
StoreDefinition newStoreDef = VoldemortUtils.getStoreDef(newStoreDefXml);
try {
adminClientPerCluster.get(clusterURL).storeMgmtOps.verifyOrAddStore(newStoreDef, "BnP config/data",
enableStoreCreation, this.storeVerificationExecutorService);
} catch (UnreachableStoreException e) {
log.info("verifyOrAddStore() failed on some nodes for clusterURL: " + clusterURL + " (this is harmless).", e);
// When we can't reach some node, we just skip it and won't create the store on it.
// Next time BnP is run while the node is up, it will get the store created.
} // Other exceptions need to bubble up!
storeDef = newStoreDef;
} | java |
public void syncInternalStateFromFileSystem(boolean alsoSyncRemoteState) {
// Make sure versions missing from the file-system are cleaned up from the internal state
for (Long version: versionToEnabledMap.keySet()) {
File[] existingVersionDirs = ReadOnlyUtils.getVersionDirs(rootDir, version, version);
if (existingVersionDirs.length == 0) {
removeVersion(version, alsoSyncRemoteState);
}
}
// Make sure we have all versions on the file-system in the internal state
File[] versionDirs = ReadOnlyUtils.getVersionDirs(rootDir);
if (versionDirs != null) {
for (File versionDir: versionDirs) {
long versionNumber = ReadOnlyUtils.getVersionId(versionDir);
boolean versionEnabled = isVersionEnabled(versionDir);
versionToEnabledMap.put(versionNumber, versionEnabled);
}
}
// Identify the current version (based on a symlink in the file-system)
File currentVersionDir = ReadOnlyUtils.getCurrentVersion(rootDir);
if (currentVersionDir != null) {
currentVersion = ReadOnlyUtils.getVersionId(currentVersionDir);
} else {
currentVersion = -1; // Should we throw instead?
}
logger.info("Successfully synced internal state from local file-system: " + this.toString());
} | java |
private void persistDisabledVersion(long version) throws PersistenceFailureException {
File disabledMarker = getDisabledMarkerFile(version);
try {
disabledMarker.createNewFile();
} catch (IOException e) {
throw new PersistenceFailureException("Failed to create the disabled marker at path: " +
disabledMarker.getAbsolutePath() + "\nThe store/version " +
"will remain disabled only until the next restart.", e);
}
} | java |
private void persistEnabledVersion(long version) throws PersistenceFailureException {
File disabledMarker = getDisabledMarkerFile(version);
if (disabledMarker.exists()) {
if (!disabledMarker.delete()) {
throw new PersistenceFailureException("Failed to create the disabled marker at path: " +
disabledMarker.getAbsolutePath() + "\nThe store/version " +
"will remain enabled only until the next restart.");
}
}
} | java |
private File getDisabledMarkerFile(long version) throws PersistenceFailureException {
File[] versionDirArray = ReadOnlyUtils.getVersionDirs(rootDir, version, version);
if (versionDirArray.length == 0) {
throw new PersistenceFailureException("getDisabledMarkerFile did not find the requested version directory" +
" on disk. Version: " + version + ", rootDir: " + rootDir);
}
File disabledMarkerFile = new File(versionDirArray[0], DISABLED_MARKER_NAME);
return disabledMarkerFile;
} | java |
public Double getAvgEventValue() {
resetIfNeeded();
synchronized(this) {
long eventsLastInterval = numEventsLastInterval - numEventsLastLastInterval;
if(eventsLastInterval > 0)
return ((totalEventValueLastInterval - totalEventValueLastLastInterval) * 1.0)
/ eventsLastInterval;
else
return 0.0;
}
} | java |
@SuppressWarnings("unchecked")
public static void executeCommand(String[] args) throws IOException {
OptionParser parser = getParser();
// declare parameters
List<String> metaKeys = null;
String url = null;
// parse command-line input
args = AdminToolUtils.copyArrayAddFirst(args, "--" + OPT_HEAD_META_CHECK);
OptionSet options = parser.parse(args);
if(options.has(AdminParserUtils.OPT_HELP)) {
printHelp(System.out);
return;
}
// check required options and/or conflicting options
AdminParserUtils.checkRequired(options, OPT_HEAD_META_CHECK);
AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_URL);
// load parameters
metaKeys = (List<String>) options.valuesOf(OPT_HEAD_META_CHECK);
url = (String) options.valueOf(AdminParserUtils.OPT_URL);
// execute command
if(metaKeys.size() == 0
|| (metaKeys.size() == 1 && metaKeys.get(0).equals(METAKEY_ALL))) {
metaKeys = Lists.newArrayList();
metaKeys.add(MetadataStore.CLUSTER_KEY);
metaKeys.add(MetadataStore.STORES_KEY);
metaKeys.add(MetadataStore.SERVER_STATE_KEY);
}
AdminClient adminClient = AdminToolUtils.getAdminClient(url);
doMetaCheck(adminClient, metaKeys);
} | java |
@SuppressWarnings("unchecked")
public static void executeCommand(String[] args) throws IOException {
OptionParser parser = getParser();
// declare parameters
String url = null;
List<Integer> nodeIds = null;
Boolean allNodes = true;
Boolean confirm = false;
// parse command-line input
OptionSet options = parser.parse(args);
if(options.has(AdminParserUtils.OPT_HELP)) {
printHelp(System.out);
return;
}
// check required options and/or conflicting options
AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_URL);
AdminParserUtils.checkOptional(options,
AdminParserUtils.OPT_NODE,
AdminParserUtils.OPT_ALL_NODES);
// load parameters
url = (String) options.valueOf(AdminParserUtils.OPT_URL);
if(options.has(AdminParserUtils.OPT_NODE)) {
nodeIds = (List<Integer>) options.valuesOf(AdminParserUtils.OPT_NODE);
allNodes = false;
}
if(options.has(AdminParserUtils.OPT_CONFIRM)) {
confirm = true;
}
// print summary
System.out.println("Remove metadata related to rebalancing");
System.out.println("Location:");
System.out.println(" bootstrap url = " + url);
if(allNodes) {
System.out.println(" node = all nodes");
} else {
System.out.println(" node = " + Joiner.on(", ").join(nodeIds));
}
// execute command
if(!AdminToolUtils.askConfirm(confirm, "remove metadata related to rebalancing")) {
return;
}
AdminClient adminClient = AdminToolUtils.getAdminClient(url);
if(allNodes) {
nodeIds = AdminToolUtils.getAllNodeIds(adminClient);
}
AdminToolUtils.assertServerNotInRebalancingState(adminClient, nodeIds);
doMetaClearRebalance(adminClient, nodeIds);
} | java |
public static void doMetaClearRebalance(AdminClient adminClient, List<Integer> nodeIds) {
AdminToolUtils.assertServerNotInOfflineState(adminClient, nodeIds);
System.out.println("Setting " + MetadataStore.SERVER_STATE_KEY + " to "
+ MetadataStore.VoldemortState.NORMAL_SERVER);
doMetaSet(adminClient,
nodeIds,
MetadataStore.SERVER_STATE_KEY,
MetadataStore.VoldemortState.NORMAL_SERVER.toString());
RebalancerState state = RebalancerState.create("[]");
System.out.println("Cleaning up " + MetadataStore.REBALANCING_STEAL_INFO + " to "
+ state.toJsonString());
doMetaSet(adminClient,
nodeIds,
MetadataStore.REBALANCING_STEAL_INFO,
state.toJsonString());
System.out.println("Cleaning up " + MetadataStore.REBALANCING_SOURCE_CLUSTER_XML
+ " to empty string");
doMetaSet(adminClient, nodeIds, MetadataStore.REBALANCING_SOURCE_CLUSTER_XML, "");
} | java |
@SuppressWarnings("unchecked")
public static void executeCommand(String[] args) throws IOException {
OptionParser parser = getParser();
// declare parameters
List<String> metaKeys = null;
String url = null;
String dir = null;
List<Integer> nodeIds = null;
Boolean allNodes = true;
Boolean verbose = false;
// parse command-line input
args = AdminToolUtils.copyArrayAddFirst(args, "--" + OPT_HEAD_META_GET);
OptionSet options = parser.parse(args);
if(options.has(AdminParserUtils.OPT_HELP)) {
printHelp(System.out);
return;
}
// check required options and/or conflicting options
AdminParserUtils.checkRequired(options, OPT_HEAD_META_GET);
AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_URL);
AdminParserUtils.checkOptional(options,
AdminParserUtils.OPT_NODE,
AdminParserUtils.OPT_ALL_NODES);
// load parameters
metaKeys = (List<String>) options.valuesOf(OPT_HEAD_META_GET);
url = (String) options.valueOf(AdminParserUtils.OPT_URL);
if(options.has(AdminParserUtils.OPT_DIR)) {
dir = (String) options.valueOf(AdminParserUtils.OPT_DIR);
}
if(options.has(AdminParserUtils.OPT_NODE)) {
nodeIds = (List<Integer>) options.valuesOf(AdminParserUtils.OPT_NODE);
allNodes = false;
}
if(options.has(OPT_VERBOSE)) {
verbose = true;
}
// execute command
File directory = AdminToolUtils.createDir(dir);
AdminClient adminClient = AdminToolUtils.getAdminClient(url);
if(allNodes) {
nodeIds = AdminToolUtils.getAllNodeIds(adminClient);
}
if(metaKeys.size() == 1 && metaKeys.get(0).equals(METAKEY_ALL)) {
metaKeys = Lists.newArrayList();
for(Object key: MetadataStore.METADATA_KEYS) {
metaKeys.add((String) key);
}
}
doMetaGet(adminClient, nodeIds, metaKeys, directory, verbose);
} | java |
@SuppressWarnings("unchecked")
public static void executeCommand(String[] args) throws IOException {
OptionParser parser = getParser();
// declare parameters
List<String> metaKeys = null;
String url = null;
List<Integer> nodeIds = null;
Boolean allNodes = true;
List<String> storeNames = null;
// parse command-line input
args = AdminToolUtils.copyArrayAddFirst(args, "--" + OPT_HEAD_META_GET_RO);
OptionSet options = parser.parse(args);
if(options.has(AdminParserUtils.OPT_HELP)) {
printHelp(System.out);
return;
}
// check required options and/or conflicting options
AdminParserUtils.checkRequired(options, OPT_HEAD_META_GET_RO);
AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_URL);
AdminParserUtils.checkOptional(options,
AdminParserUtils.OPT_NODE,
AdminParserUtils.OPT_ALL_NODES);
AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_STORE);
// load parameters
metaKeys = (List<String>) options.valuesOf(OPT_HEAD_META_GET_RO);
url = (String) options.valueOf(AdminParserUtils.OPT_URL);
if(options.has(AdminParserUtils.OPT_NODE)) {
nodeIds = (List<Integer>) options.valuesOf(AdminParserUtils.OPT_NODE);
allNodes = false;
}
storeNames = (List<String>) options.valuesOf(AdminParserUtils.OPT_STORE);
// execute command
AdminClient adminClient = AdminToolUtils.getAdminClient(url);
if(allNodes) {
nodeIds = AdminToolUtils.getAllNodeIds(adminClient);
}
if(metaKeys.size() == 1 && metaKeys.get(0).equals(METAKEY_ALL)) {
metaKeys = Lists.newArrayList();
metaKeys.add(KEY_MAX_VERSION);
metaKeys.add(KEY_CURRENT_VERSION);
metaKeys.add(KEY_STORAGE_FORMAT);
}
doMetaGetRO(adminClient, nodeIds, storeNames, metaKeys);
} | java |
public static void doMetaGetRO(AdminClient adminClient,
Collection<Integer> nodeIds,
List<String> storeNames,
List<String> metaKeys) throws IOException {
for(String key: metaKeys) {
System.out.println("Metadata: " + key);
if(!key.equals(KEY_MAX_VERSION) && !key.equals(KEY_CURRENT_VERSION)
&& !key.equals(KEY_STORAGE_FORMAT)) {
System.out.println(" Invalid read-only metadata key: " + key);
} else {
for(Integer nodeId: nodeIds) {
String hostName = adminClient.getAdminClientCluster()
.getNodeById(nodeId)
.getHost();
System.out.println(" Node: " + hostName + ":" + nodeId);
if(key.equals(KEY_MAX_VERSION)) {
Map<String, Long> mapStoreToROVersion = adminClient.readonlyOps.getROMaxVersion(nodeId,
storeNames);
for(String storeName: mapStoreToROVersion.keySet()) {
System.out.println(" " + storeName + ":"
+ mapStoreToROVersion.get(storeName));
}
} else if(key.equals(KEY_CURRENT_VERSION)) {
Map<String, Long> mapStoreToROVersion = adminClient.readonlyOps.getROCurrentVersion(nodeId,
storeNames);
for(String storeName: mapStoreToROVersion.keySet()) {
System.out.println(" " + storeName + ":"
+ mapStoreToROVersion.get(storeName));
}
} else if(key.equals(KEY_STORAGE_FORMAT)) {
Map<String, String> mapStoreToROFormat = adminClient.readonlyOps.getROStorageFormat(nodeId,
storeNames);
for(String storeName: mapStoreToROFormat.keySet()) {
System.out.println(" " + storeName + ":"
+ mapStoreToROFormat.get(storeName));
}
}
}
}
System.out.println();
}
} | java |
public static void doMetaUpdateVersionsOnStores(AdminClient adminClient,
List<StoreDefinition> oldStoreDefs,
List<StoreDefinition> newStoreDefs) {
Set<String> storeNamesUnion = new HashSet<String>();
Map<String, StoreDefinition> oldStoreDefinitionMap = new HashMap<String, StoreDefinition>();
Map<String, StoreDefinition> newStoreDefinitionMap = new HashMap<String, StoreDefinition>();
List<String> storesChanged = new ArrayList<String>();
for(StoreDefinition storeDef: oldStoreDefs) {
String storeName = storeDef.getName();
storeNamesUnion.add(storeName);
oldStoreDefinitionMap.put(storeName, storeDef);
}
for(StoreDefinition storeDef: newStoreDefs) {
String storeName = storeDef.getName();
storeNamesUnion.add(storeName);
newStoreDefinitionMap.put(storeName, storeDef);
}
for(String storeName: storeNamesUnion) {
StoreDefinition oldStoreDef = oldStoreDefinitionMap.get(storeName);
StoreDefinition newStoreDef = newStoreDefinitionMap.get(storeName);
if(oldStoreDef == null && newStoreDef != null || oldStoreDef != null
&& newStoreDef == null || oldStoreDef != null && newStoreDef != null
&& !oldStoreDef.equals(newStoreDef)) {
storesChanged.add(storeName);
}
}
System.out.println("Updating metadata version for the following stores: "
+ storesChanged);
try {
adminClient.metadataMgmtOps.updateMetadataversion(adminClient.getAdminClientCluster()
.getNodeIds(),
storesChanged);
} catch(Exception e) {
System.err.println("Error while updating metadata version for the specified store.");
}
} | java |
public static void executeCommand(String[] args) throws IOException {
OptionParser parser = getParser();
String url = null;
Boolean confirm = false;
// parse command-line input
OptionSet options = parser.parse(args);
if(options.has(AdminParserUtils.OPT_HELP)) {
printHelp(System.out);
return;
}
AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_URL);
url = (String) options.valueOf(AdminParserUtils.OPT_URL);
if(options.has(AdminParserUtils.OPT_CONFIRM)) {
confirm = true;
}
// print summary
System.out.println("Synchronize metadata versions across all nodes");
System.out.println("Location:");
System.out.println(" bootstrap url = " + url);
System.out.println(" node = all nodes");
AdminClient adminClient = AdminToolUtils.getAdminClient(url);
AdminToolUtils.assertServerNotInRebalancingState(adminClient);
Versioned<Properties> versionedProps = mergeAllVersions(adminClient);
printVersions(versionedProps);
// execute command
if(!AdminToolUtils.askConfirm(confirm,
"do you want to synchronize metadata versions to all node"))
return;
adminClient.metadataMgmtOps.setMetadataVersion(versionedProps);
} | java |
public static void executeCommand(String[] args) throws IOException {
OptionParser parser = getParser();
// declare parameters
String url = null;
// parse command-line input
OptionSet options = parser.parse(args);
if(options.has(AdminParserUtils.OPT_HELP)) {
printHelp(System.out);
return;
}
// check required options and/or conflicting options
AdminParserUtils.checkRequired(options, AdminParserUtils.OPT_URL);
// load parameters
url = (String) options.valueOf(AdminParserUtils.OPT_URL);
// execute command
AdminClient adminClient = AdminToolUtils.getAdminClient(url);
doMetaCheckVersion(adminClient);
} | java |
private Integer getKeyPartitionId(byte[] key) {
Integer keyPartitionId = storeInstance.getNodesPartitionIdForKey(nodeId, key);
Utils.notNull(keyPartitionId);
return keyPartitionId;
} | java |
protected boolean isItemAccepted(byte[] key) {
boolean entryAccepted = false;
if (!fetchOrphaned) {
if (isKeyNeeded(key)) {
entryAccepted = true;
}
} else {
if (!StoreRoutingPlan.checkKeyBelongsToNode(key, nodeId, initialCluster, storeDef)) {
entryAccepted = true;
}
}
return entryAccepted;
} | java |
protected void accountForFetchedKey(byte[] key) {
fetched++;
if (streamStats != null) {
streamStats.reportStreamingFetch(operation);
}
if (recordsPerPartition <= 0) {
return;
}
Integer keyPartitionId = getKeyPartitionId(key);
Long partitionFetch = partitionFetches.get(keyPartitionId);
Utils.notNull(partitionFetch);
partitionFetch++;
partitionFetches.put(keyPartitionId, partitionFetch);
if (partitionFetch == recordsPerPartition) {
if (partitionsToFetch.contains(keyPartitionId)) {
partitionsToFetch.remove(keyPartitionId);
} else {
logger.warn("Partitions to fetch did not contain expected partition ID: "
+ keyPartitionId);
}
} else if (partitionFetch > recordsPerPartition) {
logger.warn("Partition fetch count larger than expected for partition ID "
+ keyPartitionId + " : " + partitionFetch);
}
} | java |
protected StreamRequestHandlerState determineRequestHandlerState(String itemTag) {
if (keyIterator.hasNext() && !fetchedEnoughForAllPartitions()) {
return StreamRequestHandlerState.WRITING;
} else {
logger.info("Finished fetch " + itemTag + " for store '" + storageEngine.getName()
+ "' with partitions " + partitionIds);
progressInfoMessage("Fetch " + itemTag + " (end of scan)");
return StreamRequestHandlerState.COMPLETE;
}
} | java |
protected List<Versioned<V>> resolveAndConstructVersionsToPersist(List<Versioned<V>> valuesInStorage,
List<Versioned<V>> multiPutValues) {
List<Versioned<V>> obsoleteVals = new ArrayList<Versioned<V>>(multiPutValues.size());
// Go over all the values and determine whether the version is
// acceptable
for(Versioned<V> value: multiPutValues) {
Iterator<Versioned<V>> iter = valuesInStorage.iterator();
boolean obsolete = false;
// Compare the current version with a set of accepted versions
while(iter.hasNext()) {
Versioned<V> curr = iter.next();
Occurred occurred = value.getVersion().compare(curr.getVersion());
if(occurred == Occurred.BEFORE) {
obsolete = true;
break;
} else if(occurred == Occurred.AFTER) {
iter.remove();
}
}
if(obsolete) {
// add to return value if obsolete
obsoleteVals.add(value);
} else {
// else update the set of accepted versions
valuesInStorage.add(value);
}
}
return obsoleteVals;
} | java |
private <T> InternalProviderImpl installInternalProvider(Class<T> clazz, String bindingName, InternalProviderImpl<? extends T> internalProvider,
boolean isBound, boolean isTestProvider) {
if (bindingName == null) {
if (isBound) {
return installUnNamedProvider(mapClassesToUnNamedBoundProviders, clazz, internalProvider, isTestProvider);
} else {
return installUnNamedProvider(mapClassesToUnNamedUnBoundProviders, clazz, internalProvider, isTestProvider);
}
} else {
return installNamedProvider(mapClassesToNamedBoundProviders, clazz, bindingName, internalProvider, isTestProvider);
}
} | java |
@Override
protected void reset() {
super.reset();
mapClassesToNamedBoundProviders.clear();
mapClassesToUnNamedBoundProviders.clear();
hasTestModules = false;
installBindingForScope();
} | java |
public synchronized T get(Scope scope) {
if (instance != null) {
return instance;
}
if (providerInstance != null) {
if (isProvidingSingletonInScope) {
instance = providerInstance.get();
//gc
providerInstance = null;
return instance;
}
return providerInstance.get();
}
if (factoryClass != null && factory == null) {
factory = FactoryLocator.getFactory(factoryClass);
//gc
factoryClass = null;
}
if (factory != null) {
if (!factory.hasScopeAnnotation() && !isCreatingSingletonInScope) {
return factory.createInstance(scope);
}
instance = factory.createInstance(scope);
//gc
factory = null;
return instance;
}
if (providerFactoryClass != null && providerFactory == null) {
providerFactory = FactoryLocator.getFactory(providerFactoryClass);
//gc
providerFactoryClass = null;
}
if (providerFactory != null) {
if (providerFactory.hasProvidesSingletonInScopeAnnotation() || isProvidingSingletonInScope) {
instance = providerFactory.createInstance(scope).get();
//gc
providerFactory = null;
return instance;
}
if (providerFactory.hasScopeAnnotation() || isCreatingSingletonInScope) {
providerInstance = providerFactory.createInstance(scope);
//gc
providerFactory = null;
return providerInstance.get();
}
return providerFactory.createInstance(scope).get();
}
throw new IllegalStateException("A provider can only be used with an instance, a provider, a factory or a provider factory. Should not happen.");
} | java |
public static void closeScope(Object name) {
//we remove the scope first, so that other threads don't see it, and see the next snapshot of the tree
ScopeNode scope = (ScopeNode) MAP_KEY_TO_SCOPE.remove(name);
if (scope != null) {
ScopeNode parentScope = scope.getParentScope();
if (parentScope != null) {
parentScope.removeChild(scope);
} else {
ConfigurationHolder.configuration.onScopeForestReset();
}
removeScopeAndChildrenFromMap(scope);
}
} | java |
public static void reset() {
for (Object name : Collections.list(MAP_KEY_TO_SCOPE.keys())) {
closeScope(name);
}
ConfigurationHolder.configuration.onScopeForestReset();
ScopeImpl.resetUnBoundProviders();
} | java |
void notifyMwRevisionProcessors(MwRevision mwRevision, boolean isCurrent) {
if (mwRevision == null || mwRevision.getPageId() <= 0) {
return;
}
for (MwRevisionProcessorBroker.RevisionSubscription rs : this.revisionSubscriptions) {
if (rs.onlyCurrentRevisions == isCurrent
&& (rs.model == null || mwRevision.getModel().equals(
rs.model))) {
rs.mwRevisionProcessor.processRevision(mwRevision);
}
}
} | java |
static ItemIdValueImpl fromIri(String iri) {
int separator = iri.lastIndexOf('/') + 1;
try {
return new ItemIdValueImpl(iri.substring(separator), iri.substring(0, separator));
} catch (IllegalArgumentException e) {
throw new IllegalArgumentException("Invalid Wikibase entity IRI: " + iri, e);
}
} | java |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.