repo stringclasses 11
values | path stringlengths 41 214 | func_name stringlengths 7 82 | original_string stringlengths 77 11.9k | language stringclasses 1
value | code stringlengths 77 11.9k | code_tokens listlengths 22 1.57k | docstring stringlengths 2 2.27k | docstring_tokens listlengths 1 352 | sha stringclasses 11
values | url stringlengths 129 319 | partition stringclasses 1
value | summary stringlengths 7 191 | input_ids listlengths 502 502 | token_type_ids listlengths 502 502 | attention_mask listlengths 502 502 | labels listlengths 502 502 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
apache/flink | flink-table/flink-table-runtime-blink/src/main/java/org/apache/flink/table/runtime/hashtable/BinaryHashBucketArea.java | BinaryHashBucketArea.startLookup | void startLookup(int hashCode) {
final int posHashCode = findBucket(hashCode);
// get the bucket for the given hash code
final int bucketArrayPos = posHashCode >> table.bucketsPerSegmentBits;
final int bucketInSegmentOffset = (posHashCode & table.bucketsPerSegmentMask) << BUCKET_SIZE_BITS;
final MemorySegmen... | java | void startLookup(int hashCode) {
final int posHashCode = findBucket(hashCode);
// get the bucket for the given hash code
final int bucketArrayPos = posHashCode >> table.bucketsPerSegmentBits;
final int bucketInSegmentOffset = (posHashCode & table.bucketsPerSegmentMask) << BUCKET_SIZE_BITS;
final MemorySegmen... | [
"void",
"startLookup",
"(",
"int",
"hashCode",
")",
"{",
"final",
"int",
"posHashCode",
"=",
"findBucket",
"(",
"hashCode",
")",
";",
"// get the bucket for the given hash code",
"final",
"int",
"bucketArrayPos",
"=",
"posHashCode",
">>",
"table",
".",
"bucketsPerSe... | Probe start lookup joined build rows. | [
"Probe",
"start",
"lookup",
"joined",
"build",
"rows",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-table/flink-table-runtime-blink/src/main/java/org/apache/flink/table/runtime/hashtable/BinaryHashBucketArea.java#L528-L536 | train | Start lookup of the given hash code. | [
30522,
11675,
2707,
4135,
21940,
2361,
1006,
20014,
23325,
16044,
1007,
1063,
2345,
20014,
13433,
7377,
4095,
16044,
1027,
2424,
24204,
3388,
1006,
23325,
16044,
1007,
1025,
1013,
1013,
2131,
1996,
13610,
2005,
1996,
2445,
23325,
3642,
2345,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/util/ArrayUtil.java | ArrayUtil.setOrAppend | public static <T> T[] setOrAppend(T[] buffer, int index, T value) {
if(index < buffer.length) {
Array.set(buffer, index, value);
return buffer;
}else {
return append(buffer, value);
}
} | java | public static <T> T[] setOrAppend(T[] buffer, int index, T value) {
if(index < buffer.length) {
Array.set(buffer, index, value);
return buffer;
}else {
return append(buffer, value);
}
} | [
"public",
"static",
"<",
"T",
">",
"T",
"[",
"]",
"setOrAppend",
"(",
"T",
"[",
"]",
"buffer",
",",
"int",
"index",
",",
"T",
"value",
")",
"{",
"if",
"(",
"index",
"<",
"buffer",
".",
"length",
")",
"{",
"Array",
".",
"set",
"(",
"buffer",
","... | 将元素值设置为数组的某个位置,当给定的index大于数组长度,则追加
@param <T> 数组元素类型
@param buffer 已有数组
@param index 位置,大于长度追加,否则替换
@param value 新值
@return 新数组或原有数组
@since 4.1.2 | [
"将元素值设置为数组的某个位置,当给定的index大于数组长度,则追加"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/util/ArrayUtil.java#L417-L424 | train | Sets the value in the buffer at the given index or appends the value to the buffer. | [
30522,
2270,
10763,
1026,
1056,
1028,
1056,
1031,
1033,
2275,
6525,
21512,
4859,
1006,
1056,
1031,
1033,
17698,
1010,
20014,
5950,
1010,
1056,
3643,
1007,
1063,
2065,
1006,
5950,
1026,
17698,
1012,
3091,
1007,
1063,
9140,
1012,
2275,
1006,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/io/FileUtil.java | FileUtil.readLines | public static List<String> readLines(URL url, String charset) throws IORuntimeException {
return readLines(url, charset, new ArrayList<String>());
} | java | public static List<String> readLines(URL url, String charset) throws IORuntimeException {
return readLines(url, charset, new ArrayList<String>());
} | [
"public",
"static",
"List",
"<",
"String",
">",
"readLines",
"(",
"URL",
"url",
",",
"String",
"charset",
")",
"throws",
"IORuntimeException",
"{",
"return",
"readLines",
"(",
"url",
",",
"charset",
",",
"new",
"ArrayList",
"<",
"String",
">",
"(",
")",
... | 从文件中读取每一行数据
@param url 文件的URL
@param charset 字符集
@return 文件中的每行内容的集合List
@throws IORuntimeException IO异常 | [
"从文件中读取每一行数据"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/io/FileUtil.java#L2307-L2309 | train | Reads the contents of the URL into a list of strings. | [
30522,
2270,
10763,
2862,
1026,
5164,
1028,
3191,
12735,
1006,
24471,
2140,
24471,
2140,
1010,
5164,
25869,
13462,
1007,
11618,
22834,
15532,
7292,
10288,
24422,
1063,
2709,
3191,
12735,
1006,
24471,
2140,
1010,
25869,
13462,
1010,
2047,
9140... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | common/src/main/java/io/netty/util/internal/StringUtil.java | StringUtil.toHexStringPadded | public static <T extends Appendable> T toHexStringPadded(T dst, byte[] src, int offset, int length) {
final int end = offset + length;
for (int i = offset; i < end; i++) {
byteToHexStringPadded(dst, src[i]);
}
return dst;
} | java | public static <T extends Appendable> T toHexStringPadded(T dst, byte[] src, int offset, int length) {
final int end = offset + length;
for (int i = offset; i < end; i++) {
byteToHexStringPadded(dst, src[i]);
}
return dst;
} | [
"public",
"static",
"<",
"T",
"extends",
"Appendable",
">",
"T",
"toHexStringPadded",
"(",
"T",
"dst",
",",
"byte",
"[",
"]",
"src",
",",
"int",
"offset",
",",
"int",
"length",
")",
"{",
"final",
"int",
"end",
"=",
"offset",
"+",
"length",
";",
"for"... | Converts the specified byte array into a hexadecimal value and appends it to the specified buffer. | [
"Converts",
"the",
"specified",
"byte",
"array",
"into",
"a",
"hexadecimal",
"value",
"and",
"appends",
"it",
"to",
"the",
"specified",
"buffer",
"."
] | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/common/src/main/java/io/netty/util/internal/StringUtil.java#L130-L136 | train | Converts a byte array to a hexidecimal string. | [
30522,
2270,
10763,
1026,
1056,
8908,
10439,
10497,
3085,
1028,
1056,
2000,
5369,
2595,
3367,
4892,
15455,
5732,
1006,
1056,
16233,
2102,
1010,
24880,
1031,
1033,
5034,
2278,
1010,
20014,
16396,
1010,
20014,
3091,
1007,
1063,
2345,
20014,
2... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-log/src/main/java/cn/hutool/log/StaticLog.java | StaticLog.error | public static void error(Log log, Throwable e) {
error(log, e, e.getMessage());
} | java | public static void error(Log log, Throwable e) {
error(log, e, e.getMessage());
} | [
"public",
"static",
"void",
"error",
"(",
"Log",
"log",
",",
"Throwable",
"e",
")",
"{",
"error",
"(",
"log",
",",
"e",
",",
"e",
".",
"getMessage",
"(",
")",
")",
";",
"}"
] | Error等级日志<br>
@param log 日志对象
@param e 需在日志中堆栈打印的异常 | [
"Error等级日志<br",
">"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-log/src/main/java/cn/hutool/log/StaticLog.java#L184-L186 | train | Logs an error. | [
30522,
2270,
10763,
11675,
7561,
1006,
8833,
8833,
1010,
5466,
3085,
1041,
1007,
1063,
7561,
1006,
8833,
1010,
1041,
1010,
1041,
1012,
2131,
7834,
3736,
3351,
1006,
1007,
1007,
1025,
1065,
102,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/incubator-shardingsphere | sharding-transaction/sharding-transaction-2pc/sharding-transaction-xa/sharding-transaction-xa-core/src/main/java/org/apache/shardingsphere/transaction/xa/jta/connection/XAConnectionFactory.java | XAConnectionFactory.createXAConnection | public static XAConnection createXAConnection(final DatabaseType databaseType, final XADataSource xaDataSource, final Connection connection) {
switch (databaseType) {
case MySQL:
return new MySQLXAConnectionWrapper().wrap(xaDataSource, connection);
case PostgreSQL:
... | java | public static XAConnection createXAConnection(final DatabaseType databaseType, final XADataSource xaDataSource, final Connection connection) {
switch (databaseType) {
case MySQL:
return new MySQLXAConnectionWrapper().wrap(xaDataSource, connection);
case PostgreSQL:
... | [
"public",
"static",
"XAConnection",
"createXAConnection",
"(",
"final",
"DatabaseType",
"databaseType",
",",
"final",
"XADataSource",
"xaDataSource",
",",
"final",
"Connection",
"connection",
")",
"{",
"switch",
"(",
"databaseType",
")",
"{",
"case",
"MySQL",
":",
... | Create XA connection from normal connection.
@param databaseType database type
@param connection normal connection
@param xaDataSource XA data source
@return XA connection | [
"Create",
"XA",
"connection",
"from",
"normal",
"connection",
"."
] | f88fd29fc345dfb31fdce12e9e96cbfa0fd2402d | https://github.com/apache/incubator-shardingsphere/blob/f88fd29fc345dfb31fdce12e9e96cbfa0fd2402d/sharding-transaction/sharding-transaction-2pc/sharding-transaction-xa/sharding-transaction-xa-core/src/main/java/org/apache/shardingsphere/transaction/xa/jta/connection/XAConnectionFactory.java#L47-L58 | train | Create an XAConnection. | [
30522,
2270,
10763,
1060,
22684,
10087,
7542,
3443,
18684,
8663,
2638,
7542,
1006,
30524,
2026,
2015,
4160,
2140,
18684,
8663,
2638,
7542,
13088,
29098,
2121,
1006,
1007,
1012,
10236,
1006,
1060,
8447,
10230,
8162,
3401,
1010,
4434,
1007,
1... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-socket/src/main/java/cn/hutool/socket/aio/AioServer.java | AioServer.close | public void close() {
IoUtil.close(this.channel);
if (null != this.group && false == this.group.isShutdown()) {
try {
this.group.shutdownNow();
} catch (IOException e) {
// ignore
}
}
// 结束阻塞
synchronized (this) {
this.notify();
}
} | java | public void close() {
IoUtil.close(this.channel);
if (null != this.group && false == this.group.isShutdown()) {
try {
this.group.shutdownNow();
} catch (IOException e) {
// ignore
}
}
// 结束阻塞
synchronized (this) {
this.notify();
}
} | [
"public",
"void",
"close",
"(",
")",
"{",
"IoUtil",
".",
"close",
"(",
"this",
".",
"channel",
")",
";",
"if",
"(",
"null",
"!=",
"this",
".",
"group",
"&&",
"false",
"==",
"this",
".",
"group",
".",
"isShutdown",
"(",
")",
")",
"{",
"try",
"{",
... | 关闭服务 | [
"关闭服务"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-socket/src/main/java/cn/hutool/socket/aio/AioServer.java#L151-L166 | train | Close the connection. | [
30522,
2270,
11675,
2485,
1006,
1007,
1063,
22834,
21823,
2140,
1012,
2485,
1006,
2023,
1012,
3149,
1007,
1025,
2065,
1006,
19701,
999,
1027,
2023,
1012,
2177,
1004,
1004,
6270,
1027,
1027,
2023,
1012,
2177,
1012,
26354,
6979,
2102,
7698,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/convert/ConverterRegistry.java | ConverterRegistry.getConverter | public <T> Converter<T> getConverter(Type type, boolean isCustomFirst) {
Converter<T> converter = null;
if (isCustomFirst) {
converter = this.getCustomConverter(type);
if (null == converter) {
converter = this.getDefaultConverter(type);
}
} else {
converter = this.getDefaultConverter(type)... | java | public <T> Converter<T> getConverter(Type type, boolean isCustomFirst) {
Converter<T> converter = null;
if (isCustomFirst) {
converter = this.getCustomConverter(type);
if (null == converter) {
converter = this.getDefaultConverter(type);
}
} else {
converter = this.getDefaultConverter(type)... | [
"public",
"<",
"T",
">",
"Converter",
"<",
"T",
">",
"getConverter",
"(",
"Type",
"type",
",",
"boolean",
"isCustomFirst",
")",
"{",
"Converter",
"<",
"T",
">",
"converter",
"=",
"null",
";",
"if",
"(",
"isCustomFirst",
")",
"{",
"converter",
"=",
"thi... | 获得转换器<br>
@param <T> 转换的目标类型
@param type 类型
@param isCustomFirst 是否自定义转换器优先
@return 转换器 | [
"获得转换器<br",
">"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/convert/ConverterRegistry.java#L135-L149 | train | Returns a converter for the specified type. | [
30522,
2270,
1026,
1056,
1028,
10463,
2121,
1026,
1056,
1028,
2131,
8663,
16874,
2121,
1006,
2828,
2828,
1010,
22017,
20898,
2003,
7874,
20389,
8873,
12096,
1007,
1063,
10463,
2121,
1026,
1056,
1028,
10463,
2121,
1027,
19701,
1025,
2065,
10... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/util/NumberUtil.java | NumberUtil.roundHalfEven | public static BigDecimal roundHalfEven(BigDecimal value, int scale) {
return round(value, scale, RoundingMode.HALF_EVEN);
} | java | public static BigDecimal roundHalfEven(BigDecimal value, int scale) {
return round(value, scale, RoundingMode.HALF_EVEN);
} | [
"public",
"static",
"BigDecimal",
"roundHalfEven",
"(",
"BigDecimal",
"value",
",",
"int",
"scale",
")",
"{",
"return",
"round",
"(",
"value",
",",
"scale",
",",
"RoundingMode",
".",
"HALF_EVEN",
")",
";",
"}"
] | 四舍六入五成双计算法
<p>
四舍六入五成双是一种比较精确比较科学的计数保留法,是一种数字修约规则。
</p>
<pre>
算法规则:
四舍六入五考虑,
五后非零就进一,
五后皆零看奇偶,
五前为偶应舍去,
五前为奇要进一。
</pre>
@param value 需要科学计算的数据
@param scale 保留的小数位
@return 结果
@since 4.1.0 | [
"四舍六入五成双计算法",
"<p",
">",
"四舍六入五成双是一种比较精确比较科学的计数保留法,是一种数字修约规则。",
"<",
"/",
"p",
">"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/util/NumberUtil.java#L945-L947 | train | Round to half even. | [
30522,
2270,
10763,
2502,
3207,
6895,
9067,
2461,
8865,
7959,
8159,
1006,
2502,
3207,
6895,
9067,
3643,
1010,
20014,
4094,
1007,
1063,
2709,
2461,
1006,
3643,
1010,
4094,
1010,
26939,
5302,
3207,
1012,
2431,
1035,
2130,
1007,
1025,
1065,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/incubator-shardingsphere | sharding-proxy/sharding-proxy-frontend/sharding-proxy-frontend-postgresql/src/main/java/org/apache/shardingsphere/shardingproxy/frontend/postgresql/command/PostgreSQLCommandExecutorFactory.java | PostgreSQLCommandExecutorFactory.newInstance | public static CommandExecutor newInstance(final PostgreSQLCommandPacketType commandPacketType, final PostgreSQLCommandPacket commandPacket, final BackendConnection backendConnection) {
log.debug("Execute packet type: {}, value: {}", commandPacketType, commandPacket);
switch (commandPacketType) {
... | java | public static CommandExecutor newInstance(final PostgreSQLCommandPacketType commandPacketType, final PostgreSQLCommandPacket commandPacket, final BackendConnection backendConnection) {
log.debug("Execute packet type: {}, value: {}", commandPacketType, commandPacket);
switch (commandPacketType) {
... | [
"public",
"static",
"CommandExecutor",
"newInstance",
"(",
"final",
"PostgreSQLCommandPacketType",
"commandPacketType",
",",
"final",
"PostgreSQLCommandPacket",
"commandPacket",
",",
"final",
"BackendConnection",
"backendConnection",
")",
"{",
"log",
".",
"debug",
"(",
"\... | Create new instance of command executor.
@param commandPacketType command packet type for PostgreSQL
@param commandPacket command packet for PostgreSQL
@param backendConnection backend connection
@return command executor | [
"Create",
"new",
"instance",
"of",
"command",
"executor",
"."
] | f88fd29fc345dfb31fdce12e9e96cbfa0fd2402d | https://github.com/apache/incubator-shardingsphere/blob/f88fd29fc345dfb31fdce12e9e96cbfa0fd2402d/sharding-proxy/sharding-proxy-frontend/sharding-proxy-frontend-postgresql/src/main/java/org/apache/shardingsphere/shardingproxy/frontend/postgresql/command/PostgreSQLCommandExecutorFactory.java#L56-L76 | train | Create new instance of command executor. | [
30522,
2270,
10763,
3094,
10288,
8586,
16161,
2099,
2047,
7076,
26897,
1006,
2345,
2695,
17603,
2015,
4160,
22499,
14760,
4859,
23947,
6582,
18863,
3094,
23947,
6582,
18863,
1010,
2345,
2695,
17603,
2015,
4160,
22499,
14760,
4859,
23947,
3388... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
hankcs/HanLP | src/main/java/com/hankcs/hanlp/collection/trie/datrie/MutableDoubleArrayTrieInteger.java | MutableDoubleArrayTrieInteger.insert | public boolean insert(String key, int value, boolean overwrite)
{
if ((null == key) || key.length() == 0 || (key.indexOf(UNUSED_CHAR) != -1))
{
return false;
}
if ((value < 0) || ((value & LEAF_BIT) != 0))
{
return false;
}
value = set... | java | public boolean insert(String key, int value, boolean overwrite)
{
if ((null == key) || key.length() == 0 || (key.indexOf(UNUSED_CHAR) != -1))
{
return false;
}
if ((value < 0) || ((value & LEAF_BIT) != 0))
{
return false;
}
value = set... | [
"public",
"boolean",
"insert",
"(",
"String",
"key",
",",
"int",
"value",
",",
"boolean",
"overwrite",
")",
"{",
"if",
"(",
"(",
"null",
"==",
"key",
")",
"||",
"key",
".",
"length",
"(",
")",
"==",
"0",
"||",
"(",
"key",
".",
"indexOf",
"(",
"UN... | 插入条目
@param key 键
@param value 值
@param overwrite 是否覆盖
@return | [
"插入条目"
] | a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce | https://github.com/hankcs/HanLP/blob/a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce/src/main/java/com/hankcs/hanlp/collection/trie/datrie/MutableDoubleArrayTrieInteger.java#L251-L304 | train | Insert a value into the cache. | [
30522,
2270,
22017,
20898,
19274,
1006,
5164,
3145,
1010,
20014,
30524,
5950,
11253,
1006,
15171,
1035,
25869,
1007,
999,
1027,
1011,
1015,
1007,
1007,
1063,
2709,
6270,
1025,
1065,
2065,
1006,
1006,
3643,
1026,
1014,
1007,
1064,
1064,
1006... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
spring-projects/spring-boot | spring-boot-project/spring-boot/src/main/java/org/springframework/boot/context/properties/bind/Bindable.java | Bindable.withExistingValue | public Bindable<T> withExistingValue(T existingValue) {
Assert.isTrue(
existingValue == null || this.type.isArray()
|| this.boxedType.resolve().isInstance(existingValue),
() -> "ExistingValue must be an instance of " + this.type);
Supplier<T> value = (existingValue != null) ? () -> existingValue : nul... | java | public Bindable<T> withExistingValue(T existingValue) {
Assert.isTrue(
existingValue == null || this.type.isArray()
|| this.boxedType.resolve().isInstance(existingValue),
() -> "ExistingValue must be an instance of " + this.type);
Supplier<T> value = (existingValue != null) ? () -> existingValue : nul... | [
"public",
"Bindable",
"<",
"T",
">",
"withExistingValue",
"(",
"T",
"existingValue",
")",
"{",
"Assert",
".",
"isTrue",
"(",
"existingValue",
"==",
"null",
"||",
"this",
".",
"type",
".",
"isArray",
"(",
")",
"||",
"this",
".",
"boxedType",
".",
"resolve... | Create an updated {@link Bindable} instance with an existing value.
@param existingValue the existing value
@return an updated {@link Bindable} | [
"Create",
"an",
"updated",
"{"
] | 0b27f7c70e164b2b1a96477f1d9c1acba56790c1 | https://github.com/spring-projects/spring-boot/blob/0b27f7c70e164b2b1a96477f1d9c1acba56790c1/spring-boot-project/spring-boot/src/main/java/org/springframework/boot/context/properties/bind/Bindable.java#L161-L168 | train | Returns a new instance of this Bindable with the specified existing value. | [
30522,
2270,
14187,
3085,
1026,
1056,
1028,
2007,
10288,
2923,
2075,
10175,
5657,
1006,
1056,
4493,
10175,
5657,
1007,
1063,
20865,
1012,
21541,
6820,
2063,
1006,
4493,
10175,
5657,
1027,
1027,
19701,
1064,
1064,
2023,
1012,
2828,
1012,
180... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
hankcs/HanLP | src/main/java/com/hankcs/hanlp/mining/word2vec/DocVectorModel.java | DocVectorModel.query | public Vector query(String content)
{
if (content == null || content.length() == 0) return null;
List<Term> termList = NotionalTokenizer.segment(content);
Vector result = new Vector(dimension());
int n = 0;
for (Term term : termList)
{
Vector vector = word... | java | public Vector query(String content)
{
if (content == null || content.length() == 0) return null;
List<Term> termList = NotionalTokenizer.segment(content);
Vector result = new Vector(dimension());
int n = 0;
for (Term term : termList)
{
Vector vector = word... | [
"public",
"Vector",
"query",
"(",
"String",
"content",
")",
"{",
"if",
"(",
"content",
"==",
"null",
"||",
"content",
".",
"length",
"(",
")",
"==",
"0",
")",
"return",
"null",
";",
"List",
"<",
"Term",
">",
"termList",
"=",
"NotionalTokenizer",
".",
... | 将一个文档转为向量
@param content 文档
@return 向量 | [
"将一个文档转为向量"
] | a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce | https://github.com/hankcs/HanLP/blob/a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce/src/main/java/com/hankcs/hanlp/mining/word2vec/DocVectorModel.java#L68-L90 | train | Query a Sequence of Terms from a String. | [
30522,
2270,
9207,
23032,
1006,
5164,
4180,
1007,
1063,
2065,
1006,
4180,
1027,
1027,
19701,
1064,
1064,
4180,
1012,
3091,
1006,
1007,
1027,
1027,
1014,
1007,
2709,
19701,
1025,
2862,
1026,
2744,
1028,
2744,
9863,
1027,
9366,
2389,
18715,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
alibaba/canal | client-adapter/rdb/src/main/java/com/alibaba/otter/canal/client/adapter/rdb/RdbAdapter.java | RdbAdapter.destroy | @Override
public void destroy() {
if (rdbConfigMonitor != null) {
rdbConfigMonitor.destroy();
}
if (rdbSyncService != null) {
rdbSyncService.close();
}
if (dataSource != null) {
dataSource.close();
}
} | java | @Override
public void destroy() {
if (rdbConfigMonitor != null) {
rdbConfigMonitor.destroy();
}
if (rdbSyncService != null) {
rdbSyncService.close();
}
if (dataSource != null) {
dataSource.close();
}
} | [
"@",
"Override",
"public",
"void",
"destroy",
"(",
")",
"{",
"if",
"(",
"rdbConfigMonitor",
"!=",
"null",
")",
"{",
"rdbConfigMonitor",
".",
"destroy",
"(",
")",
";",
"}",
"if",
"(",
"rdbSyncService",
"!=",
"null",
")",
"{",
"rdbSyncService",
".",
"close... | 销毁方法 | [
"销毁方法"
] | 8f088cddc0755f4350c5aaae95c6e4002d90a40f | https://github.com/alibaba/canal/blob/8f088cddc0755f4350c5aaae95c6e4002d90a40f/client-adapter/rdb/src/main/java/com/alibaba/otter/canal/client/adapter/rdb/RdbAdapter.java#L274-L287 | train | Destroy the application. | [
30522,
1030,
2058,
15637,
2270,
11675,
6033,
1006,
1007,
1063,
2065,
1006,
16428,
9818,
2239,
8873,
21693,
10698,
4263,
999,
1027,
19701,
1007,
1063,
16428,
9818,
2239,
8873,
21693,
10698,
4263,
1012,
6033,
1006,
1007,
1025,
1065,
2065,
100... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/io/FileUtil.java | FileUtil.file | public static File file(String path) {
if (StrUtil.isBlank(path)) {
throw new NullPointerException("File path is blank!");
}
return new File(getAbsolutePath(path));
} | java | public static File file(String path) {
if (StrUtil.isBlank(path)) {
throw new NullPointerException("File path is blank!");
}
return new File(getAbsolutePath(path));
} | [
"public",
"static",
"File",
"file",
"(",
"String",
"path",
")",
"{",
"if",
"(",
"StrUtil",
".",
"isBlank",
"(",
"path",
")",
")",
"{",
"throw",
"new",
"NullPointerException",
"(",
"\"File path is blank!\"",
")",
";",
"}",
"return",
"new",
"File",
"(",
"g... | 创建File对象,自动识别相对或绝对路径,相对路径将自动从ClassPath下寻找
@param path 文件路径
@return File | [
"创建File对象,自动识别相对或绝对路径,相对路径将自动从ClassPath下寻找"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/io/FileUtil.java#L314-L319 | train | Create a File object from a file path. | [
30522,
2270,
10763,
5371,
5371,
1006,
5164,
4130,
1007,
1063,
2065,
1006,
2358,
22134,
4014,
1012,
2003,
28522,
8950,
1006,
4130,
1007,
1007,
1063,
5466,
2047,
19701,
8400,
7869,
2595,
24422,
1006,
1000,
5371,
4130,
2003,
8744,
999,
1000,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-libraries/flink-cep/src/main/java/org/apache/flink/cep/PatternProcessFunctionBuilder.java | PatternProcessFunctionBuilder.fromFlatSelect | static <IN, OUT> FlatSelectBuilder<IN, OUT> fromFlatSelect(final PatternFlatSelectFunction<IN, OUT> function) {
return new FlatSelectBuilder<>(function);
} | java | static <IN, OUT> FlatSelectBuilder<IN, OUT> fromFlatSelect(final PatternFlatSelectFunction<IN, OUT> function) {
return new FlatSelectBuilder<>(function);
} | [
"static",
"<",
"IN",
",",
"OUT",
">",
"FlatSelectBuilder",
"<",
"IN",
",",
"OUT",
">",
"fromFlatSelect",
"(",
"final",
"PatternFlatSelectFunction",
"<",
"IN",
",",
"OUT",
">",
"function",
")",
"{",
"return",
"new",
"FlatSelectBuilder",
"<>",
"(",
"function",... | Starts constructing a {@link PatternProcessFunction} from a {@link PatternFlatSelectFunction} that
emitted elements through {@link org.apache.flink.util.Collector}. | [
"Starts",
"constructing",
"a",
"{"
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-libraries/flink-cep/src/main/java/org/apache/flink/cep/PatternProcessFunctionBuilder.java#L42-L44 | train | Create a new FlatSelectBuilder from a pattern flat select function. | [
30522,
10763,
1026,
1999,
1010,
2041,
1028,
14201,
12260,
6593,
8569,
23891,
2099,
1026,
1999,
1010,
2041,
1028,
2013,
10258,
11149,
12260,
6593,
1006,
2345,
5418,
10258,
11149,
12260,
6593,
11263,
27989,
1026,
1999,
1010,
2041,
1028,
3853,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
spring-projects/spring-boot | spring-boot-project/spring-boot-cli/src/main/java/org/springframework/boot/cli/compiler/AstUtils.java | AstUtils.hasAtLeastOneAnnotation | public static boolean hasAtLeastOneAnnotation(AnnotatedNode node,
String... annotations) {
for (AnnotationNode annotationNode : node.getAnnotations()) {
for (String annotation : annotations) {
if (PatternMatchUtils.simpleMatch(annotation,
annotationNode.getClassNode().getName())) {
return true;
... | java | public static boolean hasAtLeastOneAnnotation(AnnotatedNode node,
String... annotations) {
for (AnnotationNode annotationNode : node.getAnnotations()) {
for (String annotation : annotations) {
if (PatternMatchUtils.simpleMatch(annotation,
annotationNode.getClassNode().getName())) {
return true;
... | [
"public",
"static",
"boolean",
"hasAtLeastOneAnnotation",
"(",
"AnnotatedNode",
"node",
",",
"String",
"...",
"annotations",
")",
"{",
"for",
"(",
"AnnotationNode",
"annotationNode",
":",
"node",
".",
"getAnnotations",
"(",
")",
")",
"{",
"for",
"(",
"String",
... | Determine if an {@link AnnotatedNode} has one or more of the specified annotations.
N.B. the annotation type names are not normally fully qualified.
@param node the node to examine
@param annotations the annotations to look for
@return {@code true} if at least one of the annotations is found, otherwise
{@code false} | [
"Determine",
"if",
"an",
"{"
] | 0b27f7c70e164b2b1a96477f1d9c1acba56790c1 | https://github.com/spring-projects/spring-boot/blob/0b27f7c70e164b2b1a96477f1d9c1acba56790c1/spring-boot-project/spring-boot-cli/src/main/java/org/springframework/boot/cli/compiler/AstUtils.java#L79-L90 | train | Checks if the given annotated node has at least one of the given annotations. | [
30522,
2270,
10763,
22017,
20898,
2038,
4017,
19738,
9221,
11639,
17287,
3508,
1006,
5754,
17287,
3064,
3630,
3207,
13045,
1010,
5164,
1012,
1012,
1012,
5754,
17287,
9285,
1007,
1063,
2005,
1006,
5754,
17287,
3508,
3630,
3207,
5754,
17287,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
spring-projects/spring-boot | spring-boot-project/spring-boot-autoconfigure/src/main/java/org/springframework/boot/autoconfigure/hazelcast/HazelcastClientFactory.java | HazelcastClientFactory.getHazelcastInstance | public HazelcastInstance getHazelcastInstance() {
if (StringUtils.hasText(this.clientConfig.getInstanceName())) {
return HazelcastClient
.getHazelcastClientByName(this.clientConfig.getInstanceName());
}
return HazelcastClient.newHazelcastClient(this.clientConfig);
} | java | public HazelcastInstance getHazelcastInstance() {
if (StringUtils.hasText(this.clientConfig.getInstanceName())) {
return HazelcastClient
.getHazelcastClientByName(this.clientConfig.getInstanceName());
}
return HazelcastClient.newHazelcastClient(this.clientConfig);
} | [
"public",
"HazelcastInstance",
"getHazelcastInstance",
"(",
")",
"{",
"if",
"(",
"StringUtils",
".",
"hasText",
"(",
"this",
".",
"clientConfig",
".",
"getInstanceName",
"(",
")",
")",
")",
"{",
"return",
"HazelcastClient",
".",
"getHazelcastClientByName",
"(",
... | Get the {@link HazelcastInstance}.
@return the {@link HazelcastInstance} | [
"Get",
"the",
"{"
] | 0b27f7c70e164b2b1a96477f1d9c1acba56790c1 | https://github.com/spring-projects/spring-boot/blob/0b27f7c70e164b2b1a96477f1d9c1acba56790c1/spring-boot-project/spring-boot-autoconfigure/src/main/java/org/springframework/boot/autoconfigure/hazelcast/HazelcastClientFactory.java#L74-L80 | train | Get the HazelcastInstance. | [
30522,
2270,
14015,
10526,
7076,
26897,
2131,
3270,
12638,
10526,
7076,
26897,
1006,
1007,
1063,
2065,
1006,
5164,
21823,
4877,
1012,
24748,
18413,
1006,
2023,
1012,
7396,
8663,
8873,
2290,
1012,
2131,
7076,
26897,
18442,
1006,
1007,
1007,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/io/FileUtil.java | FileUtil.appendUtf8Lines | public static <T> File appendUtf8Lines(Collection<T> list, String path) throws IORuntimeException {
return appendLines(list, path, CharsetUtil.CHARSET_UTF_8);
} | java | public static <T> File appendUtf8Lines(Collection<T> list, String path) throws IORuntimeException {
return appendLines(list, path, CharsetUtil.CHARSET_UTF_8);
} | [
"public",
"static",
"<",
"T",
">",
"File",
"appendUtf8Lines",
"(",
"Collection",
"<",
"T",
">",
"list",
",",
"String",
"path",
")",
"throws",
"IORuntimeException",
"{",
"return",
"appendLines",
"(",
"list",
",",
"path",
",",
"CharsetUtil",
".",
"CHARSET_UTF_... | 将列表写入文件,追加模式
@param <T> 集合元素类型
@param list 列表
@param path 文件路径
@return 目标文件
@throws IORuntimeException IO异常
@since 3.1.2 | [
"将列表写入文件,追加模式"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/io/FileUtil.java#L2953-L2955 | train | Append UTF - 8 lines to a file. | [
30522,
2270,
10763,
1026,
1056,
1028,
5371,
10439,
10497,
4904,
2546,
2620,
12735,
1006,
3074,
1026,
1056,
1028,
2862,
1010,
5164,
4130,
1007,
11618,
22834,
15532,
7292,
10288,
24422,
1063,
2709,
10439,
10497,
12735,
1006,
2862,
1010,
4130,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-core/src/main/java/org/apache/flink/api/common/io/FileOutputFormat.java | FileOutputFormat.initDefaultsFromConfiguration | public static void initDefaultsFromConfiguration(Configuration configuration) {
final boolean overwrite = configuration.getBoolean(CoreOptions.FILESYTEM_DEFAULT_OVERRIDE);
DEFAULT_WRITE_MODE = overwrite ? WriteMode.OVERWRITE : WriteMode.NO_OVERWRITE;
final boolean alwaysCreateDirectory = configuration.getBoo... | java | public static void initDefaultsFromConfiguration(Configuration configuration) {
final boolean overwrite = configuration.getBoolean(CoreOptions.FILESYTEM_DEFAULT_OVERRIDE);
DEFAULT_WRITE_MODE = overwrite ? WriteMode.OVERWRITE : WriteMode.NO_OVERWRITE;
final boolean alwaysCreateDirectory = configuration.getBoo... | [
"public",
"static",
"void",
"initDefaultsFromConfiguration",
"(",
"Configuration",
"configuration",
")",
"{",
"final",
"boolean",
"overwrite",
"=",
"configuration",
".",
"getBoolean",
"(",
"CoreOptions",
".",
"FILESYTEM_DEFAULT_OVERRIDE",
")",
";",
"DEFAULT_WRITE_MODE",
... | Initialize defaults for output format. Needs to be a static method because it is configured for local
cluster execution.
@param configuration The configuration to load defaults from | [
"Initialize",
"defaults",
"for",
"output",
"format",
".",
"Needs",
"to",
"be",
"a",
"static",
"method",
"because",
"it",
"is",
"configured",
"for",
"local",
"cluster",
"execution",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-core/src/main/java/org/apache/flink/api/common/io/FileOutputFormat.java#L75-L83 | train | Initialize the defaults from the configuration. | [
30522,
2270,
10763,
11675,
1999,
4183,
3207,
7011,
11314,
22747,
21716,
8663,
8873,
27390,
3370,
1006,
9563,
9563,
1007,
1063,
2345,
22017,
20898,
2058,
26373,
1027,
9563,
1012,
2131,
5092,
9890,
2319,
1006,
4563,
7361,
9285,
1012,
6764,
17... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/hooks/MasterHooks.java | MasterHooks.triggerMasterHooks | public static List<MasterState> triggerMasterHooks(
Collection<MasterTriggerRestoreHook<?>> hooks,
long checkpointId,
long timestamp,
Executor executor,
Time timeout) throws FlinkException {
final ArrayList<MasterState> states = new ArrayList<>(hooks.size());
for (MasterTriggerRestoreHook<?> hook :... | java | public static List<MasterState> triggerMasterHooks(
Collection<MasterTriggerRestoreHook<?>> hooks,
long checkpointId,
long timestamp,
Executor executor,
Time timeout) throws FlinkException {
final ArrayList<MasterState> states = new ArrayList<>(hooks.size());
for (MasterTriggerRestoreHook<?> hook :... | [
"public",
"static",
"List",
"<",
"MasterState",
">",
"triggerMasterHooks",
"(",
"Collection",
"<",
"MasterTriggerRestoreHook",
"<",
"?",
">",
">",
"hooks",
",",
"long",
"checkpointId",
",",
"long",
"timestamp",
",",
"Executor",
"executor",
",",
"Time",
"timeout"... | Triggers all given master hooks and returns state objects for each hook that
produced a state.
@param hooks The hooks to trigger
@param checkpointId The checkpoint ID of the triggering checkpoint
@param timestamp The (informational) timestamp for the triggering checkpoint
@param executor An executor that can be used f... | [
"Triggers",
"all",
"given",
"master",
"hooks",
"and",
"returns",
"state",
"objects",
"for",
"each",
"hook",
"that",
"produced",
"a",
"state",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/hooks/MasterHooks.java#L117-L135 | train | Triggers the given set of master hooks. | [
30522,
2270,
10763,
2862,
1026,
5972,
12259,
1028,
9495,
8706,
6806,
6559,
2015,
1006,
3074,
1026,
3040,
18886,
13327,
28533,
5686,
6806,
6559,
1026,
1029,
1028,
1028,
18008,
1010,
2146,
26520,
3593,
1010,
2146,
2335,
15464,
2361,
1010,
465... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | common/src/main/java/io/netty/util/ThreadDeathWatcher.java | ThreadDeathWatcher.awaitInactivity | public static boolean awaitInactivity(long timeout, TimeUnit unit) throws InterruptedException {
if (unit == null) {
throw new NullPointerException("unit");
}
Thread watcherThread = ThreadDeathWatcher.watcherThread;
if (watcherThread != null) {
watcherThread.join... | java | public static boolean awaitInactivity(long timeout, TimeUnit unit) throws InterruptedException {
if (unit == null) {
throw new NullPointerException("unit");
}
Thread watcherThread = ThreadDeathWatcher.watcherThread;
if (watcherThread != null) {
watcherThread.join... | [
"public",
"static",
"boolean",
"awaitInactivity",
"(",
"long",
"timeout",
",",
"TimeUnit",
"unit",
")",
"throws",
"InterruptedException",
"{",
"if",
"(",
"unit",
"==",
"null",
")",
"{",
"throw",
"new",
"NullPointerException",
"(",
"\"unit\"",
")",
";",
"}",
... | Waits until the thread of this watcher has no threads to watch and terminates itself.
Because a new watcher thread will be started again on {@link #watch(Thread, Runnable)},
this operation is only useful when you want to ensure that the watcher thread is terminated
<strong>after</strong> your application is shut down a... | [
"Waits",
"until",
"the",
"thread",
"of",
"this",
"watcher",
"has",
"no",
"threads",
"to",
"watch",
"and",
"terminates",
"itself",
".",
"Because",
"a",
"new",
"watcher",
"thread",
"will",
"be",
"started",
"again",
"on",
"{",
"@link",
"#watch",
"(",
"Thread"... | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/common/src/main/java/io/netty/util/ThreadDeathWatcher.java#L139-L151 | train | Await inactivity. | [
30522,
2270,
10763,
22017,
20898,
26751,
3981,
6593,
7730,
1006,
2146,
2051,
5833,
1010,
2051,
19496,
2102,
3131,
1007,
11618,
7153,
10288,
24422,
1063,
2065,
1006,
3131,
1027,
1027,
19701,
1007,
1063,
5466,
2047,
19701,
8400,
7869,
2595,
2... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AllWindowedStream.java | AllWindowedStream.sum | public SingleOutputStreamOperator<T> sum(String field) {
return aggregate(new SumAggregator<>(field, input.getType(), input.getExecutionConfig()));
} | java | public SingleOutputStreamOperator<T> sum(String field) {
return aggregate(new SumAggregator<>(field, input.getType(), input.getExecutionConfig()));
} | [
"public",
"SingleOutputStreamOperator",
"<",
"T",
">",
"sum",
"(",
"String",
"field",
")",
"{",
"return",
"aggregate",
"(",
"new",
"SumAggregator",
"<>",
"(",
"field",
",",
"input",
".",
"getType",
"(",
")",
",",
"input",
".",
"getExecutionConfig",
"(",
")... | Applies an aggregation that sums every window of the pojo data stream at
the given field for every window.
<p>A field expression is either the name of a public field or a getter method with
parentheses of the stream's underlying type. A dot can be used to drill down into objects,
as in {@code "field1.getInnerField2()"... | [
"Applies",
"an",
"aggregation",
"that",
"sums",
"every",
"window",
"of",
"the",
"pojo",
"data",
"stream",
"at",
"the",
"given",
"field",
"for",
"every",
"window",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AllWindowedStream.java#L1393-L1395 | train | Applies an aggregation that sums the values of the stream | [
30522,
2270,
2309,
5833,
18780,
21422,
25918,
8844,
1026,
1056,
1028,
7680,
1006,
5164,
2492,
1007,
1063,
2709,
9572,
1006,
2047,
7680,
8490,
17603,
20697,
2953,
1026,
1028,
1006,
2492,
1010,
7953,
1012,
2131,
13874,
1006,
1007,
1010,
7953,... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | codec-http/src/main/java/io/netty/handler/codec/http/multipart/HttpPostRequestDecoder.java | HttpPostRequestDecoder.isMultipart | public static boolean isMultipart(HttpRequest request) {
if (request.headers().contains(HttpHeaderNames.CONTENT_TYPE)) {
return getMultipartDataBoundary(request.headers().get(HttpHeaderNames.CONTENT_TYPE)) != null;
} else {
return false;
}
} | java | public static boolean isMultipart(HttpRequest request) {
if (request.headers().contains(HttpHeaderNames.CONTENT_TYPE)) {
return getMultipartDataBoundary(request.headers().get(HttpHeaderNames.CONTENT_TYPE)) != null;
} else {
return false;
}
} | [
"public",
"static",
"boolean",
"isMultipart",
"(",
"HttpRequest",
"request",
")",
"{",
"if",
"(",
"request",
".",
"headers",
"(",
")",
".",
"contains",
"(",
"HttpHeaderNames",
".",
"CONTENT_TYPE",
")",
")",
"{",
"return",
"getMultipartDataBoundary",
"(",
"requ... | Check if the given request is a multipart request
@return True if the request is a Multipart request | [
"Check",
"if",
"the",
"given",
"request",
"is",
"a",
"multipart",
"request"
] | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/codec-http/src/main/java/io/netty/handler/codec/http/multipart/HttpPostRequestDecoder.java#L142-L148 | train | Is multipart request? | [
30522,
2270,
10763,
22017,
20898,
2003,
12274,
7096,
11514,
8445,
1006,
8299,
2890,
15500,
5227,
1007,
1063,
2065,
1006,
5227,
1012,
20346,
2015,
1006,
1007,
1012,
3397,
1006,
8299,
4974,
11795,
14074,
2015,
1012,
4180,
1035,
2828,
1007,
10... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | common/src/main/java/io/netty/util/internal/logging/Log4JLogger.java | Log4JLogger.trace | @Override
public void trace(String msg, Throwable t) {
logger.log(FQCN, traceCapable ? Level.TRACE : Level.DEBUG, msg, t);
} | java | @Override
public void trace(String msg, Throwable t) {
logger.log(FQCN, traceCapable ? Level.TRACE : Level.DEBUG, msg, t);
} | [
"@",
"Override",
"public",
"void",
"trace",
"(",
"String",
"msg",
",",
"Throwable",
"t",
")",
"{",
"logger",
".",
"log",
"(",
"FQCN",
",",
"traceCapable",
"?",
"Level",
".",
"TRACE",
":",
"Level",
".",
"DEBUG",
",",
"msg",
",",
"t",
")",
";",
"}"
] | Log an exception (throwable) at level TRACE with an accompanying message.
@param msg
the message accompanying the exception
@param t
the exception (throwable) to log | [
"Log",
"an",
"exception",
"(",
"throwable",
")",
"at",
"level",
"TRACE",
"with",
"an",
"accompanying",
"message",
"."
] | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/common/src/main/java/io/netty/util/internal/logging/Log4JLogger.java#L184-L187 | train | Log a message at the TRACE level. | [
30522,
1030,
2058,
15637,
2270,
11675,
7637,
1006,
5164,
5796,
2290,
1010,
5466,
3085,
1056,
1007,
1063,
8833,
4590,
1012,
8833,
1006,
1042,
4160,
2278,
2078,
1010,
7637,
17695,
3085,
1029,
2504,
1012,
7637,
1024,
2504,
1012,
2139,
8569,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-core/src/main/java/org/apache/flink/api/java/typeutils/TypeExtractionUtils.java | TypeExtractionUtils.checkAndExtractLambda | public static LambdaExecutable checkAndExtractLambda(Function function) throws TypeExtractionException {
try {
// get serialized lambda
SerializedLambda serializedLambda = null;
for (Class<?> clazz = function.getClass(); clazz != null; clazz = clazz.getSuperclass()) {
try {
Method replaceMethod = cl... | java | public static LambdaExecutable checkAndExtractLambda(Function function) throws TypeExtractionException {
try {
// get serialized lambda
SerializedLambda serializedLambda = null;
for (Class<?> clazz = function.getClass(); clazz != null; clazz = clazz.getSuperclass()) {
try {
Method replaceMethod = cl... | [
"public",
"static",
"LambdaExecutable",
"checkAndExtractLambda",
"(",
"Function",
"function",
")",
"throws",
"TypeExtractionException",
"{",
"try",
"{",
"// get serialized lambda",
"SerializedLambda",
"serializedLambda",
"=",
"null",
";",
"for",
"(",
"Class",
"<",
"?",
... | Checks if the given function has been implemented using a Java 8 lambda. If yes, a LambdaExecutable
is returned describing the method/constructor. Otherwise null.
@throws TypeExtractionException lambda extraction is pretty hacky, it might fail for unknown JVM issues. | [
"Checks",
"if",
"the",
"given",
"function",
"has",
"been",
"implemented",
"using",
"a",
"Java",
"8",
"lambda",
".",
"If",
"yes",
"a",
"LambdaExecutable",
"is",
"returned",
"describing",
"the",
"method",
"/",
"constructor",
".",
"Otherwise",
"null",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-core/src/main/java/org/apache/flink/api/java/typeutils/TypeExtractionUtils.java#L99-L156 | train | Check if the given function is a Lambda function and if so extract it. | [
30522,
2270,
10763,
23375,
10288,
8586,
23056,
4638,
5685,
10288,
6494,
6593,
10278,
2497,
2850,
1006,
3853,
3853,
1007,
11618,
2828,
10288,
6494,
7542,
10288,
24422,
1063,
3046,
1063,
1013,
1013,
2131,
27289,
23375,
27289,
10278,
2497,
2850,... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-yarn/src/main/java/org/apache/flink/yarn/Utils.java | Utils.require | static void require(boolean condition, String message, Object... values) {
if (!condition) {
throw new RuntimeException(String.format(message, values));
}
} | java | static void require(boolean condition, String message, Object... values) {
if (!condition) {
throw new RuntimeException(String.format(message, values));
}
} | [
"static",
"void",
"require",
"(",
"boolean",
"condition",
",",
"String",
"message",
",",
"Object",
"...",
"values",
")",
"{",
"if",
"(",
"!",
"condition",
")",
"{",
"throw",
"new",
"RuntimeException",
"(",
"String",
".",
"format",
"(",
"message",
",",
"v... | Validates a condition, throwing a RuntimeException if the condition is violated.
@param condition The condition.
@param message The message for the runtime exception, with format variables as defined by
{@link String#format(String, Object...)}.
@param values The format arguments. | [
"Validates",
"a",
"condition",
"throwing",
"a",
"RuntimeException",
"if",
"the",
"condition",
"is",
"violated",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-yarn/src/main/java/org/apache/flink/yarn/Utils.java#L589-L593 | train | Require that the condition is true. | [
30522,
10763,
11675,
5478,
1006,
22017,
20898,
4650,
1010,
5164,
4471,
1010,
4874,
1012,
1012,
1012,
5300,
1007,
1063,
2065,
1006,
999,
4650,
1007,
1063,
5466,
2047,
2448,
7292,
10288,
24422,
1006,
5164,
1012,
4289,
1006,
4471,
1010,
5300,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-clients/src/main/java/org/apache/flink/client/program/JobWithJars.java | JobWithJars.getUserCodeClassLoader | public ClassLoader getUserCodeClassLoader() {
if (this.userCodeClassLoader == null) {
this.userCodeClassLoader = buildUserCodeClassLoader(jarFiles, classpaths, getClass().getClassLoader());
}
return this.userCodeClassLoader;
} | java | public ClassLoader getUserCodeClassLoader() {
if (this.userCodeClassLoader == null) {
this.userCodeClassLoader = buildUserCodeClassLoader(jarFiles, classpaths, getClass().getClassLoader());
}
return this.userCodeClassLoader;
} | [
"public",
"ClassLoader",
"getUserCodeClassLoader",
"(",
")",
"{",
"if",
"(",
"this",
".",
"userCodeClassLoader",
"==",
"null",
")",
"{",
"this",
".",
"userCodeClassLoader",
"=",
"buildUserCodeClassLoader",
"(",
"jarFiles",
",",
"classpaths",
",",
"getClass",
"(",
... | Gets the {@link java.lang.ClassLoader} that must be used to load user code classes.
@return The user code ClassLoader. | [
"Gets",
"the",
"{",
"@link",
"java",
".",
"lang",
".",
"ClassLoader",
"}",
"that",
"must",
"be",
"used",
"to",
"load",
"user",
"code",
"classes",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-clients/src/main/java/org/apache/flink/client/program/JobWithJars.java#L105-L110 | train | Gets the user code class loader. | [
30522,
2270,
2465,
11066,
2121,
2131,
20330,
16044,
26266,
11066,
2121,
1006,
1007,
1063,
2065,
1006,
2023,
1012,
5310,
16044,
26266,
11066,
2121,
1027,
1027,
19701,
1007,
1063,
2023,
1012,
5310,
16044,
26266,
11066,
2121,
1027,
3857,
20330,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/io/BOMInputStream.java | BOMInputStream.init | protected void init() throws IOException {
if (isInited) {
return;
}
byte bom[] = new byte[BOM_SIZE];
int n, unread;
n = in.read(bom, 0, bom.length);
if ((bom[0] == (byte) 0x00) && (bom[1] == (byte) 0x00) && (bom[2] == (byte) 0xFE) && (bom[3] == (byte) 0xFF)) {
charset = "UTF-32BE";
un... | java | protected void init() throws IOException {
if (isInited) {
return;
}
byte bom[] = new byte[BOM_SIZE];
int n, unread;
n = in.read(bom, 0, bom.length);
if ((bom[0] == (byte) 0x00) && (bom[1] == (byte) 0x00) && (bom[2] == (byte) 0xFE) && (bom[3] == (byte) 0xFF)) {
charset = "UTF-32BE";
un... | [
"protected",
"void",
"init",
"(",
")",
"throws",
"IOException",
"{",
"if",
"(",
"isInited",
")",
"{",
"return",
";",
"}",
"byte",
"bom",
"[",
"]",
"=",
"new",
"byte",
"[",
"BOM_SIZE",
"]",
";",
"int",
"n",
",",
"unread",
";",
"n",
"=",
"in",
".",... | Read-ahead four bytes and check for BOM marks. <br>
Extra bytes are unread back to the stream, only BOM bytes are skipped.
@throws IOException 读取引起的异常 | [
"Read",
"-",
"ahead",
"four",
"bytes",
"and",
"check",
"for",
"BOM",
"marks",
".",
"<br",
">",
"Extra",
"bytes",
"are",
"unread",
"back",
"to",
"the",
"stream",
"only",
"BOM",
"bytes",
"are",
"skipped",
"."
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/io/BOMInputStream.java#L80-L116 | train | Initializes the object. | [
30522,
5123,
11675,
1999,
4183,
1006,
1007,
11618,
22834,
10288,
24422,
1063,
2065,
1006,
2003,
5498,
3064,
1007,
1063,
2709,
1025,
1065,
24880,
8945,
2213,
1031,
1033,
1027,
2047,
24880,
1031,
8945,
2213,
1035,
2946,
1033,
1025,
20014,
105... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-core/src/main/java/org/apache/flink/types/Either.java | Either.Left | public static <L, R> Either<L, R> Left(L value) {
return new Left<L, R>(value);
} | java | public static <L, R> Either<L, R> Left(L value) {
return new Left<L, R>(value);
} | [
"public",
"static",
"<",
"L",
",",
"R",
">",
"Either",
"<",
"L",
",",
"R",
">",
"Left",
"(",
"L",
"value",
")",
"{",
"return",
"new",
"Left",
"<",
"L",
",",
"R",
">",
"(",
"value",
")",
";",
"}"
] | Create a Left value of Either | [
"Create",
"a",
"Left",
"value",
"of",
"Either"
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-core/src/main/java/org/apache/flink/types/Either.java#L44-L46 | train | Create a Left instance. | [
30522,
2270,
10763,
1026,
1048,
1010,
1054,
1028,
2593,
1026,
1048,
1010,
1054,
1028,
2187,
1006,
1048,
3643,
1007,
1063,
2709,
2047,
2187,
1026,
1048,
1010,
1054,
1028,
1006,
3643,
1007,
1025,
1065,
102,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
spring-projects/spring-boot | spring-boot-project/spring-boot-tools/spring-boot-configuration-metadata/src/main/java/org/springframework/boot/configurationmetadata/SimpleConfigurationMetadataRepository.java | SimpleConfigurationMetadataRepository.add | public void add(ConfigurationMetadataProperty property,
ConfigurationMetadataSource source) {
if (source != null) {
putIfAbsent(source.getProperties(), property.getId(), property);
}
putIfAbsent(getGroup(source).getProperties(), property.getId(), property);
} | java | public void add(ConfigurationMetadataProperty property,
ConfigurationMetadataSource source) {
if (source != null) {
putIfAbsent(source.getProperties(), property.getId(), property);
}
putIfAbsent(getGroup(source).getProperties(), property.getId(), property);
} | [
"public",
"void",
"add",
"(",
"ConfigurationMetadataProperty",
"property",
",",
"ConfigurationMetadataSource",
"source",
")",
"{",
"if",
"(",
"source",
"!=",
"null",
")",
"{",
"putIfAbsent",
"(",
"source",
".",
"getProperties",
"(",
")",
",",
"property",
".",
... | Add a {@link ConfigurationMetadataProperty} with the
{@link ConfigurationMetadataSource source} that defines it, if any.
@param property the property to add
@param source the source | [
"Add",
"a",
"{"
] | 0b27f7c70e164b2b1a96477f1d9c1acba56790c1 | https://github.com/spring-projects/spring-boot/blob/0b27f7c70e164b2b1a96477f1d9c1acba56790c1/spring-boot-project/spring-boot-tools/spring-boot-configuration-metadata/src/main/java/org/springframework/boot/configurationmetadata/SimpleConfigurationMetadataRepository.java#L76-L82 | train | Add a configuration property to the configuration. | [
30522,
2270,
11675,
5587,
1006,
9563,
11368,
8447,
2696,
21572,
4842,
3723,
3200,
1010,
9563,
11368,
8447,
10230,
8162,
3401,
3120,
1007,
1063,
2065,
1006,
3120,
999,
1027,
19701,
1007,
1063,
2404,
10128,
7875,
5054,
2102,
1006,
3120,
1012,... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-core/src/main/java/org/apache/flink/api/common/operators/base/MapPartitionOperatorBase.java | MapPartitionOperatorBase.executeOnCollections | @Override
protected List<OUT> executeOnCollections(List<IN> inputData, RuntimeContext ctx, ExecutionConfig executionConfig) throws Exception {
MapPartitionFunction<IN, OUT> function = this.userFunction.getUserCodeObject();
FunctionUtils.setFunctionRuntimeContext(function, ctx);
FunctionUtils.openFunction(func... | java | @Override
protected List<OUT> executeOnCollections(List<IN> inputData, RuntimeContext ctx, ExecutionConfig executionConfig) throws Exception {
MapPartitionFunction<IN, OUT> function = this.userFunction.getUserCodeObject();
FunctionUtils.setFunctionRuntimeContext(function, ctx);
FunctionUtils.openFunction(func... | [
"@",
"Override",
"protected",
"List",
"<",
"OUT",
">",
"executeOnCollections",
"(",
"List",
"<",
"IN",
">",
"inputData",
",",
"RuntimeContext",
"ctx",
",",
"ExecutionConfig",
"executionConfig",
")",
"throws",
"Exception",
"{",
"MapPartitionFunction",
"<",
"IN",
... | -------------------------------------------------------------------------------------------- | [
"--------------------------------------------------------------------------------------------"
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-core/src/main/java/org/apache/flink/api/common/operators/base/MapPartitionOperatorBase.java#L61-L81 | train | Execute on collections. | [
30522,
1030,
2058,
15637,
5123,
2862,
1026,
2041,
1028,
15389,
2239,
26895,
18491,
2015,
1006,
2862,
1026,
1999,
1028,
7953,
2850,
2696,
1010,
2448,
7292,
8663,
18209,
14931,
2595,
1010,
7781,
8663,
8873,
2290,
7781,
8663,
8873,
2290,
1007,... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-runtime/src/main/java/org/apache/flink/runtime/resourcemanager/slotmanager/SlotManager.java | SlotManager.findMatchingSlot | protected TaskManagerSlot findMatchingSlot(ResourceProfile requestResourceProfile) {
Iterator<Map.Entry<SlotID, TaskManagerSlot>> iterator = freeSlots.entrySet().iterator();
while (iterator.hasNext()) {
TaskManagerSlot taskManagerSlot = iterator.next().getValue();
// sanity check
Preconditions.checkState... | java | protected TaskManagerSlot findMatchingSlot(ResourceProfile requestResourceProfile) {
Iterator<Map.Entry<SlotID, TaskManagerSlot>> iterator = freeSlots.entrySet().iterator();
while (iterator.hasNext()) {
TaskManagerSlot taskManagerSlot = iterator.next().getValue();
// sanity check
Preconditions.checkState... | [
"protected",
"TaskManagerSlot",
"findMatchingSlot",
"(",
"ResourceProfile",
"requestResourceProfile",
")",
"{",
"Iterator",
"<",
"Map",
".",
"Entry",
"<",
"SlotID",
",",
"TaskManagerSlot",
">",
">",
"iterator",
"=",
"freeSlots",
".",
"entrySet",
"(",
")",
".",
"... | Finds a matching slot for a given resource profile. A matching slot has at least as many
resources available as the given resource profile. If there is no such slot available, then
the method returns null.
<p>Note: If you want to change the behaviour of the slot manager wrt slot allocation and
request fulfillment, the... | [
"Finds",
"a",
"matching",
"slot",
"for",
"a",
"given",
"resource",
"profile",
".",
"A",
"matching",
"slot",
"has",
"at",
"least",
"as",
"many",
"resources",
"available",
"as",
"the",
"given",
"resource",
"profile",
".",
"If",
"there",
"is",
"no",
"such",
... | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-runtime/src/main/java/org/apache/flink/runtime/resourcemanager/slotmanager/SlotManager.java#L503-L522 | train | Find a matching slot for the given resource profile. | [
30522,
5123,
4708,
24805,
15776,
10994,
2424,
18900,
8450,
14540,
4140,
1006,
7692,
21572,
8873,
2571,
5227,
6072,
8162,
3401,
21572,
8873,
2571,
1007,
1063,
2009,
6906,
4263,
1026,
4949,
1012,
4443,
1026,
10453,
3593,
1010,
4708,
24805,
15... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/io/FileTypeUtil.java | FileTypeUtil.getType | public static String getType(String fileStreamHexHead) {
for (Entry<String, String> fileTypeEntry : fileTypeMap.entrySet()) {
if(StrUtil.startWithIgnoreCase(fileStreamHexHead, fileTypeEntry.getKey())) {
return fileTypeEntry.getValue();
}
}
return null;
} | java | public static String getType(String fileStreamHexHead) {
for (Entry<String, String> fileTypeEntry : fileTypeMap.entrySet()) {
if(StrUtil.startWithIgnoreCase(fileStreamHexHead, fileTypeEntry.getKey())) {
return fileTypeEntry.getValue();
}
}
return null;
} | [
"public",
"static",
"String",
"getType",
"(",
"String",
"fileStreamHexHead",
")",
"{",
"for",
"(",
"Entry",
"<",
"String",
",",
"String",
">",
"fileTypeEntry",
":",
"fileTypeMap",
".",
"entrySet",
"(",
")",
")",
"{",
"if",
"(",
"StrUtil",
".",
"startWithIg... | 根据文件流的头部信息获得文件类型
@param fileStreamHexHead 文件流头部16进制字符串
@return 文件类型,未找到为<code>null</code> | [
"根据文件流的头部信息获得文件类型"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/io/FileTypeUtil.java#L109-L116 | train | Returns the type of the file stream hex head. | [
30522,
2270,
10763,
5164,
2131,
13874,
1006,
5164,
6764,
25379,
5369,
2595,
4974,
1007,
1063,
2005,
1006,
4443,
1026,
5164,
1010,
5164,
1028,
5371,
13874,
4765,
2854,
1024,
5371,
13874,
2863,
2361,
1012,
4443,
13462,
1006,
1007,
1007,
1063,... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/collection/CollUtil.java | CollUtil.removeNull | public static <T> Collection<T> removeNull(Collection<T> collection) {
return filter(collection, new Editor<T>() {
@Override
public T edit(T t) {
// 返回null便不加入集合
return t;
}
});
} | java | public static <T> Collection<T> removeNull(Collection<T> collection) {
return filter(collection, new Editor<T>() {
@Override
public T edit(T t) {
// 返回null便不加入集合
return t;
}
});
} | [
"public",
"static",
"<",
"T",
">",
"Collection",
"<",
"T",
">",
"removeNull",
"(",
"Collection",
"<",
"T",
">",
"collection",
")",
"{",
"return",
"filter",
"(",
"collection",
",",
"new",
"Editor",
"<",
"T",
">",
"(",
")",
"{",
"@",
"Override",
"publi... | 去除{@code null} 元素
@param collection 集合
@return 处理后的集合
@since 3.2.2 | [
"去除",
"{",
"@code",
"null",
"}",
"元素"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/collection/CollUtil.java#L1059-L1067 | train | Removes null elements from a collection. | [
30522,
2270,
10763,
1026,
1056,
1028,
3074,
1026,
1056,
1028,
6366,
11231,
3363,
1006,
3074,
1026,
1056,
1028,
3074,
1007,
1063,
2709,
11307,
1006,
3074,
1010,
2047,
3559,
1026,
1056,
1028,
1006,
1007,
1063,
1030,
2058,
15637,
2270,
1056,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/map/MapUtil.java | MapUtil.newHashMap | public static <K, V> HashMap<K, V> newHashMap(int size, boolean isOrder) {
int initialCapacity = (int) (size / DEFAULT_LOAD_FACTOR);
return isOrder ? new LinkedHashMap<K, V>(initialCapacity) : new HashMap<K, V>(initialCapacity);
} | java | public static <K, V> HashMap<K, V> newHashMap(int size, boolean isOrder) {
int initialCapacity = (int) (size / DEFAULT_LOAD_FACTOR);
return isOrder ? new LinkedHashMap<K, V>(initialCapacity) : new HashMap<K, V>(initialCapacity);
} | [
"public",
"static",
"<",
"K",
",",
"V",
">",
"HashMap",
"<",
"K",
",",
"V",
">",
"newHashMap",
"(",
"int",
"size",
",",
"boolean",
"isOrder",
")",
"{",
"int",
"initialCapacity",
"=",
"(",
"int",
")",
"(",
"size",
"/",
"DEFAULT_LOAD_FACTOR",
")",
";",... | 新建一个HashMap
@param <K> Key类型
@param <V> Value类型
@param size 初始大小,由于默认负载因子0.75,传入的size会实际初始大小为size / 0.75
@param isOrder Map的Key是否有序,有序返回 {@link LinkedHashMap},否则返回 {@link HashMap}
@return HashMap对象
@since 3.0.4 | [
"新建一个HashMap"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/map/MapUtil.java#L81-L84 | train | Creates a new HashMap with the specified size. | [
30522,
2270,
10763,
1026,
1047,
1010,
1058,
1028,
23325,
2863,
2361,
1026,
1047,
1010,
1058,
1028,
2047,
14949,
22444,
2361,
1006,
20014,
2946,
1010,
22017,
20898,
11163,
26764,
1007,
1063,
20014,
3988,
17695,
6305,
3012,
1027,
1006,
20014,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-core/src/main/java/org/apache/flink/api/java/typeutils/TypeExtractor.java | TypeExtractor.isValidPojoField | private boolean isValidPojoField(Field f, Class<?> clazz, ArrayList<Type> typeHierarchy) {
if(Modifier.isPublic(f.getModifiers())) {
return true;
} else {
boolean hasGetter = false, hasSetter = false;
final String fieldNameLow = f.getName().toLowerCase().replaceAll("_", "");
Type fieldType = f.getGener... | java | private boolean isValidPojoField(Field f, Class<?> clazz, ArrayList<Type> typeHierarchy) {
if(Modifier.isPublic(f.getModifiers())) {
return true;
} else {
boolean hasGetter = false, hasSetter = false;
final String fieldNameLow = f.getName().toLowerCase().replaceAll("_", "");
Type fieldType = f.getGener... | [
"private",
"boolean",
"isValidPojoField",
"(",
"Field",
"f",
",",
"Class",
"<",
"?",
">",
"clazz",
",",
"ArrayList",
"<",
"Type",
">",
"typeHierarchy",
")",
"{",
"if",
"(",
"Modifier",
".",
"isPublic",
"(",
"f",
".",
"getModifiers",
"(",
")",
")",
")",... | Checks if the given field is a valid pojo field:
- it is public
OR
- there are getter and setter methods for the field.
@param f field to check
@param clazz class of field
@param typeHierarchy type hierarchy for materializing generic types | [
"Checks",
"if",
"the",
"given",
"field",
"is",
"a",
"valid",
"pojo",
"field",
":",
"-",
"it",
"is",
"public",
"OR",
"-",
"there",
"are",
"getter",
"and",
"setter",
"methods",
"for",
"the",
"field",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-core/src/main/java/org/apache/flink/api/java/typeutils/TypeExtractor.java#L1771-L1823 | train | Checks if the given field is a POJO field. | [
30522,
2797,
22017,
20898,
2003,
10175,
3593,
6873,
5558,
3790,
1006,
2492,
1042,
1010,
2465,
1026,
1029,
1028,
18856,
10936,
2480,
1010,
9140,
9863,
1026,
2828,
1028,
2828,
4048,
6906,
29389,
1007,
1063,
2065,
1006,
16913,
18095,
1012,
200... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-runtime/src/main/java/org/apache/flink/runtime/executiongraph/ExecutionVertex.java | ExecutionVertex.connectSource | public void connectSource(int inputNumber, IntermediateResult source, JobEdge edge, int consumerNumber) {
final DistributionPattern pattern = edge.getDistributionPattern();
final IntermediateResultPartition[] sourcePartitions = source.getPartitions();
ExecutionEdge[] edges;
switch (pattern) {
case POINTWI... | java | public void connectSource(int inputNumber, IntermediateResult source, JobEdge edge, int consumerNumber) {
final DistributionPattern pattern = edge.getDistributionPattern();
final IntermediateResultPartition[] sourcePartitions = source.getPartitions();
ExecutionEdge[] edges;
switch (pattern) {
case POINTWI... | [
"public",
"void",
"connectSource",
"(",
"int",
"inputNumber",
",",
"IntermediateResult",
"source",
",",
"JobEdge",
"edge",
",",
"int",
"consumerNumber",
")",
"{",
"final",
"DistributionPattern",
"pattern",
"=",
"edge",
".",
"getDistributionPattern",
"(",
")",
";",... | -------------------------------------------------------------------------------------------- | [
"--------------------------------------------------------------------------------------------"
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-runtime/src/main/java/org/apache/flink/runtime/executiongraph/ExecutionVertex.java#L369-L398 | train | Connects the given edge to the given source. | [
30522,
2270,
11675,
8539,
8162,
3401,
1006,
20014,
7953,
19172,
5677,
1010,
7783,
6072,
11314,
3120,
1010,
3105,
24225,
3341,
1010,
20014,
7325,
19172,
5677,
1007,
1063,
2345,
4353,
4502,
12079,
2078,
5418,
1027,
3341,
1012,
2131,
10521,
18... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-streaming-java/src/main/java/org/apache/flink/streaming/api/windowing/windows/TimeWindow.java | TimeWindow.cover | public TimeWindow cover(TimeWindow other) {
return new TimeWindow(Math.min(start, other.start), Math.max(end, other.end));
} | java | public TimeWindow cover(TimeWindow other) {
return new TimeWindow(Math.min(start, other.start), Math.max(end, other.end));
} | [
"public",
"TimeWindow",
"cover",
"(",
"TimeWindow",
"other",
")",
"{",
"return",
"new",
"TimeWindow",
"(",
"Math",
".",
"min",
"(",
"start",
",",
"other",
".",
"start",
")",
",",
"Math",
".",
"max",
"(",
"end",
",",
"other",
".",
"end",
")",
")",
"... | Returns the minimal window covers both this window and the given window. | [
"Returns",
"the",
"minimal",
"window",
"covers",
"both",
"this",
"window",
"and",
"the",
"given",
"window",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/windowing/windows/TimeWindow.java#L126-L128 | train | Returns a new time window that contains the specified time window. | [
30522,
2270,
2051,
11101,
5004,
3104,
1006,
2051,
11101,
5004,
2060,
1007,
1063,
2709,
2047,
2051,
11101,
5004,
1006,
8785,
1012,
8117,
1006,
2707,
1010,
2060,
1012,
2707,
1007,
1010,
8785,
1012,
4098,
1006,
30524,
0,
0,
0,
0,
0,
0,
0... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/Graph.java | Graph.groupReduceOnEdges | public <T> DataSet<T> groupReduceOnEdges(EdgesFunction<K, EV, T> edgesFunction,
EdgeDirection direction, TypeInformation<T> typeInfo) throws IllegalArgumentException {
switch (direction) {
case IN:
return edges.map(new ProjectVertexIdMap<>(1)).name("Vertex ID")
.withForwardedFields("f1->f0")
.g... | java | public <T> DataSet<T> groupReduceOnEdges(EdgesFunction<K, EV, T> edgesFunction,
EdgeDirection direction, TypeInformation<T> typeInfo) throws IllegalArgumentException {
switch (direction) {
case IN:
return edges.map(new ProjectVertexIdMap<>(1)).name("Vertex ID")
.withForwardedFields("f1->f0")
.g... | [
"public",
"<",
"T",
">",
"DataSet",
"<",
"T",
">",
"groupReduceOnEdges",
"(",
"EdgesFunction",
"<",
"K",
",",
"EV",
",",
"T",
">",
"edgesFunction",
",",
"EdgeDirection",
"direction",
",",
"TypeInformation",
"<",
"T",
">",
"typeInfo",
")",
"throws",
"Illega... | Groups by vertex and computes a GroupReduce transformation over the edge values of each vertex.
The edgesFunction applied on the edges only has access to the vertex id (not the vertex value)
of the grouping vertex.
<p>For each vertex, the edgesFunction can iterate over all edges of this vertex
with the specified direc... | [
"Groups",
"by",
"vertex",
"and",
"computes",
"a",
"GroupReduce",
"transformation",
"over",
"the",
"edge",
"values",
"of",
"each",
"vertex",
".",
"The",
"edgesFunction",
"applied",
"on",
"the",
"edges",
"only",
"has",
"access",
"to",
"the",
"vertex",
"id",
"(... | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/Graph.java#L1103-L1124 | train | Group reduce on edges DataSet. | [
30522,
2270,
1026,
1056,
1028,
2951,
13462,
1026,
1056,
1028,
2177,
5596,
18796,
17799,
8449,
1006,
7926,
11263,
27989,
1026,
1047,
1010,
23408,
1010,
1056,
1028,
7926,
11263,
27989,
1010,
13011,
7442,
7542,
3257,
1010,
2828,
2378,
14192,
3... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | codec-http2/src/main/java/io/netty/handler/codec/http2/HpackStaticTable.java | HpackStaticTable.createMap | private static CharSequenceMap<Integer> createMap() {
int length = STATIC_TABLE.size();
@SuppressWarnings("unchecked")
CharSequenceMap<Integer> ret = new CharSequenceMap<Integer>(true,
UnsupportedValueConverter.<Integer>instance(), length);
// Iterate through the static t... | java | private static CharSequenceMap<Integer> createMap() {
int length = STATIC_TABLE.size();
@SuppressWarnings("unchecked")
CharSequenceMap<Integer> ret = new CharSequenceMap<Integer>(true,
UnsupportedValueConverter.<Integer>instance(), length);
// Iterate through the static t... | [
"private",
"static",
"CharSequenceMap",
"<",
"Integer",
">",
"createMap",
"(",
")",
"{",
"int",
"length",
"=",
"STATIC_TABLE",
".",
"size",
"(",
")",
";",
"@",
"SuppressWarnings",
"(",
"\"unchecked\"",
")",
"CharSequenceMap",
"<",
"Integer",
">",
"ret",
"=",... | create a map CharSequenceMap header name to index value to allow quick lookup | [
"create",
"a",
"map",
"CharSequenceMap",
"header",
"name",
"to",
"index",
"value",
"to",
"allow",
"quick",
"lookup"
] | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/codec-http2/src/main/java/io/netty/handler/codec/http2/HpackStaticTable.java#L170-L183 | train | Create a map of all the class names to the smallest index. | [
30522,
2797,
10763,
25869,
3366,
4226,
5897,
2863,
2361,
1026,
16109,
1028,
3443,
2863,
2361,
1006,
1007,
1063,
20014,
3091,
1027,
10763,
1035,
2795,
1012,
2946,
1006,
1007,
1025,
1030,
16081,
9028,
5582,
2015,
1006,
1000,
4895,
5403,
18141... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-table/flink-table-runtime-blink/src/main/java/org/apache/flink/table/runtime/functions/SqlDateTimeUtils.java | SqlDateTimeUtils.toTimestamp | public static Long toTimestamp(String dateStr, TimeZone tz) {
int length = dateStr.length();
String format;
if (length == 21) {
format = DEFAULT_DATETIME_FORMATS[1];
} else if (length == 22) {
format = DEFAULT_DATETIME_FORMATS[2];
} else if (length == 23) {
format = DEFAULT_DATETIME_FORMATS[3];
} e... | java | public static Long toTimestamp(String dateStr, TimeZone tz) {
int length = dateStr.length();
String format;
if (length == 21) {
format = DEFAULT_DATETIME_FORMATS[1];
} else if (length == 22) {
format = DEFAULT_DATETIME_FORMATS[2];
} else if (length == 23) {
format = DEFAULT_DATETIME_FORMATS[3];
} e... | [
"public",
"static",
"Long",
"toTimestamp",
"(",
"String",
"dateStr",
",",
"TimeZone",
"tz",
")",
"{",
"int",
"length",
"=",
"dateStr",
".",
"length",
"(",
")",
";",
"String",
"format",
";",
"if",
"(",
"length",
"==",
"21",
")",
"{",
"format",
"=",
"D... | Parse date time string to timestamp based on the given time zone and
"yyyy-MM-dd HH:mm:ss" format. Returns null if parsing failed.
@param dateStr the date time string
@param tz the time zone | [
"Parse",
"date",
"time",
"string",
"to",
"timestamp",
"based",
"on",
"the",
"given",
"time",
"zone",
"and",
"yyyy",
"-",
"MM",
"-",
"dd",
"HH",
":",
"mm",
":",
"ss",
"format",
".",
"Returns",
"null",
"if",
"parsing",
"failed",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-table/flink-table-runtime-blink/src/main/java/org/apache/flink/table/runtime/functions/SqlDateTimeUtils.java#L247-L261 | train | Converts a String to a Long. | [
30522,
2270,
10763,
2146,
2000,
7292,
9153,
8737,
1006,
5164,
5246,
16344,
1010,
2051,
15975,
1056,
2480,
1007,
1063,
20014,
3091,
1027,
5246,
16344,
1012,
3091,
1006,
1007,
1025,
5164,
4289,
1025,
2065,
1006,
3091,
1027,
1027,
2538,
1007,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
hankcs/HanLP | src/main/java/com/hankcs/hanlp/collection/trie/DoubleArrayTrie.java | DoubleArrayTrie.loadBaseAndCheck | private boolean loadBaseAndCheck(String path)
{
try
{
DataInputStream in = new DataInputStream(new BufferedInputStream(IOAdapter == null ?
new FileInputStream(path) :
IOAdapter.open(p... | java | private boolean loadBaseAndCheck(String path)
{
try
{
DataInputStream in = new DataInputStream(new BufferedInputStream(IOAdapter == null ?
new FileInputStream(path) :
IOAdapter.open(p... | [
"private",
"boolean",
"loadBaseAndCheck",
"(",
"String",
"path",
")",
"{",
"try",
"{",
"DataInputStream",
"in",
"=",
"new",
"DataInputStream",
"(",
"new",
"BufferedInputStream",
"(",
"IOAdapter",
"==",
"null",
"?",
"new",
"FileInputStream",
"(",
"path",
")",
"... | 从磁盘加载双数组
@param path
@return | [
"从磁盘加载双数组"
] | a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce | https://github.com/hankcs/HanLP/blob/a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce/src/main/java/com/hankcs/hanlp/collection/trie/DoubleArrayTrie.java#L590-L612 | train | Load base check. | [
30522,
2797,
22017,
20898,
7170,
15058,
5685,
5403,
3600,
1006,
5164,
4130,
1007,
1063,
3046,
1063,
2951,
2378,
18780,
21422,
1999,
1027,
2047,
2951,
2378,
18780,
21422,
1006,
2047,
17698,
2098,
2378,
18780,
21422,
1006,
22834,
8447,
13876,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
spring-projects/spring-boot | spring-boot-project/spring-boot/src/main/java/org/springframework/boot/web/servlet/context/ServletWebServerApplicationContext.java | ServletWebServerApplicationContext.prepareWebApplicationContext | protected void prepareWebApplicationContext(ServletContext servletContext) {
Object rootContext = servletContext.getAttribute(
WebApplicationContext.ROOT_WEB_APPLICATION_CONTEXT_ATTRIBUTE);
if (rootContext != null) {
if (rootContext == this) {
throw new IllegalStateException(
"Cannot initialize con... | java | protected void prepareWebApplicationContext(ServletContext servletContext) {
Object rootContext = servletContext.getAttribute(
WebApplicationContext.ROOT_WEB_APPLICATION_CONTEXT_ATTRIBUTE);
if (rootContext != null) {
if (rootContext == this) {
throw new IllegalStateException(
"Cannot initialize con... | [
"protected",
"void",
"prepareWebApplicationContext",
"(",
"ServletContext",
"servletContext",
")",
"{",
"Object",
"rootContext",
"=",
"servletContext",
".",
"getAttribute",
"(",
"WebApplicationContext",
".",
"ROOT_WEB_APPLICATION_CONTEXT_ATTRIBUTE",
")",
";",
"if",
"(",
"... | Prepare the {@link WebApplicationContext} with the given fully loaded
{@link ServletContext}. This method is usually called from
{@link ServletContextInitializer#onStartup(ServletContext)} and is similar to the
functionality usually provided by a {@link ContextLoaderListener}.
@param servletContext the operational serv... | [
"Prepare",
"the",
"{"
] | 0b27f7c70e164b2b1a96477f1d9c1acba56790c1 | https://github.com/spring-projects/spring-boot/blob/0b27f7c70e164b2b1a96477f1d9c1acba56790c1/spring-boot-project/spring-boot/src/main/java/org/springframework/boot/web/servlet/context/ServletWebServerApplicationContext.java#L271-L306 | train | Prepare the Spring embedded WebApplicationContext. | [
30522,
5123,
11675,
7374,
8545,
3676,
9397,
19341,
3508,
30524,
7485,
8663,
18209,
1012,
2131,
19321,
3089,
8569,
2618,
1006,
4773,
29098,
19341,
3508,
8663,
18209,
1012,
7117,
1035,
4773,
1035,
4646,
1035,
6123,
1035,
17961,
1007,
1025,
20... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-examples/flink-examples-batch/src/main/java/org/apache/flink/examples/java/clustering/KMeans.java | KMeans.getCentroidDataSet | private static DataSet<Centroid> getCentroidDataSet(ParameterTool params, ExecutionEnvironment env) {
DataSet<Centroid> centroids;
if (params.has("centroids")) {
centroids = env.readCsvFile(params.get("centroids"))
.fieldDelimiter(" ")
.pojoType(Centroid.class, "id", "x", "y");
} else {
System.out.p... | java | private static DataSet<Centroid> getCentroidDataSet(ParameterTool params, ExecutionEnvironment env) {
DataSet<Centroid> centroids;
if (params.has("centroids")) {
centroids = env.readCsvFile(params.get("centroids"))
.fieldDelimiter(" ")
.pojoType(Centroid.class, "id", "x", "y");
} else {
System.out.p... | [
"private",
"static",
"DataSet",
"<",
"Centroid",
">",
"getCentroidDataSet",
"(",
"ParameterTool",
"params",
",",
"ExecutionEnvironment",
"env",
")",
"{",
"DataSet",
"<",
"Centroid",
">",
"centroids",
";",
"if",
"(",
"params",
".",
"has",
"(",
"\"centroids\"",
... | ************************************************************************* | [
"*************************************************************************"
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-examples/flink-examples-batch/src/main/java/org/apache/flink/examples/java/clustering/KMeans.java#L125-L137 | train | Get the centroid data set from the input file. | [
30522,
2797,
10763,
2951,
13462,
1026,
18120,
3593,
1028,
2131,
13013,
22943,
2850,
18260,
2102,
1006,
16381,
3406,
4747,
11498,
5244,
1010,
7781,
2368,
21663,
2239,
3672,
4372,
2615,
1007,
1063,
2951,
13462,
1026,
18120,
3593,
1028,
18120,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-extra/src/main/java/cn/hutool/extra/template/engine/enjoy/EnjoyTemplate.java | EnjoyTemplate.wrap | public static EnjoyTemplate wrap(com.jfinal.template.Template EnjoyTemplate) {
return (null == EnjoyTemplate) ? null : new EnjoyTemplate(EnjoyTemplate);
} | java | public static EnjoyTemplate wrap(com.jfinal.template.Template EnjoyTemplate) {
return (null == EnjoyTemplate) ? null : new EnjoyTemplate(EnjoyTemplate);
} | [
"public",
"static",
"EnjoyTemplate",
"wrap",
"(",
"com",
".",
"jfinal",
".",
"template",
".",
"Template",
"EnjoyTemplate",
")",
"{",
"return",
"(",
"null",
"==",
"EnjoyTemplate",
")",
"?",
"null",
":",
"new",
"EnjoyTemplate",
"(",
"EnjoyTemplate",
")",
";",
... | 包装Enjoy模板
@param EnjoyTemplate Enjoy的模板对象 {@link com.jfinal.template.Template}
@return {@link EnjoyTemplate} | [
"包装Enjoy模板"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-extra/src/main/java/cn/hutool/extra/template/engine/enjoy/EnjoyTemplate.java#L27-L29 | train | Wraps an existing jpa. template. Template object. | [
30522,
2270,
10763,
5959,
18532,
15725,
10236,
1006,
4012,
1012,
1046,
16294,
2389,
1012,
23561,
1012,
23561,
5959,
18532,
15725,
1007,
1063,
2709,
1006,
19701,
1027,
1027,
5959,
18532,
15725,
1007,
1029,
19701,
1024,
2047,
5959,
18532,
15725... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | codec-http2/src/main/java/io/netty/handler/codec/http2/HpackHuffmanEncoder.java | HpackHuffmanEncoder.getEncodedLength | int getEncodedLength(CharSequence data) {
if (data instanceof AsciiString) {
AsciiString string = (AsciiString) data;
try {
encodedLengthProcessor.reset();
string.forEachByte(encodedLengthProcessor);
return encodedLengthProcessor.length();
... | java | int getEncodedLength(CharSequence data) {
if (data instanceof AsciiString) {
AsciiString string = (AsciiString) data;
try {
encodedLengthProcessor.reset();
string.forEachByte(encodedLengthProcessor);
return encodedLengthProcessor.length();
... | [
"int",
"getEncodedLength",
"(",
"CharSequence",
"data",
")",
"{",
"if",
"(",
"data",
"instanceof",
"AsciiString",
")",
"{",
"AsciiString",
"string",
"=",
"(",
"AsciiString",
")",
"data",
";",
"try",
"{",
"encodedLengthProcessor",
".",
"reset",
"(",
")",
";",... | Returns the number of bytes required to Huffman encode the input string literal.
@param data the string literal to be Huffman encoded
@return the number of bytes required to Huffman encode {@code data} | [
"Returns",
"the",
"number",
"of",
"bytes",
"required",
"to",
"Huffman",
"encode",
"the",
"input",
"string",
"literal",
"."
] | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/codec-http2/src/main/java/io/netty/handler/codec/http2/HpackHuffmanEncoder.java#L117-L131 | train | getEncodedLength This method is used to get the length of a char sequence. | [
30522,
20014,
2131,
2368,
16044,
10362,
3070,
2705,
1006,
25869,
3366,
4226,
5897,
2951,
1007,
1063,
2065,
1006,
2951,
6013,
11253,
2004,
6895,
2923,
4892,
1007,
1063,
2004,
6895,
2923,
4892,
5164,
1027,
1006,
2004,
6895,
2923,
4892,
1007,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
spring-projects/spring-boot | spring-boot-project/spring-boot-autoconfigure/src/main/java/org/springframework/boot/autoconfigure/domain/EntityScanner.java | EntityScanner.scan | @SafeVarargs
public final Set<Class<?>> scan(Class<? extends Annotation>... annotationTypes)
throws ClassNotFoundException {
List<String> packages = getPackages();
if (packages.isEmpty()) {
return Collections.emptySet();
}
ClassPathScanningCandidateComponentProvider scanner = new ClassPathScanningCandida... | java | @SafeVarargs
public final Set<Class<?>> scan(Class<? extends Annotation>... annotationTypes)
throws ClassNotFoundException {
List<String> packages = getPackages();
if (packages.isEmpty()) {
return Collections.emptySet();
}
ClassPathScanningCandidateComponentProvider scanner = new ClassPathScanningCandida... | [
"@",
"SafeVarargs",
"public",
"final",
"Set",
"<",
"Class",
"<",
"?",
">",
">",
"scan",
"(",
"Class",
"<",
"?",
"extends",
"Annotation",
">",
"...",
"annotationTypes",
")",
"throws",
"ClassNotFoundException",
"{",
"List",
"<",
"String",
">",
"packages",
"=... | Scan for entities with the specified annotations.
@param annotationTypes the annotation types used on the entities
@return a set of entity classes
@throws ClassNotFoundException if an entity class cannot be loaded | [
"Scan",
"for",
"entities",
"with",
"the",
"specified",
"annotations",
"."
] | 0b27f7c70e164b2b1a96477f1d9c1acba56790c1 | https://github.com/spring-projects/spring-boot/blob/0b27f7c70e164b2b1a96477f1d9c1acba56790c1/spring-boot-project/spring-boot-autoconfigure/src/main/java/org/springframework/boot/autoconfigure/domain/EntityScanner.java#L60-L85 | train | Scans the classpath for classes that are annotated with the given annotation types. | [
30522,
1030,
3647,
24516,
10623,
2015,
2270,
2345,
2275,
1026,
2465,
1026,
1029,
1028,
1028,
13594,
1006,
2465,
1026,
30524,
13923,
1006,
1007,
1025,
2065,
1006,
14555,
1012,
2003,
6633,
13876,
2100,
1006,
1007,
1007,
1063,
2709,
6407,
1012... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-table/flink-table-common/src/main/java/org/apache/flink/table/sources/tsextractors/TimestampExtractor.java | TimestampExtractor.toProperties | @Override
public Map<String, String> toProperties() {
Map<String, String> properties = new HashMap<>();
properties.put(Rowtime.ROWTIME_TIMESTAMPS_TYPE, Rowtime.ROWTIME_TIMESTAMPS_TYPE_VALUE_CUSTOM);
properties.put(Rowtime.ROWTIME_TIMESTAMPS_CLASS, this.getClass().getName());
properties.put(Rowtime.ROWTIME_TIME... | java | @Override
public Map<String, String> toProperties() {
Map<String, String> properties = new HashMap<>();
properties.put(Rowtime.ROWTIME_TIMESTAMPS_TYPE, Rowtime.ROWTIME_TIMESTAMPS_TYPE_VALUE_CUSTOM);
properties.put(Rowtime.ROWTIME_TIMESTAMPS_CLASS, this.getClass().getName());
properties.put(Rowtime.ROWTIME_TIME... | [
"@",
"Override",
"public",
"Map",
"<",
"String",
",",
"String",
">",
"toProperties",
"(",
")",
"{",
"Map",
"<",
"String",
",",
"String",
">",
"properties",
"=",
"new",
"HashMap",
"<>",
"(",
")",
";",
"properties",
".",
"put",
"(",
"Rowtime",
".",
"RO... | This method is a default implementation that uses java serialization and it is discouraged.
All implementation should provide a more specific set of properties. | [
"This",
"method",
"is",
"a",
"default",
"implementation",
"that",
"uses",
"java",
"serialization",
"and",
"it",
"is",
"discouraged",
".",
"All",
"implementation",
"should",
"provide",
"a",
"more",
"specific",
"set",
"of",
"properties",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-table/flink-table-common/src/main/java/org/apache/flink/table/sources/tsextractors/TimestampExtractor.java#L48-L55 | train | Returns a map of properties that can be used to write the rowtime. | [
30522,
1030,
2058,
15637,
2270,
4949,
1026,
5164,
1010,
5164,
1028,
2327,
18981,
8743,
3111,
1006,
1007,
1063,
4949,
1026,
5164,
1010,
5164,
1028,
5144,
1027,
2047,
23325,
2863,
2361,
1026,
1028,
1006,
1007,
1025,
5144,
1012,
2404,
1006,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/util/ArrayUtil.java | ArrayUtil.lastIndexOf | public static int lastIndexOf(short[] array, short value) {
if (null != array) {
for (int i = array.length - 1; i >= 0; i--) {
if (value == array[i]) {
return i;
}
}
}
return INDEX_NOT_FOUND;
} | java | public static int lastIndexOf(short[] array, short value) {
if (null != array) {
for (int i = array.length - 1; i >= 0; i--) {
if (value == array[i]) {
return i;
}
}
}
return INDEX_NOT_FOUND;
} | [
"public",
"static",
"int",
"lastIndexOf",
"(",
"short",
"[",
"]",
"array",
",",
"short",
"value",
")",
"{",
"if",
"(",
"null",
"!=",
"array",
")",
"{",
"for",
"(",
"int",
"i",
"=",
"array",
".",
"length",
"-",
"1",
";",
"i",
">=",
"0",
";",
"i"... | 返回数组中指定元素所在最后的位置,未找到返回{@link #INDEX_NOT_FOUND}
@param array 数组
@param value 被检查的元素
@return 数组中指定元素所在位置,未找到返回{@link #INDEX_NOT_FOUND}
@since 3.0.7 | [
"返回数组中指定元素所在最后的位置,未找到返回",
"{",
"@link",
"#INDEX_NOT_FOUND",
"}"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/util/ArrayUtil.java#L1119-L1128 | train | Returns the index of the last occurrence of the specified value in the array or INDEX_NOT_FOUND. | [
30522,
2270,
10763,
20014,
2197,
22254,
10288,
11253,
1006,
2460,
1031,
1033,
9140,
1010,
2460,
3643,
1007,
1063,
2065,
1006,
19701,
999,
1027,
9140,
1007,
1063,
2005,
1006,
20014,
1045,
1027,
9140,
1012,
3091,
1011,
1015,
1025,
1045,
1028,... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/util/PinyinUtil.java | PinyinUtil.getAllFirstLetter | public static String getAllFirstLetter(String chinese) {
if (StrUtil.isBlank(chinese)) {
return StrUtil.EMPTY;
}
int len = chinese.length();
final StrBuilder strBuilder = new StrBuilder(len);
for (int i = 0; i < len; i++) {
strBuilder.append(getFirstLetter(chinese.charAt(i)));
}
return ... | java | public static String getAllFirstLetter(String chinese) {
if (StrUtil.isBlank(chinese)) {
return StrUtil.EMPTY;
}
int len = chinese.length();
final StrBuilder strBuilder = new StrBuilder(len);
for (int i = 0; i < len; i++) {
strBuilder.append(getFirstLetter(chinese.charAt(i)));
}
return ... | [
"public",
"static",
"String",
"getAllFirstLetter",
"(",
"String",
"chinese",
")",
"{",
"if",
"(",
"StrUtil",
".",
"isBlank",
"(",
"chinese",
")",
")",
"{",
"return",
"StrUtil",
".",
"EMPTY",
";",
"}",
"int",
"len",
"=",
"chinese",
".",
"length",
"(",
"... | 获取所给中文的每个汉字首字母组成首字母字符串
@param chinese 汉字字符串
@return 首字母字符串 | [
"获取所给中文的每个汉字首字母组成首字母字符串"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/util/PinyinUtil.java#L60-L72 | train | Gets all first letter from a string. | [
30522,
2270,
10763,
5164,
2131,
8095,
8873,
12096,
27901,
2099,
1006,
5164,
2822,
1007,
1063,
2065,
1006,
2358,
22134,
4014,
1012,
2003,
28522,
8950,
1006,
2822,
1007,
1007,
1063,
2709,
2358,
22134,
4014,
1012,
4064,
1025,
1065,
20014,
1879... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java | FlinkKinesisConsumer.setPeriodicWatermarkAssigner | public void setPeriodicWatermarkAssigner(
AssignerWithPeriodicWatermarks<T> periodicWatermarkAssigner) {
this.periodicWatermarkAssigner = periodicWatermarkAssigner;
ClosureCleaner.clean(this.periodicWatermarkAssigner, true);
} | java | public void setPeriodicWatermarkAssigner(
AssignerWithPeriodicWatermarks<T> periodicWatermarkAssigner) {
this.periodicWatermarkAssigner = periodicWatermarkAssigner;
ClosureCleaner.clean(this.periodicWatermarkAssigner, true);
} | [
"public",
"void",
"setPeriodicWatermarkAssigner",
"(",
"AssignerWithPeriodicWatermarks",
"<",
"T",
">",
"periodicWatermarkAssigner",
")",
"{",
"this",
".",
"periodicWatermarkAssigner",
"=",
"periodicWatermarkAssigner",
";",
"ClosureCleaner",
".",
"clean",
"(",
"this",
"."... | Set the assigner that will extract the timestamp from {@link T} and calculate the
watermark.
@param periodicWatermarkAssigner periodic watermark assigner | [
"Set",
"the",
"assigner",
"that",
"will",
"extract",
"the",
"timestamp",
"from",
"{"
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-connectors/flink-connector-kinesis/src/main/java/org/apache/flink/streaming/connectors/kinesis/FlinkKinesisConsumer.java#L251-L255 | train | Sets the periodic watermark assigner. | [
30522,
2270,
11675,
2275,
4842,
3695,
14808,
5880,
10665,
12054,
23773,
2121,
1006,
23911,
2121,
24415,
4842,
3695,
14808,
5880,
27373,
1026,
1056,
1028,
15861,
5880,
10665,
12054,
23773,
2121,
1007,
1063,
2023,
1012,
15861,
5880,
10665,
1205... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-runtime/src/main/java/org/apache/flink/runtime/operators/hash/CompactingHashTable.java | CompactingHashTable.buildTableWithUniqueKey | public void buildTableWithUniqueKey(final MutableObjectIterator<T> input) throws IOException {
// go over the complete input and insert every element into the hash table
T value;
while (this.running && (value = input.next()) != null) {
insertOrReplaceRecord(value);
}
} | java | public void buildTableWithUniqueKey(final MutableObjectIterator<T> input) throws IOException {
// go over the complete input and insert every element into the hash table
T value;
while (this.running && (value = input.next()) != null) {
insertOrReplaceRecord(value);
}
} | [
"public",
"void",
"buildTableWithUniqueKey",
"(",
"final",
"MutableObjectIterator",
"<",
"T",
">",
"input",
")",
"throws",
"IOException",
"{",
"// go over the complete input and insert every element into the hash table",
"T",
"value",
";",
"while",
"(",
"this",
".",
"runn... | ------------------------------------------------------------------------ | [
"------------------------------------------------------------------------"
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-runtime/src/main/java/org/apache/flink/runtime/operators/hash/CompactingHashTable.java#L311-L318 | train | Build the hash table with a unique key. | [
30522,
2270,
11675,
3857,
10880,
24415,
19496,
4226,
14839,
1006,
2345,
14163,
10880,
16429,
20614,
21646,
8844,
1026,
1056,
1028,
7953,
1007,
11618,
22834,
10288,
24422,
1063,
1013,
1013,
2175,
2058,
1996,
3143,
7953,
1998,
19274,
2296,
5783... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/operations/OperationExpressionsUtils.java | OperationExpressionsUtils.extractAggregationsAndProperties | public static CategorizedExpressions extractAggregationsAndProperties(
List<Expression> expressions,
Supplier<String> uniqueAttributeGenerator) {
AggregationAndPropertiesSplitter splitter = new AggregationAndPropertiesSplitter(uniqueAttributeGenerator);
expressions.forEach(expr -> expr.accept(splitter));
L... | java | public static CategorizedExpressions extractAggregationsAndProperties(
List<Expression> expressions,
Supplier<String> uniqueAttributeGenerator) {
AggregationAndPropertiesSplitter splitter = new AggregationAndPropertiesSplitter(uniqueAttributeGenerator);
expressions.forEach(expr -> expr.accept(splitter));
L... | [
"public",
"static",
"CategorizedExpressions",
"extractAggregationsAndProperties",
"(",
"List",
"<",
"Expression",
">",
"expressions",
",",
"Supplier",
"<",
"String",
">",
"uniqueAttributeGenerator",
")",
"{",
"AggregationAndPropertiesSplitter",
"splitter",
"=",
"new",
"Ag... | Extracts and deduplicates all aggregation and window property expressions (zero, one, or more)
from the given expressions.
@param expressions a list of expressions to extract
@param uniqueAttributeGenerator a supplier that every time returns a unique attribute
@return a Tuple2, the first field contains the extracted a... | [
"Extracts",
"and",
"deduplicates",
"all",
"aggregation",
"and",
"window",
"property",
"expressions",
"(",
"zero",
"one",
"or",
"more",
")",
"from",
"the",
"given",
"expressions",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-table/flink-table-api-java/src/main/java/org/apache/flink/table/operations/OperationExpressionsUtils.java#L95-L110 | train | Extract the aggregates and properties from the given expressions. | [
30522,
2270,
10763,
20427,
10288,
20110,
8496,
14817,
8490,
17603,
12540,
8791,
18927,
18981,
30524,
19321,
3089,
8569,
2618,
6914,
6906,
4263,
1007,
1063,
28041,
5685,
21572,
4842,
7368,
13102,
15909,
3334,
3975,
3334,
1027,
2047,
28041,
568... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-runtime/src/main/java/org/apache/flink/runtime/rpc/RpcUtils.java | RpcUtils.terminateRpcEndpoint | public static void terminateRpcEndpoint(RpcEndpoint rpcEndpoint, Time timeout) throws ExecutionException, InterruptedException, TimeoutException {
rpcEndpoint.closeAsync().get(timeout.toMilliseconds(), TimeUnit.MILLISECONDS);
} | java | public static void terminateRpcEndpoint(RpcEndpoint rpcEndpoint, Time timeout) throws ExecutionException, InterruptedException, TimeoutException {
rpcEndpoint.closeAsync().get(timeout.toMilliseconds(), TimeUnit.MILLISECONDS);
} | [
"public",
"static",
"void",
"terminateRpcEndpoint",
"(",
"RpcEndpoint",
"rpcEndpoint",
",",
"Time",
"timeout",
")",
"throws",
"ExecutionException",
",",
"InterruptedException",
",",
"TimeoutException",
"{",
"rpcEndpoint",
".",
"closeAsync",
"(",
")",
".",
"get",
"("... | Shuts the given {@link RpcEndpoint} down and awaits its termination.
@param rpcEndpoint to terminate
@param timeout for this operation
@throws ExecutionException if a problem occurred
@throws InterruptedException if the operation has been interrupted
@throws TimeoutException if a timeout occurred | [
"Shuts",
"the",
"given",
"{",
"@link",
"RpcEndpoint",
"}",
"down",
"and",
"awaits",
"its",
"termination",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-runtime/src/main/java/org/apache/flink/runtime/rpc/RpcUtils.java#L76-L78 | train | Terminate the given RPC endpoint. | [
30522,
2270,
10763,
11675,
20320,
14536,
23865,
8400,
1006,
1054,
15042,
10497,
8400,
1054,
15042,
10497,
8400,
1010,
2051,
2051,
5833,
1007,
11618,
7781,
10288,
24422,
1010,
7153,
10288,
24422,
1010,
2051,
5833,
10288,
24422,
1063,
1054,
150... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
hankcs/HanLP | src/main/java/com/hankcs/hanlp/dependency/perceptron/transition/configuration/Instance.java | Instance.isNonprojective | public boolean isNonprojective()
{
for (int dep1 : goldDependencies.keySet())
{
int head1 = goldDependencies.get(dep1).headIndex;
for (int dep2 : goldDependencies.keySet())
{
int head2 = goldDependencies.get(dep2).headIndex;
if (hea... | java | public boolean isNonprojective()
{
for (int dep1 : goldDependencies.keySet())
{
int head1 = goldDependencies.get(dep1).headIndex;
for (int dep2 : goldDependencies.keySet())
{
int head2 = goldDependencies.get(dep2).headIndex;
if (hea... | [
"public",
"boolean",
"isNonprojective",
"(",
")",
"{",
"for",
"(",
"int",
"dep1",
":",
"goldDependencies",
".",
"keySet",
"(",
")",
")",
"{",
"int",
"head1",
"=",
"goldDependencies",
".",
"get",
"(",
"dep1",
")",
".",
"headIndex",
";",
"for",
"(",
"int... | Shows whether the tree to train is projective or not
@return true if the tree is non-projective | [
"Shows",
"whether",
"the",
"tree",
"to",
"train",
"is",
"projective",
"or",
"not"
] | a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce | https://github.com/hankcs/HanLP/blob/a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce/src/main/java/com/hankcs/hanlp/dependency/perceptron/transition/configuration/Instance.java#L84-L103 | train | Is the image nonprojective? | [
30522,
2270,
22017,
20898,
3475,
2239,
21572,
20614,
3512,
1006,
1007,
1063,
2005,
1006,
20014,
2139,
2361,
2487,
1024,
2751,
3207,
11837,
4181,
9243,
1012,
6309,
3388,
1006,
1007,
1007,
1063,
20014,
2132,
2487,
1027,
2751,
3207,
11837,
418... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/Graph.java | Graph.mapVertices | public <NV> Graph<K, NV, EV> mapVertices(final MapFunction<Vertex<K, VV>, NV> mapper, TypeInformation<Vertex<K, NV>> returnType) {
DataSet<Vertex<K, NV>> mappedVertices = vertices.map(
new MapFunction<Vertex<K, VV>, Vertex<K, NV>>() {
private Vertex<K, NV> output = new Vertex<>();
public Vertex<K, NV> ... | java | public <NV> Graph<K, NV, EV> mapVertices(final MapFunction<Vertex<K, VV>, NV> mapper, TypeInformation<Vertex<K, NV>> returnType) {
DataSet<Vertex<K, NV>> mappedVertices = vertices.map(
new MapFunction<Vertex<K, VV>, Vertex<K, NV>>() {
private Vertex<K, NV> output = new Vertex<>();
public Vertex<K, NV> ... | [
"public",
"<",
"NV",
">",
"Graph",
"<",
"K",
",",
"NV",
",",
"EV",
">",
"mapVertices",
"(",
"final",
"MapFunction",
"<",
"Vertex",
"<",
"K",
",",
"VV",
">",
",",
"NV",
">",
"mapper",
",",
"TypeInformation",
"<",
"Vertex",
"<",
"K",
",",
"NV",
">"... | Apply a function to the attribute of each vertex in the graph.
@param mapper the map function to apply.
@param returnType the explicit return type.
@return a new graph | [
"Apply",
"a",
"function",
"to",
"the",
"attribute",
"of",
"each",
"vertex",
"in",
"the",
"graph",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/Graph.java#L553-L569 | train | Maps the set of vertices to a graph of edges using the provided mapping function. | [
30522,
2270,
1026,
1050,
2615,
1028,
10629,
1026,
1047,
1010,
1050,
2615,
1010,
23408,
1028,
4949,
16874,
23522,
1006,
2345,
4949,
11263,
27989,
30524,
2615,
1028,
4949,
4842,
1010,
2828,
2378,
14192,
3370,
1026,
19449,
1026,
1047,
1010,
10... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | common/src/main/java/io/netty/util/AsciiString.java | AsciiString.split | public AsciiString[] split(String expr, int max) {
return toAsciiStringArray(Pattern.compile(expr).split(this, max));
} | java | public AsciiString[] split(String expr, int max) {
return toAsciiStringArray(Pattern.compile(expr).split(this, max));
} | [
"public",
"AsciiString",
"[",
"]",
"split",
"(",
"String",
"expr",
",",
"int",
"max",
")",
"{",
"return",
"toAsciiStringArray",
"(",
"Pattern",
".",
"compile",
"(",
"expr",
")",
".",
"split",
"(",
"this",
",",
"max",
")",
")",
";",
"}"
] | Splits this string using the supplied regular expression {@code expr}. The parameter {@code max} controls the
behavior how many times the pattern is applied to the string.
@param expr the regular expression used to divide the string.
@param max the number of entries in the resulting array.
@return an array of Strings ... | [
"Splits",
"this",
"string",
"using",
"the",
"supplied",
"regular",
"expression",
"{",
"@code",
"expr",
"}",
".",
"The",
"parameter",
"{",
"@code",
"max",
"}",
"controls",
"the",
"behavior",
"how",
"many",
"times",
"the",
"pattern",
"is",
"applied",
"to",
"... | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/common/src/main/java/io/netty/util/AsciiString.java#L1077-L1079 | train | Splits the given string into an array of ASCII strings. | [
30522,
2270,
2004,
6895,
2923,
4892,
1031,
1033,
3975,
1006,
5164,
4654,
18098,
1010,
20014,
4098,
1007,
1063,
2709,
2000,
3022,
6895,
2923,
4892,
2906,
9447,
1006,
5418,
1012,
4012,
22090,
1006,
4654,
18098,
1007,
1012,
3975,
1006,
2023,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CheckpointCoordinator.java | CheckpointCoordinator.receiveAcknowledgeMessage | public boolean receiveAcknowledgeMessage(AcknowledgeCheckpoint message) throws CheckpointException {
if (shutdown || message == null) {
return false;
}
if (!job.equals(message.getJob())) {
LOG.error("Received wrong AcknowledgeCheckpoint message for job {}: {}", job, message);
return false;
}
final ... | java | public boolean receiveAcknowledgeMessage(AcknowledgeCheckpoint message) throws CheckpointException {
if (shutdown || message == null) {
return false;
}
if (!job.equals(message.getJob())) {
LOG.error("Received wrong AcknowledgeCheckpoint message for job {}: {}", job, message);
return false;
}
final ... | [
"public",
"boolean",
"receiveAcknowledgeMessage",
"(",
"AcknowledgeCheckpoint",
"message",
")",
"throws",
"CheckpointException",
"{",
"if",
"(",
"shutdown",
"||",
"message",
"==",
"null",
")",
"{",
"return",
"false",
";",
"}",
"if",
"(",
"!",
"job",
".",
"equa... | Receives an AcknowledgeCheckpoint message and returns whether the
message was associated with a pending checkpoint.
@param message Checkpoint ack from the task manager
@return Flag indicating whether the ack'd checkpoint was associated
with a pending checkpoint.
@throws CheckpointException If the checkpoint cannot b... | [
"Receives",
"an",
"AcknowledgeCheckpoint",
"message",
"and",
"returns",
"whether",
"the",
"message",
"was",
"associated",
"with",
"a",
"pending",
"checkpoint",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/CheckpointCoordinator.java#L780-L862 | train | Receives an AcknowledgeCheckpoint message from the checkpoint manager. | [
30522,
2270,
22017,
20898,
4374,
8684,
19779,
3709,
3351,
7834,
3736,
3351,
1006,
13399,
5403,
3600,
8400,
4471,
1007,
11618,
26520,
10288,
24422,
1063,
2065,
1006,
3844,
7698,
1064,
1064,
4471,
1027,
1027,
19701,
1007,
1063,
2709,
6270,
10... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | codec-http/src/main/java/io/netty/handler/codec/http/multipart/HttpPostMultipartRequestDecoder.java | HttpPostMultipartRequestDecoder.readDelimiter | private static String readDelimiter(ByteBuf undecodedChunk, String delimiter) {
if (!undecodedChunk.hasArray()) {
return readDelimiterStandard(undecodedChunk, delimiter);
}
SeekAheadOptimize sao = new SeekAheadOptimize(undecodedChunk);
int readerIndex = undecodedChunk.readerI... | java | private static String readDelimiter(ByteBuf undecodedChunk, String delimiter) {
if (!undecodedChunk.hasArray()) {
return readDelimiterStandard(undecodedChunk, delimiter);
}
SeekAheadOptimize sao = new SeekAheadOptimize(undecodedChunk);
int readerIndex = undecodedChunk.readerI... | [
"private",
"static",
"String",
"readDelimiter",
"(",
"ByteBuf",
"undecodedChunk",
",",
"String",
"delimiter",
")",
"{",
"if",
"(",
"!",
"undecodedChunk",
".",
"hasArray",
"(",
")",
")",
"{",
"return",
"readDelimiterStandard",
"(",
"undecodedChunk",
",",
"delimit... | Read one line up to --delimiter or --delimiter-- and if existing the CRLF
or LF. Note that CRLF or LF are mandatory for opening delimiter
(--delimiter) but not for closing delimiter (--delimiter--) since some
clients does not include CRLF in this case.
@param delimiter
of the form --string, such that '--' is already i... | [
"Read",
"one",
"line",
"up",
"to",
"--",
"delimiter",
"or",
"--",
"delimiter",
"--",
"and",
"if",
"existing",
"the",
"CRLF",
"or",
"LF",
".",
"Note",
"that",
"CRLF",
"or",
"LF",
"are",
"mandatory",
"for",
"opening",
"delimiter",
"(",
"--",
"delimiter",
... | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/codec-http/src/main/java/io/netty/handler/codec/http/multipart/HttpPostMultipartRequestDecoder.java#L1180-L1288 | train | readDelimiter Method. | [
30522,
2797,
10763,
5164,
3191,
9247,
27605,
3334,
1006,
24880,
8569,
2546,
6151,
8586,
10244,
16409,
17157,
2243,
1010,
5164,
3972,
27605,
3334,
1007,
1063,
2065,
1006,
999,
6151,
8586,
10244,
16409,
17157,
2243,
1012,
2038,
2906,
9447,
10... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | buffer/src/main/java/io/netty/buffer/ByteBufUtil.java | ByteBufUtil.prettyHexDump | public static String prettyHexDump(ByteBuf buffer, int offset, int length) {
return HexUtil.prettyHexDump(buffer, offset, length);
} | java | public static String prettyHexDump(ByteBuf buffer, int offset, int length) {
return HexUtil.prettyHexDump(buffer, offset, length);
} | [
"public",
"static",
"String",
"prettyHexDump",
"(",
"ByteBuf",
"buffer",
",",
"int",
"offset",
",",
"int",
"length",
")",
"{",
"return",
"HexUtil",
".",
"prettyHexDump",
"(",
"buffer",
",",
"offset",
",",
"length",
")",
";",
"}"
] | Returns a multi-line hexadecimal dump of the specified {@link ByteBuf} that is easy to read by humans,
starting at the given {@code offset} using the given {@code length}. | [
"Returns",
"a",
"multi",
"-",
"line",
"hexadecimal",
"dump",
"of",
"the",
"specified",
"{"
] | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/buffer/src/main/java/io/netty/buffer/ByteBufUtil.java#L918-L920 | train | Pretty hex dump of a byte buffer. | [
30522,
2270,
10763,
5164,
3492,
5369,
2595,
8566,
8737,
1006,
24880,
8569,
2546,
17698,
1010,
20014,
16396,
1010,
20014,
3091,
1007,
1063,
2709,
2002,
2595,
21823,
2140,
1012,
3492,
5369,
2595,
8566,
8737,
1006,
17698,
1010,
16396,
1010,
30... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/spark | sql/hive-thriftserver/src/main/java/org/apache/hive/service/cli/TypeDescriptor.java | TypeDescriptor.getDecimalDigits | public Integer getDecimalDigits() {
switch (this.type) {
case BOOLEAN_TYPE:
case TINYINT_TYPE:
case SMALLINT_TYPE:
case INT_TYPE:
case BIGINT_TYPE:
return 0;
case FLOAT_TYPE:
return 7;
case DOUBLE_TYPE:
return 15;
case DECIMAL_TYPE:
return typeQualifiers.getSc... | java | public Integer getDecimalDigits() {
switch (this.type) {
case BOOLEAN_TYPE:
case TINYINT_TYPE:
case SMALLINT_TYPE:
case INT_TYPE:
case BIGINT_TYPE:
return 0;
case FLOAT_TYPE:
return 7;
case DOUBLE_TYPE:
return 15;
case DECIMAL_TYPE:
return typeQualifiers.getSc... | [
"public",
"Integer",
"getDecimalDigits",
"(",
")",
"{",
"switch",
"(",
"this",
".",
"type",
")",
"{",
"case",
"BOOLEAN_TYPE",
":",
"case",
"TINYINT_TYPE",
":",
"case",
"SMALLINT_TYPE",
":",
"case",
"INT_TYPE",
":",
"case",
"BIGINT_TYPE",
":",
"return",
"0",
... | The number of fractional digits for this type.
Null is returned for data types where this is not applicable. | [
"The",
"number",
"of",
"fractional",
"digits",
"for",
"this",
"type",
".",
"Null",
"is",
"returned",
"for",
"data",
"types",
"where",
"this",
"is",
"not",
"applicable",
"."
] | 25ee0474f47d9c30d6f553a7892d9549f91071cf | https://github.com/apache/spark/blob/25ee0474f47d9c30d6f553a7892d9549f91071cf/sql/hive-thriftserver/src/main/java/org/apache/hive/service/cli/TypeDescriptor.java#L139-L158 | train | Get the decimal digits of this type. | [
30522,
2270,
16109,
2131,
3207,
6895,
9067,
4305,
23806,
2015,
1006,
1007,
1063,
6942,
1006,
2023,
1012,
2828,
1007,
1063,
2553,
22017,
20898,
1035,
2828,
1024,
2553,
4714,
18447,
1035,
2828,
1024,
2553,
2235,
18447,
1035,
2828,
1024,
2553,... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | codec-http2/src/main/java/io/netty/handler/codec/http2/HpackDynamicTable.java | HpackDynamicTable.getEntry | public HpackHeaderField getEntry(int index) {
if (index <= 0 || index > length()) {
throw new IndexOutOfBoundsException();
}
int i = head - index;
if (i < 0) {
return hpackHeaderFields[i + hpackHeaderFields.length];
} else {
return hpackHeaderF... | java | public HpackHeaderField getEntry(int index) {
if (index <= 0 || index > length()) {
throw new IndexOutOfBoundsException();
}
int i = head - index;
if (i < 0) {
return hpackHeaderFields[i + hpackHeaderFields.length];
} else {
return hpackHeaderF... | [
"public",
"HpackHeaderField",
"getEntry",
"(",
"int",
"index",
")",
"{",
"if",
"(",
"index",
"<=",
"0",
"||",
"index",
">",
"length",
"(",
")",
")",
"{",
"throw",
"new",
"IndexOutOfBoundsException",
"(",
")",
";",
"}",
"int",
"i",
"=",
"head",
"-",
"... | Return the header field at the given index. The first and newest entry is always at index 1,
and the oldest entry is at the index length(). | [
"Return",
"the",
"header",
"field",
"at",
"the",
"given",
"index",
".",
"The",
"first",
"and",
"newest",
"entry",
"is",
"always",
"at",
"index",
"1",
"and",
"the",
"oldest",
"entry",
"is",
"at",
"the",
"index",
"length",
"()",
"."
] | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/codec-http2/src/main/java/io/netty/handler/codec/http2/HpackDynamicTable.java#L84-L94 | train | Gets the entry at the specified index in the HpackHeader. | [
30522,
2270,
6522,
8684,
4974,
2121,
3790,
2131,
4765,
2854,
1006,
20014,
5950,
1007,
1063,
2065,
1006,
5950,
1026,
1027,
1014,
1064,
1064,
5950,
1028,
3091,
1006,
1007,
1007,
1063,
5466,
2047,
5950,
5833,
11253,
15494,
3366,
2595,
24422,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/sink/SocketClientSink.java | SocketClientSink.invoke | @Override
public void invoke(IN value) throws Exception {
byte[] msg = schema.serialize(value);
try {
outputStream.write(msg);
if (autoFlush) {
outputStream.flush();
}
}
catch (IOException e) {
// if no re-tries are enable, fail immediately
if (maxNumRetries == 0) {
throw new IOExceptio... | java | @Override
public void invoke(IN value) throws Exception {
byte[] msg = schema.serialize(value);
try {
outputStream.write(msg);
if (autoFlush) {
outputStream.flush();
}
}
catch (IOException e) {
// if no re-tries are enable, fail immediately
if (maxNumRetries == 0) {
throw new IOExceptio... | [
"@",
"Override",
"public",
"void",
"invoke",
"(",
"IN",
"value",
")",
"throws",
"Exception",
"{",
"byte",
"[",
"]",
"msg",
"=",
"schema",
".",
"serialize",
"(",
"value",
")",
";",
"try",
"{",
"outputStream",
".",
"write",
"(",
"msg",
")",
";",
"if",
... | Called when new data arrives to the sink, and forwards it to Socket.
@param value The value to write to the socket. | [
"Called",
"when",
"new",
"data",
"arrives",
"to",
"the",
"sink",
"and",
"forwards",
"it",
"to",
"Socket",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/functions/sink/SocketClientSink.java#L143-L219 | train | This method is called by the client to send a message to the socket server. | [
30522,
1030,
2058,
15637,
2270,
11675,
1999,
6767,
3489,
1006,
1999,
3643,
1007,
11618,
6453,
1063,
24880,
1031,
1033,
5796,
2290,
1027,
8040,
28433,
1012,
7642,
4697,
1006,
3643,
1007,
1025,
3046,
1063,
27852,
25379,
1012,
4339,
1006,
5796... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/codec/BCD.java | BCD.bcdToStr | public static String bcdToStr(byte[] bytes) {
char temp[] = new char[bytes.length * 2], val;
for (int i = 0; i < bytes.length; i++) {
val = (char) (((bytes[i] & 0xf0) >> 4) & 0x0f);
temp[i * 2] = (char) (val > 9 ? val + 'A' - 10 : val + '0');
val = (char) (bytes[i] & 0x0f);
temp[i * 2 + 1] = (... | java | public static String bcdToStr(byte[] bytes) {
char temp[] = new char[bytes.length * 2], val;
for (int i = 0; i < bytes.length; i++) {
val = (char) (((bytes[i] & 0xf0) >> 4) & 0x0f);
temp[i * 2] = (char) (val > 9 ? val + 'A' - 10 : val + '0');
val = (char) (bytes[i] & 0x0f);
temp[i * 2 + 1] = (... | [
"public",
"static",
"String",
"bcdToStr",
"(",
"byte",
"[",
"]",
"bytes",
")",
"{",
"char",
"temp",
"[",
"]",
"=",
"new",
"char",
"[",
"bytes",
".",
"length",
"*",
"2",
"]",
",",
"val",
";",
"for",
"(",
"int",
"i",
"=",
"0",
";",
"i",
"<",
"b... | BCD转ASCII字符串
@param bytes BCD byte数组
@return ASCII字符串 | [
"BCD转ASCII字符串"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/codec/BCD.java#L84-L95 | train | Converts a byte array to a UTF - 8 string. | [
30522,
2270,
10763,
5164,
4647,
11927,
14122,
2099,
1006,
24880,
1031,
1033,
27507,
1007,
1063,
25869,
8915,
8737,
1031,
1033,
1027,
2047,
25869,
1031,
27507,
1012,
3091,
1008,
1016,
1033,
1010,
11748,
1025,
2005,
1006,
20014,
1045,
1027,
1... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-db/src/main/java/cn/hutool/db/sql/SqlExecutor.java | SqlExecutor.execute | public static boolean execute(PreparedStatement ps, Object... params) throws SQLException {
StatementUtil.fillParams(ps, params);
return ps.execute();
} | java | public static boolean execute(PreparedStatement ps, Object... params) throws SQLException {
StatementUtil.fillParams(ps, params);
return ps.execute();
} | [
"public",
"static",
"boolean",
"execute",
"(",
"PreparedStatement",
"ps",
",",
"Object",
"...",
"params",
")",
"throws",
"SQLException",
"{",
"StatementUtil",
".",
"fillParams",
"(",
"ps",
",",
"params",
")",
";",
"return",
"ps",
".",
"execute",
"(",
")",
... | 可用于执行任何SQL语句,返回一个boolean值,表明执行该SQL语句是否返回了ResultSet。<br>
如果执行后第一个结果是ResultSet,则返回true,否则返回false。<br>
此方法不会关闭PreparedStatement
@param ps PreparedStatement对象
@param params 参数
@return 如果执行后第一个结果是ResultSet,则返回true,否则返回false。
@throws SQLException SQL执行异常 | [
"可用于执行任何SQL语句,返回一个boolean值,表明执行该SQL语句是否返回了ResultSet。<br",
">",
"如果执行后第一个结果是ResultSet,则返回true,否则返回false。<br",
">",
"此方法不会关闭PreparedStatement"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-db/src/main/java/cn/hutool/db/sql/SqlExecutor.java#L296-L299 | train | Execute a prepared statement with the specified parameters. | [
30522,
2270,
10763,
22017,
20898,
15389,
1006,
4810,
9153,
18532,
4765,
8827,
1010,
4874,
1012,
1012,
1012,
11498,
5244,
1007,
11618,
29296,
10288,
24422,
1063,
4861,
21823,
2140,
1012,
6039,
28689,
5244,
1006,
8827,
1010,
11498,
5244,
1007,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/ZooKeeperCompletedCheckpointStore.java | ZooKeeperCompletedCheckpointStore.addCheckpoint | @Override
public void addCheckpoint(final CompletedCheckpoint checkpoint) throws Exception {
checkNotNull(checkpoint, "Checkpoint");
final String path = checkpointIdToPath(checkpoint.getCheckpointID());
// Now add the new one. If it fails, we don't want to loose existing data.
checkpointsInZooKeeper.addAndLo... | java | @Override
public void addCheckpoint(final CompletedCheckpoint checkpoint) throws Exception {
checkNotNull(checkpoint, "Checkpoint");
final String path = checkpointIdToPath(checkpoint.getCheckpointID());
// Now add the new one. If it fails, we don't want to loose existing data.
checkpointsInZooKeeper.addAndLo... | [
"@",
"Override",
"public",
"void",
"addCheckpoint",
"(",
"final",
"CompletedCheckpoint",
"checkpoint",
")",
"throws",
"Exception",
"{",
"checkNotNull",
"(",
"checkpoint",
",",
"\"Checkpoint\"",
")",
";",
"final",
"String",
"path",
"=",
"checkpointIdToPath",
"(",
"... | Synchronously writes the new checkpoints to ZooKeeper and asynchronously removes older ones.
@param checkpoint Completed checkpoint to add. | [
"Synchronously",
"writes",
"the",
"new",
"checkpoints",
"to",
"ZooKeeper",
"and",
"asynchronously",
"removes",
"older",
"ones",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/ZooKeeperCompletedCheckpointStore.java#L209-L227 | train | Add a checkpoint to the cluster. | [
30522,
1030,
2058,
15637,
2270,
11675,
5587,
5403,
3600,
8400,
1006,
2345,
2949,
5403,
3600,
8400,
26520,
1007,
11618,
6453,
1063,
4638,
17048,
11231,
3363,
1006,
26520,
1010,
1000,
26520,
1000,
1007,
1025,
2345,
5164,
4130,
1027,
26520,
35... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | transport/src/main/java/io/netty/channel/ChannelDuplexHandler.java | ChannelDuplexHandler.disconnect | @Skip
@Override
public void disconnect(ChannelHandlerContext ctx, ChannelPromise promise)
throws Exception {
ctx.disconnect(promise);
} | java | @Skip
@Override
public void disconnect(ChannelHandlerContext ctx, ChannelPromise promise)
throws Exception {
ctx.disconnect(promise);
} | [
"@",
"Skip",
"@",
"Override",
"public",
"void",
"disconnect",
"(",
"ChannelHandlerContext",
"ctx",
",",
"ChannelPromise",
"promise",
")",
"throws",
"Exception",
"{",
"ctx",
".",
"disconnect",
"(",
"promise",
")",
";",
"}"
] | Calls {@link ChannelHandlerContext#disconnect(ChannelPromise)} to forward
to the next {@link ChannelOutboundHandler} in the {@link ChannelPipeline}.
Sub-classes may override this method to change behavior. | [
"Calls",
"{",
"@link",
"ChannelHandlerContext#disconnect",
"(",
"ChannelPromise",
")",
"}",
"to",
"forward",
"to",
"the",
"next",
"{",
"@link",
"ChannelOutboundHandler",
"}",
"in",
"the",
"{",
"@link",
"ChannelPipeline",
"}",
"."
] | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/transport/src/main/java/io/netty/channel/ChannelDuplexHandler.java#L63-L68 | train | Disconnect from a node. | [
30522,
1030,
13558,
1030,
2058,
15637,
2270,
11675,
12532,
10087,
6593,
1006,
3149,
11774,
3917,
8663,
18209,
14931,
2595,
1010,
3149,
21572,
28732,
4872,
1007,
11618,
6453,
1063,
14931,
2595,
1012,
12532,
10087,
6593,
1006,
4872,
1007,
1025,... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | handler/src/main/java/io/netty/handler/ssl/ReferenceCountedOpenSslContext.java | ReferenceCountedOpenSslContext.sslCtxPointer | @Deprecated
public final long sslCtxPointer() {
Lock readerLock = ctxLock.readLock();
readerLock.lock();
try {
return SSLContext.getSslCtx(ctx);
} finally {
readerLock.unlock();
}
} | java | @Deprecated
public final long sslCtxPointer() {
Lock readerLock = ctxLock.readLock();
readerLock.lock();
try {
return SSLContext.getSslCtx(ctx);
} finally {
readerLock.unlock();
}
} | [
"@",
"Deprecated",
"public",
"final",
"long",
"sslCtxPointer",
"(",
")",
"{",
"Lock",
"readerLock",
"=",
"ctxLock",
".",
"readLock",
"(",
")",
";",
"readerLock",
".",
"lock",
"(",
")",
";",
"try",
"{",
"return",
"SSLContext",
".",
"getSslCtx",
"(",
"ctx"... | Returns the pointer to the {@code SSL_CTX} object for this {@link ReferenceCountedOpenSslContext}.
Be aware that it is freed as soon as the {@link #release()} method is called.
At this point {@code 0} will be returned.
@deprecated this method is considered unsafe as the returned pointer may be released later. Dont use... | [
"Returns",
"the",
"pointer",
"to",
"the",
"{",
"@code",
"SSL_CTX",
"}",
"object",
"for",
"this",
"{",
"@link",
"ReferenceCountedOpenSslContext",
"}",
".",
"Be",
"aware",
"that",
"it",
"is",
"freed",
"as",
"soon",
"as",
"the",
"{",
"@link",
"#release",
"()"... | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/handler/src/main/java/io/netty/handler/ssl/ReferenceCountedOpenSslContext.java#L513-L522 | train | Get the pointer of the SSL context. | [
30522,
1030,
2139,
28139,
12921,
2270,
2345,
2146,
7020,
15472,
2102,
2595,
8400,
2121,
30524,
1063,
8068,
7878,
1012,
19829,
1006,
1007,
1025,
1065,
1065,
102,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
hankcs/HanLP | src/main/java/com/hankcs/hanlp/mining/cluster/Cluster.java | Cluster.clear | void clear()
{
documents_.clear();
composite_.clear();
if (centroid_ != null)
centroid_.clear();
if (sectioned_clusters_ != null)
sectioned_clusters_.clear();
sectioned_gain_ = 0.0;
} | java | void clear()
{
documents_.clear();
composite_.clear();
if (centroid_ != null)
centroid_.clear();
if (sectioned_clusters_ != null)
sectioned_clusters_.clear();
sectioned_gain_ = 0.0;
} | [
"void",
"clear",
"(",
")",
"{",
"documents_",
".",
"clear",
"(",
")",
";",
"composite_",
".",
"clear",
"(",
")",
";",
"if",
"(",
"centroid_",
"!=",
"null",
")",
"centroid_",
".",
"clear",
"(",
")",
";",
"if",
"(",
"sectioned_clusters_",
"!=",
"null",... | Clear status. | [
"Clear",
"status",
"."
] | a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce | https://github.com/hankcs/HanLP/blob/a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce/src/main/java/com/hankcs/hanlp/mining/cluster/Cluster.java#L54-L63 | train | Clear all the records in the SECTYPE. | [
30522,
11675,
3154,
1006,
1007,
1063,
5491,
1035,
1012,
3154,
1006,
1007,
1025,
12490,
1035,
1012,
3154,
1006,
1007,
1025,
2065,
1006,
18120,
3593,
1035,
999,
1027,
19701,
1007,
18120,
3593,
1035,
1012,
3154,
1006,
1007,
1025,
2065,
1006,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
spring-projects/spring-boot | spring-boot-project/spring-boot/src/main/java/org/springframework/boot/BeanDefinitionLoader.java | BeanDefinitionLoader.setEnvironment | public void setEnvironment(ConfigurableEnvironment environment) {
this.annotatedReader.setEnvironment(environment);
this.xmlReader.setEnvironment(environment);
this.scanner.setEnvironment(environment);
} | java | public void setEnvironment(ConfigurableEnvironment environment) {
this.annotatedReader.setEnvironment(environment);
this.xmlReader.setEnvironment(environment);
this.scanner.setEnvironment(environment);
} | [
"public",
"void",
"setEnvironment",
"(",
"ConfigurableEnvironment",
"environment",
")",
"{",
"this",
".",
"annotatedReader",
".",
"setEnvironment",
"(",
"environment",
")",
";",
"this",
".",
"xmlReader",
".",
"setEnvironment",
"(",
"environment",
")",
";",
"this",... | Set the environment to be used by the underlying readers and scanner.
@param environment the environment | [
"Set",
"the",
"environment",
"to",
"be",
"used",
"by",
"the",
"underlying",
"readers",
"and",
"scanner",
"."
] | 0b27f7c70e164b2b1a96477f1d9c1acba56790c1 | https://github.com/spring-projects/spring-boot/blob/0b27f7c70e164b2b1a96477f1d9c1acba56790c1/spring-boot-project/spring-boot/src/main/java/org/springframework/boot/BeanDefinitionLoader.java#L115-L119 | train | Sets the environment. | [
30522,
2270,
11675,
2275,
2368,
21663,
2239,
3672,
1006,
9530,
8873,
27390,
3085,
2368,
21663,
2239,
3672,
4044,
1007,
1063,
2023,
1012,
5754,
17287,
30524,
2275,
2368,
21663,
2239,
3672,
1006,
4044,
1007,
1025,
2023,
1012,
26221,
1012,
227... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AsyncDataStream.java | AsyncDataStream.unorderedWait | public static <IN, OUT> SingleOutputStreamOperator<OUT> unorderedWait(
DataStream<IN> in,
AsyncFunction<IN, OUT> func,
long timeout,
TimeUnit timeUnit,
int capacity) {
return addOperator(in, func, timeUnit.toMillis(timeout), capacity, OutputMode.UNORDERED);
} | java | public static <IN, OUT> SingleOutputStreamOperator<OUT> unorderedWait(
DataStream<IN> in,
AsyncFunction<IN, OUT> func,
long timeout,
TimeUnit timeUnit,
int capacity) {
return addOperator(in, func, timeUnit.toMillis(timeout), capacity, OutputMode.UNORDERED);
} | [
"public",
"static",
"<",
"IN",
",",
"OUT",
">",
"SingleOutputStreamOperator",
"<",
"OUT",
">",
"unorderedWait",
"(",
"DataStream",
"<",
"IN",
">",
"in",
",",
"AsyncFunction",
"<",
"IN",
",",
"OUT",
">",
"func",
",",
"long",
"timeout",
",",
"TimeUnit",
"t... | Add an AsyncWaitOperator. The order of output stream records may be reordered.
@param in Input {@link DataStream}
@param func {@link AsyncFunction}
@param timeout for the asynchronous operation to complete
@param timeUnit of the given timeout
@param capacity The max number of async i/o operation that can be triggered
... | [
"Add",
"an",
"AsyncWaitOperator",
".",
"The",
"order",
"of",
"output",
"stream",
"records",
"may",
"be",
"reordered",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/AsyncDataStream.java#L102-L109 | train | Create an unordered wait operator. | [
30522,
2270,
10763,
1026,
1999,
1010,
2041,
1028,
2309,
5833,
18780,
21422,
25918,
8844,
1026,
2041,
1028,
27776,
26764,
2098,
21547,
2102,
1006,
2951,
21422,
1026,
1999,
1028,
1999,
1010,
2004,
6038,
2278,
11263,
27989,
1026,
1999,
1010,
2... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-db/src/main/java/cn/hutool/db/ds/GlobalDSFactory.java | GlobalDSFactory.get | public static DSFactory get() {
if (null == factory) {
synchronized (lock) {
if (null == factory) {
factory = DSFactory.create(null);
}
}
}
return factory;
} | java | public static DSFactory get() {
if (null == factory) {
synchronized (lock) {
if (null == factory) {
factory = DSFactory.create(null);
}
}
}
return factory;
} | [
"public",
"static",
"DSFactory",
"get",
"(",
")",
"{",
"if",
"(",
"null",
"==",
"factory",
")",
"{",
"synchronized",
"(",
"lock",
")",
"{",
"if",
"(",
"null",
"==",
"factory",
")",
"{",
"factory",
"=",
"DSFactory",
".",
"create",
"(",
"null",
")",
... | 获取默认的数据源工厂,读取默认数据库配置文件<br>
此处使用懒加载模式,在第一次调用此方法时才创建默认数据源工厂<br>
如果想自定义全局的数据源工厂,请在第一次调用此方法前调用{@link #set(DSFactory)} 方法自行定义
@return 当前使用的数据源工厂 | [
"获取默认的数据源工厂,读取默认数据库配置文件<br",
">",
"此处使用懒加载模式,在第一次调用此方法时才创建默认数据源工厂<br",
">",
"如果想自定义全局的数据源工厂,请在第一次调用此方法前调用",
"{",
"@link",
"#set",
"(",
"DSFactory",
")",
"}",
"方法自行定义"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-db/src/main/java/cn/hutool/db/ds/GlobalDSFactory.java#L42-L51 | train | Gets the reference to the DSFactory. | [
30522,
2270,
10763,
16233,
21450,
2131,
1006,
1007,
1063,
2065,
1006,
19701,
1027,
1027,
4713,
1007,
1063,
25549,
1006,
5843,
1007,
1063,
2065,
1006,
19701,
1027,
1027,
4713,
1007,
1063,
4713,
1027,
16233,
21450,
1012,
3443,
1006,
19701,
10... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-runtime/src/main/java/org/apache/flink/runtime/rest/handler/router/Router.java | Router.aggregateRoutes | private static <T> void aggregateRoutes(
String method, Map<PathPattern, T> routes,
List<String> accMethods, List<String> accPatterns, List<String> accTargets) {
for (Map.Entry<PathPattern, T> entry : routes.entrySet()) {
accMethods.add(method);
accPatterns.add("/" + entry.getKey().pattern());
accTargets... | java | private static <T> void aggregateRoutes(
String method, Map<PathPattern, T> routes,
List<String> accMethods, List<String> accPatterns, List<String> accTargets) {
for (Map.Entry<PathPattern, T> entry : routes.entrySet()) {
accMethods.add(method);
accPatterns.add("/" + entry.getKey().pattern());
accTargets... | [
"private",
"static",
"<",
"T",
">",
"void",
"aggregateRoutes",
"(",
"String",
"method",
",",
"Map",
"<",
"PathPattern",
",",
"T",
">",
"routes",
",",
"List",
"<",
"String",
">",
"accMethods",
",",
"List",
"<",
"String",
">",
"accPatterns",
",",
"List",
... | Helper for toString. | [
"Helper",
"for",
"toString",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-runtime/src/main/java/org/apache/flink/runtime/rest/handler/router/Router.java#L106-L114 | train | Aggregate the routes. | [
30522,
2797,
10763,
1026,
1056,
1028,
11675,
9572,
22494,
4570,
1006,
5164,
4118,
1010,
4949,
1026,
4130,
4502,
12079,
2078,
1010,
1056,
1028,
5847,
1010,
2862,
1026,
5164,
1028,
16222,
11368,
6806,
5104,
1010,
2862,
1026,
5164,
1028,
16222... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-extra/src/main/java/cn/hutool/extra/qrcode/QrCodeUtil.java | QrCodeUtil.generate | public static BufferedImage generate(String content, BarcodeFormat format, int width, int height) {
return generate(content, format, new QrConfig(width, height));
} | java | public static BufferedImage generate(String content, BarcodeFormat format, int width, int height) {
return generate(content, format, new QrConfig(width, height));
} | [
"public",
"static",
"BufferedImage",
"generate",
"(",
"String",
"content",
",",
"BarcodeFormat",
"format",
",",
"int",
"width",
",",
"int",
"height",
")",
"{",
"return",
"generate",
"(",
"content",
",",
"format",
",",
"new",
"QrConfig",
"(",
"width",
",",
... | 生成二维码或条形码图片
@param content 文本内容
@param format 格式,可选二维码或者条形码
@param width 宽度
@param height 高度
@return 二维码图片(黑白) | [
"生成二维码或条形码图片"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-extra/src/main/java/cn/hutool/extra/qrcode/QrCodeUtil.java#L146-L148 | train | Generates a image of the specified size using the specified format. | [
30522,
2270,
10763,
17698,
2098,
9581,
3351,
9699,
1006,
5164,
4180,
1010,
3347,
16044,
14192,
4017,
4289,
1010,
20014,
9381,
1010,
20014,
4578,
1007,
1063,
2709,
9699,
1006,
4180,
1010,
4289,
1010,
2047,
1053,
29566,
2078,
8873,
2290,
1006... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
hankcs/HanLP | src/main/java/com/hankcs/hanlp/algorithm/ahocorasick/trie/State.java | State.addEmit | public void addEmit(String keyword)
{
if (this.emits == null)
{
this.emits = new TreeSet<String>();
}
this.emits.add(keyword);
} | java | public void addEmit(String keyword)
{
if (this.emits == null)
{
this.emits = new TreeSet<String>();
}
this.emits.add(keyword);
} | [
"public",
"void",
"addEmit",
"(",
"String",
"keyword",
")",
"{",
"if",
"(",
"this",
".",
"emits",
"==",
"null",
")",
"{",
"this",
".",
"emits",
"=",
"new",
"TreeSet",
"<",
"String",
">",
"(",
")",
";",
"}",
"this",
".",
"emits",
".",
"add",
"(",
... | 添加一个匹配到的模式串(这个状态对应着这个模式串)
@param keyword | [
"添加一个匹配到的模式串(这个状态对应着这个模式串",
")"
] | a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce | https://github.com/hankcs/HanLP/blob/a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce/src/main/java/com/hankcs/hanlp/algorithm/ahocorasick/trie/State.java#L74-L81 | train | Add an entity to the set of entity emission keywords. | [
30522,
2270,
11675,
5587,
23238,
2102,
1006,
5164,
3145,
18351,
1007,
1063,
2065,
1006,
2023,
1012,
12495,
3215,
1027,
1027,
19701,
1007,
1063,
2023,
1012,
12495,
3215,
1027,
2047,
3628,
3388,
1026,
5164,
1028,
1006,
1007,
1025,
1065,
2023,... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | buffer/src/main/java/io/netty/buffer/ByteBufUtil.java | ByteBufUtil.compare | public static int compare(ByteBuf bufferA, ByteBuf bufferB) {
final int aLen = bufferA.readableBytes();
final int bLen = bufferB.readableBytes();
final int minLength = Math.min(aLen, bLen);
final int uintCount = minLength >>> 2;
final int byteCount = minLength & 3;
int aI... | java | public static int compare(ByteBuf bufferA, ByteBuf bufferB) {
final int aLen = bufferA.readableBytes();
final int bLen = bufferB.readableBytes();
final int minLength = Math.min(aLen, bLen);
final int uintCount = minLength >>> 2;
final int byteCount = minLength & 3;
int aI... | [
"public",
"static",
"int",
"compare",
"(",
"ByteBuf",
"bufferA",
",",
"ByteBuf",
"bufferB",
")",
"{",
"final",
"int",
"aLen",
"=",
"bufferA",
".",
"readableBytes",
"(",
")",
";",
"final",
"int",
"bLen",
"=",
"bufferB",
".",
"readableBytes",
"(",
")",
";"... | Compares the two specified buffers as described in {@link ByteBuf#compareTo(ByteBuf)}.
This method is useful when implementing a new buffer type. | [
"Compares",
"the",
"two",
"specified",
"buffers",
"as",
"described",
"in",
"{"
] | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/buffer/src/main/java/io/netty/buffer/ByteBufUtil.java#L288-L325 | train | Compare two byte buffers. | [
30522,
2270,
10763,
20014,
12826,
1006,
24880,
8569,
2546,
17698,
2050,
1010,
24880,
8569,
2546,
17698,
2497,
1007,
1063,
2345,
20014,
15669,
2078,
1027,
17698,
2050,
1012,
3191,
3085,
3762,
4570,
1006,
1007,
1025,
2345,
20014,
1038,
7770,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/util/ZipUtil.java | ZipUtil.zip | public static File zip(File zipFile, String path, InputStream in) throws UtilException {
return zip(zipFile, path, in, DEFAULT_CHARSET);
} | java | public static File zip(File zipFile, String path, InputStream in) throws UtilException {
return zip(zipFile, path, in, DEFAULT_CHARSET);
} | [
"public",
"static",
"File",
"zip",
"(",
"File",
"zipFile",
",",
"String",
"path",
",",
"InputStream",
"in",
")",
"throws",
"UtilException",
"{",
"return",
"zip",
"(",
"zipFile",
",",
"path",
",",
"in",
",",
"DEFAULT_CHARSET",
")",
";",
"}"
] | 对流中的数据加入到压缩文件<br>
使用默认编码UTF-8
@param zipFile 生成的Zip文件,包括文件名。注意:zipPath不能是srcPath路径下的子文件夹
@param path 流数据在压缩文件中的路径或文件名
@param in 要压缩的源
@return 压缩文件
@throws UtilException IO异常
@since 3.0.6 | [
"对流中的数据加入到压缩文件<br",
">",
"使用默认编码UTF",
"-",
"8"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/util/ZipUtil.java#L221-L223 | train | Creates a zip file from the given input stream. | [
30522,
2270,
10763,
5371,
14101,
1006,
5371,
14101,
8873,
2571,
1010,
5164,
4130,
1010,
20407,
25379,
1999,
1007,
11618,
21183,
9463,
2595,
24422,
1063,
2709,
14101,
1006,
14101,
8873,
2571,
1010,
4130,
1010,
1999,
1010,
12398,
1035,
25869,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
hankcs/HanLP | src/main/java/com/hankcs/hanlp/seg/common/Graph.java | Graph.connect | public void connect(int from, int to, double weight)
{
edgesTo[to].add(new EdgeFrom(from, weight, vertexes[from].word + '@' + vertexes[to].word));
} | java | public void connect(int from, int to, double weight)
{
edgesTo[to].add(new EdgeFrom(from, weight, vertexes[from].word + '@' + vertexes[to].word));
} | [
"public",
"void",
"connect",
"(",
"int",
"from",
",",
"int",
"to",
",",
"double",
"weight",
")",
"{",
"edgesTo",
"[",
"to",
"]",
".",
"add",
"(",
"new",
"EdgeFrom",
"(",
"from",
",",
"weight",
",",
"vertexes",
"[",
"from",
"]",
".",
"word",
"+",
... | 连接两个节点
@param from 起点
@param to 终点
@param weight 花费 | [
"连接两个节点"
] | a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce | https://github.com/hankcs/HanLP/blob/a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce/src/main/java/com/hankcs/hanlp/seg/common/Graph.java#L54-L57 | train | Connect two words. | [
30522,
2270,
11675,
7532,
1006,
20014,
2013,
1010,
20014,
2000,
1010,
3313,
3635,
1007,
1063,
7926,
3406,
1031,
2000,
1033,
1012,
5587,
1006,
2047,
3341,
19699,
5358,
1006,
2013,
1010,
3635,
1010,
19449,
2229,
1031,
2013,
1033,
1012,
2773,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
netty/netty | codec-memcache/src/main/java/io/netty/handler/codec/memcache/binary/AbstractBinaryMemcacheEncoder.java | AbstractBinaryMemcacheEncoder.encodeKey | private static void encodeKey(ByteBuf buf, ByteBuf key) {
if (key == null || !key.isReadable()) {
return;
}
buf.writeBytes(key);
} | java | private static void encodeKey(ByteBuf buf, ByteBuf key) {
if (key == null || !key.isReadable()) {
return;
}
buf.writeBytes(key);
} | [
"private",
"static",
"void",
"encodeKey",
"(",
"ByteBuf",
"buf",
",",
"ByteBuf",
"key",
")",
"{",
"if",
"(",
"key",
"==",
"null",
"||",
"!",
"key",
".",
"isReadable",
"(",
")",
")",
"{",
"return",
";",
"}",
"buf",
".",
"writeBytes",
"(",
"key",
")"... | Encode the key.
@param buf the {@link ByteBuf} to write into.
@param key the key to encode. | [
"Encode",
"the",
"key",
"."
] | ba06eafa1c1824bd154f1a380019e7ea2edf3c4c | https://github.com/netty/netty/blob/ba06eafa1c1824bd154f1a380019e7ea2edf3c4c/codec-memcache/src/main/java/io/netty/handler/codec/memcache/binary/AbstractBinaryMemcacheEncoder.java#L68-L74 | train | Encode a single key into the buffer. | [
30522,
2797,
10763,
11675,
4372,
16044,
14839,
1006,
24880,
8569,
2546,
20934,
2546,
1010,
24880,
8569,
2546,
3145,
1007,
1063,
2065,
1006,
3145,
1027,
1027,
19701,
1064,
1064,
999,
3145,
1012,
2003,
16416,
20782,
1006,
1007,
1007,
1063,
27... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
spring-projects/spring-boot | spring-boot-project/spring-boot-actuator-autoconfigure/src/main/java/org/springframework/boot/actuate/autoconfigure/metrics/orm/jpa/HibernateMetricsAutoConfiguration.java | HibernateMetricsAutoConfiguration.getEntityManagerFactoryName | private String getEntityManagerFactoryName(String beanName) {
if (beanName.length() > ENTITY_MANAGER_FACTORY_SUFFIX.length() && StringUtils
.endsWithIgnoreCase(beanName, ENTITY_MANAGER_FACTORY_SUFFIX)) {
return beanName.substring(0,
beanName.length() - ENTITY_MANAGER_FACTORY_SUFFIX.length());
}
return... | java | private String getEntityManagerFactoryName(String beanName) {
if (beanName.length() > ENTITY_MANAGER_FACTORY_SUFFIX.length() && StringUtils
.endsWithIgnoreCase(beanName, ENTITY_MANAGER_FACTORY_SUFFIX)) {
return beanName.substring(0,
beanName.length() - ENTITY_MANAGER_FACTORY_SUFFIX.length());
}
return... | [
"private",
"String",
"getEntityManagerFactoryName",
"(",
"String",
"beanName",
")",
"{",
"if",
"(",
"beanName",
".",
"length",
"(",
")",
">",
"ENTITY_MANAGER_FACTORY_SUFFIX",
".",
"length",
"(",
")",
"&&",
"StringUtils",
".",
"endsWithIgnoreCase",
"(",
"beanName",... | Get the name of an {@link EntityManagerFactory} based on its {@code beanName}.
@param beanName the name of the {@link EntityManagerFactory} bean
@return a name for the given entity manager factory | [
"Get",
"the",
"name",
"of",
"an",
"{"
] | 0b27f7c70e164b2b1a96477f1d9c1acba56790c1 | https://github.com/spring-projects/spring-boot/blob/0b27f7c70e164b2b1a96477f1d9c1acba56790c1/spring-boot-project/spring-boot-actuator-autoconfigure/src/main/java/org/springframework/boot/actuate/autoconfigure/metrics/orm/jpa/HibernateMetricsAutoConfiguration.java#L83-L90 | train | Gets the entity manager factory name. | [
30522,
2797,
5164,
2131,
4765,
3012,
24805,
4590,
21450,
18442,
1006,
5164,
14068,
18442,
1007,
1063,
2065,
1006,
14068,
18442,
1012,
3091,
1006,
1007,
1028,
9178,
1035,
3208,
1035,
4713,
1035,
16809,
1012,
3091,
1006,
1007,
1004,
1004,
516... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-java/src/main/java/org/apache/flink/api/java/DataSet.java | DataSet.distinct | public <K> DistinctOperator<T> distinct(KeySelector<T, K> keyExtractor) {
TypeInformation<K> keyType = TypeExtractor.getKeySelectorTypes(keyExtractor, getType());
return new DistinctOperator<>(this, new Keys.SelectorFunctionKeys<>(keyExtractor, getType(), keyType), Utils.getCallLocationName());
} | java | public <K> DistinctOperator<T> distinct(KeySelector<T, K> keyExtractor) {
TypeInformation<K> keyType = TypeExtractor.getKeySelectorTypes(keyExtractor, getType());
return new DistinctOperator<>(this, new Keys.SelectorFunctionKeys<>(keyExtractor, getType(), keyType), Utils.getCallLocationName());
} | [
"public",
"<",
"K",
">",
"DistinctOperator",
"<",
"T",
">",
"distinct",
"(",
"KeySelector",
"<",
"T",
",",
"K",
">",
"keyExtractor",
")",
"{",
"TypeInformation",
"<",
"K",
">",
"keyType",
"=",
"TypeExtractor",
".",
"getKeySelectorTypes",
"(",
"keyExtractor",... | Returns a distinct set of a {@link DataSet} using a {@link KeySelector} function.
<p>The KeySelector function is called for each element of the DataSet and extracts a single key value on which the
decision is made if two items are distinct or not.
@param keyExtractor The KeySelector function which extracts the key va... | [
"Returns",
"a",
"distinct",
"set",
"of",
"a",
"{",
"@link",
"DataSet",
"}",
"using",
"a",
"{",
"@link",
"KeySelector",
"}",
"function",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-java/src/main/java/org/apache/flink/api/java/DataSet.java#L596-L599 | train | Create a distinct operator. | [
30522,
2270,
1026,
1047,
1028,
5664,
25918,
8844,
1026,
1056,
1028,
5664,
1006,
6309,
12260,
16761,
1026,
1056,
1010,
1047,
1028,
3145,
10288,
6494,
16761,
1007,
1063,
2828,
2378,
14192,
3370,
1026,
1047,
1028,
3145,
13874,
1027,
2828,
1028... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/restore/RocksDBFullRestoreOperation.java | RocksDBFullRestoreOperation.restore | @Override
public RocksDBRestoreResult restore()
throws IOException, StateMigrationException, RocksDBException {
openDB();
for (KeyedStateHandle keyedStateHandle : restoreStateHandles) {
if (keyedStateHandle != null) {
if (!(keyedStateHandle instanceof KeyGroupsStateHandle)) {
throw new IllegalStateE... | java | @Override
public RocksDBRestoreResult restore()
throws IOException, StateMigrationException, RocksDBException {
openDB();
for (KeyedStateHandle keyedStateHandle : restoreStateHandles) {
if (keyedStateHandle != null) {
if (!(keyedStateHandle instanceof KeyGroupsStateHandle)) {
throw new IllegalStateE... | [
"@",
"Override",
"public",
"RocksDBRestoreResult",
"restore",
"(",
")",
"throws",
"IOException",
",",
"StateMigrationException",
",",
"RocksDBException",
"{",
"openDB",
"(",
")",
";",
"for",
"(",
"KeyedStateHandle",
"keyedStateHandle",
":",
"restoreStateHandles",
")",... | Restores all key-groups data that is referenced by the passed state handles. | [
"Restores",
"all",
"key",
"-",
"groups",
"data",
"that",
"is",
"referenced",
"by",
"the",
"passed",
"state",
"handles",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/restore/RocksDBFullRestoreOperation.java#L128-L146 | train | Restore the RocksDB database. | [
30522,
1030,
2058,
15637,
2270,
5749,
18939,
28533,
5686,
6072,
11314,
9239,
1006,
1007,
11618,
22834,
10288,
24422,
1010,
2110,
4328,
29397,
10288,
24422,
1010,
5749,
18939,
10288,
24422,
1063,
2330,
18939,
1006,
1007,
1025,
2005,
1006,
3145... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-libraries/flink-cep/src/main/java/org/apache/flink/cep/nfa/sharedbuffer/SharedBuffer.java | SharedBuffer.flushCache | void flushCache() throws Exception {
if (!entryCache.isEmpty()) {
entries.putAll(entryCache);
entryCache.clear();
}
if (!eventsBufferCache.isEmpty()) {
eventsBuffer.putAll(eventsBufferCache);
eventsBufferCache.clear();
}
} | java | void flushCache() throws Exception {
if (!entryCache.isEmpty()) {
entries.putAll(entryCache);
entryCache.clear();
}
if (!eventsBufferCache.isEmpty()) {
eventsBuffer.putAll(eventsBufferCache);
eventsBufferCache.clear();
}
} | [
"void",
"flushCache",
"(",
")",
"throws",
"Exception",
"{",
"if",
"(",
"!",
"entryCache",
".",
"isEmpty",
"(",
")",
")",
"{",
"entries",
".",
"putAll",
"(",
"entryCache",
")",
";",
"entryCache",
".",
"clear",
"(",
")",
";",
"}",
"if",
"(",
"!",
"ev... | Flush the event and node from cache to state.
@throws Exception Thrown if the system cannot access the state. | [
"Flush",
"the",
"event",
"and",
"node",
"from",
"cache",
"to",
"state",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-libraries/flink-cep/src/main/java/org/apache/flink/cep/nfa/sharedbuffer/SharedBuffer.java#L233-L242 | train | Flushes the cache. | [
30522,
11675,
13862,
3540,
5403,
1006,
1007,
11618,
6453,
1063,
2065,
1006,
999,
4443,
3540,
5403,
1012,
2003,
6633,
13876,
2100,
1006,
1007,
1007,
1063,
10445,
1012,
2404,
8095,
1006,
4443,
3540,
5403,
1007,
1025,
4443,
3540,
5403,
1012,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
alibaba/canal | dbsync/src/main/java/com/taobao/tddl/dbsync/binlog/LogBuffer.java | LogBuffer.getUint8 | public final int getUint8(final int pos) {
if (pos >= limit || pos < 0) throw new IllegalArgumentException("limit excceed: " + pos);
return 0xff & buffer[origin + pos];
} | java | public final int getUint8(final int pos) {
if (pos >= limit || pos < 0) throw new IllegalArgumentException("limit excceed: " + pos);
return 0xff & buffer[origin + pos];
} | [
"public",
"final",
"int",
"getUint8",
"(",
"final",
"int",
"pos",
")",
"{",
"if",
"(",
"pos",
">=",
"limit",
"||",
"pos",
"<",
"0",
")",
"throw",
"new",
"IllegalArgumentException",
"(",
"\"limit excceed: \"",
"+",
"pos",
")",
";",
"return",
"0xff",
"&",
... | Return 8-bit unsigned int from buffer. | [
"Return",
"8",
"-",
"bit",
"unsigned",
"int",
"from",
"buffer",
"."
] | 8f088cddc0755f4350c5aaae95c6e4002d90a40f | https://github.com/alibaba/canal/blob/8f088cddc0755f4350c5aaae95c6e4002d90a40f/dbsync/src/main/java/com/taobao/tddl/dbsync/binlog/LogBuffer.java#L225-L229 | train | Gets an unsigned 8 - bit value from the buffer. | [
30522,
2270,
2345,
20014,
2131,
20023,
2102,
2620,
1006,
2345,
20014,
13433,
2015,
1007,
1063,
2065,
1006,
13433,
2015,
1028,
1027,
5787,
1064,
1064,
13433,
2015,
1026,
1014,
1007,
5466,
2047,
6206,
2906,
22850,
15781,
2595,
24422,
1006,
10... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-connectors/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/HadoopInputs.java | HadoopInputs.readHadoopFile | public static <K, V> org.apache.flink.api.java.hadoop.mapreduce.HadoopInputFormat<K, V> readHadoopFile(
org.apache.hadoop.mapreduce.lib.input.FileInputFormat<K, V> mapreduceInputFormat, Class<K> key, Class<V> value, String inputPath) throws IOException {
return readHadoopFile(mapreduceInputFormat, key, value, inpu... | java | public static <K, V> org.apache.flink.api.java.hadoop.mapreduce.HadoopInputFormat<K, V> readHadoopFile(
org.apache.hadoop.mapreduce.lib.input.FileInputFormat<K, V> mapreduceInputFormat, Class<K> key, Class<V> value, String inputPath) throws IOException {
return readHadoopFile(mapreduceInputFormat, key, value, inpu... | [
"public",
"static",
"<",
"K",
",",
"V",
">",
"org",
".",
"apache",
".",
"flink",
".",
"api",
".",
"java",
".",
"hadoop",
".",
"mapreduce",
".",
"HadoopInputFormat",
"<",
"K",
",",
"V",
">",
"readHadoopFile",
"(",
"org",
".",
"apache",
".",
"hadoop",
... | Creates a Flink {@link InputFormat} that wraps the given Hadoop {@link org.apache.hadoop.mapreduce.lib.input.FileInputFormat}.
@return A Flink InputFormat that wraps the Hadoop FileInputFormat. | [
"Creates",
"a",
"Flink",
"{",
"@link",
"InputFormat",
"}",
"that",
"wraps",
"the",
"given",
"Hadoop",
"{",
"@link",
"org",
".",
"apache",
".",
"hadoop",
".",
"mapreduce",
".",
"lib",
".",
"input",
".",
"FileInputFormat",
"}",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-connectors/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/HadoopInputs.java#L102-L105 | train | Read hadoop file. | [
30522,
2270,
10763,
1026,
1047,
1010,
1058,
1028,
8917,
1012,
15895,
1012,
13109,
19839,
1012,
17928,
1012,
9262,
1012,
2018,
18589,
1012,
4949,
5596,
18796,
1012,
2018,
18589,
2378,
18780,
14192,
4017,
1026,
1047,
1010,
1058,
1028,
3191,
1... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-formats/flink-avro/src/main/java/org/apache/flink/formats/avro/typeutils/AvroSchemaConverter.java | AvroSchemaConverter.convertToTypeInfo | @SuppressWarnings("unchecked")
public static <T extends SpecificRecord> TypeInformation<Row> convertToTypeInfo(Class<T> avroClass) {
Preconditions.checkNotNull(avroClass, "Avro specific record class must not be null.");
// determine schema to retrieve deterministic field order
final Schema schema = SpecificData.... | java | @SuppressWarnings("unchecked")
public static <T extends SpecificRecord> TypeInformation<Row> convertToTypeInfo(Class<T> avroClass) {
Preconditions.checkNotNull(avroClass, "Avro specific record class must not be null.");
// determine schema to retrieve deterministic field order
final Schema schema = SpecificData.... | [
"@",
"SuppressWarnings",
"(",
"\"unchecked\"",
")",
"public",
"static",
"<",
"T",
"extends",
"SpecificRecord",
">",
"TypeInformation",
"<",
"Row",
">",
"convertToTypeInfo",
"(",
"Class",
"<",
"T",
">",
"avroClass",
")",
"{",
"Preconditions",
".",
"checkNotNull",... | Converts an Avro class into a nested row structure with deterministic field order and data
types that are compatible with Flink's Table & SQL API.
@param avroClass Avro specific record that contains schema information
@return type information matching the schema | [
"Converts",
"an",
"Avro",
"class",
"into",
"a",
"nested",
"row",
"structure",
"with",
"deterministic",
"field",
"order",
"and",
"data",
"types",
"that",
"are",
"compatible",
"with",
"Flink",
"s",
"Table",
"&",
"SQL",
"API",
"."
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-formats/flink-avro/src/main/java/org/apache/flink/formats/avro/typeutils/AvroSchemaConverter.java#L58-L64 | train | Converts an Avro specific record class to a type information. | [
30522,
1030,
16081,
9028,
5582,
2015,
1006,
1000,
4895,
5403,
18141,
1000,
1007,
2270,
10763,
1026,
1056,
8908,
3563,
2890,
27108,
2094,
1028,
2828,
2378,
14192,
3370,
1026,
5216,
1028,
10463,
3406,
13874,
2378,
14876,
1006,
2465,
1026,
105... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
spring-projects/spring-boot | spring-boot-project/spring-boot-tools/spring-boot-loader-tools/src/main/java/org/springframework/boot/loader/tools/Repackager.java | Repackager.repackage | public void repackage(File destination, Libraries libraries,
LaunchScript launchScript) throws IOException {
if (destination == null || destination.isDirectory()) {
throw new IllegalArgumentException("Invalid destination");
}
if (libraries == null) {
throw new IllegalArgumentException("Libraries must not... | java | public void repackage(File destination, Libraries libraries,
LaunchScript launchScript) throws IOException {
if (destination == null || destination.isDirectory()) {
throw new IllegalArgumentException("Invalid destination");
}
if (libraries == null) {
throw new IllegalArgumentException("Libraries must not... | [
"public",
"void",
"repackage",
"(",
"File",
"destination",
",",
"Libraries",
"libraries",
",",
"LaunchScript",
"launchScript",
")",
"throws",
"IOException",
"{",
"if",
"(",
"destination",
"==",
"null",
"||",
"destination",
".",
"isDirectory",
"(",
")",
")",
"{... | Repackage to the given destination so that it can be launched using '
{@literal java -jar}'.
@param destination the destination file (may be the same as the source)
@param libraries the libraries required to run the archive
@param launchScript an optional launch script prepended to the front of the jar
@throws IOExcept... | [
"Repackage",
"to",
"the",
"given",
"destination",
"so",
"that",
"it",
"can",
"be",
"launched",
"using",
"{"
] | 0b27f7c70e164b2b1a96477f1d9c1acba56790c1 | https://github.com/spring-projects/spring-boot/blob/0b27f7c70e164b2b1a96477f1d9c1acba56790c1/spring-boot-project/spring-boot-tools/spring-boot-loader-tools/src/main/java/org/springframework/boot/loader/tools/Repackager.java#L171-L203 | train | Repackage the source file to the destination file. | [
30522,
2270,
11675,
16360,
8684,
4270,
1006,
5371,
7688,
1010,
8860,
8860,
1010,
4888,
22483,
4888,
22483,
1007,
11618,
22834,
10288,
24422,
1063,
2065,
1006,
7688,
1027,
1027,
19701,
1064,
1064,
7688,
1012,
2003,
4305,
2890,
16761,
2100,
1... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-connectors/flink-connector-kafka-base/src/main/java/org/apache/flink/streaming/connectors/kafka/internals/AbstractFetcher.java | AbstractFetcher.commitInternalOffsetsToKafka | public final void commitInternalOffsetsToKafka(
Map<KafkaTopicPartition, Long> offsets,
@Nonnull KafkaCommitCallback commitCallback) throws Exception {
// Ignore sentinels. They might appear here if snapshot has started before actual offsets values
// replaced sentinels
doCommitInternalOffsetsToKafka(filter... | java | public final void commitInternalOffsetsToKafka(
Map<KafkaTopicPartition, Long> offsets,
@Nonnull KafkaCommitCallback commitCallback) throws Exception {
// Ignore sentinels. They might appear here if snapshot has started before actual offsets values
// replaced sentinels
doCommitInternalOffsetsToKafka(filter... | [
"public",
"final",
"void",
"commitInternalOffsetsToKafka",
"(",
"Map",
"<",
"KafkaTopicPartition",
",",
"Long",
">",
"offsets",
",",
"@",
"Nonnull",
"KafkaCommitCallback",
"commitCallback",
")",
"throws",
"Exception",
"{",
"// Ignore sentinels. They might appear here if sna... | Commits the given partition offsets to the Kafka brokers (or to ZooKeeper for
older Kafka versions). This method is only ever called when the offset commit mode of
the consumer is {@link OffsetCommitMode#ON_CHECKPOINTS}.
<p>The given offsets are the internal checkpointed offsets, representing
the last processed record... | [
"Commits",
"the",
"given",
"partition",
"offsets",
"to",
"the",
"Kafka",
"brokers",
"(",
"or",
"to",
"ZooKeeper",
"for",
"older",
"Kafka",
"versions",
")",
".",
"This",
"method",
"is",
"only",
"ever",
"called",
"when",
"the",
"offset",
"commit",
"mode",
"o... | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-connectors/flink-connector-kafka-base/src/main/java/org/apache/flink/streaming/connectors/kafka/internals/AbstractFetcher.java#L289-L295 | train | Commit the offsets to Kafka. | [
30522,
2270,
2345,
11675,
10797,
18447,
11795,
23067,
21807,
8454,
18715,
10354,
2912,
1006,
4949,
1026,
10556,
24316,
10610,
24330,
19362,
3775,
3508,
1010,
2146,
1028,
16396,
2015,
1010,
1030,
30524,
3711,
2182,
2065,
20057,
12326,
2038,
23... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/img/ImgUtil.java | ImgUtil.cut | public static BufferedImage cut(Image srcImage, int x, int y, int radius) {
return Img.from(srcImage).cut(x, y, radius).getImg();
} | java | public static BufferedImage cut(Image srcImage, int x, int y, int radius) {
return Img.from(srcImage).cut(x, y, radius).getImg();
} | [
"public",
"static",
"BufferedImage",
"cut",
"(",
"Image",
"srcImage",
",",
"int",
"x",
",",
"int",
"y",
",",
"int",
"radius",
")",
"{",
"return",
"Img",
".",
"from",
"(",
"srcImage",
")",
".",
"cut",
"(",
"x",
",",
"y",
",",
"radius",
")",
".",
"... | 图像切割(按指定起点坐标和宽高切割)
@param srcImage 源图像
@param x 原图的x坐标起始位置
@param y 原图的y坐标起始位置
@param radius 半径,小于0表示填充满整个图片(直径取长宽最小值)
@return {@link BufferedImage}
@since 4.1.15 | [
"图像切割",
"(",
"按指定起点坐标和宽高切割",
")"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/img/ImgUtil.java#L358-L360 | train | Cut the image at the specified coordinates. | [
30522,
2270,
10763,
17698,
2098,
9581,
3351,
3013,
1006,
3746,
5034,
6895,
26860,
1010,
20014,
1060,
1010,
20014,
1061,
1010,
20014,
12177,
1007,
1063,
2709,
10047,
2290,
1012,
2013,
1006,
5034,
6895,
26860,
1007,
1012,
3013,
1006,
1060,
10... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/spark | common/network-yarn/src/main/java/org/apache/spark/network/yarn/YarnShuffleServiceMetrics.java | YarnShuffleServiceMetrics.collectMetric | public static void collectMetric(
MetricsRecordBuilder metricsRecordBuilder, String name, Metric metric) {
if (metric instanceof Timer) {
Timer t = (Timer) metric;
metricsRecordBuilder
.addCounter(new ShuffleServiceMetricsInfo(name + "_count", "Count of timer " + name),
t.getCount... | java | public static void collectMetric(
MetricsRecordBuilder metricsRecordBuilder, String name, Metric metric) {
if (metric instanceof Timer) {
Timer t = (Timer) metric;
metricsRecordBuilder
.addCounter(new ShuffleServiceMetricsInfo(name + "_count", "Count of timer " + name),
t.getCount... | [
"public",
"static",
"void",
"collectMetric",
"(",
"MetricsRecordBuilder",
"metricsRecordBuilder",
",",
"String",
"name",
",",
"Metric",
"metric",
")",
"{",
"if",
"(",
"metric",
"instanceof",
"Timer",
")",
"{",
"Timer",
"t",
"=",
"(",
"Timer",
")",
"metric",
... | The metric types used in
{@link org.apache.spark.network.shuffle.ExternalShuffleBlockHandler.ShuffleMetrics}.
Visible for testing. | [
"The",
"metric",
"types",
"used",
"in",
"{"
] | 25ee0474f47d9c30d6f553a7892d9549f91071cf | https://github.com/apache/spark/blob/25ee0474f47d9c30d6f553a7892d9549f91071cf/common/network-yarn/src/main/java/org/apache/spark/network/yarn/YarnShuffleServiceMetrics.java#L61-L116 | train | Collect a metric. | [
30522,
2270,
10763,
11675,
8145,
12589,
1006,
12046,
21338,
8586,
8551,
8569,
23891,
2099,
12046,
21338,
8586,
8551,
8569,
23891,
2099,
1010,
5164,
2171,
1010,
12046,
12046,
1007,
1063,
2065,
1006,
12046,
6013,
11253,
25309,
1007,
30524,
2530... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/spark | sql/core/src/main/java/org/apache/spark/sql/execution/datasources/orc/OrcColumnarBatchReader.java | OrcColumnarBatchReader.initBatch | public void initBatch(
TypeDescription orcSchema,
StructField[] requiredFields,
int[] requestedDataColIds,
int[] requestedPartitionColIds,
InternalRow partitionValues) {
wrap = new VectorizedRowBatchWrap(orcSchema.createRowBatch(capacity));
assert(!wrap.batch().selectedInUse); // `... | java | public void initBatch(
TypeDescription orcSchema,
StructField[] requiredFields,
int[] requestedDataColIds,
int[] requestedPartitionColIds,
InternalRow partitionValues) {
wrap = new VectorizedRowBatchWrap(orcSchema.createRowBatch(capacity));
assert(!wrap.batch().selectedInUse); // `... | [
"public",
"void",
"initBatch",
"(",
"TypeDescription",
"orcSchema",
",",
"StructField",
"[",
"]",
"requiredFields",
",",
"int",
"[",
"]",
"requestedDataColIds",
",",
"int",
"[",
"]",
"requestedPartitionColIds",
",",
"InternalRow",
"partitionValues",
")",
"{",
"wra... | Initialize columnar batch by setting required schema and partition information.
With this information, this creates ColumnarBatch with the full schema.
@param orcSchema Schema from ORC file reader.
@param requiredFields All the fields that are required to return, including partition fields.
@param requestedDataColIds ... | [
"Initialize",
"columnar",
"batch",
"by",
"setting",
"required",
"schema",
"and",
"partition",
"information",
".",
"With",
"this",
"information",
"this",
"creates",
"ColumnarBatch",
"with",
"the",
"full",
"schema",
"."
] | 25ee0474f47d9c30d6f553a7892d9549f91071cf | https://github.com/apache/spark/blob/25ee0474f47d9c30d6f553a7892d9549f91071cf/sql/core/src/main/java/org/apache/spark/sql/execution/datasources/orc/OrcColumnarBatchReader.java#L143-L189 | train | Initialize the batch. | [
30522,
2270,
11675,
1999,
4183,
14479,
2818,
1006,
21189,
2229,
23235,
3258,
2030,
6169,
5403,
2863,
1010,
2358,
6820,
6593,
3790,
1031,
1033,
3223,
15155,
1010,
20014,
1031,
1033,
7303,
2850,
2696,
25778,
9821,
1010,
20014,
1031,
1033,
730... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-examples/flink-examples-batch/src/main/java/org/apache/flink/examples/java/graph/EnumTriangles.java | EnumTriangles.main | public static void main(String[] args) throws Exception {
// Checking input parameters
final ParameterTool params = ParameterTool.fromArgs(args);
// set up execution environment
final ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
// make parameters available in the web interfac... | java | public static void main(String[] args) throws Exception {
// Checking input parameters
final ParameterTool params = ParameterTool.fromArgs(args);
// set up execution environment
final ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
// make parameters available in the web interfac... | [
"public",
"static",
"void",
"main",
"(",
"String",
"[",
"]",
"args",
")",
"throws",
"Exception",
"{",
"// Checking input parameters",
"final",
"ParameterTool",
"params",
"=",
"ParameterTool",
".",
"fromArgs",
"(",
"args",
")",
";",
"// set up execution environment",... | ************************************************************************* | [
"*************************************************************************"
] | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-examples/flink-examples-batch/src/main/java/org/apache/flink/examples/java/graph/EnumTriangles.java#L77-L121 | train | Main method to run the Sequence Enumeration example. | [
30522,
2270,
10763,
11675,
2364,
1006,
5164,
1031,
1033,
12098,
5620,
1007,
11618,
6453,
1063,
1013,
1013,
9361,
7953,
11709,
2345,
16381,
3406,
4747,
11498,
5244,
1027,
16381,
3406,
4747,
1012,
2013,
2906,
5620,
1006,
12098,
5620,
1007,
10... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-db/src/main/java/cn/hutool/db/Session.java | Session.quietRollback | public void quietRollback() {
try {
getConnection().rollback();
} catch (Exception e) {
log.error(e);
} finally {
try {
getConnection().setAutoCommit(true); // 事务结束,恢复自动提交
} catch (SQLException e) {
log.error(e);
}
}
} | java | public void quietRollback() {
try {
getConnection().rollback();
} catch (Exception e) {
log.error(e);
} finally {
try {
getConnection().setAutoCommit(true); // 事务结束,恢复自动提交
} catch (SQLException e) {
log.error(e);
}
}
} | [
"public",
"void",
"quietRollback",
"(",
")",
"{",
"try",
"{",
"getConnection",
"(",
")",
".",
"rollback",
"(",
")",
";",
"}",
"catch",
"(",
"Exception",
"e",
")",
"{",
"log",
".",
"error",
"(",
"e",
")",
";",
"}",
"finally",
"{",
"try",
"{",
"get... | 静默回滚事务<br>
回滚事务 | [
"静默回滚事务<br",
">",
"回滚事务"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-db/src/main/java/cn/hutool/db/Session.java#L159-L171 | train | Rollback the transaction. | [
30522,
2270,
11675,
4251,
28402,
5963,
1006,
1007,
1063,
3046,
1063,
2131,
8663,
2638,
7542,
1006,
1007,
1012,
4897,
5963,
1006,
1007,
1025,
1065,
4608,
1006,
6453,
1041,
1007,
1063,
8833,
1012,
7561,
1006,
1041,
1007,
1025,
1065,
2633,
1... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-core/src/main/java/cn/hutool/core/util/HexUtil.java | HexUtil.decodeHex | public static byte[] decodeHex(char[] hexData) {
int len = hexData.length;
if ((len & 0x01) != 0) {
throw new RuntimeException("Odd number of characters.");
}
byte[] out = new byte[len >> 1];
// two characters form the hex value.
for (int i = 0, j = 0; j < len; i++) {
int f = toDigit(h... | java | public static byte[] decodeHex(char[] hexData) {
int len = hexData.length;
if ((len & 0x01) != 0) {
throw new RuntimeException("Odd number of characters.");
}
byte[] out = new byte[len >> 1];
// two characters form the hex value.
for (int i = 0, j = 0; j < len; i++) {
int f = toDigit(h... | [
"public",
"static",
"byte",
"[",
"]",
"decodeHex",
"(",
"char",
"[",
"]",
"hexData",
")",
"{",
"int",
"len",
"=",
"hexData",
".",
"length",
";",
"if",
"(",
"(",
"len",
"&",
"0x01",
")",
"!=",
"0",
")",
"{",
"throw",
"new",
"RuntimeException",
"(",
... | 将十六进制字符数组转换为字节数组
@param hexData 十六进制char[]
@return byte[]
@throws RuntimeException 如果源十六进制字符数组是一个奇怪的长度,将抛出运行时异常 | [
"将十六进制字符数组转换为字节数组"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-core/src/main/java/cn/hutool/core/util/HexUtil.java#L167-L187 | train | Decodes a byte array of hex characters from the given char array. | [
30522,
2270,
10763,
24880,
1031,
1033,
21933,
25383,
10288,
1006,
25869,
1031,
1033,
2002,
2595,
2850,
2696,
1007,
1063,
20014,
18798,
1027,
2002,
2595,
2850,
2696,
1012,
3091,
1025,
2065,
1006,
1006,
18798,
1004,
1014,
2595,
24096,
1007,
9... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
looly/hutool | hutool-cron/src/main/java/cn/hutool/cron/pattern/CronPattern.java | CronPattern.isMatchDayOfMonth | private static boolean isMatchDayOfMonth(ValueMatcher matcher, int dayOfMonth, int month, boolean isLeapYear) {
return ((matcher instanceof DayOfMonthValueMatcher) //
? ((DayOfMonthValueMatcher) matcher).match(dayOfMonth, month, isLeapYear) //
: matcher.match(dayOfMonth));
} | java | private static boolean isMatchDayOfMonth(ValueMatcher matcher, int dayOfMonth, int month, boolean isLeapYear) {
return ((matcher instanceof DayOfMonthValueMatcher) //
? ((DayOfMonthValueMatcher) matcher).match(dayOfMonth, month, isLeapYear) //
: matcher.match(dayOfMonth));
} | [
"private",
"static",
"boolean",
"isMatchDayOfMonth",
"(",
"ValueMatcher",
"matcher",
",",
"int",
"dayOfMonth",
",",
"int",
"month",
",",
"boolean",
"isLeapYear",
")",
"{",
"return",
"(",
"(",
"matcher",
"instanceof",
"DayOfMonthValueMatcher",
")",
"//\r",
"?",
"... | 是否匹配日(指定月份的第几天)
@param matcher {@link ValueMatcher}
@param dayOfMonth 日
@param month 月
@param isLeapYear 是否闰年
@return 是否匹配 | [
"是否匹配日(指定月份的第几天)"
] | bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a | https://github.com/looly/hutool/blob/bbd74eda4c7e8a81fe7a991fa6c2276eec062e6a/hutool-cron/src/main/java/cn/hutool/cron/pattern/CronPattern.java#L198-L202 | train | Checks if the given matcher is a match for a day of month. | [
30522,
2797,
10763,
22017,
20898,
2003,
18900,
2818,
10259,
11253,
9629,
2232,
1006,
30524,
11253,
9629,
2232,
10175,
5657,
18900,
7474,
1007,
1013,
1013,
1029,
1006,
1006,
2154,
11253,
9629,
2232,
10175,
5657,
18900,
7474,
1007,
2674,
2121,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
hankcs/HanLP | src/main/java/com/hankcs/hanlp/algorithm/ahocorasick/interval/IntervalTree.java | IntervalTree.removeOverlaps | public List<Intervalable> removeOverlaps(List<Intervalable> intervals)
{
// 排序,按照先大小后左端点的顺序
Collections.sort(intervals, new IntervalableComparatorBySize());
Set<Intervalable> removeIntervals = new TreeSet<Intervalable>();
for (Intervalable interval : intervals)
{
... | java | public List<Intervalable> removeOverlaps(List<Intervalable> intervals)
{
// 排序,按照先大小后左端点的顺序
Collections.sort(intervals, new IntervalableComparatorBySize());
Set<Intervalable> removeIntervals = new TreeSet<Intervalable>();
for (Intervalable interval : intervals)
{
... | [
"public",
"List",
"<",
"Intervalable",
">",
"removeOverlaps",
"(",
"List",
"<",
"Intervalable",
">",
"intervals",
")",
"{",
"// 排序,按照先大小后左端点的顺序",
"Collections",
".",
"sort",
"(",
"intervals",
",",
"new",
"IntervalableComparatorBySize",
"(",
")",
")",
";",
"Set",... | 从区间列表中移除重叠的区间
@param intervals
@return | [
"从区间列表中移除重叠的区间"
] | a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce | https://github.com/hankcs/HanLP/blob/a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce/src/main/java/com/hankcs/hanlp/algorithm/ahocorasick/interval/IntervalTree.java#L34-L64 | train | Remove overlapping intervals. | [
30522,
2270,
2862,
1026,
13483,
3085,
1028,
6366,
7840,
2721,
4523,
1006,
2862,
30524,
1025,
2275,
1026,
13483,
3085,
1028,
6366,
18447,
2121,
10175,
2015,
1027,
2047,
3628,
3388,
1026,
13483,
3085,
1028,
1006,
1007,
1025,
2005,
1006,
13483... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
hankcs/HanLP | src/main/java/com/hankcs/hanlp/corpus/dictionary/TFDictionary.java | TFDictionary.saveKeyTo | public boolean saveKeyTo(String path)
{
LinkedList<String> keyList = new LinkedList<String>();
for (Map.Entry<String, TermFrequency> entry : trie.entrySet())
{
keyList.add(entry.getKey());
}
return IOUtil.saveCollectionToTxt(keyList, path);
} | java | public boolean saveKeyTo(String path)
{
LinkedList<String> keyList = new LinkedList<String>();
for (Map.Entry<String, TermFrequency> entry : trie.entrySet())
{
keyList.add(entry.getKey());
}
return IOUtil.saveCollectionToTxt(keyList, path);
} | [
"public",
"boolean",
"saveKeyTo",
"(",
"String",
"path",
")",
"{",
"LinkedList",
"<",
"String",
">",
"keyList",
"=",
"new",
"LinkedList",
"<",
"String",
">",
"(",
")",
";",
"for",
"(",
"Map",
".",
"Entry",
"<",
"String",
",",
"TermFrequency",
">",
"ent... | 仅仅将值保存到文件
@param path
@return | [
"仅仅将值保存到文件"
] | a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce | https://github.com/hankcs/HanLP/blob/a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce/src/main/java/com/hankcs/hanlp/corpus/dictionary/TFDictionary.java#L167-L175 | train | Save the key to a file. | [
30522,
2270,
22017,
20898,
3828,
14839,
3406,
1006,
5164,
4130,
1007,
1063,
5799,
9863,
1026,
5164,
1028,
3145,
9863,
1027,
2047,
5799,
9863,
1026,
5164,
1028,
1006,
1007,
1025,
2005,
1006,
4949,
1012,
4443,
1026,
5164,
1010,
2744,
19699,
... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
apache/flink | flink-java/src/main/java/org/apache/flink/api/java/ExecutionEnvironment.java | ExecutionEnvironment.createRemoteEnvironment | public static ExecutionEnvironment createRemoteEnvironment(String host, int port, String... jarFiles) {
return new RemoteEnvironment(host, port, jarFiles);
} | java | public static ExecutionEnvironment createRemoteEnvironment(String host, int port, String... jarFiles) {
return new RemoteEnvironment(host, port, jarFiles);
} | [
"public",
"static",
"ExecutionEnvironment",
"createRemoteEnvironment",
"(",
"String",
"host",
",",
"int",
"port",
",",
"String",
"...",
"jarFiles",
")",
"{",
"return",
"new",
"RemoteEnvironment",
"(",
"host",
",",
"port",
",",
"jarFiles",
")",
";",
"}"
] | Creates a {@link RemoteEnvironment}. The remote environment sends (parts of) the program
to a cluster for execution. Note that all file paths used in the program must be accessible from the
cluster. The execution will use the cluster's default parallelism, unless the parallelism is
set explicitly via {@link ExecutionEn... | [
"Creates",
"a",
"{",
"@link",
"RemoteEnvironment",
"}",
".",
"The",
"remote",
"environment",
"sends",
"(",
"parts",
"of",
")",
"the",
"program",
"to",
"a",
"cluster",
"for",
"execution",
".",
"Note",
"that",
"all",
"file",
"paths",
"used",
"in",
"the",
"... | b62db93bf63cb3bb34dd03d611a779d9e3fc61ac | https://github.com/apache/flink/blob/b62db93bf63cb3bb34dd03d611a779d9e3fc61ac/flink-java/src/main/java/org/apache/flink/api/java/ExecutionEnvironment.java#L1172-L1174 | train | Creates a remote environment. | [
30522,
2270,
10763,
7781,
2368,
21663,
2239,
3672,
3443,
28578,
12184,
2368,
21663,
2239,
3672,
1006,
5164,
3677,
1010,
20014,
3417,
1010,
5164,
1012,
1012,
1012,
15723,
8873,
4244,
1007,
1063,
2709,
2047,
6556,
2368,
21663,
2239,
3672,
100... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
hankcs/HanLP | src/main/java/com/hankcs/hanlp/collection/trie/DoubleArrayTrie.java | DoubleArrayTrie.insert | private int insert(List<Node> siblings, BitSet used)
{
if (error_ < 0)
return 0;
int begin = 0;
int pos = Math.max(siblings.get(0).code + 1, nextCheckPos) - 1;
int nonzero_num = 0;
int first = 0;
if (allocSize <= pos)
resize(pos + 1);
... | java | private int insert(List<Node> siblings, BitSet used)
{
if (error_ < 0)
return 0;
int begin = 0;
int pos = Math.max(siblings.get(0).code + 1, nextCheckPos) - 1;
int nonzero_num = 0;
int first = 0;
if (allocSize <= pos)
resize(pos + 1);
... | [
"private",
"int",
"insert",
"(",
"List",
"<",
"Node",
">",
"siblings",
",",
"BitSet",
"used",
")",
"{",
"if",
"(",
"error_",
"<",
"0",
")",
"return",
"0",
";",
"int",
"begin",
"=",
"0",
";",
"int",
"pos",
"=",
"Math",
".",
"max",
"(",
"siblings",... | 插入节点
@param siblings 等待插入的兄弟节点
@return 插入位置 | [
"插入节点"
] | a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce | https://github.com/hankcs/HanLP/blob/a538d0722ab2e4980a9dcd9ea40324fc3ddba7ce/src/main/java/com/hankcs/hanlp/collection/trie/DoubleArrayTrie.java#L161-L262 | train | inserts a new entry in the log. | [
30522,
2797,
20014,
19274,
1006,
2862,
1026,
13045,
1028,
9504,
1010,
9017,
3388,
2109,
1007,
1063,
2065,
1006,
7561,
1035,
1026,
1014,
1007,
2709,
1014,
1025,
20014,
4088,
1027,
1014,
1025,
20014,
13433,
2015,
1027,
8785,
1012,
4098,
1006,... | [
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.