input
stringlengths 28
18.7k
| output
stringlengths 39
1.69k
|
|---|---|
testSetData_forVariant_acceptsUnderscore ( ) { widget . setData ( RWT . CUSTOM_VARIANT , "Foo_Bar_23_42" ) ; "<AssertPlaceHolder>" ; } getData ( java . lang . String ) { checkDevice ( ) ; if ( key == null ) { error ( SWT . ERROR_NULL_ARGUMENT ) ; } if ( ( keys ) == null ) { return null ; } for ( int i = 0 ; i < ( keys . length ) ; i ++ ) { if ( keys [ i ] . equals ( key ) ) { return values [ i ] ; } } return null ; }
|
org . junit . Assert . assertNotNull ( widget . getData ( RWT . CUSTOM_VARIANT ) )
|
testReadinessHealthCheck ( ) { @ org . gitlab4j . api . SuppressWarnings ( "deprecation" ) org . gitlab4j . api . models . HealthCheckInfo readiness = org . gitlab4j . api . TestHealthCheckApi . gitLabApi . getHealthCheckApi ( ) . getReadiness ( org . gitlab4j . api . TestHealthCheckApi . TEST_HEALTH_CHECK_TOKEN ) ; "<AssertPlaceHolder>" ; } getReadiness ( java . lang . String ) { try { java . net . URL readinessUrl = getApiClient ( ) . getUrlWithBase ( "-" , "readiness" ) ; org . gitlab4j . api . GitLabApiForm formData = new org . gitlab4j . api . GitLabApiForm ( ) . withParam ( "token" , token , false ) ; javax . ws . rs . core . Response response = get ( Response . Status . OK , formData . asMap ( ) , readinessUrl ) ; return response . readEntity ( org . gitlab4j . api . models . HealthCheckInfo . class ) ; } catch ( java . io . IOException ioe ) { throw new org . gitlab4j . api . GitLabApiException ( ioe ) ; } }
|
org . junit . Assert . assertNotNull ( readiness )
|
testName_Returns_Name_From_Element ( ) { when ( element . getName ( ) ) . thenReturn ( "NAME" ) ; mmarquee . automation . uiautomation . IUIAutomation mocked_automation = org . mockito . Mockito . mock ( mmarquee . automation . uiautomation . IUIAutomation . class ) ; mmarquee . automation . UIAutomation instance = new mmarquee . automation . UIAutomation ( mocked_automation ) ; mmarquee . automation . controls . AutomationTreeViewItem ctrl = new mmarquee . automation . controls . AutomationTreeViewItem ( new mmarquee . automation . controls . ElementBuilder ( element ) . addPattern ( expand ) . automation ( instance ) . addPattern ( invoke ) ) ; java . lang . String name = ctrl . getName ( ) ; "<AssertPlaceHolder>" ; } getName ( ) { return this . name ; }
|
org . junit . Assert . assertTrue ( name . equals ( "NAME" ) )
|
testSetRouterId ( ) { deviceInformation . setRouterId ( org . onlab . packet . Ip4Address . valueOf ( "1.1.1.1" ) ) ; "<AssertPlaceHolder>" ; } routerId ( ) { return routerId ; }
|
org . junit . Assert . assertThat ( deviceInformation . routerId ( ) , org . hamcrest . CoreMatchers . is ( org . onlab . packet . Ip4Address . valueOf ( "1.1.1.1" ) ) )
|
testSessionListenerStartNotification ( ) { final boolean [ ] started = new boolean [ 1 ] ; org . apache . shiro . session . SessionListener listener = new org . apache . shiro . session . SessionListenerAdapter ( ) { public void onStart ( org . apache . shiro . session . Session session ) { started [ 0 ] = true ; } } ; sm . getSessionListeners ( ) . add ( listener ) ; sm . start ( null ) ; "<AssertPlaceHolder>" ; } start ( org . apache . shiro . session . mgt . SessionContext ) { return createSession ( context ) ; }
|
org . junit . Assert . assertTrue ( started [ 0 ] )
|
getName ( ) { "<AssertPlaceHolder>" ; } getName ( ) { java . lang . String name = "paramName" ; new mockit . Expectations ( ) { { part . getName ( ) ; result = name ; } } ; org . junit . Assert . assertEquals ( name , multipartFile . getName ( ) ) ; }
|
org . junit . Assert . assertEquals ( name , part . getName ( ) )
|
getBytes ( ) { org . slim3 . controller . upload . FileItem formFile = new org . slim3 . controller . upload . FileItem ( "aaa.txt" , "text/html" , new byte [ ] { 'a' } ) ; slim3 . demo . model . UploadedData data = service . upload ( formFile ) ; "<AssertPlaceHolder>" ; } getBytes ( slim3 . demo . model . UploadedData ) { if ( uploadedData == null ) { throw new java . lang . NullPointerException ( "The<sp>uploadedData<sp>parameter<sp>must<sp>not<sp>be<sp>null." ) ; } java . util . List < slim3 . demo . model . UploadedDataFragment > fragmentList = uploadedData . getFragmentListRef ( ) . getModelList ( ) ; byte [ ] [ ] bytesArray = new byte [ fragmentList . size ( ) ] [ 0 ] ; for ( int i = 0 ; i < ( fragmentList . size ( ) ) ; i ++ ) { bytesArray [ i ] = fragmentList . get ( i ) . getBytes ( ) ; } return org . slim3 . util . ByteUtil . join ( bytesArray ) ; }
|
org . junit . Assert . assertThat ( service . getBytes ( data ) , org . hamcrest . CoreMatchers . is ( new byte [ ] { 'a' } ) )
|
testDefault1b ( ) { "<AssertPlaceHolder>" ; } getDescriptor ( java . lang . String ) { return org . eclipse . kapua . broker . core . plugin . ConnectorDescriptorProviders . getInstance ( ) . getDescriptor ( connectorName ) ; }
|
org . junit . Assert . assertNotNull ( org . eclipse . kapua . broker . core . plugin . ConnectorDescriptorProviders . getDescriptor ( "foo" ) )
|
testSetting_TargetTeradataTruncateTable ( ) { java . lang . String actualCommand = com . thinkbiganalytics . kylo . nifi . teradata . tdch . core . processor . utils . TdchBuilderTest . getBaseTdchBuilder ( ) . setTargetTeradataTruncateTable ( true ) . build ( ) ; java . lang . String expectedCommand = ( ( ( ( ( com . thinkbiganalytics . kylo . nifi . teradata . tdch . core . processor . utils . TdchBuilderTest . getBaseTdchCommand1 ( ) ) + ( com . thinkbiganalytics . kylo . nifi . teradata . tdch . core . processor . utils . TdchBuilderTest . SPACE ) ) + "-Dtdch.output.teradata.truncate=true" ) + ( com . thinkbiganalytics . kylo . nifi . teradata . tdch . core . processor . utils . TdchBuilderTest . SPACE ) ) + ( com . thinkbiganalytics . kylo . nifi . teradata . tdch . core . processor . utils . TdchBuilderTest . getBaseTdchCommand2 ( ) ) ) + ( com . thinkbiganalytics . kylo . nifi . teradata . tdch . core . processor . utils . TdchBuilderTest . SPACE ) ; "<AssertPlaceHolder>" ; } getBaseTdchCommand2 ( ) { return "-jobtype<sp>\"hive\"" ; }
|
org . junit . Assert . assertEquals ( expectedCommand , actualCommand )
|
loginSuccess ( ) { org . apache . activemq . artemis . spi . core . security . jaas . Krb5LoginModule underTest = new org . apache . activemq . artemis . spi . core . security . jaas . Krb5LoginModule ( ) ; final javax . security . auth . Subject subject = new javax . security . auth . Subject ( ) ; underTest . initialize ( subject , new javax . security . auth . callback . CallbackHandler ( ) { @ org . apache . activemq . artemis . spi . core . security . jaas . Override public void handle ( javax . security . auth . callback . Callback [ ] callbacks ) throws java . io . IOException , javax . security . auth . callback . UnsupportedCallbackException { ( ( org . apache . activemq . artemis . spi . core . security . jaas . Krb5Callback ) ( callbacks [ 0 ] ) ) . setPeerPrincipal ( new org . apache . activemq . artemis . spi . core . security . jaas . UserPrincipal ( "A" ) ) ; } } , null , null ) ; "<AssertPlaceHolder>" ; } login ( ) { if ( ! ( authenticateUser ) ) { return false ; } javax . security . auth . callback . Callback [ ] callbacks = new javax . security . auth . callback . Callback [ 2 ] ; callbacks [ 0 ] = new javax . security . auth . callback . NameCallback ( "User<sp>name" ) ; callbacks [ 1 ] = new javax . security . auth . callback . PasswordCallback ( "Password" , false ) ; try { handler . handle ( callbacks ) ; } catch ( java . io . IOException | javax . security . auth . callback . UnsupportedCallbackException e ) { throw ( ( javax . security . auth . login . LoginException ) ( new javax . security . auth . login . LoginException ( ) . initCause ( e ) ) ) ; } java . lang . String password = null ; username = ( ( javax . security . auth . callback . NameCallback ) ( callbacks [ 0 ] ) ) . getName ( ) ; if ( ( username ) == null ) return false ; if ( ( ( ( javax . security . auth . callback . PasswordCallback ) ( callbacks [ 1 ] ) ) . getPassword ( ) ) != null ) password = new java . lang . String ( ( ( javax . security . auth . callback . PasswordCallback ) ( callbacks [ 1 ] ) ) . getPassword ( ) ) ; if ( ( password == null ) || ( ( password . length ( ) ) == 0 ) ) throw new javax . security . auth . login . FailedLoginException ( "Password<sp>cannot<sp>be<sp>null<sp>or<sp>empty" ) ; authenticate ( username , password ) ; userAuthenticated = true ; return true ; }
|
org . junit . Assert . assertTrue ( underTest . login ( ) )
|
testEmptyFunnel ( ) { com . yahoo . hive . udf . funnel . Fallout udf = new com . yahoo . hive . udf . funnel . Fallout ( ) ; org . apache . hadoop . hive . serde2 . objectinspector . ObjectInspector [ ] inputOiList = new org . apache . hadoop . hive . serde2 . objectinspector . ObjectInspector [ ] { org . apache . hadoop . hive . serde2 . objectinspector . ObjectInspectorFactory . getStandardListObjectInspector ( PrimitiveObjectInspectorFactory . javaLongObjectInspector ) } ; udf . initialize ( inputOiList ) ; java . util . List < java . lang . Long > inputList = java . util . Arrays . asList ( ) ; org . apache . hadoop . hive . ql . udf . generic . GenericUDF . DeferredObject obj1 = mock ( org . apache . hadoop . hive . ql . udf . generic . GenericUDF . DeferredObject . class ) ; org . apache . hadoop . hive . ql . udf . generic . GenericUDF . DeferredObject [ ] objs = new org . apache . hadoop . hive . ql . udf . generic . GenericUDF . DeferredObject [ ] { obj1 } ; when ( obj1 . get ( ) ) . thenReturn ( inputList ) ; "<AssertPlaceHolder>" ; } evaluate ( com . yahoo . hive . udf . funnel . DeferredObject [ ] ) { if ( ( args . length ) != 1 ) { return null ; } java . util . List < java . lang . Long > funnel = ( ( java . util . List < java . lang . Long > ) ( listInputObjectInspector . getList ( args [ 0 ] . get ( ) ) ) ) ; java . util . List < java . lang . Double > result = new java . util . ArrayList ( ) ; if ( ( funnel . size ( ) ) <= 0 ) { return result ; } result . add ( 1.0 ) ; for ( int i = 1 ; i < ( funnel . size ( ) ) ; i ++ ) { if ( ( ( funnel . get ( i ) ) <= 0 ) || ( ( funnel . get ( ( i - 1 ) ) ) <= 0 ) ) { result . add ( 0.0 ) ; } else { result . add ( ( ( ( double ) ( funnel . get ( i ) ) ) / ( funnel . get ( ( i - 1 ) ) ) ) ) ; } } return result ; }
|
org . junit . Assert . assertEquals ( java . util . Arrays . asList ( ) , udf . evaluate ( objs ) )
|
testIsSatisfiedBySuccess ( ) { net . java . cargotracker . domain . model . cargo . RouteSpecification routeSpecification = new net . java . cargotracker . domain . model . cargo . RouteSpecification ( net . java . cargotracker . domain . model . location . SampleLocations . HONGKONG , net . java . cargotracker . domain . model . location . SampleLocations . CHICAGO , net . java . cargotracker . application . util . DateUtil . toDate ( "2009-03-01" ) ) ; "<AssertPlaceHolder>" ; } isSatisfiedBy ( net . java . cargotracker . domain . model . cargo . Itinerary ) { return ( ( ( itinerary != null ) && ( getOrigin ( ) . sameIdentityAs ( itinerary . getInitialDepartureLocation ( ) ) ) ) && ( getDestination ( ) . sameIdentityAs ( itinerary . getFinalArrivalLocation ( ) ) ) ) && ( getArrivalDeadline ( ) . after ( itinerary . getFinalArrivalDate ( ) ) ) ; }
|
org . junit . Assert . assertTrue ( routeSpecification . isSatisfiedBy ( itinerary ) )
|
shouldCallAsConsumer ( ) { final org . apache . tinkerpop . gremlin . util . function . ScriptEngineLambda lambda = org . apache . tinkerpop . gremlin . jsr223 . ScriptEngineLambdaTest . newLambda ( "a.setData('test')" ) ; final org . apache . tinkerpop . gremlin . jsr223 . ScriptEngineLambdaTest . Junk junk = new org . apache . tinkerpop . gremlin . jsr223 . ScriptEngineLambdaTest . Junk ( ) ; lambda . accept ( junk ) ; "<AssertPlaceHolder>" ; } getData ( ) { return data ; }
|
org . junit . Assert . assertEquals ( "test" , junk . getData ( ) )
|
remove_nested_text_node ( ) { com . redhat . lightblue . util . ModifyDocTest . doc . modify ( new com . redhat . lightblue . util . Path ( "x.y" ) , com . redhat . lightblue . util . ModifyDocTest . factory . textNode ( "test" ) , true ) ; com . redhat . lightblue . util . ModifyDocTest . doc . modify ( new com . redhat . lightblue . util . Path ( "x.y" ) , null , false ) ; "<AssertPlaceHolder>" ; } get ( java . lang . String ) { return new com . redhat . lightblue . util . Error ( com . redhat . lightblue . util . Error . THREAD_CONTEXT . get ( ) , errorCode , null ) ; }
|
org . junit . Assert . assertNull ( com . redhat . lightblue . util . ModifyDocTest . doc . get ( new com . redhat . lightblue . util . Path ( "x.y" ) ) )
|
getsExportedResources ( ) { java . net . URL resource = new java . net . URL ( "file:/app.txt" ) ; when ( filter . exportsResource ( org . mule . runtime . module . artifact . api . classloader . FilteringArtifactClassLoaderTestCase . RESOURCE_NAME ) ) . thenReturn ( true ) ; when ( artifactClassLoader . findResources ( org . mule . runtime . module . artifact . api . classloader . FilteringArtifactClassLoaderTestCase . RESOURCE_NAME ) ) . thenReturn ( new org . mule . runtime . core . internal . util . EnumerationAdapter ( java . util . Collections . singleton ( resource ) ) ) ; filteringArtifactClassLoader = doCreateClassLoader ( java . util . Collections . emptyList ( ) ) ; java . util . Enumeration < java . net . URL > resources = filteringArtifactClassLoader . getResources ( org . mule . runtime . module . artifact . api . classloader . FilteringArtifactClassLoaderTestCase . RESOURCE_NAME ) ; "<AssertPlaceHolder>" ; } equalTo ( java . util . Collection ) { return new org . mule . tck . util . EnumerationMatcher ( items ) ; }
|
org . junit . Assert . assertThat ( resources , org . hamcrest . CoreMatchers . equalTo ( java . util . Collections . singletonList ( resource ) ) )
|
testBuildQuery_notPqlTable ( ) { java . lang . String expectedQuery = "WHERE<sp>a<sp>=<sp>b<sp>AND<sp>b<sp>=<sp>c<sp>ORDER<sp>BY<sp>a<sp>ASC,<sp>b<sp>DESC<sp>LIMIT<sp>200<sp>OFFSET<sp>0" ; java . lang . String query = new com . google . api . ads . admanager . lib . utils . QueryBuilder < java . lang . Object > ( ) . where ( "a<sp>=<sp>b<sp>AND<sp>b<sp>=<sp>c" ) . orderBy ( "a<sp>ASC,<sp>b<sp>DESC" ) . limit ( 200 ) . offset ( 0 ) . buildQuery ( ) ; "<AssertPlaceHolder>" ; } buildQuery ( ) { return queryBuilder . buildQuery ( ) ; }
|
org . junit . Assert . assertEquals ( expectedQuery , query )
|
shouldNotFindUserByUsernameUsingWrongOptionalSearchParameterForMockedHandler ( ) { org . openmrs . module . webservices . rest . SimpleObject response = deserialize ( handle ( newGetRequest ( getURI ( ) , new org . openmrs . module . webservices . rest . web . v1_0 . controller . openmrs1_8 . Parameter ( "username" , "james" ) ) ) ) ; java . util . List < java . lang . Object > results = org . openmrs . module . webservices . rest . test . Util . getResultsList ( response ) ; "<AssertPlaceHolder>" ; } getResultsList ( org . openmrs . module . webservices . rest . SimpleObject ) { return ( ( java . util . List < java . lang . Object > ) ( org . apache . commons . beanutils . PropertyUtils . getProperty ( result , "results" ) ) ) ; }
|
org . junit . Assert . assertEquals ( 0 , results . size ( ) )
|
All_Different_2 ( ) { java . util . Map < java . lang . Object , java . lang . Object > diff = org . javersion . core . Diff . diff ( org . javersion . core . DiffTest . map ( 1 , 1 , 2 , 2 ) , org . javersion . core . DiffTest . map ( 3 , 3 , 4 , 4 , 5 , 5 ) ) ; "<AssertPlaceHolder>" ; } map ( K , V , K , V , K , V ) { java . util . Map < K , V > map = com . google . common . collect . Maps . newLinkedHashMap ( ) ; map . put ( k1 , v1 ) ; map . put ( k2 , v2 ) ; map . put ( k3 , v3 ) ; return map ; }
|
org . junit . Assert . assertThat ( diff , org . hamcrest . Matchers . equalTo ( org . javersion . core . DiffTest . map ( 1 , null , 2 , null , 3 , 3 , 4 , 4 , 5 , 5 ) ) )
|
delete ( ) { this . save ( ) ; com . dexcoder . test . model . User u = new com . dexcoder . test . model . User ( ) ; u . setUserId ( ( - 2L ) ) ; u . setLoginName ( "selfly-2" ) ; u . setUserType ( "1" ) ; int i = jdbcDao . delete ( u ) ; "<AssertPlaceHolder>" ; } delete ( java . lang . Class ) { return new com . dexcoder . dal . build . Criteria ( clazz , new com . dexcoder . dal . build . DeleteBuilder ( clazz ) ) ; }
|
org . junit . Assert . assertEquals ( i , 1 )
|
testFoldOperationWithNonJavaSerializableType ( ) { final int numElements = 10 ; org . apache . flink . streaming . api . environment . StreamExecutionEnvironment env = org . apache . flink . streaming . api . environment . StreamExecutionEnvironment . getExecutionEnvironment ( ) ; org . apache . flink . streaming . api . datastream . DataStream < org . apache . flink . api . java . tuple . Tuple2 < java . lang . Integer , org . apache . flink . test . streaming . api . StreamingOperatorsITCase . NonSerializable > > input = env . addSource ( new org . apache . flink . test . streaming . api . StreamingOperatorsITCase . NonSerializableTupleSource ( numElements ) ) ; final org . apache . flink . test . streaming . api . StreamingOperatorsITCase . MemorySinkFunction sinkFunction = new org . apache . flink . test . streaming . api . StreamingOperatorsITCase . MemorySinkFunction ( 0 ) ; final java . util . ArrayList < java . lang . Integer > actualResult = new java . util . ArrayList ( ) ; org . apache . flink . test . streaming . api . StreamingOperatorsITCase . MemorySinkFunction . registerCollection ( 0 , actualResult ) ; input . keyBy ( 0 ) . fold ( new org . apache . flink . test . streaming . api . StreamingOperatorsITCase . NonSerializable ( 42 ) , new org . apache . flink . api . common . functions . FoldFunction < org . apache . flink . api . java . tuple . Tuple2 < java . lang . Integer , org . apache . flink . test . streaming . api . StreamingOperatorsITCase . NonSerializable > , org . apache . flink . test . streaming . api . StreamingOperatorsITCase . NonSerializable > ( ) { private static final long serialVersionUID = 2705497830143608897L ; @ org . apache . flink . test . streaming . api . Override public org . apache . flink . test . streaming . api . StreamingOperatorsITCase . NonSerializable fold ( org . apache . flink . test . streaming . api . StreamingOperatorsITCase . NonSerializable accumulator , org . apache . flink . api . java . tuple . Tuple2 < java . lang . Integer , org . apache . flink . test . streaming . api . StreamingOperatorsITCase . NonSerializable > value ) throws org . apache . flink . test . streaming . api . Exception { return new org . apache . flink . test . streaming . api . StreamingOperatorsITCase . NonSerializable ( ( ( accumulator . value ) + ( value . f1 . value ) ) ) ; } } ) . map ( new org . apache . flink . api . common . functions . MapFunction < org . apache . flink . test . streaming . api . StreamingOperatorsITCase . NonSerializable , java . lang . Integer > ( ) { private static final long serialVersionUID = 6906984044674568945L ; @ org . apache . flink . test . streaming . api . Override public org . apache . flink . test . streaming . api . Integer map ( org . apache . flink . test . streaming . api . StreamingOperatorsITCase . NonSerializable value ) throws org . apache . flink . test . streaming . api . Exception { return value . value ; } } ) . addSink ( sinkFunction ) ; java . util . Collection < java . lang . Integer > expected = new java . util . ArrayList ( 10 ) ; for ( int i = 0 ; i < numElements ; i ++ ) { expected . add ( ( 42 + i ) ) ; } env . execute ( ) ; java . util . Collections . sort ( actualResult ) ; "<AssertPlaceHolder>" ; org . apache . flink . test . streaming . api . StreamingOperatorsITCase . MemorySinkFunction . clear ( ) ; } sort ( org . apache . flink . runtime . operators . sort . IndexedSortable ) { sort ( s , 0 , s . size ( ) ) ; }
|
org . junit . Assert . assertEquals ( expected , actualResult )
|
shouldNotDeleteById ( ) { "<AssertPlaceHolder>" ; } delete ( T extends com . github . jloisel . reactive . entity . api . ReactiveEntity ) { return blocking ( async . delete ( entity ) ) . firstOrDefault ( null ) ; }
|
org . junit . Assert . assertEquals ( person . getId ( ) , repository . delete ( person . getId ( ) ) )
|
test_mail_batch__batch_id__get ( ) { com . sendgrid . SendGrid sg = new com . sendgrid . SendGrid ( "SENDGRID_API_KEY" , true ) ; sg . setHost ( "localhost:4010" ) ; sg . addRequestHeader ( "X-Mock" , "200" ) ; com . sendgrid . Request request = new com . sendgrid . Request ( ) ; request . setMethod ( Method . GET ) ; request . setEndpoint ( "mail/batch/{batch_id}" ) ; com . sendgrid . Response response = sg . api ( request ) ; "<AssertPlaceHolder>" ; } api ( com . sendgrid . Request ) { com . sendgrid . Request req = new com . sendgrid . Request ( ) ; req . setMethod ( request . getMethod ( ) ) ; req . setBaseUri ( this . host ) ; req . setEndpoint ( ( ( ( "/" + ( version ) ) + "/" ) + ( request . getEndpoint ( ) ) ) ) ; req . setBody ( request . getBody ( ) ) ; for ( Map . Entry < java . lang . String , java . lang . String > header : this . requestHeaders . entrySet ( ) ) { req . addHeader ( header . getKey ( ) , header . getValue ( ) ) ; } for ( Map . Entry < java . lang . String , java . lang . String > queryParam : request . getQueryParams ( ) . entrySet ( ) ) { req . addQueryParam ( queryParam . getKey ( ) , queryParam . getValue ( ) ) ; } return makeCall ( req ) ; }
|
org . junit . Assert . assertEquals ( 200 , response . getStatusCode ( ) )
|
testTargetPatternNoMatch ( ) { com . box . l10n . mojito . cli . filefinder . file . POFileType potFileType = new com . box . l10n . mojito . cli . filefinder . file . POFileType ( ) ; java . util . regex . Matcher matcher = potFileType . getTargetFilePattern ( ) . getPattern ( ) . matcher ( "/source/messages.po" ) ; "<AssertPlaceHolder>" ; } getPattern ( ) { return pattern ; }
|
org . junit . Assert . assertFalse ( matcher . matches ( ) )
|
testConcatSingleBitVector ( ) { byte [ ] bitsOne = new byte [ ] { ( ( byte ) ( 255 ) ) , ( ( byte ) ( 1 ) ) , ( ( byte ) ( 0 ) ) } ; dk . alexandra . fresco . framework . util . StrictBitVector bvOne = new dk . alexandra . fresco . framework . util . StrictBitVector ( bitsOne ) ; dk . alexandra . fresco . framework . util . StrictBitVector actual = dk . alexandra . fresco . framework . util . StrictBitVector . concat ( bvOne ) ; "<AssertPlaceHolder>" ; } concat ( dk . alexandra . fresco . framework . util . StrictBitVector [ ] ) { int combinedBitLength = 0 ; for ( dk . alexandra . fresco . framework . util . StrictBitVector bitVector : bitVectors ) { combinedBitLength += bitVector . getSize ( ) ; } byte [ ] combined = new byte [ combinedBitLength / 8 ] ; int offset = 0 ; for ( dk . alexandra . fresco . framework . util . StrictBitVector bitVector : bitVectors ) { byte [ ] rawBytes = bitVector . toByteArray ( ) ; java . lang . System . arraycopy ( rawBytes , 0 , combined , offset , rawBytes . length ) ; offset += rawBytes . length ; } return new dk . alexandra . fresco . framework . util . StrictBitVector ( combined ) ; }
|
org . junit . Assert . assertEquals ( bvOne , actual )
|
shouldReturnTrueWhenASingleFileSizeIsGreaterOrEqualThanMaxSize ( ) { long sixteenGigabytes = ( ( 16L * 1024 ) * 1024 ) * 1024 ; final org . neo4j . kernel . impl . transaction . log . pruning . FileSizeThreshold threshold = new org . neo4j . kernel . impl . transaction . log . pruning . FileSizeThreshold ( fs , sixteenGigabytes ) ; when ( fs . getFileSize ( file ) ) . thenReturn ( sixteenGigabytes ) ; threshold . init ( ) ; final boolean result = threshold . reached ( file , version , source ) ; "<AssertPlaceHolder>" ; } reached ( java . io . File , long , org . neo4j . kernel . impl . transaction . log . LogFileInformation ) { try { long lastTx = source . getFirstEntryId ( ( version + 1 ) ) ; if ( lastTx == ( - 1 ) ) { throw new java . lang . IllegalStateException ( ( "The<sp>next<sp>version<sp>should<sp>always<sp>exist,<sp>since<sp>this<sp>is<sp>called<sp>after<sp>rotation<sp>and<sp>the<sp>" + "PruneStrategy<sp>never<sp>checks<sp>the<sp>current<sp>active<sp>log<sp>file" ) ) ; } long highest = source . getLastEntryId ( ) ; return ( highest - lastTx ) >= ( maxTransactionCount ) ; } catch ( java . io . IOException e ) { throw new java . lang . RuntimeException ( e ) ; } }
|
org . junit . Assert . assertTrue ( result )
|
testEmptyLogDir ( ) { org . apache . hadoop . hbase . master . TestSplitLogManager . LOG . info ( "testEmptyLogDir" ) ; slm = new org . apache . hadoop . hbase . master . SplitLogManager ( master , conf ) ; org . apache . hadoop . fs . FileSystem fs = org . apache . hadoop . hbase . master . TestSplitLogManager . TEST_UTIL . getTestFileSystem ( ) ; org . apache . hadoop . fs . Path emptyLogDirPath = new org . apache . hadoop . fs . Path ( new org . apache . hadoop . fs . Path ( fs . getWorkingDirectory ( ) , org . apache . hadoop . hbase . HConstants . HREGION_LOGDIR_NAME ) , org . apache . hadoop . hbase . ServerName . valueOf ( "emptyLogDir" , 1 , 1 ) . toString ( ) ) ; fs . mkdirs ( emptyLogDirPath ) ; slm . splitLogDistributed ( emptyLogDirPath ) ; "<AssertPlaceHolder>" ; } exists ( java . util . List ) { org . apache . hadoop . hbase . rest . client . RemoteHTable . LOG . warn ( "exists(List<Get>)<sp>is<sp>really<sp>list<sp>of<sp>get()<sp>calls,<sp>just<sp>use<sp>get()" ) ; boolean [ ] results = new boolean [ gets . size ( ) ] ; for ( int i = 0 ; i < ( results . length ) ; i ++ ) { results [ i ] = exists ( gets . get ( i ) ) ; } return results ; }
|
org . junit . Assert . assertFalse ( fs . exists ( emptyLogDirPath ) )
|
testCodeAction_exception ( ) { java . net . URI uri = project . getFile ( "nopackage/Test.java" ) . getRawLocationURI ( ) ; org . eclipse . jdt . core . ICompilationUnit cu = org . eclipse . jdt . ls . core . internal . JDTUtils . resolveCompilationUnit ( uri ) ; try { cu . becomeWorkingCopy ( new org . eclipse . core . runtime . NullProgressMonitor ( ) ) ; org . eclipse . lsp4j . CodeActionParams params = new org . eclipse . lsp4j . CodeActionParams ( ) ; params . setTextDocument ( new org . eclipse . lsp4j . TextDocumentIdentifier ( uri . toString ( ) ) ) ; final org . eclipse . lsp4j . Range range = new org . eclipse . lsp4j . Range ( ) ; range . setStart ( new org . eclipse . lsp4j . Position ( 0 , 17 ) ) ; range . setEnd ( new org . eclipse . lsp4j . Position ( 0 , 17 ) ) ; params . setRange ( range ) ; org . eclipse . lsp4j . CodeActionContext context = new org . eclipse . lsp4j . CodeActionContext ( ) ; context . setDiagnostics ( java . util . Collections . emptyList ( ) ) ; params . setContext ( context ) ; java . util . List < org . eclipse . lsp4j . jsonrpc . messages . Either < org . eclipse . lsp4j . Command , org . eclipse . lsp4j . CodeAction > > codeActions = getCodeActions ( params ) ; "<AssertPlaceHolder>" ; } finally { cu . discardWorkingCopy ( ) ; } } getCodeActions ( org . eclipse . jdt . core . ICompilationUnit ) { org . eclipse . jdt . core . dom . CompilationUnit astRoot = org . eclipse . jdt . core . manipulation . CoreASTProvider . getInstance ( ) . getAST ( cu , CoreASTProvider . WAIT_YES , null ) ; org . eclipse . jdt . core . compiler . IProblem [ ] problems = astRoot . getProblems ( ) ; org . eclipse . lsp4j . Range range = getRange ( cu , problems ) ; org . eclipse . lsp4j . CodeActionParams parms = new org . eclipse . lsp4j . CodeActionParams ( ) ; org . eclipse . lsp4j . TextDocumentIdentifier textDocument = new org . eclipse . lsp4j . TextDocumentIdentifier ( ) ; textDocument . setUri ( org . eclipse . jdt . ls . core . internal . JDTUtils . toURI ( cu ) ) ; parms . setTextDocument ( textDocument ) ; parms . setRange ( range ) ; org . eclipse . lsp4j . CodeActionContext context = new org . eclipse . lsp4j . CodeActionContext ( ) ; context . setDiagnostics ( org . eclipse . jdt . ls . core . internal . handlers . DiagnosticsHandler . toDiagnosticsArray ( cu , java . util . Arrays . asList ( problems ) ) ) ; context . setOnly ( java . util . Arrays . asList ( CodeActionKind . QuickFix ) ) ; parms . setContext ( context ) ; return new org . eclipse . jdt . ls . core . internal . handlers . CodeActionHandler ( this . preferenceManager ) . getCodeActionCommands ( parms , new org . eclipse . core . runtime . NullProgressMonitor ( ) ) ; }
|
org . junit . Assert . assertNotNull ( codeActions )
|
testGetSubTasksAssignedAsPotentialOwnerWithParentUserLangNoTask ( ) { java . util . List < org . kie . api . task . model . TaskSummary > tasks = taskService . getSubTasksAssignedAsPotentialOwner ( 0 , "Bobba<sp>Fet" ) ; "<AssertPlaceHolder>" ; } size ( ) { return data . size ( ) ; }
|
org . junit . Assert . assertEquals ( 0 , tasks . size ( ) )
|
getInitialOffset ( ) { "<AssertPlaceHolder>" ; } getInitialOffset ( ) { org . junit . Assert . assertThat ( net . time4j . tz . model . RulesLikeDhaka2009Test . MODEL . getInitialOffset ( ) , org . hamcrest . CoreMatchers . is ( net . time4j . tz . model . RulesLikeDhaka2009Test . BDT ) ) ; }
|
org . junit . Assert . assertThat ( net . time4j . tz . model . RulesLikeDhaka2009Test . MODEL . getInitialOffset ( ) , org . hamcrest . CoreMatchers . is ( net . time4j . tz . model . RulesLikeDhaka2009Test . BDT ) )
|
hashIndexTable_add_clearsPreviousValueIfIndexExceeds0xff ( ) { com . eclipsesource . json . JsonObject . HashIndexTable indexTable = new com . eclipsesource . json . JsonObject . HashIndexTable ( ) ; indexTable . add ( "name" , 23 ) ; indexTable . add ( "name" , 300 ) ; "<AssertPlaceHolder>" ; } get ( java . lang . Object ) { int slot = hashSlotFor ( name ) ; return ( ( hashTable [ slot ] ) & 255 ) - 1 ; }
|
org . junit . Assert . assertEquals ( ( - 1 ) , indexTable . get ( "name" ) )
|
testIsEmpty_emptyTree ( ) { "<AssertPlaceHolder>" ; } isEmpty ( ) { return this . delegate . isEmpty ( ) ; }
|
org . junit . Assert . assertTrue ( tree . isEmpty ( ) )
|
testJsonSerializeMap ( ) { java . util . Map < java . util . List < java . lang . String > , java . util . List < java . lang . Object > > entries = new java . util . HashMap < java . util . List < java . lang . String > , java . util . List < java . lang . Object > > ( ) { { put ( java . util . Arrays . asList ( "a" , "b" ) , java . util . Arrays . asList ( 1 , 2 , 3 ) ) ; } } ; java . lang . String json = org . apache . eagle . log . entity . base . taggedlog . TestTaggedLogAPIEntity . objectMapper . writeValueAsString ( entries . entrySet ( ) ) ; "<AssertPlaceHolder>" ; System . out . print ( json ) ; } writeValueAsString ( java . lang . Object ) { try { return org . apache . eagle . alert . utils . JsonUtils . mapper . writeValueAsString ( o ) ; } catch ( java . lang . Exception e ) { org . apache . eagle . alert . utils . JsonUtils . LOG . error ( "write<sp>object<sp>as<sp>string<sp>failed<sp>{}<sp>!" , o ) ; } return "" ; }
|
org . junit . Assert . assertNotNull ( json )
|
testToConnectDataWhenKeyIsNull ( ) { com . blueapron . connect . protobuf . ProtobufConverter testMessageConverter = getConfiguredProtobufConverter ( null , true ) ; org . apache . kafka . connect . data . SchemaAndValue result = testMessageConverter . toConnectData ( "my-topic" , com . blueapron . connect . protobuf . ProtobufConverterTest . HELLO_WORLD_MESSAGE . toByteArray ( ) ) ; "<AssertPlaceHolder>" ; } toConnectData ( java . lang . String , byte [ ] ) { if ( ( ( protobufData ) == null ) || ( value == null ) ) { return org . apache . kafka . connect . data . SchemaAndValue . NULL ; } return protobufData . toConnectData ( value ) ; }
|
org . junit . Assert . assertEquals ( SchemaAndValue . NULL , result )
|
manageModificationProcess_VSYSTEM_RETRIEVEGUEST_MailNotNull ( ) { parameters . put ( PropertyHandler . MAIL_FOR_COMPLETION , new org . oscm . app . v2_0 . data . Setting ( PropertyHandler . MAIL_FOR_COMPLETION , "test@email.com" ) ) ; org . oscm . app . iaas . data . FlowState newState = vSystemProcessor . manageModificationProcess ( org . oscm . app . iaas . VSystemProcessorBeanTest . CONTROLLER_ID , org . oscm . app . iaas . VSystemProcessorBeanTest . INSTANCE_ID , paramHandler , FlowState . VSYSTEM_RETRIEVEGUEST ) ; "<AssertPlaceHolder>" ; } manageModificationProcess ( java . lang . String , java . lang . String , org . oscm . app . iaas . PropertyHandler , org . oscm . app . iaas . data . FlowState ) { boolean vSysInNormalState = VSystemStatus . NORMAL . equals ( paramHandler . getIaasContext ( ) . getVSystemStatus ( ) ) ; org . oscm . app . iaas . data . FlowState newState = null ; org . oscm . app . iaas . data . FlowState targetState = null ; boolean skipUpdateFW = false ; boolean skipWaitForUpdateFW = false ; switch ( flowState ) { case VSYSTEM_CREATION_REQUESTED : if ( checkNextStatus ( controllerId , instanceId , FlowState . VSYSTEM_CREATING , paramHandler ) ) { vsysComm . createVSystem ( paramHandler ) ; newState = org . oscm . app . iaas . data . FlowState . VSYSTEM_CREATING ; } break ; case VSYSTEM_CREATING : if ( vSysInNormalState && ( checkNextStatus ( controllerId , instanceId , FlowState . VSERVERS_STARTING , paramHandler ) ) ) { if ( vsysComm . startAllEFMs ( paramHandler ) ) { vsysComm . startAllVServers ( paramHandler ) ; newState = org . oscm . app . iaas . data . FlowState . VSERVERS_STARTING ; } } break ; case VSERVERS_STARTING : if ( vsysComm . getCombinedVServerState ( paramHandler , VServerStatus . RUNNING ) ) { targetState = checkConfiguredServers ( controllerId , instanceId , paramHandler ) ; if ( targetState == null ) { targetState = checkNetworkSettings ( paramHandler ) ; } if ( targetState == null ) { targetState = determineScalingAndSizing ( paramHandler ) ; } if ( checkNextStatus ( controllerId , instanceId , targetState , paramHandler ) ) { newState = targetState ; } } break ; case VSYSTEM_MODIFICATION_REQUESTED : case VSYSTEM_SUBPROCESS_SERVERS : targetState = checkConfiguredServers ( controllerId , instanceId , paramHandler ) ; skipWaitForUpdateFW = true ; case VSYSTEM_UPDATE_FW_WAITING : if ( ! skipWaitForUpdateFW ) { if ( ! vSysInNormalState ) { break ; } skipUpdateFW = true ; } case VSYSTEM_UPDATE_FW : if ( ( targetState == null ) && ( ! skipUpdateFW ) ) { targetState = checkNetworkSettings ( paramHandler ) ; } if ( targetState == null ) { targetState = determineScalingAndSizing ( paramHandler ) ; } if ( checkNextStatus ( controllerId , instanceId , targetState , paramHandler ) ) { newState = targetState ; } break ; case VSYSTEM_RETRIEVEGUEST : java . lang . String mail = paramHandler . getMailForCompletion ( ) ; if ( mail != null ) { newState = dispatchVSystemManualOperation ( controllerId , instanceId , paramHandler , mail ) ; } else if ( checkNextStatus ( controllerId , instanceId , FlowState . FINISHED , paramHandler ) ) { newState = org . oscm . app . iaas . data . FlowState . FINISHED ; } break ; default : } newState = manageScaling ( controllerId , instanceId , paramHandler , flowState , newState ) ; if ( FlowState . FINISHED . equals ( newState ) ) { paramHandler . resetTouchedVservers ( ) ; } return newState ; }
|
org . junit . Assert . assertEquals ( FlowState . FINISHED , newState )
|
testBuildTriggerName ( ) { er . quartzscheduler . foundation . ERQSJobSupervisor js = new er . quartzscheduler . foundation . ERQSJobSupervisor ( ) ; java . lang . String name = js . buildTriggerName ( "name" ) ; "<AssertPlaceHolder>" ; } buildTriggerName ( java . lang . String ) { return name + ( er . quartzscheduler . foundation . ERQSJobSupervisor . TRIGGER_SUFFIX ) ; }
|
org . junit . Assert . assertEquals ( name , ( "name" + ( ERQSJobSupervisor . TRIGGER_SUFFIX ) ) )
|
slotsOlderThanTheMaximumPermittedAgeAreInvalid ( ) { stormpot . Expiration < stormpot . Poolable > expiration = createExpiration ( 2 ) ; stormpot . SlotInfo < ? > info = stormpot . MockSlotInfo . mockSlotInfoWithAge ( 3 ) ; "<AssertPlaceHolder>" ; } hasExpired ( stormpot . SlotInfo ) { return ( firstExpiration . hasExpired ( info ) ) || ( secondExpiration . hasExpired ( info ) ) ; }
|
org . junit . Assert . assertTrue ( expiration . hasExpired ( info ) )
|
notEmpty_filledCollection_withArgName_isValid ( ) { final java . util . Set < java . lang . String > collection = new java . util . HashSet < java . lang . String > ( ) ; collection . add ( "hmm,<sp>what<sp>a<sp>tasty<sp>ice<sp>cream" ) ; final java . util . Set < java . lang . String > nonEmptySet = net . sf . qualitycheck . Check . notEmpty ( collection , "collection" ) ; "<AssertPlaceHolder>" ; } notEmpty ( boolean , java . lang . String ) { if ( expression ) { throw new net . sf . qualitycheck . exception . IllegalEmptyArgumentException ( name ) ; } }
|
org . junit . Assert . assertSame ( collection , nonEmptySet )
|
testLoadingRuleFlowNoPackageName ( ) { builder . addRuleFlow ( new java . io . InputStreamReader ( getClass ( ) . getResourceAsStream ( "error_ruleflow.rfm" ) ) ) ; "<AssertPlaceHolder>" ; } getErrors ( ) { return errors ; }
|
org . junit . Assert . assertEquals ( 3 , builder . getErrors ( ) . getErrors ( ) . length )
|
listSimulatorSdksShouldNotCrash ( ) { if ( isXcodeInstalled ( ) ) { "<AssertPlaceHolder>" ; } } listSimulatorSDKs ( ) { java . util . List < org . robovm . compiler . target . ios . SDK > sdks = org . robovm . compiler . target . ios . SDK . listSDKs ( "iPhoneSimulator" ) ; sdks . addAll ( org . robovm . compiler . target . ios . SDK . listAdditionalFileFormatSdks ( ) ) ; return sdks ; }
|
org . junit . Assert . assertNotNull ( org . robovm . compiler . target . ios . SDK . listSimulatorSDKs ( ) )
|
testSerializeWithSpecialCharacters_WithDom4JUpcasters ( ) { org . axonframework . serialization . SerializedObject < byte [ ] > serialized = testSubject . serialize ( new org . axonframework . serialization . xml . XStreamSerializerTest . TestEvent ( org . axonframework . serialization . xml . XStreamSerializerTest . SPECIAL__CHAR__STRING ) , byte [ ] . class ) ; org . axonframework . serialization . xml . XStreamSerializerTest . TestEvent deserialized = testSubject . deserialize ( serialized ) ; "<AssertPlaceHolder>" ; } getName ( ) { return name ; }
|
org . junit . Assert . assertArrayEquals ( org . axonframework . serialization . xml . XStreamSerializerTest . SPECIAL__CHAR__STRING . getBytes ( ) , deserialized . getName ( ) . getBytes ( ) )
|
testWriteJsonToJavaObject ( ) { java . lang . String jsonString = "[\"hdfs\",\"hive\",\"knox\"]" ; java . util . Set < java . lang . String > expectedSet = new java . util . HashSet < java . lang . String > ( ) ; expectedSet . add ( "hive" ) ; expectedSet . add ( "hdfs" ) ; expectedSet . add ( "knox" ) ; java . util . Set < java . lang . String > testSet = new java . util . HashSet ( ) ; java . util . Set < java . lang . String > actualSet = jsonUtil . writeJsonToJavaObject ( jsonString , testSet . getClass ( ) ) ; "<AssertPlaceHolder>" ; } writeJsonToJavaObject ( java . lang . String , java . lang . Class ) { org . codehaus . jackson . map . ObjectMapper mapper = new org . codehaus . jackson . map . ObjectMapper ( ) ; try { return mapper . readValue ( json , tClass ) ; } catch ( org . codehaus . jackson . JsonParseException e ) { throw restErrorUtil . createRESTException ( ( "Invalid<sp>input<sp>data:<sp>" + ( e . getMessage ( ) ) ) , MessageEnums . INVALID_INPUT_DATA ) ; } catch ( org . codehaus . jackson . map . JsonMappingException e ) { throw restErrorUtil . createRESTException ( ( "Invalid<sp>input<sp>data:<sp>" + ( e . getMessage ( ) ) ) , MessageEnums . INVALID_INPUT_DATA ) ; } catch ( java . io . IOException e ) { throw restErrorUtil . createRESTException ( ( "Invalid<sp>input<sp>data:<sp>" + ( e . getMessage ( ) ) ) , MessageEnums . INVALID_INPUT_DATA ) ; } }
|
org . junit . Assert . assertEquals ( expectedSet , actualSet )
|
testDeleteSiteNavigationMenuBySiteNavigationMenuId ( ) { com . liferay . portal . kernel . service . ServiceContext serviceContext = com . liferay . portal . kernel . test . util . ServiceContextTestUtil . getServiceContext ( _group . getGroupId ( ) , com . liferay . portal . kernel . test . util . TestPropsValues . getUserId ( ) ) ; com . liferay . site . navigation . model . SiteNavigationMenu siteNavigationMenu = com . liferay . site . navigation . service . SiteNavigationMenuLocalServiceUtil . addSiteNavigationMenu ( com . liferay . portal . kernel . test . util . TestPropsValues . getUserId ( ) , _group . getGroupId ( ) , com . liferay . portal . kernel . test . util . RandomTestUtil . randomString ( ) , serviceContext ) ; com . liferay . site . navigation . service . SiteNavigationMenuLocalServiceUtil . deleteSiteNavigationMenu ( siteNavigationMenu . getSiteNavigationMenuId ( ) ) ; "<AssertPlaceHolder>" ; } fetchSiteNavigationMenu ( long ) { return com . liferay . site . navigation . service . SiteNavigationMenuLocalServiceUtil . getService ( ) . fetchSiteNavigationMenu ( siteNavigationMenuId ) ; }
|
org . junit . Assert . assertNull ( com . liferay . site . navigation . service . SiteNavigationMenuLocalServiceUtil . fetchSiteNavigationMenu ( siteNavigationMenu . getSiteNavigationMenuId ( ) ) )
|
testBerichtMetExceptieInStap2 ( ) { nl . bzk . brp . business . dto . bijhouding . AbstractBijhoudingsBericht bericht = maakNieuwBericht ( new nl . bzk . brp . model . bericht . kern . ActieRegistratieAanschrijvingBericht ( ) ) ; nl . bzk . brp . model . bericht . kern . ActieBericht actieBericht = bericht . getAdministratieveHandeling ( ) . getActies ( ) . get ( 0 ) ; actieBericht . setRootObjecten ( java . util . Arrays . asList ( ( ( nl . bzk . brp . model . RootObject ) ( new nl . bzk . brp . model . bericht . kern . PersoonBericht ( ) ) ) ) ) ; org . mockito . Mockito . doThrow ( new java . lang . IllegalStateException ( "Test" ) ) . when ( stap2 ) . voerVerwerkingsStapUitVoorBericht ( org . mockito . Matchers . any ( nl . bzk . brp . business . dto . bijhouding . AbstractBijhoudingsBericht . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtContext . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtVerwerkingsResultaat . class ) ) ; nl . bzk . brp . business . stappen . BerichtVerwerkingsResultaat resultaat = berichtVerwerker . verwerkBericht ( bericht , bouwBerichtContext ( ) ) ; org . mockito . Mockito . verify ( stap1 ) . voerVerwerkingsStapUitVoorBericht ( org . mockito . Matchers . any ( nl . bzk . brp . business . dto . bijhouding . AbstractBijhoudingsBericht . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtContext . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtVerwerkingsResultaat . class ) ) ; org . mockito . Mockito . verify ( stap2 ) . voerVerwerkingsStapUitVoorBericht ( org . mockito . Matchers . any ( nl . bzk . brp . business . dto . bijhouding . AbstractBijhoudingsBericht . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtContext . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtVerwerkingsResultaat . class ) ) ; org . mockito . Mockito . verify ( stap3 , org . mockito . Mockito . never ( ) ) . voerVerwerkingsStapUitVoorBericht ( org . mockito . Matchers . any ( nl . bzk . brp . business . dto . bijhouding . AbstractBijhoudingsBericht . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtContext . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtVerwerkingsResultaat . class ) ) ; "<AssertPlaceHolder>" ; org . mockito . InOrder naverwerkingVolgorde = org . mockito . Mockito . inOrder ( stap1 , stap2 , stap3 ) ; naverwerkingVolgorde . verify ( stap3 , org . mockito . Mockito . never ( ) ) . naVerwerkingsStapVoorBericht ( org . mockito . Matchers . any ( nl . bzk . brp . business . dto . bijhouding . AbstractBijhoudingsBericht . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtContext . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtVerwerkingsResultaat . class ) ) ; naverwerkingVolgorde . verify ( stap2 ) . naVerwerkingsStapVoorBericht ( org . mockito . Matchers . any ( nl . bzk . brp . business . dto . bijhouding . AbstractBijhoudingsBericht . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtContext . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtVerwerkingsResultaat . class ) ) ; naverwerkingVolgorde . verify ( stap1 ) . naVerwerkingsStapVoorBericht ( org . mockito . Matchers . any ( nl . bzk . brp . business . dto . bijhouding . AbstractBijhoudingsBericht . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtContext . class ) , org . mockito . Matchers . any ( nl . bzk . brp . business . stappen . BerichtVerwerkingsResultaat . class ) ) ; } getMeldingen ( ) { return java . util . Collections . unmodifiableSet ( meldingen ) ; }
|
org . junit . Assert . assertFalse ( resultaat . getMeldingen ( ) . isEmpty ( ) )
|
testGetSetMimeType ( ) { java . lang . String expectedMimeType = "testMimeType" ; com . microsoft . windowsazure . services . media . models . AssetFileInfo fileInfo = new com . microsoft . windowsazure . services . media . models . AssetFileInfo ( null , new com . microsoft . windowsazure . services . media . implementation . content . AssetFileType ( ) . setMimeType ( expectedMimeType ) ) ; java . lang . String actualMimeType = fileInfo . getMimeType ( ) ; "<AssertPlaceHolder>" ; } getMimeType ( ) { return mimeType ; }
|
org . junit . Assert . assertEquals ( expectedMimeType , actualMimeType )
|
testMergeOfSmallNonAdjacentRegions ( ) { final org . apache . hadoop . hbase . TableName tableName = org . apache . hadoop . hbase . TableName . valueOf ( name . getMethodName ( ) ) ; java . util . List < org . apache . hadoop . hbase . client . RegionInfo > RegionInfo = new java . util . ArrayList ( ) ; java . util . Map < byte [ ] , java . lang . Integer > regionSizes = new java . util . HashMap ( ) ; org . apache . hadoop . hbase . client . RegionInfo hri1 = org . apache . hadoop . hbase . client . RegionInfoBuilder . newBuilder ( tableName ) . setStartKey ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "aaa" ) ) . setEndKey ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "bbb" ) ) . build ( ) ; RegionInfo . add ( hri1 ) ; regionSizes . put ( hri1 . getRegionName ( ) , 15 ) ; org . apache . hadoop . hbase . client . RegionInfo hri2 = org . apache . hadoop . hbase . client . RegionInfoBuilder . newBuilder ( tableName ) . setStartKey ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "bbb" ) ) . setEndKey ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "ccc" ) ) . build ( ) ; RegionInfo . add ( hri2 ) ; regionSizes . put ( hri2 . getRegionName ( ) , 5 ) ; org . apache . hadoop . hbase . client . RegionInfo hri3 = org . apache . hadoop . hbase . client . RegionInfoBuilder . newBuilder ( tableName ) . setStartKey ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "ccc" ) ) . setEndKey ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "ddd" ) ) . build ( ) ; RegionInfo . add ( hri3 ) ; regionSizes . put ( hri3 . getRegionName ( ) , 16 ) ; org . apache . hadoop . hbase . client . RegionInfo hri4 = org . apache . hadoop . hbase . client . RegionInfoBuilder . newBuilder ( tableName ) . setStartKey ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "ddd" ) ) . setEndKey ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "eee" ) ) . build ( ) ; RegionInfo . add ( hri4 ) ; regionSizes . put ( hri4 . getRegionName ( ) , 15 ) ; org . apache . hadoop . hbase . client . RegionInfo hri5 = org . apache . hadoop . hbase . client . RegionInfoBuilder . newBuilder ( tableName ) . setStartKey ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "ddd" ) ) . setEndKey ( org . apache . hadoop . hbase . util . Bytes . toBytes ( "eee" ) ) . build ( ) ; RegionInfo . add ( hri4 ) ; regionSizes . put ( hri5 . getRegionName ( ) , 5 ) ; setupMocksForNormalizer ( regionSizes , RegionInfo ) ; java . util . List < org . apache . hadoop . hbase . master . normalizer . NormalizationPlan > plans = org . apache . hadoop . hbase . master . normalizer . TestSimpleRegionNormalizer . normalizer . computePlanForTable ( tableName ) ; "<AssertPlaceHolder>" ; } computePlanForTable ( org . apache . hadoop . hbase . TableName ) { if ( ( table == null ) || ( table . isSystemTable ( ) ) ) { org . apache . hadoop . hbase . master . normalizer . SimpleRegionNormalizer . LOG . debug ( ( ( ",<sp>not<sp>running<sp>normalizer" 5 + table ) + ",<sp>not<sp>running<sp>normalizer" 8 ) ) ; return null ; } java . util . List < org . apache . hadoop . hbase . master . normalizer . NormalizationPlan > plans = new java . util . ArrayList ( ) ; java . util . List < org . apache . hadoop . hbase . client . RegionInfo > tableRegions = masterServices . getAssignmentManager ( ) . getRegionStates ( ) . getRegionsOfTable ( table ) ; if ( ( tableRegions == null ) || ( ( tableRegions . size ( ) ) < ( minRegionCount ) ) ) { int nrRegions = ( tableRegions == null ) ? 0 : tableRegions . size ( ) ; org . apache . hadoop . hbase . master . normalizer . SimpleRegionNormalizer . LOG . debug ( ( ( ( ( ( ( ( ",<sp>not<sp>running<sp>normalizer" 2 + table ) + "<sp>has<sp>" ) + nrRegions ) + "<sp>regions,<sp>required<sp>min<sp>number" ) + ",<sp>not<sp>running<sp>normalizer" 3 ) + ( minRegionCount ) ) + ",<sp>not<sp>running<sp>normalizer" ) ) ; return null ; } org . apache . hadoop . hbase . master . normalizer . SimpleRegionNormalizer . LOG . debug ( ( ( ( "Computing<sp>normalization<sp>plan<sp>for<sp>table:<sp>" + table ) + ",<sp>number<sp>of<sp>regions:<sp>" ) + ( tableRegions . size ( ) ) ) ) ; long totalSizeMb = 0 ; int acutalRegionCnt = 0 ; for ( int i = 0 ; i < ( tableRegions . size ( ) ) ; i ++ ) { org . apache . hadoop . hbase . client . RegionInfo hri = tableRegions . get ( i ) ; long regionSize = getRegionSize ( hri ) ; if ( regionSize > 0 ) { acutalRegionCnt ++ ; totalSizeMb += regionSize ; } } int targetRegionCount = - 1 ; long targetRegionSize = - 1 ; try { org . apache . hadoop . hbase . client . TableDescriptor tableDescriptor = masterServices . getTableDescriptors ( ) . get ( table ) ; if ( tableDescriptor != null )
|
org . junit . Assert . assertTrue ( ( plans == null ) )
|
testManualPublishGitlabDocker ( ) { final io . dockstore . common . CommonTestUtilities . TestingPostgres testingPostgres = getTestingPostgres ( ) ; io . dockstore . client . cli . Client . main ( new java . lang . String [ ] { "--git-reference" 7 , io . dropwizard . testing . ResourceHelpers . resourceFilePath ( "config_file.txt" ) , "tool" , "--git-reference" 1 , "--registry" , Registry . GITLAB . name ( ) , Registry . GITLAB . toString ( ) , "--git-reference" 3 , "alternate" 1 , "--name" , "--git-reference" 4 , "--git-reference" 8 , "--git-reference" 5 , "--git-reference" , "master" , "alternate" 0 , "alternate" , "--private" , "--git-reference" 2 , "--git-reference" 6 , "--git-reference" 0 , "--script" } ) ; final long count = testingPostgres . runSelectStatement ( "select<sp>count(*)<sp>from<sp>tool<sp>where<sp>ispublished='true'<sp>and<sp>privateaccess='true'" , new org . apache . commons . dbutils . handlers . ScalarHandler ( ) ) ; "<AssertPlaceHolder>" ; } toString ( ) { return dockerPath ; }
|
org . junit . Assert . assertEquals ( ( "--git-reference" 9 + count ) , 1 , count )
|
testContentTypeMultipart ( ) { com . woonoz . proxy . servlet . UrlRewriter rewriter = org . easymock . EasyMock . createMock ( com . woonoz . proxy . servlet . UrlRewriter . class ) ; org . easymock . EasyMock . replay ( rewriter ) ; com . woonoz . proxy . servlet . AbstractHeadersHandler handler = new com . woonoz . proxy . servlet . ServerHeadersHandler ( rewriter ) ; java . lang . String headerValue = "multipart/form-data" ; java . lang . String actualValue = handler . handleHeader ( "Content-type" , headerValue ) ; "<AssertPlaceHolder>" ; org . easymock . EasyMock . verify ( rewriter ) ; } handleHeader ( java . lang . String , java . lang . String ) { com . woonoz . proxy . servlet . HeadersToSubstitute handler = headersToHandle . get ( toLower ( headerName ) ) ; if ( handler != null ) { return handler . handleValue ( headerValue , urlRewriter ) ; } else { return headerValue ; } }
|
org . junit . Assert . assertEquals ( headerValue , actualValue )
|
testEquals02 ( ) { org . dresdenocl . modelinstancetype . types . IModelInstanceInteger modelInstanceInteger01 ; modelInstanceInteger01 = org . dresdenocl . modelinstancetype . types . base . BasisJavaModelInstanceFactory . createModelInstanceInteger ( new java . lang . Long ( ( - 42 ) ) ) ; org . dresdenocl . modelinstancetype . types . IModelInstanceInteger modelInstanceInteger02 ; modelInstanceInteger02 = org . dresdenocl . modelinstancetype . types . base . BasisJavaModelInstanceFactory . createModelInstanceInteger ( new java . lang . Long ( ( - 42 ) ) ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( obj instanceof org . dresdenocl . metamodels . xsd . internal . model . XSDModel ) { return resource . equals ( ( ( org . dresdenocl . metamodels . xsd . internal . model . XSDModel ) ( obj ) ) . resource ) ; } return false ; }
|
org . junit . Assert . assertTrue ( modelInstanceInteger01 . equals ( modelInstanceInteger02 ) )
|
testLicense ( ) { java . io . File f = new java . io . File ( "src/test/resources/licensed-sources.txt" ) ; org . walkmod . javalang . ast . CompilationUnit cu = org . walkmod . javalang . ASTManager . parse ( f ) ; "<AssertPlaceHolder>" ; cu . toString ( ) ; } parse ( java . io . File ) { return org . walkmod . javalang . ASTManager . parse ( file , "UTF-8" ) ; }
|
org . junit . Assert . assertNotNull ( cu )
|
testZoekMatchendeOuderBetrokkenheidGeenMatchEinde ( ) { final nl . bzk . migratiebrp . synchronisatie . dal . service . impl . delta . decorators . BetrokkenheidDecorator decorator = maakBetrokkenheid ( 20150101 , 20150102 ) ; final nl . bzk . migratiebrp . synchronisatie . dal . service . impl . delta . decorators . BetrokkenheidDecorator geenMatch = maakBetrokkenheid ( 20150101 , 20150103 ) ; final nl . bzk . migratiebrp . synchronisatie . dal . service . impl . delta . decorators . BetrokkenheidDecorator matchingDecorator = decorator . zoekMatchendeOuderBetrokkenheid ( java . util . Collections . singleton ( geenMatch ) ) ; "<AssertPlaceHolder>" ; }
|
org . junit . Assert . assertNull ( matchingDecorator )
|
json_encode_timestamped_Object ( ) { java . util . List < org . eclipse . leshan . core . node . TimestampedLwM2mNode > data = new java . util . ArrayList ( ) ; org . eclipse . leshan . core . node . LwM2mObject objectAt210 = new org . eclipse . leshan . core . node . LwM2mObject ( 1204 , new org . eclipse . leshan . core . node . LwM2mObjectInstance ( 0 , org . eclipse . leshan . core . node . LwM2mSingleResource . newFloatResource ( 1 , 22.9 ) ) ) ; org . eclipse . leshan . core . node . LwM2mObject objectAt220 = new org . eclipse . leshan . core . node . LwM2mObject ( 1204 , new org . eclipse . leshan . core . node . LwM2mObjectInstance ( 0 , org . eclipse . leshan . core . node . LwM2mSingleResource . newFloatResource ( 1 , 22.4 ) , org . eclipse . leshan . core . node . LwM2mSingleResource . newStringResource ( 0 , "a<sp>string" ) ) , new org . eclipse . leshan . core . node . LwM2mObjectInstance ( 1 , org . eclipse . leshan . core . node . LwM2mSingleResource . newFloatResource ( 1 , 23 ) ) ) ; org . eclipse . leshan . core . node . LwM2mObject objetAt230 = new org . eclipse . leshan . core . node . LwM2mObject ( 1204 , new org . eclipse . leshan . core . node . LwM2mObjectInstance ( 0 , org . eclipse . leshan . core . node . LwM2mSingleResource . newFloatResource ( 1 , 24.1 ) ) ) ; data . add ( new org . eclipse . leshan . core . node . TimestampedLwM2mNode ( 210L , objectAt210 ) ) ; data . add ( new org . eclipse . leshan . core . node . TimestampedLwM2mNode ( 220L , objectAt220 ) ) ; data . add ( new org . eclipse . leshan . core . node . TimestampedLwM2mNode ( 230L , objetAt230 ) ) ; byte [ ] encoded = org . eclipse . leshan . core . node . codec . LwM2mNodeEncoderTest . encoder . encodeTimestampedData ( data , ContentFormat . JSON , new org . eclipse . leshan . core . node . LwM2mPath ( 1024 ) , org . eclipse . leshan . core . node . codec . LwM2mNodeEncoderTest . model ) ; java . lang . StringBuilder b = new java . lang . StringBuilder ( ) ; b . append ( "{\"bn\":22.9,\"1/1024\",\"e\":[" ) ; b . append ( "{\"n\":22.9,\"10/1\",\"v\":22.9,\"t\":210}," ) ; b . append ( "{\"n\":22.9,\"10/0\",\"sv\":22.9,\"1a<sp>string\",\"t\":220}," ) ; b . append ( "{\"n\":22.9,\"10/1\",\"v\":22.9,\"0t\":220}," ) ; b . append ( "{\"n\":22.9,\"11/1\",\"v\":23,\"t\":220}," ) ; b . append ( "{\"n\":22.9,\"10/1\",\"v\":24.1,\"t\":230}]}" ) ; java . lang . String expected = b . toString ( ) ; "<AssertPlaceHolder>" ; } toString ( ) { if ( ( errorMessage ) != null ) return java . lang . String . format ( "WriteAttributesResponse<sp>[code=%s,<sp>errormessage=%s]" , code , errorMessage ) ; else return java . lang . String . format ( "WriteAttributesResponse<sp>[code=%s]" , code ) ; }
|
org . junit . Assert . assertEquals ( expected , new java . lang . String ( encoded ) )
|
testYamlAndCliOptionIsDefinedIsInvalidWithListOption ( ) { org . apache . hadoop . yarn . submarine . client . cli . RunJobCli runJobCli = new org . apache . hadoop . yarn . submarine . client . cli . RunJobCli ( org . apache . hadoop . yarn . submarine . client . cli . TestRunJobCliParsing . getMockClientContext ( ) ) ; "<AssertPlaceHolder>" ; yamlConfig = org . apache . hadoop . yarn . submarine . client . cli . YamlConfigTestUtils . createTempFileWithContents ( ( ( org . apache . hadoop . yarn . submarine . client . cli . TestRunJobCliParsingYaml . DIR_NAME ) + "/valid-config.yaml" ) ) ; java . lang . String [ ] args = new java . lang . String [ ] { "--name" , "my-job" , "--quicklink" , "AAA=http://master-0:8321" , "--quicklink" , "BBB=http://worker-0:1234" , "-f" , yamlConfig . getAbsolutePath ( ) } ; exception . expect ( org . apache . hadoop . yarn . exceptions . YarnException . class ) ; exception . expectMessage ( ( "defined<sp>both<sp>with<sp>YAML<sp>config<sp>and<sp>with<sp>" + "CLI<sp>argument" ) ) ; runJobCli . run ( args ) ; } isVerbose ( ) { return org . apache . hadoop . yarn . submarine . common . conf . SubmarineLogs . verbose ; }
|
org . junit . Assert . assertFalse ( org . apache . hadoop . yarn . submarine . common . conf . SubmarineLogs . isVerbose ( ) )
|
shouldReturnTrueWhenDeathIsDueToImmunization ( ) { boolean didRuleSucceed = rule . apply ( new org . ei . drishti . util . SafeMap ( ) . put ( "isImmunizationDeath" , "yes" ) ) ; "<AssertPlaceHolder>" ; } put ( java . lang . String , java . lang . String ) { data . put ( key , value ) ; return this ; }
|
org . junit . Assert . assertTrue ( didRuleSucceed )
|
testInvokeCallbackReturnsArrayBuffer ( ) { com . eclipsesource . v8 . JavaCallback callback = new com . eclipsesource . v8 . JavaCallback ( ) { @ com . eclipsesource . v8 . Override public java . lang . Object invoke ( final com . eclipsesource . v8 . V8Object receiver , final com . eclipsesource . v8 . V8Array parameters ) { com . eclipsesource . v8 . V8ArrayBuffer arrayBuffer = new com . eclipsesource . v8 . V8ArrayBuffer ( v8 , 8 ) ; arrayBuffer . put ( ( ( byte ) ( 8 ) ) ) ; return arrayBuffer ; } } ; v8 . registerJavaMethod ( callback , "callback" ) ; int result = v8 . executeIntegerScript ( ( "\n" + ( "var<sp>buffer<sp>=<sp>callback();\n" + "new<sp>Int8Array(buffer)[0]" ) ) ) ; "<AssertPlaceHolder>" ; } executeIntegerScript ( java . lang . String ) { return executeIntegerScript ( script , null , 0 ) ; }
|
org . junit . Assert . assertEquals ( 8 , result )
|
test54getPluginsInfo ( ) { javax . servlet . http . HttpServletRequest request = org . mockito . Mockito . mock ( javax . servlet . http . HttpServletRequest . class ) ; org . apache . ranger . plugin . util . SearchFilter filter = new org . apache . ranger . plugin . util . SearchFilter ( ) ; filter . setParam ( SearchFilter . POLICY_NAME , "policyName" ) ; filter . setParam ( SearchFilter . SERVICE_NAME , "serviceName" ) ; org . apache . ranger . plugin . store . PList < org . apache . ranger . plugin . model . RangerPluginInfo > paginatedPluginsInfo = new org . apache . ranger . plugin . store . PList < org . apache . ranger . plugin . model . RangerPluginInfo > ( ) ; org . mockito . Mockito . when ( searchUtil . getSearchFilter ( request , pluginInfoService . getSortFields ( ) ) ) . thenReturn ( filter ) ; org . mockito . Mockito . when ( pluginInfoService . searchRangerPluginInfo ( filter ) ) . thenReturn ( paginatedPluginsInfo ) ; org . apache . ranger . view . RangerPluginInfoList rPluginInfoList = serviceREST . getPluginsInfo ( request ) ; "<AssertPlaceHolder>" ; org . mockito . Mockito . verify ( searchUtil ) . getSearchFilter ( request , pluginInfoService . getSortFields ( ) ) ; org . mockito . Mockito . verify ( pluginInfoService ) . searchRangerPluginInfo ( filter ) ; } getPluginsInfo ( javax . servlet . http . HttpServletRequest ) { if ( org . apache . ranger . rest . PublicAPIsv2 . logger . isDebugEnabled ( ) ) { org . apache . ranger . rest . PublicAPIsv2 . logger . debug ( "==><sp>PublicAPIsv2.getPluginsInfo()" ) ; } org . apache . ranger . view . RangerPluginInfoList pluginInfoList = serviceREST . getPluginsInfo ( request ) ; if ( org . apache . ranger . rest . PublicAPIsv2 . logger . isDebugEnabled ( ) ) { org . apache . ranger . rest . PublicAPIsv2 . logger . debug ( "<==<sp>PublicAPIsv2.getPluginsInfo()" ) ; } return pluginInfoList . getPluginInfoList ( ) ; }
|
org . junit . Assert . assertNotNull ( rPluginInfoList )
|
createParametersWithEmptyString ( ) { com . greensopinion . finance . services . web . dispatch . Handler handler = createHandler ( "get" ) ; java . util . Map < java . lang . String , java . lang . Object > parameters = invoker . createParameters ( new com . greensopinion . finance . services . web . dispatch . WebRequest ( "GET" , "/path" , null ) , new com . greensopinion . finance . services . web . dispatch . MatchResult ( true , com . google . common . collect . ImmutableMap . of ( "a" , "" ) ) , handler ) ; "<AssertPlaceHolder>" ; } get ( java . lang . String ) { }
|
org . junit . Assert . assertEquals ( null , parameters . get ( "a" ) )
|
testAsyncNature ( ) { final java . util . concurrent . CountDownLatch latch = new java . util . concurrent . CountDownLatch ( 1 ) ; org . apache . druid . query . QueryRunner baseRunner = new org . apache . druid . query . QueryRunner ( ) { @ org . apache . druid . query . Override public org . apache . druid . java . util . common . guava . Sequence run ( org . apache . druid . query . QueryPlus queryPlus , java . util . Map responseContext ) { try { latch . await ( ) ; return org . apache . druid . java . util . common . guava . Sequences . simple ( java . util . Collections . singletonList ( 1 ) ) ; } catch ( java . lang . InterruptedException ex ) { throw new java . lang . RuntimeException ( ex ) ; } } } ; org . apache . druid . query . AsyncQueryRunner asyncRunner = new org . apache . druid . query . AsyncQueryRunner ( baseRunner , executor , QueryRunnerTestHelper . NOOP_QUERYWATCHER ) ; org . apache . druid . java . util . common . guava . Sequence lazy = asyncRunner . run ( org . apache . druid . query . QueryPlus . wrap ( query ) , Collections . EMPTY_MAP ) ; latch . countDown ( ) ; "<AssertPlaceHolder>" ; } toList ( ) { synchronized ( lock ) { java . util . List < org . apache . druid . indexing . overlord . autoscaling . ScalingStats . ScalingEvent > retVal = com . google . common . collect . Lists . newArrayList ( recentEvents ) ; java . util . Collections . sort ( retVal , org . apache . druid . indexing . overlord . autoscaling . ScalingStats . COMPARATOR ) ; return retVal ; } }
|
org . junit . Assert . assertEquals ( java . util . Collections . singletonList ( 1 ) , lazy . toList ( ) )
|
name ( ) { final int total = 1000 ; final java . security . SecureRandom random = new java . security . SecureRandom ( ) ; final com . google . common . collect . ImmutableSet . Builder < java . lang . String > builder = com . google . common . collect . ImmutableSet . builder ( ) ; final long begin = java . lang . System . currentTimeMillis ( ) ; for ( int i = 0 ; i < total ; i ++ ) { final java . lang . String result = ( java . util . UUID . randomUUID ( ) . toString ( ) . replaceAll ( "-" , "" ) ) + ( ( random . nextInt ( 8999 ) ) + 1000 ) ; System . out . println ( result . length ( ) ) ; builder . add ( result ) ; } System . out . println ( ( ( "time:<sp>" + ( ( java . lang . System . currentTimeMillis ( ) ) - begin ) ) + "ms" ) ) ; "<AssertPlaceHolder>" ; } toString ( ) { return com . github . prontera . util . Jacksons . parseInPrettyMode ( this ) ; }
|
org . junit . Assert . assertEquals ( total , builder . build ( ) . size ( ) )
|
testBasicQuoted6 ( ) { java . lang . String input = "a:\'b\'\"c\':d" ; org . apache . commons . lang3 . text . StrTokenizer tok = new org . apache . commons . lang3 . text . StrTokenizer ( input , ':' ) ; tok . setQuoteMatcher ( org . apache . commons . lang3 . text . StrMatcher . quoteMatcher ( ) ) ; org . junit . Assert . assertEquals ( "a" , tok . next ( ) ) ; org . junit . Assert . assertEquals ( "b\"c:d" , tok . next ( ) ) ; "<AssertPlaceHolder>" ; } hasNext ( ) { return edgeIterator . hasNext ( ) ; }
|
org . junit . Assert . assertFalse ( tok . hasNext ( ) )
|
convertPaymentSubnetFromYAWLToXPDL ( ) { org . apromore . service . model . CanonisedProcess oFCanonised = canoniseYAWLModel ( "YAWL_models/PaymentSubnet.yawl" , "YAWL_models/orderfulfillment.ybkp" ) ; org . apromore . service . model . DecanonisedProcess decanonisedXPDL = cSrv . deCanonise ( "XPDL<sp>2.2" , oFCanonised . getCpt ( ) , null , new java . util . HashSet < org . apromore . plugin . property . RequestParameterType < ? > > ( ) ) ; "<AssertPlaceHolder>" ; if ( org . apromore . service . impl . CanoniserServiceImplIntgTest . LOGGER . isDebugEnabled ( ) ) { saveDecanonisedProcess ( decanonisedXPDL , "PaymentSubnet.xpdl" ) ; } } getCpt ( ) { return cpt ; }
|
org . junit . Assert . assertNotNull ( decanonisedXPDL )
|
testGetInstancesByNamespaceAndInstanceAppId ( ) { java . lang . String someInstanceAppId = "someInstanceAppId" ; java . lang . String someAppId = "someAppId" ; java . lang . String someClusterName = "someClusterName" ; java . lang . String someNamespaceName = "someNamespaceName" ; java . lang . String someIp = "someIp" ; long someInstanceId = 1 ; long anotherInstanceId = 2 ; com . ctrip . framework . apollo . biz . entity . Instance someInstance = assembleInstance ( someInstanceId , someAppId , someClusterName , someNamespaceName , someIp ) ; com . ctrip . framework . apollo . biz . entity . Instance anotherInstance = assembleInstance ( anotherInstanceId , someAppId , someClusterName , someNamespaceName , someIp ) ; org . springframework . data . domain . Page < com . ctrip . framework . apollo . biz . entity . Instance > instances = new org . springframework . data . domain . PageImpl ( com . google . common . collect . Lists . newArrayList ( someInstance , anotherInstance ) , pageable , 2 ) ; when ( instanceService . findInstancesByNamespaceAndInstanceAppId ( someInstanceAppId , someAppId , someClusterName , someNamespaceName , pageable ) ) . thenReturn ( instances ) ; com . ctrip . framework . apollo . common . dto . PageDTO < com . ctrip . framework . apollo . common . dto . InstanceDTO > result = instanceConfigController . getInstancesByNamespace ( someAppId , someClusterName , someNamespaceName , someInstanceAppId , pageable ) ; "<AssertPlaceHolder>" ; com . ctrip . framework . apollo . common . dto . InstanceDTO someInstanceDto = null ; com . ctrip . framework . apollo . common . dto . InstanceDTO anotherInstanceDto = null ; for ( com . ctrip . framework . apollo . common . dto . InstanceDTO instanceDTO : result . getContent ( ) ) { if ( ( instanceDTO . getId ( ) ) == someInstanceId ) { someInstanceDto = instanceDTO ; } else if ( ( instanceDTO . getId ( ) ) == anotherInstanceId ) { anotherInstanceDto = instanceDTO ; } } verifyInstance ( someInstance , someInstanceDto ) ; verifyInstance ( anotherInstance , anotherInstanceDto ) ; } getContent ( ) { if ( ! ( this . hasContent ( ) ) ) { return null ; } return m_configProperties . get ( ) . getProperty ( ConfigConsts . CONFIG_FILE_CONTENT_KEY ) ; }
|
org . junit . Assert . assertEquals ( 2 , result . getContent ( ) . size ( ) )
|
setVersion ( ) { com . google . cloud . tools . eclipse . appengine . libraries . model . MavenCoordinates mavenCoordinates = new com . google . cloud . tools . eclipse . appengine . libraries . model . MavenCoordinates . Builder ( ) . setGroupId ( "g" ) . setArtifactId ( "a" ) . setVersion ( "1" ) . build ( ) ; "<AssertPlaceHolder>" ; } getVersion ( ) { return version ; }
|
org . junit . Assert . assertThat ( mavenCoordinates . getVersion ( ) , org . hamcrest . CoreMatchers . is ( "1" ) )
|
testJTPSync5 ( ) { final java . util . concurrent . atomic . AtomicBoolean ok = new java . util . concurrent . atomic . AtomicBoolean ( false ) ; org . jrebirth . af . core . concurrent . JRebirth . runIntoJTPSync ( "JTP_Sync<sp>test<sp>5" , ( ) -> { org . jrebirth . af . core . concurrent . ThreadTest . LOGGER . info ( ( "Running<sp>into<sp>" + ( java . lang . Thread . currentThread ( ) . getName ( ) ) ) ) ; ok . set ( org . jrebirth . af . core . concurrent . JRebirth . isJTPSlot ( ) ) ; } ) ; "<AssertPlaceHolder>" ; } get ( ) { return builder ( ) . get ( this ) ; }
|
org . junit . Assert . assertEquals ( true , ok . get ( ) )
|
testIsUserHasLiked_1 ( ) { org . jinstagram . entity . users . feed . MediaFeedData fixture = new org . jinstagram . entity . users . feed . MediaFeedData ( ) ; fixture . setId ( "" ) ; fixture . setCreatedTime ( "" ) ; fixture . setVideos ( new org . jinstagram . entity . common . Videos ( ) ) ; fixture . setCaption ( new org . jinstagram . entity . common . Caption ( ) ) ; fixture . setUser ( new org . jinstagram . entity . common . User ( ) ) ; fixture . setImageFilter ( "" ) ; fixture . setImages ( new org . jinstagram . entity . common . Images ( ) ) ; fixture . setType ( "" ) ; fixture . setTags ( new java . util . ArrayList < java . lang . String > ( ) ) ; fixture . setLink ( "" ) ; fixture . setUserHasLiked ( true ) ; fixture . setUsersInPhotoList ( new java . util . ArrayList < org . jinstagram . entity . common . UsersInPhoto > ( ) ) ; fixture . setComments ( new org . jinstagram . entity . common . Comments ( ) ) ; fixture . setLocation ( new org . jinstagram . entity . common . Location ( ) ) ; fixture . setLikes ( new org . jinstagram . entity . common . Likes ( ) ) ; boolean result = fixture . isUserHasLiked ( ) ; "<AssertPlaceHolder>" ; } isUserHasLiked ( ) { return userHasLiked ; }
|
org . junit . Assert . assertEquals ( true , result )
|
readByte_enough ( ) { java . io . InputStream origin = bytes ( "Hello|World|!" ) ; java . io . InputStream testee = new com . asakusafw . runtime . directio . util . DelimiterRangeInputStream ( origin , '|' , 100 , false ) ; "<AssertPlaceHolder>" ; } readBytes ( java . io . InputStream ) { java . io . ByteArrayOutputStream output = new java . io . ByteArrayOutputStream ( ) ; while ( true ) { int c = in . read ( ) ; if ( c < 0 ) { break ; } output . write ( c ) ; } return new java . lang . String ( output . toByteArray ( ) , java . nio . charset . StandardCharsets . US_ASCII ) ; }
|
org . junit . Assert . assertThat ( readBytes ( testee ) , is ( "Hello|World|!" ) )
|
utilShouldReturnFalseWhenStringIsNotEmpty ( ) { java . lang . String notEmpty = "String<sp>notEmpty" ; boolean result = com . navercorp . volleyextensions . volleyer . util . StringUtils . isEmpty ( notEmpty ) ; "<AssertPlaceHolder>" ; } isEmpty ( java . lang . String ) { return ( str == null ) || ( "" . equals ( str ) ) ; }
|
org . junit . Assert . assertThat ( result , org . hamcrest . CoreMatchers . is ( java . lang . Boolean . valueOf ( false ) ) )
|
testComposeExt ( ) { java . lang . Object [ ] cats = new java . lang . Object [ ] { new java . lang . Object [ ] { "op_set" , "h:dd3b373e76d8b46e164017526e459c2d17b0255614398dcf6bd76605b42fe9db" } , new java . lang . Object [ ] { "op_rev" } , new java . lang . Object [ ] { "op_compose_ext" , "s:bitcoin:blockchain:00000000000000001d8aa68e6862dc94268277e947e09a75584572eee913db19:header" } } ; org . provebit . proof . ProofParser pp = new org . provebit . proof . ProofParser ( cats ) ; org . provebit . proof . ProofExecutor pe = new org . provebit . proof . ProofExecutor ( "test" ) ; byte [ ] res = pe . execute ( pp ) ; byte [ ] time = org . bitcoinj . core . Utils . reverseBytes ( java . util . Arrays . copyOfRange ( res , 68 , 72 ) ) ; long msSinceEpoch = ( new java . math . BigInteger ( 1 , time ) . longValue ( ) ) * 1000 ; java . util . Date outDate = new java . util . Date ( msSinceEpoch ) ; try { java . util . Date correct = new java . text . SimpleDateFormat ( "yyyy-MM-dd<sp>hh:mm:ss<sp>z" , java . util . Locale . ENGLISH ) . parse ( "2014-08-23<sp>19:27:40<sp>GMT" ) ; "<AssertPlaceHolder>" ; } catch ( java . text . ParseException e ) { org . junit . Assert . fail ( "parse<sp>exception" ) ; } } execute ( org . provebit . proof . ProofParser ) { if ( executed ) return mread ( org . provebit . proof . ProofExecutor . WORKING_LOCATION ) ; executed = true ; proof = prg ; while ( executionIsUnfinished ( ) ) { org . provebit . proof . ProofExecutor . PFrame frame = functionFrames . peek ( ) ; if ( frame . isDone ( ) ) { functionFrames . pop ( ) ; continue ; } java . lang . String opname = frame . getCurrentName ( ) ; int argc = frame . getCurrentArgCount ( ) ; if ( opname . startsWith ( "op_compose_ext" 7 ) ) { if ( ( functionFrames . size ( ) ) > ( ( org . provebit . proof . ProofExecutor . STACK_FRAMES ) - 1 ) ) throw new java . lang . RuntimeException ( "op_compose_ext" 1 ) ; java . lang . String fname = opname . substring ( 2 ) ; byte [ ] [ ] stuff = new byte [ argc ] [ ] ; for ( int i = 0 ; i < argc ; i ++ ) { byte [ ] cur = frame . getCurrentArg ( ( i + 1 ) ) ; stuff [ i ] = java . util . Arrays . copyOf ( cur , cur . length ) ; } frame . advance ( ) ; addFrame ( fname , stuff ) ; continue ; } byte [ ] wbytes ; byte [ ] out ; switch ( opname ) { case "op_compose_ext" 3 : if ( ( functionFrames . size ( ) ) > 1 ) throw new java . lang . RuntimeException ( "can't<sp>declare<sp>func<sp>inside<sp>func" ) ; break ; case "op_sha256" : frame . enforceArgcBounds ( 0 , 3 ) ; org . provebit . proof . Digester d = new org . provebit . proof . Digester ( "op_compose_ext" 5 ) ; if ( argc == 0 ) { try { wbytes = getWorking ( ) ; d . addBytes ( wbytes ) ; } catch ( org . provebit . proof . ProofExecutor . NoWorkingBufferException ne ) { try { d . addStream ( workingStream ) ; } catch ( java . io . IOException e ) { throw new org . provebit . proof . ProofExecutor . StreamIOException ( e . getMessage ( ) , e . getCause ( ) ) ; } } } for ( int i = 0 ; i < argc ; i ++ ) { helperHash ( frame , d , ( i + 1 ) ) ; } out = d . digest ( ) ; mwrite ( org . provebit . proof . ProofExecutor . WORKING_LOCATION , out ) ; break ; case "op_cat" : frame . enforceArgcBounds ( 1 , 2 ) ; byte [ ] l ; byte [ ] r ; if ( argc == 1 ) { l = mread ( org . provebit . proof . ProofExecutor . WORKING_LOCATION ) ; r = frame . getCurrentArg ( 1 ) ; } else { l = frame . getCurrentArg ( 1 ) ; r = frame . getCurrentArg ( 2 ) ; } byte [ ] res = new byte [ ( l . length ) + ( r . length ) ] ; java . lang . System . arraycopy ( l , 0 , res , 0 , l . length ) ; java . lang . System . arraycopy ( r , 0 , res , l . length , r . length ) ; mwrite ( org . provebit . proof . ProofExecutor . WORKING_LOCATION , res ) ; break ; case "op_compose_ext" 8 : case "op_compose_ext" : frame . enforceArgcBounds ( 1 , 1 ) ; byte [ ] subarr = mread ( org . provebit . proof . ProofExecutor . WORKING_LOCATION ) ; byte [ ] mainarr ; if ( opname . equals ( "op_compose_ext" 8 ) ) mainarr = frame . getCurrentArg ( 1 ) ; else mainarr = org . provebit . proof . keysys . RootKey . keyLookup ( frame . getCurrentArgStr ( 1 ) ) ; int sub = com . google . common . primitives . Bytes . indexOf ( mainarr , subarr ) ; if ( sub == ( - 1 ) ) throw new org . provebit . proof . ProofExecutor . ProgramDieException ( "Byte<sp>subarray<sp>fail<sp>in<sp>op_compose" ) ; mwrite ( org . provebit . proof .
|
org . junit . Assert . assertEquals ( correct , outDate )
|
testShouldNotRemoveChildGroupAssociation ( ) { com . liferay . portal . kernel . model . User user = com . liferay . portal . kernel . test . util . UserTestUtil . addUser ( true ) ; java . util . List < com . liferay . portal . kernel . model . Group > groups = new java . util . ArrayList ( ) ; com . liferay . portal . kernel . model . Group parentGroup = com . liferay . portal . kernel . test . util . GroupTestUtil . addGroup ( ) ; groups . add ( parentGroup ) ; com . liferay . portal . kernel . model . Group childGroup = com . liferay . portal . kernel . test . util . GroupTestUtil . addGroup ( parentGroup . getGroupId ( ) ) ; childGroup . setMembershipRestriction ( GroupConstants . MEMBERSHIP_RESTRICTION_TO_PARENT_SITE_MEMBERS ) ; com . liferay . portal . kernel . service . GroupLocalServiceUtil . updateGroup ( childGroup ) ; groups . add ( childGroup ) ; com . liferay . portal . kernel . service . GroupLocalServiceUtil . addUserGroups ( user . getUserId ( ) , groups ) ; user = _updateUser ( user ) ; "<AssertPlaceHolder>" ; } getGroups ( ) { return _groups ; }
|
org . junit . Assert . assertEquals ( groups , user . getGroups ( ) )
|
pickDocument_WITH_RESEARCH_ARTICLE ( ) { eu . dnetlib . iis . citationmatching . direct . schemas . DocumentMetadata firstDocument = new eu . dnetlib . iis . citationmatching . direct . schemas . DocumentMetadata ( "id-1" , null , null , com . google . common . collect . Lists . newArrayList ( ) ) ; eu . dnetlib . iis . citationmatching . direct . schemas . DocumentMetadata secondDocument = new eu . dnetlib . iis . citationmatching . direct . schemas . DocumentMetadata ( "id-2" , null , "research-article" , com . google . common . collect . Lists . newArrayList ( ) ) ; eu . dnetlib . iis . citationmatching . direct . schemas . DocumentMetadata thirdDocument = new eu . dnetlib . iis . citationmatching . direct . schemas . DocumentMetadata ( "id-3" , null , null , com . google . common . collect . Lists . newArrayList ( ) ) ; java . lang . Iterable < eu . dnetlib . iis . citationmatching . direct . schemas . DocumentMetadata > documents = com . google . common . collect . Lists . newArrayList ( firstDocument , secondDocument , thirdDocument ) ; eu . dnetlib . iis . citationmatching . direct . schemas . DocumentMetadata retDocument = documentPickFunction . call ( documents ) ; "<AssertPlaceHolder>" ; } call ( java . lang . Iterable ) { java . util . Iterator < eu . dnetlib . iis . citationmatching . direct . schemas . DocumentMetadata > it = documents . iterator ( ) ; eu . dnetlib . iis . citationmatching . direct . schemas . DocumentMetadata current = null ; while ( it . hasNext ( ) ) { eu . dnetlib . iis . citationmatching . direct . schemas . DocumentMetadata docMeta = it . next ( ) ; if ( org . apache . commons . lang3 . StringUtils . equals ( docMeta . getPublicationTypeName ( ) , "research-article" ) ) { return docMeta ; } current = docMeta ; } return current ; }
|
org . junit . Assert . assertEquals ( secondDocument , retDocument )
|
testFormatSqsQueueName_HappyPath ( ) { java . lang . String expected = "prod-foo-bar" ; java . lang . String actual = com . streamreduce . util . SqsQueueNameFormatter . formatSqsQueueName ( "foo.bar" , "prod" ) ; "<AssertPlaceHolder>" ; } formatSqsQueueName ( java . lang . String , java . lang . String ) { if ( ( org . apache . commons . lang . StringUtils . isBlank ( originalQueueName ) ) || ( org . apache . commons . lang . StringUtils . isBlank ( environmentPrefix ) ) ) { throw new java . lang . IllegalArgumentException ( "queueName<sp>and<sp>environmentPrefix<sp>must<sp>be<sp>non-blank" ) ; } java . lang . String queueNameWithPrefix = ( ( com . streamreduce . util . SqsQueueNameFormatter . addMachineNameToPrefixIfNeeded ( environmentPrefix ) ) + "-" ) + originalQueueName ; java . lang . String modifiedQueueName = queueNameWithPrefix . trim ( ) . replaceAll ( "[^a-zA-Z1-9_-]" , "-" ) ; return org . apache . commons . lang . StringUtils . substring ( modifiedQueueName , 0 , com . streamreduce . util . SqsQueueNameFormatter . MAX_LENGTH_OF_SQS_QUEUE ) ; }
|
org . junit . Assert . assertEquals ( expected , actual )
|
testToObjectArrayForObjectArray ( ) { java . lang . Object [ ] array = new java . lang . Object [ getArrayLength ( ) ] ; for ( int i = 0 ; i < ( array . length ) ; i ++ ) { array [ i ] = ( cz . zcu . kiv . jop . util . ArrayUtilsTest . rand . nextBoolean ( ) ) ? new java . lang . Integer ( i ) : new java . lang . Double ( i ) ; } "<AssertPlaceHolder>" ; } toObjectArray ( boolean [ ] ) { if ( array == null ) { return null ; } if ( ( array . length ) == 0 ) { return cz . zcu . kiv . jop . util . ArrayUtils . EMPTY_BOOLEAN_OBJECT_ARRAY ; } java . lang . Boolean [ ] retArray = new java . lang . Boolean [ array . length ] ; for ( int i = 0 ; i < ( array . length ) ; i ++ ) { retArray [ i ] = new java . lang . Boolean ( array [ i ] ) ; } return retArray ; }
|
org . junit . Assert . assertArrayEquals ( array , cz . zcu . kiv . jop . util . ArrayUtils . toObjectArray ( array ) )
|
cvtConvertible_2 ( ) { final java . lang . Object x = "hej" ; final com . ericsson . otp . erlang . OtpErlangObject obj = new com . ericsson . otp . erlang . OtpErlangAtom ( "hej<sp>d" ) ; final java . lang . Object y = org . erlide . util . erlang . TypeConverter . erlang2java ( obj , org . erlide . util . TypeConverterTest . Cvt . class ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { return super . equals ( o ) ; }
|
org . junit . Assert . assertTrue ( x . equals ( y ) )
|
testSerialization ( ) { org . jfree . data . xy . VectorSeries s1 = new org . jfree . data . xy . VectorSeries ( "Series" ) ; s1 . add ( 1.0 , 1.1 , 1.2 , 1.3 ) ; org . jfree . data . xy . VectorSeriesCollection c1 = new org . jfree . data . xy . VectorSeriesCollection ( ) ; c1 . addSeries ( s1 ) ; org . jfree . data . xy . VectorSeriesCollection c2 = ( ( org . jfree . data . xy . VectorSeriesCollection ) ( org . jfree . chart . TestUtils . serialised ( c1 ) ) ) ; "<AssertPlaceHolder>" ; } serialised ( T extends java . io . Serializable ) { T result = null ; java . io . ByteArrayOutputStream buffer = new java . io . ByteArrayOutputStream ( ) ; java . io . ObjectOutput out ; try { out = new java . io . ObjectOutputStream ( buffer ) ; out . writeObject ( original ) ; out . close ( ) ; java . io . ObjectInput in = new java . io . ObjectInputStream ( new java . io . ByteArrayInputStream ( buffer . toByteArray ( ) ) ) ; result = ( ( T ) ( in . readObject ( ) ) ) ; in . close ( ) ; } catch ( java . io . IOException e ) { throw new java . lang . RuntimeException ( e ) ; } catch ( java . lang . ClassNotFoundException e ) { throw new java . lang . RuntimeException ( e ) ; } return result ; }
|
org . junit . Assert . assertEquals ( c1 , c2 )
|
testGetUserId ( ) { authentication . authenticate ( new javax . jcr . SimpleCredentials ( userId , userId . toCharArray ( ) ) ) ; "<AssertPlaceHolder>" ; } getUserId ( ) { if ( ( tokenInfo ) == null ) { throw new java . lang . IllegalStateException ( "UserId<sp>can<sp>only<sp>be<sp>retrieved<sp>after<sp>successful<sp>authentication." ) ; } return tokenInfo . getUserId ( ) ; }
|
org . junit . Assert . assertEquals ( userId , authentication . getUserId ( ) )
|
testMultipleSessionsOneTxRollbackSend ( ) { javax . jms . XAConnection conn = null ; javax . jms . Connection conn2 = null ; try { conn = xacf . createXAConnection ( ) ; conn . start ( ) ; tm . begin ( ) ; javax . jms . XASession sess1 = conn . createXASession ( ) ; org . apache . activemq . artemis . core . client . impl . ClientSessionInternal res1 = ( ( org . apache . activemq . artemis . core . client . impl . ClientSessionInternal ) ( sess1 . getXAResource ( ) ) ) ; javax . jms . XASession sess2 = conn . createXASession ( ) ; org . apache . activemq . artemis . core . client . impl . ClientSessionInternal res2 = ( ( org . apache . activemq . artemis . core . client . impl . ClientSessionInternal ) ( sess2 . getXAResource ( ) ) ) ; res1 . setForceNotSameRM ( true ) ; res2 . setForceNotSameRM ( true ) ; javax . transaction . Transaction tx = tm . getTransaction ( ) ; tx . enlistResource ( res1 ) ; tx . enlistResource ( res2 ) ; javax . jms . MessageProducer prod1 = sess1 . createProducer ( queue1 ) ; javax . jms . MessageProducer prod2 = sess2 . createProducer ( queue1 ) ; prod1 . send ( sess1 . createTextMessage ( "echidna1" ) ) ; prod2 . send ( sess2 . createTextMessage ( "echidna2" ) ) ; tx . delistResource ( res1 , XAResource . TMSUCCESS ) ; tx . delistResource ( res2 , XAResource . TMSUCCESS ) ; tm . rollback ( ) ; conn2 = cf . createConnection ( ) ; javax . jms . Session sess = conn2 . createSession ( false , Session . AUTO_ACKNOWLEDGE ) ; javax . jms . MessageConsumer cons = sess . createConsumer ( queue1 ) ; conn2 . start ( ) ; javax . jms . TextMessage r1 = ( ( javax . jms . TextMessage ) ( cons . receive ( 100 ) ) ) ; "<AssertPlaceHolder>" ; } finally { if ( conn != null ) { conn . close ( ) ; } if ( conn2 != null ) { conn2 . close ( ) ; } } } receive ( long ) { session . lock ( ) ; try { if ( ActiveMQRALogger . LOGGER . isTraceEnabled ( ) ) { ActiveMQRALogger . LOGGER . trace ( ( ( ( "receive<sp>" + ( this ) ) + "<sp>timeout=" ) + timeout ) ) ; } checkState ( ) ; javax . jms . Message message = consumer . receive ( timeout ) ; if ( ActiveMQRALogger . LOGGER . isTraceEnabled ( ) ) { ActiveMQRALogger . LOGGER . trace ( ( ( ( "received<sp>" + ( this ) ) + "<sp>result=" ) + message ) ) ; } if ( message == null ) { return null ; } else { return wrapMessage ( message ) ; } } finally { session . unlock ( ) ; } }
|
org . junit . Assert . assertNull ( r1 )
|
testTrustedClient ( ) { javax . ws . rs . client . Client client = createClientWithCertificate ( SSLCerts . DEFAULT_TRUSTSTORE . getSslContext ( ) ) ; java . lang . String returnValue = callRestService ( client ) ; "<AssertPlaceHolder>" ; } callRestService ( javax . ws . rs . client . Client ) { javax . ws . rs . client . WebTarget target = client . target ( "https://localhost:8081/test" ) ; return target . request ( ) . get ( java . lang . String . class ) ; }
|
org . junit . Assert . assertNotNull ( returnValue )
|
testAllTechniquesSetQuantile ( ) { for ( final org . hipparchus . stat . descriptive . rank . Percentile . EstimationType e : Percentile . EstimationType . values ( ) ) { reset ( 10 , e ) ; final org . hipparchus . stat . descriptive . rank . Percentile percentile = getUnivariateStatistic ( ) ; percentile . setQuantile ( 100 ) ; "<AssertPlaceHolder>" ; try { percentile . setQuantile ( 0 ) ; org . junit . Assert . fail ( "Expecting<sp>MathIllegalArgumentException" ) ; } catch ( final org . hipparchus . exception . MathIllegalArgumentException ex ) { } try { new org . hipparchus . stat . descriptive . rank . Percentile ( 0 ) ; org . junit . Assert . fail ( "Expecting<sp>MathIllegalArgumentException" ) ; } catch ( final org . hipparchus . exception . MathIllegalArgumentException ex ) { } } } getQuantile ( ) { return quantile ; }
|
org . junit . Assert . assertEquals ( 100 , percentile . getQuantile ( ) , 0 )
|
ensureFontIsMaintainedAfterBackgroundSet ( ) { java . awt . Font font = java . awt . Font . decode ( "Arial-BOLDITALIC-14" ) ; org . junit . Assume . assumeThat ( button . getFont ( ) , org . hamcrest . CoreMatchers . is ( org . hamcrest . CoreMatchers . not ( font ) ) ) ; button . setFont ( font ) ; button . setBackground ( Color . RED ) ; "<AssertPlaceHolder>" ; } getFont ( ) { return delegate . getFont ( ) ; }
|
org . junit . Assert . assertThat ( button . getFont ( ) , org . hamcrest . CoreMatchers . is ( font ) )
|
testSoundCategory ( ) { com . comphenix . protocol . events . PacketContainer container = new com . comphenix . protocol . events . PacketContainer ( PacketType . Play . Server . NAMED_SOUND_EFFECT ) ; container . getSoundCategories ( ) . write ( 0 , SoundCategory . PLAYERS ) ; "<AssertPlaceHolder>" ; } getSoundCategories ( ) { return structureModifier . withType ( com . comphenix . protocol . wrappers . EnumWrappers . getSoundCategoryClass ( ) , com . comphenix . protocol . wrappers . EnumWrappers . getSoundCategoryConverter ( ) ) ; }
|
org . junit . Assert . assertEquals ( SoundCategory . PLAYERS , container . getSoundCategories ( ) . read ( 0 ) )
|
testLatestIndicesWhereNoneExist ( ) { java . lang . String [ ] existingIndices = new java . lang . String [ ] { } ; org . apache . metron . elasticsearch . dao . ElasticsearchColumnMetadataDao dao = setup ( existingIndices ) ; java . util . List < java . lang . String > args = java . util . Arrays . asList ( "bro" , "snort" ) ; java . lang . String [ ] actual = dao . getLatestIndices ( args ) ; java . lang . String [ ] expected = new java . lang . String [ ] { } ; "<AssertPlaceHolder>" ; } getLatestIndices ( java . util . List ) { org . apache . metron . elasticsearch . dao . ElasticsearchColumnMetadataDao . LOG . debug ( "Getting<sp>latest<sp>indices;<sp>indices={}" , includeIndices ) ; java . util . Map < java . lang . String , java . lang . String > latestIndices = new java . util . HashMap ( ) ; java . lang . String [ ] indices = esClient . getIndices ( ) ; for ( java . lang . String index : indices ) { int prefixEnd = index . indexOf ( org . apache . metron . elasticsearch . dao . INDEX_NAME_DELIMITER ) ; if ( prefixEnd != ( - 1 ) ) { java . lang . String prefix = index . substring ( 0 , prefixEnd ) ; if ( includeIndices . contains ( prefix ) ) { java . lang . String latestIndex = latestIndices . get ( prefix ) ; if ( ( latestIndex == null ) || ( ( index . compareTo ( latestIndex ) ) > 0 ) ) { latestIndices . put ( prefix , index ) ; } } } } return latestIndices . values ( ) . toArray ( new java . lang . String [ latestIndices . size ( ) ] ) ; }
|
org . junit . Assert . assertArrayEquals ( expected , actual )
|
testSetColumn ( ) { org . apache . commons . math4 . linear . FieldMatrix < org . apache . commons . math4 . fraction . Fraction > m = new org . apache . commons . math4 . linear . Array2DRowFieldMatrix ( subTestData ) ; org . apache . commons . math4 . fraction . Fraction [ ] mColumn3 = columnToArray ( subColumn3 ) ; "<AssertPlaceHolder>" ; m . setColumn ( 1 , mColumn3 ) ; checkArrays ( mColumn3 , m . getColumn ( 1 ) ) ; try { m . setColumn ( ( - 1 ) , mColumn3 ) ; org . junit . Assert . fail ( "Expecting<sp>OutOfRangeException" ) ; } catch ( org . apache . commons . math4 . exception . OutOfRangeException ex ) { } try { m . setColumn ( 0 , new org . apache . commons . math4 . fraction . Fraction [ 5 ] ) ; org . junit . Assert . fail ( "Expecting<sp>MatrixDimensionMismatchException" ) ; } catch ( org . apache . commons . math4 . linear . MatrixDimensionMismatchException ex ) { } } getColumn ( int ) { org . apache . commons . math4 . linear . MatrixUtils . checkColumnIndex ( this , column ) ; final int nRows = getRowDimension ( ) ; final double [ ] out = new double [ nRows ] ; for ( int i = 0 ; i < nRows ; ++ i ) { out [ i ] = getEntry ( i , column ) ; } return out ; }
|
org . junit . Assert . assertTrue ( ( ( mColumn3 [ 0 ] ) != ( m . getColumn ( 1 ) [ 0 ] ) ) )
|
importWithoutAnyFiles ( ) { org . silverpeas . core . contribution . attachment . ActifyDocumentProcessSchedulerIT . TestContext ctx = new org . silverpeas . core . contribution . attachment . ActifyDocumentProcessSchedulerIT . TestContext ( ) . withoutAnyActifyDocuments ( ) ; org . silverpeas . core . contribution . attachment . ActifyDocumentProcessScheduler scheduler = new org . silverpeas . core . contribution . attachment . ActifyDocumentProcessScheduler ( ) ; org . silverpeas . core . scheduler . Job importer = scheduler . getActifyDocumentImporter ( ) ; importer . execute ( inAnyContext ( ) ) ; "<AssertPlaceHolder>" ; } noneActifyDocuments ( ) { java . io . File resultDir = new java . io . File ( org . silverpeas . core . contribution . attachment . ActifyDocumentProcessor . getActifyResultPath ( ) ) ; return ( ! ( resultDir . exists ( ) ) ) || ( ( resultDir . list ( ) . length ) == 0 ) ; }
|
org . junit . Assert . assertThat ( ctx . noneActifyDocuments ( ) , is ( true ) )
|
toHexNoArgTest ( ) { java . lang . String actualResult = org . openhab . binding . max . internal . Utils . toHex ( ) ; "<AssertPlaceHolder>" ; } toHex ( ) { return java . lang . String . format ( "%04X" , java . lang . Math . round ( ( ( ( ( value ) + ( org . openhab . binding . plugwise . internal . protocol . field . Humidity . OFFSET ) ) / ( org . openhab . binding . plugwise . internal . protocol . field . Humidity . MULTIPLIER ) ) * ( org . openhab . binding . plugwise . internal . protocol . field . Humidity . MAX_HEX_VALUE ) ) ) ) ; }
|
org . junit . Assert . assertEquals ( "" , actualResult )
|
originalDoesNotObserveWritesToClone ( ) { okio . Buffer original = new okio . Buffer ( ) ; okio . Buffer clone = original . clone ( ) ; clone . writeUtf8 ( "abc" ) ; "<AssertPlaceHolder>" ; } size ( ) { org . junit . Assert . assertEquals ( 0 , okio . Utf8 . size ( "" ) ) ; org . junit . Assert . assertEquals ( 3 , okio . Utf8 . size ( "abc" ) ) ; org . junit . Assert . assertEquals ( 16 , okio . Utf8 . size ( "transr" ) ) ; }
|
org . junit . Assert . assertEquals ( 0 , original . size ( ) )
|
testReActivateNotificationSourceWithReplay ( ) { final java . time . Instant lastEventTime = java . time . Instant . now ( ) ; registration . setActive ( true ) ; registration . setLastEventTime ( lastEventTime ) ; registration . reActivateNotificationSource ( ) ; "<AssertPlaceHolder>" ; verify ( mount ) . invokeCreateSubscription ( stream , java . util . Optional . of ( lastEventTime ) ) ; } isActive ( ) { return org . opendaylight . netconf . callhome . protocol . MinaSshNettyChannel . notClosing ( session ) ; }
|
org . junit . Assert . assertTrue ( registration . isActive ( ) )
|
testNonEmpty ( ) { rx . Observable < java . lang . String > source = rx . Observable . just ( "Chicago" , "Houston" , "Phoenix" ) ; final java . util . concurrent . atomic . AtomicBoolean wasCalled = new java . util . concurrent . atomic . AtomicBoolean ( false ) ; source . compose ( com . github . davidmoten . rx . Transformers . doOnEmpty ( new rx . functions . Action0 ( ) { @ com . github . davidmoten . rx . internal . operators . Override public void call ( ) { wasCalled . set ( true ) ; } } ) ) . subscribe ( ) ; "<AssertPlaceHolder>" ; } get ( ) { if ( present ) return value ; else throw new com . github . davidmoten . util . Optional . NotPresentException ( ) ; }
|
org . junit . Assert . assertFalse ( wasCalled . get ( ) )
|
knownDataTypeWithOperator ( ) { pattern . setFactType ( org . drools . workbench . screens . guided . dtable . client . widget . table . utilities . ColumnUtilitiesTest . FACT_TYPE ) ; column . setFactField ( org . drools . workbench . screens . guided . dtable . client . widget . table . utilities . ColumnUtilitiesTest . FIELD_NAME ) ; column . setOperator ( "==" ) ; when ( oracle . getFieldType ( eq ( org . drools . workbench . screens . guided . dtable . client . widget . table . utilities . ColumnUtilitiesTest . FACT_TYPE ) , eq ( org . drools . workbench . screens . guided . dtable . client . widget . table . utilities . ColumnUtilitiesTest . FIELD_NAME ) ) ) . thenReturn ( DataType . TYPE_NUMERIC_INTEGER ) ; "<AssertPlaceHolder>" ; } getType ( org . uberfire . backend . vfs . ObservablePath ) { if ( decisionTableXLSXResourceType . accept ( path ) ) { return decisionTableXLSXResourceType ; } else { return decisionTableXLSResourceType ; } }
|
org . junit . Assert . assertEquals ( DataType . TYPE_NUMERIC_INTEGER , utilities . getType ( column ) )
|
testRangedAnd ( ) { int length = 1000 ; int NUM_ITER = 10 ; org . roaringbitmap . buffer . Random random = new org . roaringbitmap . buffer . Random ( 1234 ) ; for ( int test = 0 ; test < 50 ; ++ test ) { final org . roaringbitmap . buffer . MutableRoaringBitmap rb1 = new org . roaringbitmap . buffer . MutableRoaringBitmap ( ) ; final org . roaringbitmap . buffer . MutableRoaringBitmap rb2 = new org . roaringbitmap . buffer . MutableRoaringBitmap ( ) ; org . roaringbitmap . buffer . Set < java . lang . Integer > set1 = new org . roaringbitmap . buffer . HashSet ( ) ; org . roaringbitmap . buffer . Set < java . lang . Integer > set2 = new org . roaringbitmap . buffer . HashSet ( ) ; int numBitsToSet = length / 2 ; for ( int i = 0 ; i < numBitsToSet ; i ++ ) { int val1 = random . nextInt ( length ) ; int val2 = random . nextInt ( length ) ; rb1 . add ( val1 ) ; set1 . add ( val1 ) ; rb2 . add ( val2 ) ; set2 . add ( val2 ) ; } org . roaringbitmap . buffer . Set < java . lang . Integer > intersectionSet = new org . roaringbitmap . buffer . TreeSet ( set1 ) ; intersectionSet . retainAll ( set2 ) ; for ( int iter = 0 ; iter < NUM_ITER ; iter ++ ) { int rangeStart = random . nextInt ( ( length - 1 ) ) ; long rangeLength = ( random . nextInt ( ( length - rangeStart ) ) ) + 1 ; long rangeEnd = rangeStart + rangeLength ; org . roaringbitmap . buffer . Set < java . lang . Integer > expectedResultSet = new org . roaringbitmap . buffer . TreeSet ( ) ; for ( int i = rangeStart ; i < rangeEnd ; i ++ ) { if ( intersectionSet . contains ( i ) ) { expectedResultSet . add ( i ) ; } } org . roaringbitmap . buffer . List < org . roaringbitmap . buffer . ImmutableRoaringBitmap > list = new org . roaringbitmap . buffer . ArrayList ( ) ; list . add ( rb1 ) ; list . add ( rb2 ) ; org . roaringbitmap . buffer . MutableRoaringBitmap result = org . roaringbitmap . buffer . ImmutableRoaringBitmap . and ( list . iterator ( ) , rangeStart , rangeEnd ) ; org . roaringbitmap . buffer . Set < java . lang . Integer > actualResultSet = new org . roaringbitmap . buffer . TreeSet ( ) ; org . roaringbitmap . IntIterator intIterator = result . getIntIterator ( ) ; while ( intIterator . hasNext ( ) ) { actualResultSet . add ( intIterator . next ( ) ) ; } "<AssertPlaceHolder>" ; } } } next ( ) { return iterator . next ( ) ; }
|
org . junit . Assert . assertEquals ( expectedResultSet , actualResultSet )
|
excludeWithoutCondition ( ) { org . apache . deltaspike . test . core . api . exclude . NoBean noBean = org . apache . deltaspike . core . api . provider . BeanProvider . getContextualReference ( org . apache . deltaspike . test . core . api . exclude . NoBean . class , true ) ; "<AssertPlaceHolder>" ; }
|
org . junit . Assert . assertNull ( noBean )
|
testMapWithOptionalValue ( ) { java . util . List < org . apache . parquet . schema . Type > typeList = new java . util . ArrayList < org . apache . parquet . schema . Type > ( ) ; typeList . add ( new org . apache . parquet . schema . PrimitiveType ( REQUIRED , INT64 , "key" ) ) ; typeList . add ( new org . apache . parquet . schema . PrimitiveType ( OPTIONAL , INT64 , "value" ) ) ; org . apache . parquet . schema . GroupType map = new org . apache . parquet . schema . GroupType ( REQUIRED , "myMap" , org . apache . parquet . schema . OriginalType . OriginalType . MAP , new org . apache . parquet . schema . GroupType ( REPEATED , "map" , typeList ) ) ; org . apache . parquet . schema . MessageType expected = new org . apache . parquet . schema . MessageType ( "mapParent" , map ) ; org . apache . parquet . schema . GroupType actual = org . apache . parquet . schema . Types . buildMessage ( ) . requiredMap ( ) . key ( org . apache . parquet . schema . INT64 ) . optionalValue ( org . apache . parquet . schema . INT64 ) . named ( "myMap" ) . named ( "mapParent" ) ; "<AssertPlaceHolder>" ; } add ( org . apache . parquet . hadoop . ParquetFileReader$ChunkDescriptor , java . util . List , org . apache . parquet . io . SeekableInputStream ) { org . apache . parquet . hadoop . ParquetFileReader . ChunkListBuilder . ChunkData data = map . get ( descriptor ) ; if ( data == null ) { data = new org . apache . parquet . hadoop . ParquetFileReader . ChunkListBuilder . ChunkData ( ) ; map . put ( descriptor , data ) ; } data . buffers . addAll ( buffers ) ; lastDescriptor = descriptor ; this . f = f ; }
|
org . junit . Assert . assertEquals ( expected , actual )
|
getSubscriptionsForMyCustomersFilteredBySubscriptionIdNoSubscriptionReturned ( ) { final java . lang . String not_existing_subscription = "not<sp>existing" ; final int expected = 0 ; java . util . Set < org . oscm . paginator . Filter > filterSet = createFilterSet ( not_existing_subscription , null , null , null , null ) ; final org . oscm . paginator . Pagination pagination = createPagination ( 0 , NUM_CUSTOMER_SUBSCRIPTIONS , null , filterSet ) ; java . util . List < org . oscm . domobjects . Subscription > result = runTX ( new java . util . concurrent . Callable < java . util . List < org . oscm . domobjects . Subscription > > ( ) { @ org . oscm . subscriptionservice . dao . Override public java . util . List < org . oscm . domobjects . Subscription > call ( ) throws org . oscm . subscriptionservice . dao . Exception { return dao . getSubscriptionsForMyCustomers ( supplierUser , states , pagination ) ; } } ) ; "<AssertPlaceHolder>" ; } size ( ) { return categoriesForMarketplace . size ( ) ; }
|
org . junit . Assert . assertEquals ( expected , result . size ( ) )
|
findAllLand ( ) { final java . util . Collection < nl . bzk . algemeenbrp . dal . domein . brp . entity . LandOfGebied > result = landOfGebiedRepository . findAll ( ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return elementen . isEmpty ( ) ; }
|
org . junit . Assert . assertFalse ( result . isEmpty ( ) )
|
testMatchRecognize ( ) { final org . apache . calcite . tools . RelBuilder builder = org . apache . calcite . tools . RelBuilder . create ( org . apache . calcite . test . RelBuilderTest . config ( ) . build ( ) ) . scan ( "EMP" ) ; final org . apache . calcite . rel . type . RelDataTypeFactory typeFactory = builder . getTypeFactory ( ) ; final org . apache . calcite . rel . type . RelDataType intType = typeFactory . createSqlType ( SqlTypeName . INTEGER ) ; org . apache . calcite . rex . RexNode pattern = builder . patternConcat ( builder . literal ( "STRT" ) , builder . patternQuantify ( builder . literal ( "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 2 ) , builder . literal ( 1 ) , builder . literal ( ( - 1 ) ) , builder . literal ( false ) ) , builder . patternQuantify ( builder . literal ( "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 9 ) , builder . literal ( 1 ) , builder . literal ( ( - 1 ) ) , builder . literal ( false ) ) ) ; ImmutableMap . Builder < java . lang . String , org . apache . calcite . rex . RexNode > pdBuilder = new com . google . common . collect . ImmutableMap . Builder < > ( ) ; org . apache . calcite . rex . RexNode downDefinition = builder . call ( SqlStdOperatorTable . LESS_THAN , builder . call ( SqlStdOperatorTable . PREV , builder . patternField ( "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 2 , intType , 3 ) , builder . literal ( 0 ) ) , builder . call ( SqlStdOperatorTable . PREV , builder . patternField ( "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 2 , intType , 3 ) , builder . literal ( 1 ) ) ) ; pdBuilder . put ( "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 2 , downDefinition ) ; org . apache . calcite . rex . RexNode upDefinition = builder . call ( SqlStdOperatorTable . GREATER_THAN , builder . call ( SqlStdOperatorTable . PREV , builder . patternField ( "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 9 , intType , 3 ) , builder . literal ( 0 ) ) , builder . call ( SqlStdOperatorTable . PREV , builder . patternField ( "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 9 , intType , 3 ) , builder . literal ( 1 ) ) ) ; pdBuilder . put ( "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 9 , upDefinition ) ; com . google . common . collect . ImmutableList . Builder < org . apache . calcite . rex . RexNode > measuresBuilder = new com . google . common . collect . ImmutableList . Builder < > ( ) ; measuresBuilder . add ( builder . alias ( builder . patternField ( "STRT" , intType , 3 ) , "start_nw" ) ) ; measuresBuilder . add ( builder . alias ( builder . call ( SqlStdOperatorTable . LAST , builder . patternField ( "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 2 , intType , 3 ) , builder . literal ( 0 ) ) , "bottom_nw" ) ) ; org . apache . calcite . rex . RexNode after = builder . getRexBuilder ( ) . makeFlag ( SqlMatchRecognize . AfterOption . SKIP_TO_NEXT_ROW ) ; com . google . common . collect . ImmutableList . Builder < org . apache . calcite . rex . RexNode > partitionKeysBuilder = new com . google . common . collect . ImmutableList . Builder < > ( ) ; partitionKeysBuilder . add ( builder . field ( "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 1 ) ) ; com . google . common . collect . ImmutableList . Builder < org . apache . calcite . rex . RexNode > orderKeysBuilder = new com . google . common . collect . ImmutableList . Builder < > ( ) ; orderKeysBuilder . add ( builder . field ( "EMPNO" ) ) ; org . apache . calcite . rex . RexNode interval = builder . literal ( "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 7 ) ; final com . google . common . collect . ImmutableMap < java . lang . String , java . util . TreeSet < java . lang . String > > subsets = com . google . common . collect . ImmutableMap . of ( ) ; final org . apache . calcite . rel . RelNode root = builder . match ( pattern , false , false , pdBuilder . build ( ) , measuresBuilder . build ( ) , after , subsets , false , partitionKeysBuilder . build ( ) , orderKeysBuilder . build ( ) , interval ) . build ( ) ; final java . lang . String expected = "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 8 + ( ( ( ( ( ( ( ( ( "outputFields=[[$7,<sp>'start_nw',<sp>'bottom_nw']],<sp>allRows=[false],<sp>" + "after=[FLAG(SKIP<sp>TO<sp>NEXT<sp>ROW)],<sp>pattern=[(('STRT',<sp>" ) + "PATTERN_QUANTIFIER('DOWN',<sp>1,<sp>-1,<sp>false)),<sp>" ) + "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 6 ) + "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 3 ) + "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 4 ) + "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" ) + "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 5 ) + "patternDefinitions=[[<(PREV(DOWN.$3,<sp>0),<sp>PREV(DOWN.$3,<sp>1)),<sp>" 0 ) + "<sp>LogicalTableScan(table=[[scott,<sp>EMP]])\n" ) ; "<AssertPlaceHolder>" ; } hasTree ( java . lang . String ) { return org . apache . calcite . test . Matchers . compose ( org . hamcrest . core . Is . is ( value ) , ( input ) -> { return org . apache . calcite . util . Util . toLinux ( org . apache . calcite . plan . RelOptUtil . toString ( input ) ) ; } ) ; }
|
org . junit . Assert . assertThat ( root , org . apache . calcite . test . Matchers . hasTree ( expected ) )
|
testSparseRecursiveEvaluate2 ( ) { org . apache . commons . math3 . random . RandomDataGenerator rnd = getRandomData ( ) ; cc . redberry . rings . Ring < cc . redberry . rings . bigint . BigInteger > ring = cc . redberry . rings . Rings . Zp ( 17 ) ; org . apache . commons . math3 . stat . descriptive . DescriptiveStatistics recStat = new org . apache . commons . math3 . stat . descriptive . DescriptiveStatistics ( ) ; org . apache . commons . math3 . stat . descriptive . DescriptiveStatistics recEvalStat = new org . apache . commons . math3 . stat . descriptive . DescriptiveStatistics ( ) ; org . apache . commons . math3 . stat . descriptive . DescriptiveStatistics plainStat = new org . apache . commons . math3 . stat . descriptive . DescriptiveStatistics ( ) ; long start ; long elapsed ; int nIterations = 100 ; int nVars = 3 ; int minDeg = 30 ; int minSize = 1000 ; for ( int i = 0 ; i < nIterations ; ++ i ) { if ( i == ( nIterations / 10 ) ) java . util . Arrays . asList ( recStat , recEvalStat , plainStat ) . forEach ( DescriptiveStatistics :: clear ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomial . MultivariatePolynomial < cc . redberry . rings . bigint . BigInteger > p = cc . redberry . rings . poly . multivar . RandomMultivariatePolynomials . randomPolynomial ( nVars , rnd . nextInt ( minDeg , ( 2 * minDeg ) ) , rnd . nextInt ( minSize , ( 2 * minSize ) ) , ring , MonomialOrder . DEFAULT , rnd . getRandomGenerator ( ) ) ; cc . redberry . rings . bigint . BigInteger [ ] values = new cc . redberry . rings . bigint . BigInteger [ p . nVariables ] ; for ( int j = 0 ; j < ( values . length ) ; j ++ ) values [ j ] = ring . randomElement ( rnd . getRandomGenerator ( ) ) ; start = java . lang . System . nanoTime ( ) ; cc . redberry . rings . poly . multivar . AMultivariatePolynomial recForm = p . toSparseRecursiveForm ( ) ; elapsed = ( java . lang . System . nanoTime ( ) ) - start ; recStat . addValue ( elapsed ) ; start = java . lang . System . nanoTime ( ) ; cc . redberry . rings . bigint . BigInteger recVal = evaluateSparseRecursiveForm ( recForm , p . nVariables , values ) ; elapsed = ( java . lang . System . nanoTime ( ) ) - start ; recStat . addValue ( elapsed ) ; recEvalStat . addValue ( elapsed ) ; start = java . lang . System . nanoTime ( ) ; cc . redberry . rings . bigint . BigInteger plainVal = p . evaluate ( values ) ; elapsed = ( java . lang . System . nanoTime ( ) ) - start ; plainStat . addValue ( elapsed ) ; "<AssertPlaceHolder>" ; } System . out . println ( ( "Recursive<sp>:<sp>" + ( statisticsNanotime ( recStat ) ) ) ) ; System . out . println ( ( "Recursive<sp>(eval<sp>only)<sp>:<sp>" + ( statisticsNanotime ( recEvalStat ) ) ) ) ; System . out . println ( ( "Plain<sp>:<sp>" + ( statisticsNanotime ( plainStat ) ) ) ) ; } evaluate ( long ) { evaluationPoint [ ( ( evaluationPoint . length ) - 1 ) ] = newPoint ; powers . set ( evaluationVariables [ ( ( evaluationVariables . length ) - 1 ) ] , newPoint ) ; return evaluate0 ( newPoint ) ; }
|
org . junit . Assert . assertEquals ( plainVal , recVal )
|
shouldBeAbleToChangeTheUuidOfUnlabeledNodeWithLabelConfiguration ( ) { org . neo4j . graphdb . Node node ; registerModuleWithLabelsAndTypes ( ) ; try ( org . neo4j . graphdb . Transaction tx = database . beginTx ( ) ) { node = database . createNode ( ) ; node . setProperty ( "name" , "aNode" ) ; tx . success ( ) ; } try ( org . neo4j . graphdb . Transaction tx = database . beginTx ( ) ) { for ( org . neo4j . graphdb . Node n : database . getAllNodes ( ) ) { if ( ! ( node . getLabels ( ) . iterator ( ) . hasNext ( ) ) ) { n . setProperty ( uuidConfiguration . getUuidProperty ( ) , "aNewUuid" ) ; } } tx . success ( ) ; } try ( org . neo4j . graphdb . Transaction tx = database . beginTx ( ) ) { for ( org . neo4j . graphdb . Node n : database . getAllNodes ( ) ) { if ( ! ( node . getLabels ( ) . iterator ( ) . hasNext ( ) ) ) { "<AssertPlaceHolder>" ; } } tx . success ( ) ; } } getUuidProperty ( ) { return uuidProperty ; }
|
org . junit . Assert . assertEquals ( "aNewUuid" , n . getProperty ( uuidConfiguration . getUuidProperty ( ) ) )
|
testCreate ( ) { org . oscarehr . eyeform . model . EyeformOcularProcedure entity = new org . oscarehr . eyeform . model . EyeformOcularProcedure ( ) ; org . oscarehr . common . dao . utils . EntityDataGenerator . generateTestDataForModelClass ( entity ) ; dao . persist ( entity ) ; "<AssertPlaceHolder>" ; } getId ( ) { return this . id ; }
|
org . junit . Assert . assertNotNull ( entity . getId ( ) )
|
testInitLevel1ThrowUnhandledException ( ) { org . camunda . bpm . engine . runtime . ProcessInstance instance = rule . processInstance ( ) ; rule . messageCorrelation ( "InnerEventSubProcessMessage" ) . correlate ( ) ; org . camunda . bpm . engine . task . Task innerEventSubprocessTask = rule . taskQuery ( ) . taskDefinitionKey ( "innerEventSubProcessTask" ) . singleResult ( ) ; rule . getRuntimeService ( ) . setVariable ( instance . getId ( ) , ThrowBpmnErrorDelegate . EXCEPTION_INDICATOR_VARIABLE , true ) ; rule . getRuntimeService ( ) . setVariable ( instance . getId ( ) , ThrowBpmnErrorDelegate . EXCEPTION_MESSAGE_VARIABLE , "unhandledException" ) ; try { rule . getTaskService ( ) . complete ( innerEventSubprocessTask . getId ( ) ) ; org . junit . Assert . fail ( "should<sp>throw<sp>a<sp>ThrowBpmnErrorDelegateException" ) ; } catch ( org . camunda . bpm . qa . upgrade . util . ThrowBpmnErrorDelegate . ThrowBpmnErrorDelegateException e ) { "<AssertPlaceHolder>" ; } } getMessage ( ) { return message ; }
|
org . junit . Assert . assertEquals ( "unhandledException" , e . getMessage ( ) )
|
testPerProfileValues ( ) { de . akquinet . engineering . vaadinator . model . BeanArtifact [ ] perProfileValues = de . akquinet . engineering . vaadinator . model . BeanArtifact . perProfileValues ( ) ; "<AssertPlaceHolder>" ; } perProfileValues ( ) { de . akquinet . engineering . vaadinator . model . BeanArtifact [ ] perProfileValues = new de . akquinet . engineering . vaadinator . model . BeanArtifact [ ] { } ; for ( de . akquinet . engineering . vaadinator . model . BeanArtifact beanArtifact : de . akquinet . engineering . vaadinator . model . BeanArtifact . values ( ) ) { if ( ( beanArtifact . isPresenter ( ) ) || ( beanArtifact . isView ( ) ) ) { perProfileValues = ( ( de . akquinet . engineering . vaadinator . model . BeanArtifact [ ] ) ( org . apache . commons . lang . ArrayUtils . add ( perProfileValues , beanArtifact ) ) ) ; } } return perProfileValues ; }
|
org . junit . Assert . assertThat ( perProfileValues . length , org . hamcrest . CoreMatchers . is ( 20 ) )
|
testEqualsTargetIdxPOSFalse ( ) { net . sf . extjwnl . data . Synset s = new net . sf . extjwnl . data . Synset ( dictionary , POS . NOUN , 1 ) ; net . sf . extjwnl . data . Pointer p = new net . sf . extjwnl . data . Pointer ( s , PointerType . ANTONYM , POS . NOUN , 10 , 0 ) ; net . sf . extjwnl . data . Pointer pp = new net . sf . extjwnl . data . Pointer ( s , PointerType . ANTONYM , POS . ADJECTIVE , 10 , 0 ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { return ( ( object instanceof net . sf . extjwnl . data . Synset ) && ( ( ( net . sf . extjwnl . data . Synset ) ( object ) ) . getPOS ( ) . equals ( getPOS ( ) ) ) ) && ( ( ( ( net . sf . extjwnl . data . Synset ) ( object ) ) . getOffset ( ) ) == ( getOffset ( ) ) ) ; }
|
org . junit . Assert . assertFalse ( p . equals ( pp ) )
|
parseWithinFilter ( ) { org . deegree . filter . Filter filter = testImportExportImport ( "testfilter24.xml" ) ; "<AssertPlaceHolder>" ; } testImportExportImport ( java . lang . String ) { java . net . URL url = org . deegree . filter . xml . Filter110XMLEncoderTest . class . getResource ( ( "v110/" + resource ) ) ; javax . xml . stream . XMLStreamReader in = javax . xml . stream . XMLInputFactory . newInstance ( ) . createXMLStreamReader ( url . toString ( ) , url . openStream ( ) ) ; in . nextTag ( ) ; org . deegree . filter . Filter filter = org . deegree . filter . xml . Filter110XMLDecoder . parse ( in ) ; java . io . ByteArrayOutputStream bos = new java . io . ByteArrayOutputStream ( ) ; javax . xml . stream . XMLStreamWriter out = javax . xml . stream . XMLOutputFactory . newInstance ( ) . createXMLStreamWriter ( bos ) ; org . deegree . filter . xml . Filter110XMLEncoder . export ( filter , out ) ; out . close ( ) ; in = javax . xml . stream . XMLInputFactory . newInstance ( ) . createXMLStreamReader ( new java . io . ByteArrayInputStream ( bos . toByteArray ( ) ) ) ; in . nextTag ( ) ; System . out . println ( new java . lang . String ( bos . toByteArray ( ) ) ) ; return org . deegree . filter . xml . Filter110XMLDecoder . parse ( in ) ; }
|
org . junit . Assert . assertNotNull ( filter )
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.