input
stringlengths 28
18.7k
| output
stringlengths 39
1.69k
|
|---|---|
getOffsetTime ( ) { final java . time . OffsetTime value = java . time . OffsetTime . now ( ) ; final io . trane . ndbc . postgres . value . OffsetTimeValue wrapper = new io . trane . ndbc . postgres . value . OffsetTimeValue ( value ) ; "<AssertPlaceHolder>" ; } getOffsetTime ( ) { return cantRead ( "OffsetTime" ) ; }
|
org . junit . Assert . assertEquals ( value , wrapper . getOffsetTime ( ) )
|
test_BrowserFunction_callback_afterPageReload ( ) { org . junit . Assume . assumeFalse ( webkit1SkipMsg ( ) , isWebkit1 ) ; java . util . concurrent . atomic . AtomicBoolean javaCallbackExecuted = new java . util . concurrent . atomic . AtomicBoolean ( false ) ; java . util . concurrent . atomic . AtomicInteger callCount = new java . util . concurrent . atomic . AtomicInteger ( 0 ) ; class JavascriptCallback extends org . eclipse . swt . browser . BrowserFunction { JavascriptCallback ( org . eclipse . swt . browser . Browser browser , java . lang . String name ) { ( browser , name ) ; } @ org . eclipse . swt . tests . junit . Override public java . lang . Object function ( java . lang . Object [ ] arguments ) { if ( ( callCount . get ( ) ) == 0 ) { callCount . set ( 1 ) ; browser . setText ( "2nd<sp>page<sp>load" ) ; } else { javaCallbackExecuted . set ( true ) ; } return null ; } } browser . setText ( "1st<sp>(initial)<sp>page<sp>load" ) ; new JavascriptCallback ( browser , "jsCallbackToJava" ) ; browser . execute ( "jsCallbackToJava()" ) ; browser . addProgressListener ( org . eclipse . swt . browser . ProgressListener . completedAdapter ( ( e ) -> browser . execute ( "jsCallbackToJava()" ) ) ) ; shell . open ( ) ; boolean passed = waitForPassCondition ( javaCallbackExecuted :: get ) ; java . lang . String message = "A<sp>javascript<sp>callback<sp>should<sp>work<sp>after<sp>a<sp>page<sp>has<sp>been<sp>reloaded.<sp>But<sp>something<sp>went<sp>wrong" ; "<AssertPlaceHolder>" ; } waitForPassCondition ( java . util . function . Supplier ) { return waitForPassCondition ( passTest , ( 1000 * ( secondsToWaitTillFail ) ) ) ; }
|
org . junit . Assert . assertTrue ( message , passed )
|
shouldApplyNeoStoreCommandToTheStoreInRecovery ( ) { final org . neo4j . kernel . impl . api . BatchTransactionApplier applier = newApplier ( true ) ; final org . neo4j . kernel . impl . store . record . NeoStoreRecord before = new org . neo4j . kernel . impl . store . record . NeoStoreRecord ( ) ; final org . neo4j . kernel . impl . store . record . NeoStoreRecord after = new org . neo4j . kernel . impl . store . record . NeoStoreRecord ( ) ; after . setNextProp ( 42 ) ; final org . neo4j . kernel . impl . transaction . command . Command command = new org . neo4j . kernel . impl . transaction . command . Command . NeoStoreCommand ( before , after ) ; boolean result = apply ( applier , command :: handle , transactionToApply ) ; "<AssertPlaceHolder>" ; verify ( metaDataStore , times ( 1 ) ) . setGraphNextProp ( after . getNextProp ( ) ) ; } apply ( org . neo4j . kernel . api . proc . Context , java . lang . Object [ ] , org . neo4j . kernel . api . ResourceTracker ) { try { java . lang . Thread . sleep ( 50 ) ; } catch ( java . lang . InterruptedException e ) { throw new org . neo4j . internal . kernel . api . exceptions . ProcedureException ( Status . General . UnknownError , e , "Interrupted" ) ; } return org . neo4j . collection . RawIterator . empty ( ) ; }
|
org . junit . Assert . assertFalse ( result )
|
testEmptyFunnel ( ) { com . yahoo . hive . udf . funnel . Conversion udf = new com . yahoo . hive . udf . funnel . Conversion ( ) ; org . apache . hadoop . hive . serde2 . objectinspector . ObjectInspector [ ] inputOiList = new org . apache . hadoop . hive . serde2 . objectinspector . ObjectInspector [ ] { org . apache . hadoop . hive . serde2 . objectinspector . ObjectInspectorFactory . getStandardListObjectInspector ( PrimitiveObjectInspectorFactory . javaLongObjectInspector ) } ; udf . initialize ( inputOiList ) ; java . util . List < java . lang . Long > inputList = java . util . Arrays . asList ( ) ; org . apache . hadoop . hive . ql . udf . generic . GenericUDF . DeferredObject obj1 = mock ( org . apache . hadoop . hive . ql . udf . generic . GenericUDF . DeferredObject . class ) ; org . apache . hadoop . hive . ql . udf . generic . GenericUDF . DeferredObject [ ] objs = new org . apache . hadoop . hive . ql . udf . generic . GenericUDF . DeferredObject [ ] { obj1 } ; when ( obj1 . get ( ) ) . thenReturn ( inputList ) ; "<AssertPlaceHolder>" ; } evaluate ( com . yahoo . hive . udf . funnel . DeferredObject [ ] ) { if ( ( args . length ) != 1 ) { return null ; } java . util . List < java . lang . Long > funnel = ( ( java . util . List < java . lang . Long > ) ( listInputObjectInspector . getList ( args [ 0 ] . get ( ) ) ) ) ; java . util . List < java . lang . Double > result = new java . util . ArrayList ( ) ; if ( ( funnel . size ( ) ) <= 0 ) { return result ; } result . add ( 1.0 ) ; for ( int i = 1 ; i < ( funnel . size ( ) ) ; i ++ ) { if ( ( ( funnel . get ( i ) ) <= 0 ) || ( ( funnel . get ( ( i - 1 ) ) ) <= 0 ) ) { result . add ( 0.0 ) ; } else { result . add ( ( ( ( double ) ( funnel . get ( i ) ) ) / ( funnel . get ( ( i - 1 ) ) ) ) ) ; } } return result ; }
|
org . junit . Assert . assertEquals ( java . util . Arrays . asList ( ) , udf . evaluate ( objs ) )
|
shouldBeAbleToInjectInitialContextContextQualified ( ) { javax . naming . Context contextZ = new javax . naming . InitialContext ( ) ; javax . naming . Context contextX = new javax . naming . InitialContext ( ) ; org . jboss . arquillian . container . test . impl . enricher . resource . InitialContextProviderTestCase . InitialContextClassQualifed test = execute ( org . jboss . arquillian . container . test . impl . enricher . resource . InitialContextProviderTestCase . InitialContextClassQualifed . class , javax . naming . Context . class , contextZ , contextX ) ; "<AssertPlaceHolder>" ; }
|
org . junit . Assert . assertEquals ( contextX , test . context )
|
testGetXmlProlog_NoEncoding ( ) { final java . io . InputStream input = com . rometools . rome . io . XmlReaderTest . stringToStream ( "<?xml>" , "UTF-8" ) ; final java . lang . String guessedEncoding = "UTF-8" ; "<AssertPlaceHolder>" ; } getXmlProlog ( java . io . InputStream , java . lang . String ) { java . lang . String encoding = null ; if ( guessedEnc != null ) { final byte [ ] bytes = new byte [ com . rometools . rome . io . XmlReader . BUFFER_SIZE ] ; is . mark ( com . rometools . rome . io . XmlReader . BUFFER_SIZE ) ; int offset = 0 ; int max = com . rometools . rome . io . XmlReader . BUFFER_SIZE ; int c = is . read ( bytes , offset , max ) ; int firstGT = - 1 ; while ( ( ( c != ( - 1 ) ) && ( firstGT == ( - 1 ) ) ) && ( offset < ( com . rometools . rome . io . XmlReader . BUFFER_SIZE ) ) ) { offset += c ; max -= c ; c = is . read ( bytes , offset , max ) ; firstGT = new java . lang . String ( bytes , 0 , offset , guessedEnc ) . indexOf ( ">" ) ; } if ( firstGT == ( - 1 ) ) { if ( c == ( - 1 ) ) { throw new java . io . IOException ( "Unexpected<sp>end<sp>of<sp>XML<sp>stream" ) ; } else { throw new java . io . IOException ( ( ( "XML<sp>prolog<sp>or<sp>ROOT<sp>element<sp>not<sp>found<sp>on<sp>first<sp>" + offset ) + "<sp>bytes" ) ) ; } } final int bytesRead = offset ; if ( bytesRead > 0 ) { is . reset ( ) ; java . lang . String prolog = new java . lang . String ( bytes , guessedEnc ) . substring ( 0 , firstGT ) ; final java . util . regex . Matcher m = com . rometools . rome . io . XmlReader . ENCODING_PATTERN . matcher ( prolog ) ; if ( m . find ( ) ) { encoding = m . group ( 1 ) . toUpperCase ( Locale . ENGLISH ) ; encoding = encoding . substring ( 1 , ( ( encoding . length ( ) ) - 1 ) ) ; } } } return encoding ; }
|
org . junit . Assert . assertNull ( com . rometools . rome . io . XmlReader . getXmlProlog ( input , guessedEncoding ) )
|
testMongoEventMapper ( ) { org . ff4j . audit . Event evt = new org . ff4j . audit . Event ( "JAVA" , org . ff4j . audit . EventConstants . TARGET_FEATURE , "toto" , org . ff4j . audit . EventConstants . ACTION_CHECK_OFF ) ; org . ff4j . mongo . mapper . MongoEventMapper mem = new org . ff4j . mongo . mapper . MongoEventMapper ( ) ; org . ff4j . audit . Event evt2 = mem . fromStore ( mem . toStore ( evt ) ) ; "<AssertPlaceHolder>" ; } getUuid ( ) { return uuid ; }
|
org . junit . Assert . assertEquals ( evt . getUuid ( ) , evt2 . getUuid ( ) )
|
findAll ( ) { java . util . List < com . querydsl . example . dto . Order > orders = orderDao . findAll ( ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return ( values . isEmpty ( ) ) && ( batches . isEmpty ( ) ) ; }
|
org . junit . Assert . assertFalse ( orders . isEmpty ( ) )
|
testSetDonor_boolean ( ) { boolean donor = true ; org . openscience . cdk . interfaces . IAtomType a = ( ( org . openscience . cdk . interfaces . IAtomType ) ( newChemObject ( ) ) ) ; a . setFlag ( CDKConstants . IS_HYDROGENBOND_DONOR , donor ) ; "<AssertPlaceHolder>" ; } getFlag ( int ) { boolean val = false ; boolean leftFlag = left . getFlag ( flagType ) ; if ( ( right ) != null ) { if ( ( "and" . equals ( operator ) ) && leftFlag ) { boolean rightFlag = right . getFlag ( flagType ) ; val = leftFlag && rightFlag ; } else if ( "or" . equals ( operator ) ) { boolean rightFlag = right . getFlag ( flagType ) ; val = leftFlag || rightFlag ; } } else { if ( "not" . equals ( operator ) ) { val = ! leftFlag ; } else { val = leftFlag ; } } return val ; }
|
org . junit . Assert . assertTrue ( a . getFlag ( CDKConstants . IS_HYDROGENBOND_DONOR ) )
|
testRemovePhaseListener_withUnknownPhaseListener ( ) { phaseListenerManager . addPhaseListener ( new org . eclipse . rap . rwt . testfixture . internal . LoggingPhaseListener ( null ) ) ; phaseListenerManager . removePhaseListener ( new org . eclipse . rap . rwt . testfixture . internal . LoggingPhaseListener ( null ) ) ; "<AssertPlaceHolder>" ; } getPhaseListeners ( ) { return asList ( phaseListenerManager . getPhaseListeners ( ) ) ; }
|
org . junit . Assert . assertEquals ( 1 , getPhaseListeners ( ) . size ( ) )
|
runTestMultidimensionalArray1 ( ) { soot . jimple . infoflow . results . InfoflowResults res = analyzeAPKFile ( "ArraysAndLists/MultidimensionalArray1.apk" ) ; "<AssertPlaceHolder>" ; } size ( ) { return set . size ( ) ; }
|
org . junit . Assert . assertEquals ( 1 , res . size ( ) )
|
deleteHistoricDecisionInstanceByInstanceIdWithAuthenticatedTenant ( ) { testRule . deployForTenant ( org . camunda . bpm . engine . test . api . multitenancy . tenantcheck . MultiTenancyHistoricDataCmdsTenantCheckTest . TENANT_ONE , org . camunda . bpm . engine . test . api . multitenancy . tenantcheck . MultiTenancyHistoricDataCmdsTenantCheckTest . DMN ) ; evaluateDecisionTable ( null ) ; org . camunda . bpm . engine . history . HistoricDecisionInstanceQuery query = historyService . createHistoricDecisionInstanceQuery ( ) ; org . camunda . bpm . engine . history . HistoricDecisionInstance historicDecisionInstance = query . includeInputs ( ) . includeOutputs ( ) . singleResult ( ) ; identityService . setAuthentication ( "user" , null , java . util . Arrays . asList ( org . camunda . bpm . engine . test . api . multitenancy . tenantcheck . MultiTenancyHistoricDataCmdsTenantCheckTest . TENANT_ONE ) ) ; historyService . deleteHistoricDecisionInstanceByInstanceId ( historicDecisionInstance . getId ( ) ) ; identityService . clearAuthentication ( ) ; "<AssertPlaceHolder>" ; } count ( ) { this . resultType = org . camunda . bpm . engine . impl . AbstractNativeQuery . ResultType . COUNT ; if ( ( commandExecutor ) != null ) { return ( ( java . lang . Long ) ( commandExecutor . execute ( this ) ) ) ; } return executeCount ( org . camunda . bpm . engine . impl . context . Context . getCommandContext ( ) , getParameterMap ( ) ) ; }
|
org . junit . Assert . assertThat ( query . count ( ) , org . hamcrest . CoreMatchers . is ( 0L ) )
|
getServletConfigClassesTest ( ) { System . out . print ( "-><sp>getServletConfigClasses()<sp>-<sp>" ) ; ua . com . alexcoffee . config . AppInitializer appInitializer = new ua . com . alexcoffee . config . AppInitializer ( ) ; "<AssertPlaceHolder>" ; System . out . println ( "OK!" ) ; } getServletConfigClasses ( ) { return new java . lang . Class < ? > [ ] { ua . com . alexcoffee . config . WebConfig . class } ; }
|
org . junit . Assert . assertNotNull ( appInitializer . getServletConfigClasses ( ) )
|
convertToLongNull ( ) { java . lang . Long l = converter . toLong ( null ) ; "<AssertPlaceHolder>" ; } toLong ( java . lang . Long ) { return value != null ? new java . lang . Long ( value ) : null ; }
|
org . junit . Assert . assertNull ( l )
|
testToStringTransformation ( ) { final java . util . Collection < java . lang . Integer > origin = create ( closed ( 1 , 3 ) , integers ( ) ) ; final java . lang . Integer [ ] ints = origin . toArray ( new java . lang . Integer [ origin . size ( ) ] ) ; final java . util . Collection < java . lang . String > actual = newHashSet ( org . eclipse . n4js . utils . collections . Arrays2 . transform ( ints , ( input ) -> input . toString ( ) ) ) ; final java . util . Collection < java . lang . String > expected = newHashSet ( "1" , "2" , "3" ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( ( ( name ) + "(" ) + ( start ) ) + "-" ) + ( end ) ) + ")" ; }
|
org . junit . Assert . assertEquals ( expected , actual )
|
getByQueryTest ( ) { for ( int i = 0 ; i < 3 ; i ++ ) { com . datastax . driver . mapping . EntityWithIndexes obj = new com . datastax . driver . mapping . EntityWithIndexes ( ) ; obj . setCount ( 100 ) ; obj . setEmail ( "email@test" ) ; obj . setName ( ( "test" + i ) ) ; obj . setTimeStamp ( new com . datastax . driver . mapping . Date ( ) ) ; obj . setUuid ( com . datastax . driver . mapping . UUID . randomUUID ( ) ) ; target . save ( obj ) ; } com . datastax . driver . mapping . meta . EntityTypeMetadata emeta = com . datastax . driver . mapping . EntityTypeParser . getEntityMetadata ( com . datastax . driver . mapping . EntityWithIndexes . class ) ; com . datastax . driver . mapping . meta . EntityFieldMetaData fdata = emeta . getFieldMetadata ( "email" ) ; com . datastax . driver . mapping . Statement query = com . datastax . driver . core . querybuilder . QueryBuilder . select ( ) . all ( ) . from ( com . datastax . driver . mapping . MappingSessionTest . keyspace , emeta . getTableName ( ) ) . where ( eq ( fdata . getColumnName ( ) , "email@test" ) ) ; com . datastax . driver . mapping . List < com . datastax . driver . mapping . EntityWithIndexes > items = target . getByQuery ( com . datastax . driver . mapping . EntityWithIndexes . class , query ) ; "<AssertPlaceHolder>" ; } getByQuery ( java . lang . Class , java . lang . String ) { maybeSync ( clazz ) ; return getFromResultSet ( clazz , session . execute ( query ) ) ; }
|
org . junit . Assert . assertEquals ( 3 , items . size ( ) )
|
isChangesetFullyCloaked_manyToMany ( ) { final java . util . List < java . lang . String > changesetPaths = java . util . Arrays . asList ( "$/foo/bar.baz" , "$/bar/foo.baz" ) ; final java . util . List < java . lang . String > cloakedPaths = java . util . Arrays . asList ( "$/foo" , "$/bar" ) ; final boolean actual = hudson . plugins . tfs . model . Project . isChangesetFullyCloaked ( changesetPaths , cloakedPaths ) ; "<AssertPlaceHolder>" ; } isChangesetFullyCloaked ( java . util . Collection , java . util . Collection ) { if ( cloakedPaths == null ) { return false ; } for ( final java . lang . String tfsPath : changesetPaths ) { boolean isPathCloaked = false ; for ( final java . lang . String cloakedPath : cloakedPaths ) { if ( tfsPath . regionMatches ( true , 0 , cloakedPath , 0 , cloakedPath . length ( ) ) ) { isPathCloaked = true ; break ; } } if ( ! isPathCloaked ) { return false ; } } return true ; }
|
org . junit . Assert . assertEquals ( true , actual )
|
testTowards ( ) { org . kie . workbench . common . stunner . lienzo . toolbox . items . tooltip . ToolboxTextTooltip cascade = tested . towards ( Direction . NORTH ) ; "<AssertPlaceHolder>" ; verify ( delegate , times ( 1 ) ) . towards ( eq ( Direction . NORTH ) ) ; } towards ( com . ait . lienzo . shared . core . types . Direction ) { this . towards = towards ; return this ; }
|
org . junit . Assert . assertEquals ( tested , cascade )
|
testDecodeString ( ) { org . imageterrier . locfile . PositionSpec spec = new org . imageterrier . locfile . PositionSpec ( org . imageterrier . locfile . PositionSpec . PositionSpecMode . SPATIAL , new int [ ] { 2 , 3 } , new double [ ] { 0 , 1 } , new double [ ] { 10 , 11 } ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( ( term ) + ":" ) + ( splitno ) ) + ":" ) + ( flushno ) ; }
|
org . junit . Assert . assertEquals ( spec , org . imageterrier . locfile . PositionSpec . decode ( spec . toString ( ) ) )
|
testFetchByPrimaryKeysWithNoPrimaryKeys ( ) { java . util . Set < java . io . Serializable > primaryKeys = new java . util . HashSet < java . io . Serializable > ( ) ; java . util . Map < java . io . Serializable , com . liferay . portal . workflow . kaleo . model . KaleoNotificationRecipient > kaleoNotificationRecipients = _persistence . fetchByPrimaryKeys ( primaryKeys ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return _portalCacheListeners . isEmpty ( ) ; }
|
org . junit . Assert . assertTrue ( kaleoNotificationRecipients . isEmpty ( ) )
|
testReadIntegers ( ) { org . apache . activemq . artemis . core . client . impl . LargeMessageControllerImpl buffer = createBufferWithIntegers ( 3 , 1 , 2 , 3 , 4 , 5 , 6 , 7 , 8 , 9 , 10 , 11 , 12 , 13 , 14 , 15 ) ; for ( int i = 1 ; i <= 15 ; i ++ ) { "<AssertPlaceHolder>" ; } try { buffer . readByte ( ) ; org . junit . Assert . fail ( "supposed<sp>to<sp>throw<sp>an<sp>exception" ) ; } catch ( java . lang . IndexOutOfBoundsException e ) { } } readInt ( ) { return buffer . readInt ( ) ; }
|
org . junit . Assert . assertEquals ( i , buffer . readInt ( ) )
|
testWriteCandlepinQueryContents ( ) { java . util . List < org . candlepin . model . Owner > owners = this . ownerCurator . listAll ( ) . list ( ) ; org . candlepin . resteasy . filter . CandlepinQueryInterceptor cqi = new org . candlepin . resteasy . filter . CandlepinQueryInterceptor ( this . mockJsonProvider , this . emProvider ) ; org . jboss . resteasy . core . ServerResponse response = new org . jboss . resteasy . core . ServerResponse ( ) ; response . setEntity ( this . ownerCurator . listAll ( ) ) ; cqi . postProcess ( response ) ; "<AssertPlaceHolder>" ; ( ( javax . ws . rs . core . StreamingOutput ) ( response . getEntity ( ) ) ) . write ( this . mockOutputStream ) ; verify ( this . mockJsonGenerator , times ( 1 ) ) . writeStartArray ( ) ; for ( org . candlepin . model . Owner owner : owners ) { verify ( this . mockObjectMapper , times ( 1 ) ) . writeValue ( eq ( this . mockJsonGenerator ) , eq ( owner ) ) ; } verify ( this . mockJsonGenerator , times ( 1 ) ) . writeEndArray ( ) ; } postProcess ( org . jboss . resteasy . core . ServerResponse ) { java . lang . Object entity = response . getEntity ( ) ; if ( entity instanceof org . quartz . JobDetail ) { org . quartz . JobDetail jobDetail = ( ( org . quartz . JobDetail ) ( entity ) ) ; setJobPrincipal ( jobDetail ) ; org . candlepin . pinsetter . core . model . JobStatus status = this . scheduleJob ( jobDetail ) ; response . setEntity ( this . translator . translate ( status , org . candlepin . dto . api . v1 . JobStatusDTO . class ) ) ; response . setStatus ( HttpResponseCodes . SC_ACCEPTED ) ; } else if ( entity instanceof org . quartz . JobDetail [ ] ) { org . quartz . JobDetail [ ] details = ( ( org . quartz . JobDetail [ ] ) ( entity ) ) ; org . candlepin . pinsetter . core . model . JobStatus [ ] statuses = new org . candlepin . pinsetter . core . model . JobStatus [ details . length ] ; int i = 0 ; for ( org . quartz . JobDetail jobDetail : details ) { setJobPrincipal ( jobDetail ) ; org . candlepin . pinsetter . core . model . JobStatus status = this . scheduleJob ( jobDetail ) ; statuses [ ( i ++ ) ] = status ; } org . candlepin . dto . api . v1 . JobStatusDTO [ ] dtoStatuses = new org . candlepin . dto . api . v1 . JobStatusDTO [ statuses . length ] ; for ( int j = 0 ; j < ( statuses . length ) ; j ++ ) { dtoStatuses [ j ] = this . translator . translate ( statuses [ j ] , org . candlepin . dto . api . v1 . JobStatusDTO . class ) ; } response . setEntity ( dtoStatuses ) ; response . setStatus ( HttpResponseCodes . SC_ACCEPTED ) ; } }
|
org . junit . Assert . assertTrue ( ( ( response . getEntity ( ) ) instanceof javax . ws . rs . core . StreamingOutput ) )
|
test_hasJs_caseInsensitive ( ) { java . lang . String search = "hAsJS" ; to . noc . devicefp . server . util . DeviceSearchString dss = new to . noc . devicefp . server . util . DeviceSearchString ( search ) ; "<AssertPlaceHolder>" ; } hasJs ( ) { return to . noc . devicefp . server . util . DeviceSearchString . hasJsPattern . matcher ( search ) . matches ( ) ; }
|
org . junit . Assert . assertTrue ( dss . hasJs ( ) )
|
testGetNodeDiskInfoFunction ( ) { "<AssertPlaceHolder>" ; assert ( provider . getNodeDiskInfos ( ) ) != null ; } getNodeDiskInfos ( ) { java . util . List < java . util . Map < java . lang . Object , java . lang . Object > > nodeInfoList = new java . util . ArrayList < java . util . Map < java . lang . Object , java . lang . Object > > ( ) ; java . util . Map < java . lang . Object , java . lang . Object > nodeDiskInfo ; org . hyperic . sigar . FileSystem [ ] fileSystemList = null ; org . hyperic . sigar . FileSystemUsage fileSystemUsage = null ; try { fileSystemList = sigar . getFileSystemList ( ) ; } catch ( java . lang . Exception e ) { com . impetus . ankush . agent . sigar . SigarNodeInfoProvider . LOGGER . error ( e . getMessage ( ) , e ) ; } for ( org . hyperic . sigar . FileSystem fileSystem : fileSystemList ) { nodeDiskInfo = new java . util . HashMap < java . lang . Object , java . lang . Object > ( ) ; nodeDiskInfo . put ( "deviceName" , fileSystem . getDevName ( ) ) ; nodeDiskInfo . put ( "dirName" , fileSystem . getDirName ( ) ) ; nodeDiskInfo . put ( "reads" 1 , fileSystem . getTypeName ( ) ) ; nodeDiskInfo . put ( "fileSystemType" , fileSystem . getSysTypeName ( ) ) ; nodeDiskInfo . put ( "partitionFlags" , fileSystem . getFlags ( ) ) ; nodeDiskInfo . put ( "reads" 3 , fileSystem . getOptions ( ) ) ; fileSystemUsage = new org . hyperic . sigar . FileSystemUsage ( ) ; try { fileSystemUsage . gather ( sigar , fileSystem . getDirName ( ) ) ; } catch ( org . hyperic . sigar . SigarException e ) { com . impetus . ankush . agent . sigar . SigarNodeInfoProvider . LOGGER . error ( e . getMessage ( ) , e ) ; } nodeDiskInfo . put ( "availableMemory" , fileSystemUsage . getAvail ( ) ) ; nodeDiskInfo . put ( "reads" 2 , fileSystemUsage . getFree ( ) ) ; nodeDiskInfo . put ( "totalMemory" , fileSystemUsage . getTotal ( ) ) ; nodeDiskInfo . put ( "usedMemory" , fileSystemUsage . getUsed ( ) ) ; nodeDiskInfo . put ( "readBytes" , fileSystemUsage . getDiskReadBytes ( ) ) ; nodeDiskInfo . put ( "reads" , fileSystemUsage . getDiskReads ( ) ) ; nodeDiskInfo . put ( "reads" 0 , fileSystemUsage . getDiskWriteBytes ( ) ) ; nodeDiskInfo . put ( "writes" , fileSystemUsage . getDiskWrites ( ) ) ; nodeInfoList . add ( nodeDiskInfo ) ; } return nodeInfoList ; }
|
org . junit . Assert . assertEquals ( java . util . ArrayList . class , provider . getNodeDiskInfos ( ) . getClass ( ) )
|
getAll ( ) { com . xpn . xwiki . doc . XWikiDocument document1 = mock ( com . xpn . xwiki . doc . XWikiDocument . class ) ; com . xpn . xwiki . doc . XWikiDocument document2 = mock ( com . xpn . xwiki . doc . XWikiDocument . class ) ; com . xpn . xwiki . doc . XWikiDocument maindocument = mock ( com . xpn . xwiki . doc . XWikiDocument . class ) ; org . xwiki . wiki . internal . descriptor . DefaultWikiDescriptor descriptor3 = new org . xwiki . wiki . internal . descriptor . DefaultWikiDescriptor ( "wikiid3" , "wikialias3" ) ; when ( descriptorDocumentHelper . getAllXWikiServerClassDocumentNames ( ) ) . thenReturn ( java . util . Arrays . asList ( "XWiki.XWikiServerWikiid1" , "XWiki.XWikiServerWikiid2" , "XWiki.XWikiServerWikiid3" ) ) ; when ( descriptorDocumentHelper . getWikiIdFromDocumentFullname ( "XWiki.XWikiServerWikiid1" ) ) . thenReturn ( "wikiid1" ) ; when ( descriptorDocumentHelper . getWikiIdFromDocumentFullname ( "XWiki.XWikiServerWikiid2" ) ) . thenReturn ( "wikiid2" ) ; when ( descriptorDocumentHelper . getWikiIdFromDocumentFullname ( "XWiki.XWikiServerWikiid3" ) ) . thenReturn ( "wikiid3" ) ; when ( cache . getFromId ( "wikiid3" ) ) . thenReturn ( descriptor3 ) ; when ( descriptorDocumentHelper . getDocumentFromWikiId ( "wikiid1" ) ) . thenReturn ( document1 ) ; when ( descriptorDocumentHelper . getDocumentFromWikiId ( "wikiid2" ) ) . thenReturn ( document2 ) ; when ( descriptorDocumentHelper . getDocumentFromWikiId ( "xwiki" ) ) . thenReturn ( maindocument ) ; when ( maindocument . isNew ( ) ) . thenReturn ( true ) ; java . util . List < com . xpn . xwiki . objects . BaseObject > baseObjects = java . util . Arrays . asList ( mock ( com . xpn . xwiki . objects . BaseObject . class ) ) ; when ( document1 . getXObjects ( any ( org . xwiki . model . reference . EntityReference . class ) ) ) . thenReturn ( baseObjects ) ; when ( document2 . getXObjects ( any ( org . xwiki . model . reference . EntityReference . class ) ) ) . thenReturn ( baseObjects ) ; org . xwiki . wiki . internal . descriptor . DefaultWikiDescriptor descriptor1 = new org . xwiki . wiki . internal . descriptor . DefaultWikiDescriptor ( "wikiid1" , "wikialias1" ) ; org . xwiki . wiki . internal . descriptor . DefaultWikiDescriptor descriptor2 = new org . xwiki . wiki . internal . descriptor . DefaultWikiDescriptor ( "wikiid2" , "wikialias2" ) ; when ( wikiDescriptorBuilder . buildDescriptorObject ( anyListOf ( com . xpn . xwiki . objects . BaseObject . class ) , any ( com . xpn . xwiki . doc . XWikiDocument . class ) ) ) . thenReturn ( descriptor1 , descriptor2 ) ; java . util . Collection < org . xwiki . wiki . descriptor . WikiDescriptor > descriptors = this . mocker . getComponentUnderTest ( ) . getAll ( ) ; "<AssertPlaceHolder>" ; verify ( descriptorDocumentHelper , never ( ) ) . getDocumentFromWikiId ( "wikiid3" ) ; verify ( cache ) . add ( descriptor1 ) ; verify ( cache ) . add ( descriptor2 ) ; verify ( cache , never ( ) ) . add ( descriptor3 ) ; } size ( ) { return groupNames . size ( ) ; }
|
org . junit . Assert . assertEquals ( 4 , descriptors . size ( ) )
|
testTemporalFilterAbsoluteSearch ( ) { java . lang . String startDate = "2011-10-4T05:48:27.891-07:00" ; java . lang . String endDate = "2011-10-4T06:18:27.581-07:00" ; org . codice . ddf . opensearch . endpoint . query . OpenSearchQuery query = new org . codice . ddf . opensearch . endpoint . query . OpenSearchQuery ( 0 , 10 , "relevance" , "desc" , 30000 , org . codice . ddf . opensearch . endpoint . query . OpenSearchQueryTest . FILTER_BUILDER ) ; query . addStartEndTemporalFilter ( startDate , endDate ) ; org . opengis . filter . Filter filter = query . getFilter ( ) ; org . codice . ddf . opensearch . endpoint . query . VerificationVisitor verificationVisitor = new org . codice . ddf . opensearch . endpoint . query . VerificationVisitor ( ) ; filter . accept ( verificationVisitor , null ) ; java . util . HashMap < java . lang . String , org . codice . ddf . opensearch . endpoint . query . FilterStatus > map = ( ( java . util . HashMap < java . lang . String , org . codice . ddf . opensearch . endpoint . query . FilterStatus > ) ( verificationVisitor . getMap ( ) ) ) ; printFilterStatusMap ( map ) ; java . util . List < org . opengis . filter . Filter > filters = getFilters ( map , org . geotools . filter . temporal . DuringImpl . class . getName ( ) ) ; "<AssertPlaceHolder>" ; verifyTemporalFilter ( filters . get ( 0 ) , startDate , endDate ) ; } size ( ) { return map . size ( ) ; }
|
org . junit . Assert . assertEquals ( 1 , filters . size ( ) )
|
createApi ( ) { wizardAction . openNewLiferayModuleWizard ( ) ; wizardAction . newModule . prepareGradle ( project . getName ( ) , com . liferay . ide . ui . module . tests . API ) ; wizardAction . finish ( ) ; jobAction . waitForNoRunningJobs ( ) ; "<AssertPlaceHolder>" ; viewAction . project . closeAndDelete ( project . getName ( ) ) ; } visibleFileTry ( java . lang . String [ ] ) { try { return _getProjects ( ) . isVisible ( files ) ; } catch ( java . lang . Exception e ) { _getProjects ( ) . setFocus ( ) ; try { java . lang . String [ ] parents = java . util . Arrays . copyOfRange ( files , 0 , ( ( files . length ) - 1 ) ) ; _getProjects ( ) . expand ( parents ) ; _getProjects ( ) . contextMenu ( com . liferay . ide . ui . liferay . action . REFRESH , parents ) ; ide . sleep ( 2000 ) ; } catch ( java . lang . Exception e1 ) { } for ( int i = ( files . length ) - 1 ; i > 0 ; i -- ) { java . lang . String [ ] parents = java . util . Arrays . copyOfRange ( files , 0 , ( ( files . length ) - i ) ) ; org . eclipse . swtbot . swt . finder . widgets . SWTBotTreeItem parent = _getProjects ( ) . getTreeItem ( parents ) ; _getProjects ( ) . expand ( parents ) ; java . lang . String subnode = files [ ( ( files . length ) - i ) ] ; _jobAction . waitForSubnode ( parent , subnode , com . liferay . ide . ui . liferay . action . REFRESH ) ; } return _getProjects ( ) . isVisible ( files ) ; } }
|
org . junit . Assert . assertTrue ( viewAction . project . visibleFileTry ( project . getName ( ) ) )
|
testSerialization ( ) { com . orsoncharts . renderer . category . StandardCategoryColorSource s1 = new com . orsoncharts . renderer . category . StandardCategoryColorSource ( ) ; com . orsoncharts . renderer . category . StandardCategoryColorSource s2 = ( ( com . orsoncharts . renderer . category . StandardCategoryColorSource ) ( com . orsoncharts . TestUtils . serialized ( s1 ) ) ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( obj == ( this ) ) { return true ; } if ( ! ( obj instanceof com . orsoncharts . graphics3d . Offset3D ) ) { return false ; } com . orsoncharts . graphics3d . Offset3D that = ( ( com . orsoncharts . graphics3d . Offset3D ) ( obj ) ) ; if ( ( this . dx ) != ( that . dx ) ) { return false ; } if ( ( this . dy ) != ( that . dy ) ) { return false ; } if ( ( this . dz ) != ( that . dz ) ) { return false ; } return true ; }
|
org . junit . Assert . assertTrue ( s1 . equals ( s2 ) )
|
testNederlandseNationaliteitBronAnders ( ) { final nl . bzk . brp . model . bericht . kern . ActieBericht actie = bouwActie ( ) ; org . springframework . test . util . ReflectionTestUtils . setField ( actie . getBronnen ( ) . iterator ( ) . next ( ) . getDocument ( ) . getSoort ( ) . getWaarde ( ) , "naam" , new nl . bzk . brp . model . algemeen . attribuuttype . kern . NaamEnumeratiewaardeAttribuut ( "Ander<sp>document" ) ) ; final java . util . List < nl . bzk . brp . model . basis . BerichtEntiteit > overtreders = brby0158 . voerRegelUit ( null , bouwNieuweSituatie ( NationaliteitcodeAttribuut . NL_NATIONALITEIT_CODE ) , actie , null ) ; "<AssertPlaceHolder>" ; } size ( ) { return elementen . size ( ) ; }
|
org . junit . Assert . assertEquals ( 0 , overtreders . size ( ) )
|
propertyNotPresentInArguments ( ) { org . watertemplate . interpreter . parser . AbstractSyntaxTree abs = new org . watertemplate . interpreter . parser . AbstractSyntaxTree . Id ( "prop_key" ) ; java . lang . Object result = abs . string ( new org . watertemplate . TemplateMap . Arguments ( ) , org . watertemplate . interpreter . parser . AbstractSyntaxTreeIdTest . locale ) ; "<AssertPlaceHolder>" ; } string ( org . watertemplate . TemplateMap$Arguments , java . util . Locale ) { return org . watertemplate . interpreter . DefaultWaterInterpreter . cache . computeIfAbsent ( cacheKey ( locale ) , ( key ) -> parse ( lex ( templateFileWith ( locale ) ) ) ) . string ( arguments , locale ) ; }
|
org . junit . Assert . assertEquals ( "prop_key" , result )
|
testCanceled ( ) { org . geotools . data . util . DefaultProgressListener listener = new org . geotools . data . util . DefaultProgressListener ( ) ; listener . setCanceled ( true ) ; org . geotools . renderer . lite . MarkFeatureIterator iterator = org . geotools . renderer . lite . MarkFeatureIterator . create ( zroads . getFeatures ( ) , 1000 , listener ) ; "<AssertPlaceHolder>" ; } getFeatures ( ) { return features = nonNullSet ( features , org . opengis . feature . type . FeatureType . class ) ; }
|
org . junit . Assert . assertNull ( iterator )
|
testSetOverTime ( ) { com . j256 . simplemetrics . metric . ControlledMetricValue metric = new com . j256 . simplemetrics . metric . ControlledMetricValue ( "comp" , "mod" , "name" , "desc" , "unit" ) ; long duration = 500 ; int numIncrements = 10 ; long bucketDelay = duration / numIncrements ; int total = 0 ; java . util . Random random = new java . util . Random ( ) ; for ( int x = 0 ; x < numIncrements ; x ++ ) { int value = ( random . nextInt ( 1000000000 ) ) / numIncrements ; metric . adjustValue ( value ) ; total += value ; java . lang . Thread . sleep ( ( bucketDelay + 1 ) ) ; } long value = metric . getValue ( ) . longValue ( ) ; long expected = total / numIncrements ; "<AssertPlaceHolder>" ; } getValue ( ) { double value ; double adjustedDenominator = ( denominator ) * ( count ) ; if ( adjustedDenominator == 0 ) { value = 0 ; } else { value = ( numerator ) / adjustedDenominator ; } return java . lang . Double . valueOf ( value ) ; }
|
org . junit . Assert . assertEquals ( expected , value )
|
choiceB ( ) { java . util . List < java . lang . String > options = new java . util . ArrayList < java . lang . String > ( ) ; options . add ( "A" ) ; options . add ( "B" ) ; ch11 . Randoms sut = new ch11 . Randoms ( ) ; sut . generator = new ch11 . RandomNumberGenerator ( ) { @ ch11 . Override public int nextInt ( ) { return 1 ; } } ; "<AssertPlaceHolder>" ; } choice ( java . util . List ) { if ( ( options . size ( ) ) == 0 ) return null ; int idx = ( generator . nextInt ( ) ) % ( options . size ( ) ) ; return options . get ( idx ) ; }
|
org . junit . Assert . assertThat ( sut . choice ( options ) , org . hamcrest . CoreMatchers . is ( "B" ) )
|
testDispose ( ) { store . dispose ( ) ; try { ( ( org . geotools . arcsde . data . ArcSDEDataStore ) ( store ) ) . getSession ( Transaction . AUTO_COMMIT ) ; org . junit . Assert . fail ( "Expected<sp>IllegalStateException<sp>when<sp>the<sp>datastore<sp>has<sp>been<sp>disposed" ) ; } catch ( java . lang . IllegalStateException e ) { "<AssertPlaceHolder>" ; } finally { org . geotools . arcsde . data . ArcSDEDataStoreTest . testData . tearDown ( false , true ) ; org . geotools . arcsde . data . ArcSDEDataStoreTest . testData = null ; } } fail ( java . lang . Exception ) { e . printStackTrace ( ) ; fail ( ( "Should<sp>not<sp>have<sp>thrown:<sp>" + ( e . getLocalizedMessage ( ) ) ) ) ; }
|
org . junit . Assert . assertTrue ( true )
|
testGetReceiver ( ) { org . eclipse . rap . fileupload . FileUploadReceiver receiver = new org . eclipse . rap . fileupload . test . TestFileUploadReceiver ( ) ; org . eclipse . rap . fileupload . FileUploadHandler handler = new org . eclipse . rap . fileupload . FileUploadHandler ( receiver ) ; "<AssertPlaceHolder>" ; } getReceiver ( ) { return receiver ; }
|
org . junit . Assert . assertSame ( receiver , handler . getReceiver ( ) )
|
testGetBondCount_IChemSequence ( ) { int count = org . openscience . cdk . tools . manipulator . ChemSequenceManipulator . getBondCount ( chemSequence ) ; "<AssertPlaceHolder>" ; } getBondCount ( org . openscience . cdk . interfaces . IChemSequence ) { int count = 0 ; for ( int i = 0 ; i < ( sequence . getChemModelCount ( ) ) ; i ++ ) { count += org . openscience . cdk . tools . manipulator . ChemModelManipulator . getBondCount ( sequence . getChemModel ( i ) ) ; } return count ; }
|
org . junit . Assert . assertEquals ( 2 , count )
|
testDelegationTokenSecretManager ( ) { org . apache . hadoop . security . token . Token < org . apache . hadoop . hdfs . security . token . delegation . DelegationTokenIdentifier > token = generateDelegationToken ( "SomeUser" , "JobTracker" ) ; try { dtSecretManager . renewToken ( token , "FakeRenewer" ) ; org . junit . Assert . fail ( "should<sp>have<sp>failed" ) ; } catch ( org . apache . hadoop . security . AccessControlException ace ) { } dtSecretManager . renewToken ( token , "JobTracker" ) ; org . apache . hadoop . hdfs . security . token . delegation . DelegationTokenIdentifier identifier = new org . apache . hadoop . hdfs . security . token . delegation . DelegationTokenIdentifier ( ) ; byte [ ] tokenId = token . getIdentifier ( ) ; identifier . readFields ( new java . io . DataInputStream ( new java . io . ByteArrayInputStream ( tokenId ) ) ) ; "<AssertPlaceHolder>" ; org . apache . hadoop . hdfs . security . TestDelegationToken . LOG . info ( "Sleep<sp>to<sp>expire<sp>the<sp>token" ) ; java . lang . Thread . sleep ( 6000 ) ; try { dtSecretManager . retrievePassword ( identifier ) ; org . junit . Assert . fail ( "Token<sp>should<sp>have<sp>expired" ) ; } catch ( org . apache . hadoop . security . token . SecretManager . InvalidToken e ) { } dtSecretManager . renewToken ( token , "JobTracker" ) ; org . apache . hadoop . hdfs . security . TestDelegationToken . LOG . info ( "Sleep<sp>beyond<sp>the<sp>max<sp>lifetime" ) ; java . lang . Thread . sleep ( 5000 ) ; try { dtSecretManager . renewToken ( token , "JobTracker" ) ; org . junit . Assert . fail ( "should<sp>have<sp>been<sp>expired" ) ; } catch ( org . apache . hadoop . security . token . SecretManager . InvalidToken it ) { } } retrievePassword ( org . apache . hadoop . hdfs . security . token . block . BlockTokenIdentifier ) { if ( isExpired ( identifier . getExpiryDate ( ) ) ) { throw new io . hops . metadata . security . token . block . InvalidToken ( ( ( "Block<sp>token<sp>with<sp>" + ( identifier . toString ( ) ) ) + "<sp>is<sp>expired." ) ) ; } org . apache . hadoop . hdfs . security . token . block . BlockKey key = null ; try { key = getBlockKeyById ( identifier . getKeyId ( ) ) ; } catch ( java . io . IOException ex ) { } if ( key == null ) { throw new io . hops . metadata . security . token . block . InvalidToken ( ( ( ( ( "Can't<sp>re-compute<sp>password<sp>for<sp>" + ( identifier . toString ( ) ) ) + ",<sp>since<sp>the<sp>required<sp>block<sp>key<sp>(keyID=" ) + ( identifier . getKeyId ( ) ) ) + ")<sp>doesn't<sp>exist." ) ) ; } return createPassword ( identifier . getBytes ( ) , key . getKey ( ) ) ; }
|
org . junit . Assert . assertTrue ( ( null != ( dtSecretManager . retrievePassword ( identifier ) ) ) )
|
undeployNothing ( ) { final org . jboss . msc . service . util . TestTask completeTask = new org . jboss . msc . service . util . TestTask ( ) ; org . jboss . msc . service . ServiceUtils . undeployAll ( completeTask ) ; "<AssertPlaceHolder>" ; } get ( ) { return dependency . getValue ( ) ; }
|
org . junit . Assert . assertTrue ( completeTask . get ( ) )
|
shouldFindGlobalValueIfDefined ( ) { givenDefaults ( ) . with ( "property.name" , "default<sp>value" ) ; whenOracleIsLoadedAndExec ( ) ; "<AssertPlaceHolder>" ; } globalScopedProperty ( java . lang . String ) { java . lang . String args = ( "'" + key ) + "'" ; return lookupArgs ( args ) ; }
|
org . junit . Assert . assertThat ( globalScopedProperty ( "property.name" ) , org . hamcrest . Matchers . is ( "default<sp>value" ) )
|
testInvertedBasicNonMatch ( ) { org . graylog2 . plugin . streams . StreamRule rule = getSampleRule ( ) ; rule . setField ( "nonexistentField" ) ; rule . setType ( StreamRuleType . PRESENCE ) ; rule . setInverted ( true ) ; org . graylog2 . plugin . Message message = getSampleMessage ( ) ; org . graylog2 . streams . matchers . StreamRuleMatcher matcher = getMatcher ( rule ) ; java . lang . Boolean result = matcher . match ( message , rule ) ; "<AssertPlaceHolder>" ; } match ( org . graylog2 . plugin . Message , org . graylog2 . plugin . streams . StreamRule ) { java . lang . Double msgVal = getDouble ( msg . getField ( rule . getField ( ) ) ) ; if ( msgVal == null ) { return false ; } java . lang . Double ruleVal = getDouble ( rule . getValue ( ) ) ; if ( ruleVal == null ) { return false ; } return ( rule . getInverted ( ) ) ^ ( msgVal > ruleVal ) ; }
|
org . junit . Assert . assertTrue ( result )
|
forall_A$Function1_Nil ( ) { com . m3 . scalaflavor4j . Seq < java . lang . Integer > seq = com . m3 . scalaflavor4j . Seq . apply ( ) ; boolean actual = seq . forall ( new com . m3 . scalaflavor4j . F1 < java . lang . Integer , java . lang . Boolean > ( ) { public com . m3 . scalaflavor4j . Boolean apply ( java . lang . Integer v1 ) { return v1 < 3 ; } } ) ; boolean expected = true ; "<AssertPlaceHolder>" ; } apply ( ) { com . m3 . scalaflavor4j . MainFunction main = new com . m3 . scalaflavor4j . MainFunction ( ) { public void apply ( java . lang . String [ ] args ) throws com . m3 . scalaflavor4j . Exception { print . apply ( args . length ) ; com . m3 . scalaflavor4j . Seq . apply ( args ) . foreach ( new com . m3 . scalaflavor4j . VoidF1 < java . lang . String > ( ) { public void apply ( java . lang . String arg ) throws com . m3 . scalaflavor4j . Exception { print . apply ( arg ) ; } } ) ; } } ; main . apply ( new java . lang . String [ ] { "a" , "b" } ) ; }
|
org . junit . Assert . assertThat ( actual , org . hamcrest . CoreMatchers . is ( org . hamcrest . CoreMatchers . equalTo ( expected ) ) )
|
testSingleMenu ( ) { final java . lang . String caption = "test" ; final org . uberfire . workbench . model . menu . Menus menus = org . uberfire . workbench . model . menu . MenuFactory . newTopLevelMenu ( caption ) . respondsWith ( ( ) -> { } ) . endMenu ( ) . build ( ) ; final com . google . gwt . user . client . ui . Widget widget = listBar . makeItem ( menus . getItems ( ) . get ( 0 ) , true ) ; "<AssertPlaceHolder>" ; verify ( ( ( org . gwtbootstrap3 . client . ui . Button ) ( widget ) ) ) . setText ( caption ) ; } get ( java . lang . String ) { return new org . dashbuilder . navigation . impl . NavItemContextImpl ( ctx ) ; }
|
org . junit . Assert . assertTrue ( ( widget instanceof org . gwtbootstrap3 . client . ui . Button ) )
|
testParseSuperErr ( ) { actions = new java . util . HashMap < java . lang . String , java . lang . String > ( ) { { put ( "aaa" , "bbb" ) ; } } ; target = new org . o3project . odenos . core . component . network . flow . query . OFPFlowActionSetFieldQuery ( actions ) ; "<AssertPlaceHolder>" ; } parse ( ) { if ( ! ( super . parse ( ) ) ) { return false ; } if ( ! ( org . o3project . odenos . core . component . network . BasicQuery . checkMapExactly ( this . actions , new java . lang . String [ ] { } ) ) ) { return false ; } return true ; }
|
org . junit . Assert . assertThat ( target . parse ( ) , org . hamcrest . CoreMatchers . is ( false ) )
|
testUniqueness ( ) { sonia . scm . security . DefaultKeyGenerator generator = new sonia . scm . security . DefaultKeyGenerator ( ) ; java . util . Set < java . lang . String > keys = com . google . common . collect . Sets . newHashSet ( ) ; for ( int i = 0 ; i < 10000 ; i ++ ) { java . lang . String key = generator . createKey ( ) ; if ( keys . contains ( key ) ) { org . junit . Assert . fail ( "dublicate<sp>key" ) ; } keys . add ( key ) ; } "<AssertPlaceHolder>" ; } size ( ) { return ( ( int ) ( cache . size ( ) ) ) ; }
|
org . junit . Assert . assertEquals ( 10000 , keys . size ( ) )
|
testIsValidationEnabled ( ) { try { org . eclipse . jdt . core . IJavaProject _createJavaProject = org . eclipse . xtext . ui . testing . util . JavaProjectSetupUtil . createJavaProject ( "testProject" ) ; final org . eclipse . xtext . xbase . lib . Procedures . Procedure1 < org . eclipse . jdt . core . IJavaProject > _function = ( org . eclipse . jdt . core . IJavaProject it ) -> { try { org . eclipse . xtext . ui . testing . util . JavaProjectSetupUtil . addSourceFolder ( it , "filtered-src" , null , new java . lang . String [ ] { "**.xtend" } ) ; } catch ( _e ) { throw org . eclipse . xtext . xbase . lib . Exceptions . sneakyThrow ( org . eclipse . xtend . ide . tests . editor . _e ) ; } } ; org . eclipse . xtext . xbase . lib . ObjectExtensions . < org . eclipse . jdt . core . IJavaProject > operator_doubleArrow ( _createJavaProject , _function ) ; final org . eclipse . core . resources . IFile file = org . eclipse . xtext . ui . testing . util . IResourcesSetupUtil . createFile ( "testProject/filtered-src/Foo.xtend" , "class<sp>Foo<sp>{}" ) ; "<AssertPlaceHolder>" ; } catch ( java . lang . Throwable _e ) { throw org . eclipse . xtext . xbase . lib . Exceptions . sneakyThrow ( _e ) ; } } isValidationDisabled ( org . eclipse . core . resources . IStorage ) { try { final java . lang . reflect . Method method = this . resourceForEditorInputFactory . getClass ( ) . getDeclaredMethod ( "isValidationDisabled" , org . eclipse . core . resources . IStorage . class ) ; method . setAccessible ( true ) ; java . lang . Object _invoke = method . invoke ( this . resourceForEditorInputFactory , storage ) ; return ( ( java . lang . Boolean ) ( _invoke ) ) . booleanValue ( ) ; } catch ( java . lang . Throwable _e ) { throw org . eclipse . xtext . xbase . lib . Exceptions . sneakyThrow ( _e ) ; } }
|
org . junit . Assert . assertTrue ( this . isValidationDisabled ( file ) )
|
shouldDoNothing_whenTryToTurnSnakeUpAfterGameOver ( ) { shouldGameOver_whenSnakeEatItself ( ) ; com . codenjoy . dojo . snake . model . Direction direction = hero . getDirection ( ) ; hero . up ( ) ; hero . down ( ) ; hero . left ( ) ; hero . right ( ) ; hero . up ( ) ; "<AssertPlaceHolder>" ; } getDirection ( ) { return direction ; }
|
org . junit . Assert . assertEquals ( direction , hero . getDirection ( ) )
|
testSubtractYear1 ( ) { java . lang . String str1 = "2015-02-23<sp>18:54:00" ; java . lang . String str2 = "2016-08-25<sp>21:55:22" ; int rs = com . opslab . util . DateUtil . subtractYear ( str1 , str2 ) ; "<AssertPlaceHolder>" ; } subtractYear ( java . lang . String , java . lang . String ) { int result ; java . util . Calendar c1 = java . util . Calendar . getInstance ( ) ; java . util . Calendar c2 = java . util . Calendar . getInstance ( ) ; try { c1 . setTime ( com . opslab . util . DateUtil . DateInstance ( ) . parse ( date1 ) ) ; c2 . setTime ( com . opslab . util . DateUtil . DateInstance ( ) . parse ( date2 ) ) ; int year1 = c1 . get ( Calendar . YEAR ) ; int year2 = c2 . get ( Calendar . YEAR ) ; result = year2 - year1 ; } catch ( java . text . ParseException e ) { e . printStackTrace ( ) ; result = - 1 ; } return result ; }
|
org . junit . Assert . assertEquals ( "" , 1 , rs )
|
getEventsForwardsCall ( ) { java . util . Calendar from = java . util . Calendar . getInstance ( ) ; from . clear ( ) ; from . set ( 2010 , 0 , 1 , 0 , 0 , 0 ) ; java . util . Calendar to = java . util . Calendar . getInstance ( ) ; to . clear ( ) ; to . set ( 2011 , 0 , 1 , 0 , 0 , 0 ) ; org . phenotips . security . audit . AuditEvent template = new org . phenotips . security . audit . AuditEvent ( this . user , "ip" , "action" , null , this . doc , null ) ; when ( this . store . getEvents ( template , from , to , 20 , 10 ) ) . thenReturn ( this . events ) ; when ( this . auth . hasAccess ( this . user , Right . ADMIN , this . xwikiPreferences ) ) . thenReturn ( true ) ; java . util . List < org . phenotips . security . audit . AuditEvent > result = this . scriptService . getEvents ( 20 , 10 , "action" , "user" , "ip" , "Space.Page" , "01/01/2010" , "01/01/2011" ) ; org . mockito . Mockito . verify ( this . store ) . getEvents ( template , from , to , 20 , 10 ) ; "<AssertPlaceHolder>" ; } getEvents ( org . phenotips . security . audit . AuditEvent , java . util . Calendar , java . util . Calendar , int , int ) { try { org . hibernate . Session session = this . sessionFactory . getSessionFactory ( ) . openSession ( ) ; org . hibernate . Criteria c = session . createCriteria ( org . phenotips . security . audit . AuditEvent . class ) ; if ( eventTemplate != null ) { c . add ( org . hibernate . criterion . Example . create ( eventTemplate ) ) ; } setTimeInterval ( c , fromTime , toTime ) ; c . addOrder ( org . hibernate . criterion . Order . desc ( org . phenotips . security . audit . internal . HibernateAuditStore . TIME_FIELD_NAME ) ) ; if ( start > 0 ) { c . setFirstResult ( start ) ; } if ( maxResults > 0 ) { c . setMaxResults ( maxResults ) ; } c . setReadOnly ( true ) ; @ org . phenotips . security . audit . internal . SuppressWarnings ( "unchecked" ) java . util . List < org . phenotips . security . audit . AuditEvent > foundEntries = c . list ( ) ; return foundEntries ; } catch ( org . hibernate . HibernateException ex ) { this . logger . error ( "Failed<sp>to<sp>load<sp>audit<sp>event<sp>documents:<sp>{}" , ex . getMessage ( ) , ex ) ; } return java . util . Collections . emptyList ( ) ; }
|
org . junit . Assert . assertSame ( this . events , result )
|
given2EqualValueTypesWithBuilder_whenEqual_thenCorrect ( ) { com . baeldung . autovalue . AutoValueMoneyWithBuilder m1 = com . baeldung . autovalue . AutoValueMoneyWithBuilder . builder ( ) . setAmount ( 5000 ) . setCurrency ( "USD" ) . build ( ) ; com . baeldung . autovalue . AutoValueMoneyWithBuilder m2 = com . baeldung . autovalue . AutoValueMoneyWithBuilder . builder ( ) . setAmount ( 5000 ) . setCurrency ( "USD" ) . build ( ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ( this ) == obj ) { return true ; } if ( obj == null ) { return false ; } if ( ! ( obj instanceof com . baeldung . jackson . entities . MyPair ) ) { return false ; } com . baeldung . jackson . entities . MyPair other = ( ( com . baeldung . jackson . entities . MyPair ) ( obj ) ) ; if ( ( first ) == null ) { if ( ( other . first ) != null ) { return false ; } } else if ( ! ( first . equals ( other . first ) ) ) { return false ; } if ( ( second ) == null ) { if ( ( other . second ) != null ) { return false ; } } else if ( ! ( second . equals ( other . second ) ) ) { return false ; } return true ; }
|
org . junit . Assert . assertTrue ( m1 . equals ( m2 ) )
|
growALittle ( ) { int t = createTable ( "s" , "t" , "string<sp>varchar(100)" ) ; com . foundationdb . server . rowdata . RowDef rowDef = getRowDef ( t ) ; int initialBufferSize = ( ( com . foundationdb . server . rowdata . RowData . MINIMUM_RECORD_LENGTH ) + 1 ) + 5 ; com . foundationdb . server . rowdata . RowData rowData = new com . foundationdb . server . rowdata . RowData ( new byte [ initialBufferSize ] ) ; rowData . createRow ( rowDef , new java . lang . Object [ ] { "abcdefghijklmno" } , true ) ; "<AssertPlaceHolder>" ; } getBytes ( ) { return data ; }
|
org . junit . Assert . assertEquals ( ( initialBufferSize * 2 ) , rowData . getBytes ( ) . length )
|
testRangeCardinality2 ( ) { org . roaringbitmap . RoaringBitmap r = new org . roaringbitmap . RoaringBitmap ( ) ; long Min = 1L << 16 ; long Max = 1L << 18 ; r . add ( Min , Max ) ; for ( long s = Min ; s <= Max ; s += 1024 ) { for ( long e = s ; e <= Max ; e += 1024 ) { "<AssertPlaceHolder>" ; } } } rangeCardinality ( long , long ) { if ( ( java . lang . Long . compareUnsigned ( start , end ) ) >= 0 ) { return 0 ; } long size = 0 ; int startIndex = this . highLowContainer . getIndex ( org . roaringbitmap . Util . highbits ( start ) ) ; if ( startIndex < 0 ) { startIndex = ( - startIndex ) - 1 ; } else { int inContainerStart = org . roaringbitmap . Util . toIntUnsigned ( org . roaringbitmap . Util . lowbits ( start ) ) ; if ( inContainerStart != 0 ) { size -= this . highLowContainer . getContainerAtIndex ( startIndex ) . rank ( ( ( short ) ( inContainerStart - 1 ) ) ) ; } } short xhigh = org . roaringbitmap . Util . highbits ( ( end - 1 ) ) ; for ( int i = startIndex ; i < ( this . highLowContainer . size ( ) ) ; i ++ ) { short key = this . highLowContainer . getKeyAtIndex ( i ) ; int comparison = org . roaringbitmap . Util . compareUnsigned ( key , xhigh ) ; if ( comparison < 0 ) { size += this . highLowContainer . getContainerAtIndex ( i ) . getCardinality ( ) ; } else if ( comparison == 0 ) { return size + ( this . highLowContainer . getContainerAtIndex ( i ) . rank ( org . roaringbitmap . Util . lowbits ( ( ( int ) ( end - 1 ) ) ) ) ) ; } } return size ; }
|
org . junit . Assert . assertEquals ( ( e - s ) , r . rangeCardinality ( s , e ) )
|
shouldAcceptValidRateLimitObject ( ) { org . openstack . atlas . api . validation . results . ValidatorResult result = validator . validate ( rateLimit , org . openstack . atlas . api . mgmt . validation . validators . PUT ) ; "<AssertPlaceHolder>" ; } passedValidation ( ) { return expectationResultList . isEmpty ( ) ; }
|
org . junit . Assert . assertTrue ( result . passedValidation ( ) )
|
shouldBeRelatedToEntityIdWhenSourceAndVertexAreNull ( ) { final java . lang . String source = null ; final java . lang . String destination = "destination" ; final boolean directed = true ; final uk . gov . gchq . gaffer . data . element . id . EdgeId seed = new uk . gov . gchq . gaffer . operation . data . EdgeSeed ( source , destination , directed ) ; final uk . gov . gchq . gaffer . data . element . id . EntityId relatedSeed = mock ( uk . gov . gchq . gaffer . data . element . id . EntityId . class ) ; given ( relatedSeed . getVertex ( ) ) . willReturn ( source ) ; final boolean isRelated = seed . isRelated ( ( ( uk . gov . gchq . gaffer . data . element . id . ElementId ) ( relatedSeed ) ) ) . isMatch ( ) ; "<AssertPlaceHolder>" ; } getVertex ( ) { return getIdentifierTypeName ( IdentifierType . VERTEX ) ; }
|
org . junit . Assert . assertTrue ( isRelated )
|
testExecuteBatch ( ) { java . sql . PreparedStatement preparedStatement = sharedConnection . prepareStatement ( "INSERT<sp>INTO<sp>table1<sp>VALUE<sp>?" ) ; try { int [ ] result = preparedStatement . executeBatch ( ) ; "<AssertPlaceHolder>" ; } catch ( java . sql . SQLException sqle ) { org . junit . Assert . fail ( "Must<sp>not<sp>throw<sp>error" ) ; } } executeBatch ( ) { checkClose ( ) ; int size ; if ( ( ( batchQueries ) == null ) || ( ( size = batchQueries . size ( ) ) == 0 ) ) { return new int [ 0 ] ; } lock . lock ( ) ; try { internalBatchExecution ( size ) ; return results . getCmdInformation ( ) . getUpdateCounts ( ) ; } catch ( java . sql . SQLException initialSqlEx ) { throw executeBatchExceptionEpilogue ( initialSqlEx , size ) ; } finally { executeBatchEpilogue ( ) ; lock . unlock ( ) ; } }
|
org . junit . Assert . assertEquals ( 0 , result . length )
|
testOnErrorConsumerIsCalled ( ) { java . lang . RuntimeException throwable = new java . lang . RuntimeException ( ) ; java . util . Collection < java . lang . Throwable > onErrorWasCalled = new java . util . ArrayList ( ) ; org . apache . beam . runners . dataflow . harness . test . TestStreams . withOnNext ( null ) . withOnError ( onErrorWasCalled :: add ) . build ( ) . onError ( throwable ) ; "<AssertPlaceHolder>" ; } contains ( org . apache . beam . sdk . values . PCollectionView ) { return delegate . contains ( view ) ; }
|
org . junit . Assert . assertThat ( onErrorWasCalled , org . hamcrest . Matchers . contains ( throwable ) )
|
testResolved ( ) { org . apache . tuscany . sca . contribution . processor . ProcessorContext context = new org . apache . tuscany . sca . contribution . processor . ProcessorContext ( ) ; org . apache . tuscany . sca . contribution . resolver . DefaultModelResolverTestCase . Model a = new org . apache . tuscany . sca . contribution . resolver . DefaultModelResolverTestCase . Model ( "a" ) ; resolver . addModel ( a , context ) ; org . apache . tuscany . sca . contribution . resolver . DefaultModelResolverTestCase . Model x = new org . apache . tuscany . sca . contribution . resolver . DefaultModelResolverTestCase . Model ( "a" ) ; x = resolver . resolveModel ( org . apache . tuscany . sca . contribution . resolver . DefaultModelResolverTestCase . Model . class , x , context ) ; "<AssertPlaceHolder>" ; } resolveModel ( java . lang . Class , T , org . apache . tuscany . sca . contribution . processor . ProcessorContext ) { java . lang . Object resolved = map . get ( unresolved ) ; if ( resolved != null ) { return modelClass . cast ( resolved ) ; } return unresolved ; }
|
org . junit . Assert . assertTrue ( ( x == a ) )
|
testGetLocatorsForReRollHigherLevelToStorageGranularity ( ) { boolean isReroll = true ; com . rackspacecloud . blueflood . rollup . SlotKey destSlotKey = com . rackspacecloud . blueflood . rollup . SlotKey . of ( Granularity . MIN_60 , 0 , TEST_SHARD ) ; com . rackspacecloud . blueflood . rollup . Granularity delayedMetricsRerollGranularity = com . rackspacecloud . blueflood . rollup . Granularity . MIN_60 ; com . rackspacecloud . blueflood . rollup . Granularity delayedMetricsStorageGranularity = com . rackspacecloud . blueflood . rollup . Granularity . MIN_20 ; com . rackspacecloud . blueflood . service . LocatorFetchRunnable lfrunnable = new com . rackspacecloud . blueflood . service . LocatorFetchRunnable ( scheduleCtx , destSlotKey , rollupReadExecutor , rollupWriteExecutor ) ; java . util . HashSet < com . rackspacecloud . blueflood . types . Locator > delayedLocators1 = new java . util . HashSet < com . rackspacecloud . blueflood . types . Locator > ( ) { { add ( locators . get ( 0 ) ) ; } } ; java . util . HashSet < com . rackspacecloud . blueflood . types . Locator > delayedLocators2 = new java . util . HashSet < com . rackspacecloud . blueflood . types . Locator > ( ) { { add ( locators . get ( 1 ) ) ; } } ; when ( delayedLocatorIO . getLocators ( com . rackspacecloud . blueflood . rollup . SlotKey . of ( Granularity . MIN_20 , 0 , TEST_SHARD ) ) ) . thenReturn ( delayedLocators1 ) ; when ( delayedLocatorIO . getLocators ( com . rackspacecloud . blueflood . rollup . SlotKey . of ( Granularity . MIN_20 , 1 , TEST_SHARD ) ) ) . thenReturn ( delayedLocators1 ) ; when ( delayedLocatorIO . getLocators ( com . rackspacecloud . blueflood . rollup . SlotKey . of ( Granularity . MIN_20 , 2 , TEST_SHARD ) ) ) . thenReturn ( delayedLocators2 ) ; java . util . Set < com . rackspacecloud . blueflood . types . Locator > locatorsForRollup = lfrunnable . getLocators ( executionContext , isReroll , delayedMetricsRerollGranularity , delayedMetricsStorageGranularity ) ; "<AssertPlaceHolder>" ; } size ( ) { return array . length ; }
|
org . junit . Assert . assertEquals ( ( ( delayedLocators1 . size ( ) ) + ( delayedLocators2 . size ( ) ) ) , locatorsForRollup . size ( ) )
|
getBucketObjectSummariesTest ( ) { java . util . List < com . amazonaws . services . s3 . model . S3ObjectSummary > bucketObjectSummariesWithPrefix = fr . d2si . ooso . tests . Commons . getBucketObjectSummaries ( fr . d2si . ooso . tests . CommonUtilitiesTest . DUMMY_BUCKET_NAME , "pref1/" ) ; java . util . Set < java . lang . String > storedPrefixedKeys = bucketObjectSummariesWithPrefix . stream ( ) . map ( S3ObjectSummary :: getKey ) . collect ( toSet ( ) ) ; java . util . Set < java . lang . String > expectedPrefixedKeys = fr . d2si . ooso . tests . CommonUtilitiesTest . KEY_CONTENT_MAPPING . keySet ( ) . stream ( ) . filter ( ( o ) -> o . startsWith ( "pref1/" ) ) . collect ( toSet ( ) ) ; "<AssertPlaceHolder>" ; } getBucketObjectSummaries ( java . lang . String , java . lang . String ) { java . lang . String realBucket = fr . d2si . ooso . utils . Commons . getBucketFromFullPath ( bucket ) ; java . lang . String preprefix = fr . d2si . ooso . utils . Commons . getPrefixFromFullPath ( bucket ) ; com . amazonaws . services . s3 . AmazonS3 s3Client = fr . d2si . ooso . utils . AmazonS3Provider . getS3Client ( ) ; fr . d2si . ooso . utils . ListObjectsRequest req = new fr . d2si . ooso . utils . ListObjectsRequest ( ) . withBucketName ( realBucket ) . withPrefix ( ( ! ( preprefix . equals ( "" ) ) ? ( preprefix + "/" ) + prefix : prefix ) ) ; fr . d2si . ooso . utils . ObjectListing objectListing = s3Client . listObjects ( req ) ; java . util . List < fr . d2si . ooso . utils . S3ObjectSummary > summaries = objectListing . getObjectSummaries ( ) ; while ( objectListing . isTruncated ( ) ) { objectListing = s3Client . listNextBatchOfObjects ( objectListing ) ; summaries . addAll ( objectListing . getObjectSummaries ( ) ) ; } return summaries . stream ( ) . filter ( ( obj ) -> ( ! ( obj . getKey ( ) . equals ( "" ) ) ) && ( ! ( obj . getKey ( ) . endsWith ( "/" ) ) ) ) . collect ( toList ( ) ) ; }
|
org . junit . Assert . assertEquals ( storedPrefixedKeys , expectedPrefixedKeys )
|
testUsesUIConfigurationOfUI ( ) { com . eclipsesource . tabris . ui . UI ui = mock ( com . eclipsesource . tabris . ui . UI . class ) ; com . eclipsesource . tabris . ui . UIConfiguration config = mock ( com . eclipsesource . tabris . ui . UIConfiguration . class ) ; when ( ui . getConfiguration ( ) ) . thenReturn ( config ) ; com . eclipsesource . tabris . ui . PageOperator operator = mock ( com . eclipsesource . tabris . ui . PageOperator . class ) ; com . eclipsesource . tabris . ui . PageData data = new com . eclipsesource . tabris . ui . PageData ( ) ; when ( operator . getCurrentPageData ( ) ) . thenReturn ( data ) ; when ( ui . getPageOperator ( ) ) . thenReturn ( operator ) ; com . eclipsesource . tabris . ui . AbstractPageTest . TestAbstractPage page = new com . eclipsesource . tabris . ui . AbstractPageTest . TestAbstractPage ( ) ; page . createContent ( shell , ui ) ; com . eclipsesource . tabris . ui . UIConfiguration actualConfiguration = page . getUIConfiguration ( ) ; "<AssertPlaceHolder>" ; } createContent ( org . eclipse . swt . widgets . Composite , com . eclipsesource . tabris . ui . UI ) { this . ui = ui ; createContent ( parent , ui . getPageOperator ( ) . getCurrentPageData ( ) ) ; }
|
org . junit . Assert . assertSame ( config , actualConfiguration )
|
precedesPast ( ) { net . time4j . PlainDate endA = net . time4j . PlainDate . axis ( ) . getMinimum ( ) ; net . time4j . PlainDate endB = net . time4j . PlainDate . of ( 2014 , 5 , 17 ) ; net . time4j . range . DateInterval a = net . time4j . range . DateInterval . until ( endA ) ; net . time4j . range . DateInterval b = net . time4j . range . DateInterval . until ( endB ) ; "<AssertPlaceHolder>" ; } precedes ( I extends net . time4j . range . IsoInterval ) { if ( ( other . getStart ( ) . isInfinite ( ) ) || ( this . end . isInfinite ( ) ) ) { return false ; } T endA = this . end . getTemporal ( ) ; if ( this . end . isClosed ( ) ) { endA = this . getTimeLine ( ) . stepForward ( endA ) ; if ( endA == null ) { return false ; } } return endA . isBefore ( other . getClosedFiniteStart ( ) ) ; }
|
org . junit . Assert . assertThat ( a . precedes ( b ) , org . hamcrest . CoreMatchers . is ( false ) )
|
testFieldCharacter ( ) { @ org . simpleflatmapper . reflect . test . asm . SuppressWarnings ( "unchecked" ) org . simpleflatmapper . reflect . primitive . CharacterGetter < org . simpleflatmapper . test . beans . DbPrimitiveObjectFields > getter = ( ( org . simpleflatmapper . reflect . primitive . CharacterGetter < org . simpleflatmapper . test . beans . DbPrimitiveObjectFields > ) ( factory . createGetter ( org . simpleflatmapper . test . beans . DbPrimitiveObjectFields . class . getDeclaredField ( "pCharacter" ) ) ) ) ; objectField . pCharacter = 'g' ; "<AssertPlaceHolder>" ; } getCharacter ( java . sql . ResultSet ) { return ( ( char ) ( target . getInt ( column ) ) ) ; }
|
org . junit . Assert . assertEquals ( 'g' , getter . getCharacter ( objectField ) )
|
testGet_SkipParentIfAccumuloNS ( ) { c = new org . apache . accumulo . server . conf . NamespaceConfiguration ( Namespace . ACCUMULO . id ( ) , context , parent ) ; c . setZooCacheFactory ( zcf ) ; org . apache . accumulo . core . conf . Property p = org . apache . accumulo . core . conf . Property . INSTANCE_SECRET ; expect ( zc . get ( ( ( ( ( ( ( ( org . apache . accumulo . fate . zookeeper . ZooUtil . getRoot ( iid ) ) + ( org . apache . accumulo . core . Constants . ZNAMESPACES ) ) + "/" ) + ( Namespace . ACCUMULO . id ( ) ) ) + ( org . apache . accumulo . core . Constants . ZNAMESPACE_CONF ) ) + "/" ) + ( p . getKey ( ) ) ) ) ) . andReturn ( null ) ; replay ( zc ) ; "<AssertPlaceHolder>" ; } get ( org . apache . accumulo . core . conf . Property ) { java . lang . String v = props . get ( property . getKey ( ) ) ; if ( ( v == null ) & ( ( parent ) != null ) ) { v = parent . get ( property ) ; } return v ; }
|
org . junit . Assert . assertNull ( c . get ( Property . INSTANCE_SECRET ) )
|
testConfigurator ( ) { final com . github . rinde . rinsim . scenario . gendreau06 . Gendreau06Scenario scenario = com . github . rinde . rinsim . scenario . gendreau06 . Gendreau06Parser . parse ( new java . io . File ( ScenarioPaths . GENDREAU ) ) ; final com . github . rinde . rinsim . util . StochasticSupplier < com . github . rinde . rinsim . central . Solver > s = new com . github . rinde . rinsim . util . StochasticSupplier < com . github . rinde . rinsim . central . Solver > ( ) { @ com . github . rinde . rinsim . central . Override public com . github . rinde . rinsim . central . Solver get ( long seed ) { return com . github . rinde . rinsim . central . SolverValidator . wrap ( new com . github . rinde . rinsim . central . arrays . MultiVehicleSolverAdapter ( com . github . rinde . rinsim . central . arrays . ArraysSolverValidator . wrap ( new com . github . rinde . rinsim . central . arrays . RandomMVArraysSolver ( new org . apache . commons . math3 . random . MersenneTwister ( seed ) ) ) , javax . measure . unit . SI . MILLI ( SI . SECOND ) ) ) ; } } ; final com . github . rinde . rinsim . experiment . Experiment . Builder builder = com . github . rinde . rinsim . experiment . Experiment . builder ( ) . addScenario ( scenario ) . addConfiguration ( com . github . rinde . rinsim . central . Central . solverConfiguration ( s ) ) . withRandomSeed ( 123 ) ; final com . github . rinde . rinsim . experiment . ExperimentResults res1 = builder . perform ( ) ; final com . github . rinde . rinsim . experiment . ExperimentResults res2 = builder . perform ( ) ; "<AssertPlaceHolder>" ; } getResults ( ) { return com . google . common . collect . ImmutableSet . copyOf ( convertedPaths ) ; }
|
org . junit . Assert . assertEquals ( res1 . getResults ( ) , res2 . getResults ( ) )
|
getOctaves ( ) { org . smurn . jsift . ScaleSpace target = new org . smurn . jsift . ScaleSpace ( java . util . Arrays . asList ( dummy1 , dummy2 ) ) ; "<AssertPlaceHolder>" ; } getOctaves ( ) { return octaves ; }
|
org . junit . Assert . assertEquals ( java . util . Arrays . asList ( dummy1 , dummy2 ) , target . getOctaves ( ) )
|
testConfigureEnableDistanceAndEnableDistanceHistoric ( ) { System . out . println ( ( ( ( getTestTraceHead ( "enable_distance_historic" 0 ) ) + "--------<sp>When<sp>enable_distance<sp>and<sp>enable_distance_historic<sp>are<sp>configured,<sp>the<sp>second<sp>one<sp>value<sp>" ) + "is<sp>used" ) ) ; org . apache . flume . Context context = new org . apache . flume . Context ( ) ; context . put ( "enable_distance_historic" 2 , "true" ) ; context . put ( "enable_distance_historic" , "enable_distance_historic" 1 ) ; context . put ( "keys_conf_file" , "" ) ; com . telefonica . iot . cygnus . sinks . NGSICartoDBSink sink = new com . telefonica . iot . cygnus . sinks . NGSICartoDBSink ( ) ; sink . configure ( context ) ; try { "<AssertPlaceHolder>" ; System . out . println ( ( ( ( getTestTraceHead ( "enable_distance_historic" 0 ) ) + "-<sp>OK<sp>-<sp>Both<sp>'enable_distance'<sp>and<sp>'enable_distance_historic'<sp>where<sp>configured,<sp>but<sp>" ) + "'enable_distance_historic'<sp>value<sp>is<sp>used" ) ) ; } catch ( java . lang . AssertionError e ) { System . out . println ( ( ( ( getTestTraceHead ( "enable_distance_historic" 0 ) ) + "-<sp>FAIL<sp>-<sp>Both<sp>'enable_distance'<sp>and<sp>'enable_distnace_historic'<sp>where<sp>configured,<sp>but<sp>" ) + "'enable_distance'<sp>value<sp>is<sp>used" ) ) ; throw e ; } getEnableDistanceHistoric ( ) { return enableDistanceHistoric ; }
|
org . junit . Assert . assertTrue ( ( ! ( sink . getEnableDistanceHistoric ( ) ) ) )
|
testInvoke_lotsOfArgs ( ) { java . lang . Object firstArg = new java . lang . Object ( ) ; java . lang . Object [ ] args = new java . lang . Object [ ] { firstArg , new java . lang . Object [ ] { } , new java . lang . Object ( ) , new java . lang . Object ( ) } ; when ( soapCall . getSoapClientMethod ( ) ) . thenReturn ( com . google . api . ads . common . lib . soap . testing . MockSoapClient . class . getMethod ( "lotsOfArgsCall" , java . lang . Object . class , java . lang . Object [ ] . class , java . lang . Object . class , java . lang . Object . class ) ) ; when ( soapCall . getSoapClient ( ) ) . thenReturn ( new com . google . api . ads . common . lib . soap . testing . MockSoapClient ( ) ) ; when ( soapCall . getSoapArgs ( ) ) . thenReturn ( args ) ; java . lang . Object result = soapClientHandler . invoke ( soapCall ) ; "<AssertPlaceHolder>" ; } invoke ( org . apache . axis . MessageContext ) { if ( msgContext == null ) { throw org . apache . axis . AxisFault . makeFault ( new java . lang . NullPointerException ( "Null<sp>message<sp>context" ) ) ; } try { com . google . api . client . http . HttpResponse response = null ; com . google . api . client . http . HttpRequest postRequest = createHttpRequest ( msgContext ) ; response = postRequest . execute ( ) ; msgContext . setResponseMessage ( createResponseMessage ( response ) ) ; } catch ( java . lang . RuntimeException | javax . xml . soap . SOAPException | java . io . IOException e ) { throw org . apache . axis . AxisFault . makeFault ( e ) ; } }
|
org . junit . Assert . assertSame ( firstArg , result )
|
testSFBQ_3 ( ) { final lombok . val queue = new org . deeplearning4j . optimize . solvers . accumulation . SmartFancyBlockingQueue ( 1285601 , org . nd4j . linalg . factory . Nd4j . create ( 5 , 5 ) ) ; lombok . val threads = new java . util . ArrayList < java . lang . Thread > ( ) ; for ( int e = 0 ; e < 4 ; e ++ ) { lombok . val f = e ; lombok . val t = new java . lang . Thread ( new java . lang . Runnable ( ) { @ org . deeplearning4j . optimize . solvers . accumulation . Override public void run ( ) { int cnt = 0 ; while ( true ) { while ( cnt < 1000 ) { if ( ! ( queue . isEmpty ( ) ) ) { if ( ( cnt % 50 ) == 0 ) log . info ( "Thread<sp>{}:<sp>[{}]" , f , cnt ) ; lombok . val arr = queue . poll ( ) ; "<AssertPlaceHolder>" ; lombok . val local = arr . unsafeDuplication ( true ) ; cnt ++ ; } } break ; } } } ) ; t . start ( ) ; threads . add ( t ) ; } lombok . val b = new java . lang . Thread ( new java . lang . Runnable ( ) { @ org . deeplearning4j . optimize . solvers . accumulation . Override public void run ( ) { while ( true ) { queue . registerConsumers ( 4 ) ; org . deeplearning4j . util . ThreadUtils . uncheckedSleep ( 30 ) ; } } } ) ; b . setDaemon ( true ) ; b . start ( ) ; lombok . val writers = new java . util . ArrayList < java . lang . Thread > ( ) ; for ( int e = 0 ; e < 4 ; e ++ ) { lombok . val t = new java . lang . Thread ( new java . lang . Runnable ( ) { @ org . deeplearning4j . optimize . solvers . accumulation . Override public void run ( ) { for ( int e = 0 ; e < 250 ; e ++ ) { try { queue . put ( org . nd4j . linalg . factory . Nd4j . createUninitialized ( 5 , 5 ) . assign ( e ) ) ; java . lang . Thread . sleep ( 30 ) ; } catch ( java . lang . Exception ex ) { throw new java . lang . RuntimeException ( ex ) ; } } } } ) ; writers . add ( t ) ; t . start ( ) ; } for ( lombok . val t : writers ) t . join ( ) ; for ( lombok . val t : threads ) t . join ( ) ; } poll ( ) { lombok . val pointer = queue . poll ( ) ; if ( pointer != null ) ( counter ) -- ; return pointer ; }
|
org . junit . Assert . assertNotNull ( arr )
|
countShouldReturnZeroIfUnknownSubscriptionId ( ) { createConsumer ( owner ) ; java . util . List < java . lang . String > ids = new java . util . ArrayList ( ) ; ids . add ( "unknown-subId" ) ; int count = consumerCurator . countConsumers ( owner . getKey ( ) , typeLabels , skus , ids , contracts ) ; "<AssertPlaceHolder>" ; } getKey ( ) { return key ; }
|
org . junit . Assert . assertEquals ( 0 , count )
|
loadDrivers_InstallFails ( ) { java . util . List < org . osgi . service . device . DriverLocator > locators = new java . util . ArrayList < org . osgi . service . device . DriverLocator > ( ) ; org . osgi . service . device . DriverLocator dl = org . mockito . Mockito . mock ( org . osgi . service . device . DriverLocator . class , "dl" ) ; locators . add ( dl ) ; java . lang . String [ ] driverIds = new java . lang . String [ ] { "org.apache.felix.driver-1.0" , "org.apache.felix.driver-1.1" } ; for ( java . lang . String string : driverIds ) { org . mockito . Mockito . when ( dl . loadDriver ( string ) ) . thenReturn ( null ) ; org . mockito . Mockito . when ( m_context . installBundle ( ( ( DriverLoader . DRIVER_LOCATION_PREFIX ) + string ) , null ) ) . thenThrow ( new org . osgi . framework . BundleException ( "test<sp>exception" ) ) ; } java . util . List < org . osgi . framework . ServiceReference > refs = m_loader . loadDrivers ( locators , driverIds ) ; "<AssertPlaceHolder>" ; } size ( ) { return data . length ; }
|
org . junit . Assert . assertEquals ( "" , 0 , refs . size ( ) )
|
testJsonEncoderNewlineDelimited ( ) { java . io . OutputStream out = new java . io . ByteArrayOutputStream ( ) ; org . apache . avro . Schema ints = org . apache . avro . Schema . create ( Type . INT ) ; org . apache . avro . io . Encoder e = org . apache . avro . io . TestEncoders . factory . jsonEncoder ( ints , out ) ; java . lang . String separator = java . lang . System . getProperty ( "line.separator" ) ; org . apache . avro . generic . GenericDatumWriter < java . lang . Integer > writer = new org . apache . avro . generic . GenericDatumWriter ( ints ) ; writer . write ( 1 , e ) ; writer . write ( 2 , e ) ; e . flush ( ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( ( ( ( "ResolvingVisitor{" + "replace=" ) + ( replace ) ) + ",<sp>symbolTable=" ) + ( symbolTable ) ) + ",<sp>root=" ) + ( root ) ) + '}' ; }
|
org . junit . Assert . assertEquals ( ( ( "1" + separator ) + "2" ) , out . toString ( ) )
|
loadCatchesExceptionFromDocumentAccess ( ) { java . lang . Exception exception = new java . lang . RuntimeException ( ) ; doThrow ( exception ) . when ( this . patient ) . getXDocument ( ) ; org . phenotips . data . PatientData < org . phenotips . data . Gene > result = this . component . load ( this . patient ) ; "<AssertPlaceHolder>" ; verify ( this . mocker . getMockedLogger ( ) ) . error ( eq ( PatientDataController . ERROR_MESSAGE_LOAD_FAILED ) , anyString ( ) ) ; } load ( org . xwiki . bridge . DocumentModelBridge ) { try { return getEntityConstructor ( ) . newInstance ( document ) ; } catch ( java . lang . IllegalArgumentException | java . lang . reflect . InvocationTargetException ex ) { this . logger . info ( "Tried<sp>to<sp>load<sp>invalid<sp>entity<sp>of<sp>type<sp>[{}]<sp>from<sp>document<sp>[{}]" , getEntityXClassReference ( ) , ( document == null ? null : document . getDocumentReference ( ) ) ) ; } catch ( java . lang . InstantiationException | java . lang . IllegalAccessException ex ) { this . logger . error ( "Failed<sp>to<sp>instantiate<sp>primary<sp>entity<sp>of<sp>type<sp>[{}]<sp>from<sp>document<sp>[{}]:<sp>{}" , getEntityXClassReference ( ) , ( document == null ? null : document . getDocumentReference ( ) ) , ex . getMessage ( ) ) ; } return null ; }
|
org . junit . Assert . assertNull ( result )
|
testBug54602c ( ) { org . apache . tomcat . util . buf . B2CConverter conv = new org . apache . tomcat . util . buf . B2CConverter ( "UTF-8" ) ; org . apache . tomcat . util . buf . ByteChunk bc = new org . apache . tomcat . util . buf . ByteChunk ( ) ; org . apache . tomcat . util . buf . CharChunk cc = new org . apache . tomcat . util . buf . CharChunk ( ) ; bc . append ( org . apache . tomcat . util . buf . TestB2CConverter . UTF8_PARTIAL , 0 , org . apache . tomcat . util . buf . TestB2CConverter . UTF8_PARTIAL . length ) ; cc . allocate ( bc . getLength ( ) , ( - 1 ) ) ; conv . convert ( bc , cc , false ) ; java . lang . Exception e = null ; try { conv . convert ( bc , cc , true ) ; } catch ( java . nio . charset . MalformedInputException mie ) { e = mie ; } "<AssertPlaceHolder>" ; }
|
org . junit . Assert . assertNotNull ( e )
|
testEnqueue ( ) { final tlc2 . tool . TLCState expected = new tlc2 . tool . queue . DummyTLCState ( ) ; sQueue . enqueue ( expected ) ; tlc2 . tool . TLCState actual = sQueue . sDequeue ( ) ; "<AssertPlaceHolder>" ; } sDequeue ( ) { final byte [ ] bytes = sDequeueRaw ( ) ; if ( bytes != null ) { return toState ( bytes ) ; } return null ; }
|
org . junit . Assert . assertEquals ( "" , expected , actual )
|
testSetRoot ( ) { org . apache . ivy . util . ConfiguratorTest . City city = new org . apache . ivy . util . ConfiguratorTest . City ( ) ; conf . setRoot ( city ) ; "<AssertPlaceHolder>" ; } getCurrent ( ) { return objectStack . isEmpty ( ) ? null : objectStack . peek ( ) . getObject ( ) ; }
|
org . junit . Assert . assertEquals ( city , conf . getCurrent ( ) )
|
shouldCreateAndOverAddWellStructuredTreeModel ( ) { int size = storedObjectList . size ( ) ; "<AssertPlaceHolder>" ; int first = size / 3 ; int second = 2 * first ; int third = size ; org . swiftexplorer . gui . StoredObjectsTreeModel treeModel = new org . swiftexplorer . gui . StoredObjectsTreeModel ( rootContainer , storedObjectList . subList ( 0 , first ) ) ; treeModel . addAll ( storedObjectList . subList ( second , third ) ) ; treeModel . addAll ( storedObjectList . subList ( second , third ) ) ; treeModel . addAll ( storedObjectList . subList ( second , third ) ) ; treeModel . addAll ( storedObjectList . subList ( first , third ) ) ; verifyTreeStructure ( treeModel ) ; }
|
org . junit . Assert . assertTrue ( ( size > 10 ) )
|
testNewPseudoAtom_IAtom ( ) { org . openscience . cdk . interfaces . IChemObjectBuilder builder = org . openscience . cdk . AbstractChemObjectBuilderTest . rootObject . getBuilder ( ) ; org . openscience . cdk . interfaces . IPseudoAtom atom = builder . newInstance ( org . openscience . cdk . interfaces . IPseudoAtom . class , builder . newInstance ( org . openscience . cdk . interfaces . IAtom . class ) ) ; "<AssertPlaceHolder>" ; } newInstance ( java . lang . Class , java . lang . Object [ ] ) { return factory . ofClass ( clazz , params ) ; }
|
org . junit . Assert . assertNotNull ( atom )
|
testFindBySubmissionId ( ) { int facilityId1 = 100 ; int facilityId2 = 200 ; int submissionId1 = 101 ; int submissionId2 = 202 ; java . util . Date created1 = new java . util . Date ( dfm . parse ( "20110110" ) . getTime ( ) ) ; java . util . Date created2 = new java . util . Date ( dfm . parse ( "20131110" ) . getTime ( ) ) ; java . util . Date created3 = new java . util . Date ( dfm . parse ( "20120410" ) . getTime ( ) ) ; org . oscarehr . common . model . OcanStaffForm ocanStaffForm1 = new org . oscarehr . common . model . OcanStaffForm ( ) ; org . oscarehr . common . dao . utils . EntityDataGenerator . generateTestDataForModelClass ( ocanStaffForm1 ) ; ocanStaffForm1 . setFacilityId ( facilityId1 ) ; ocanStaffForm1 . setSubmissionId ( submissionId1 ) ; ocanStaffForm1 . setCreated ( created1 ) ; dao . persist ( ocanStaffForm1 ) ; org . oscarehr . common . model . OcanStaffForm ocanStaffForm2 = new org . oscarehr . common . model . OcanStaffForm ( ) ; org . oscarehr . common . dao . utils . EntityDataGenerator . generateTestDataForModelClass ( ocanStaffForm2 ) ; ocanStaffForm2 . setFacilityId ( facilityId2 ) ; ocanStaffForm2 . setSubmissionId ( submissionId2 ) ; ocanStaffForm2 . setCreated ( created2 ) ; dao . persist ( ocanStaffForm2 ) ; org . oscarehr . common . model . OcanStaffForm ocanStaffForm3 = new org . oscarehr . common . model . OcanStaffForm ( ) ; org . oscarehr . common . dao . utils . EntityDataGenerator . generateTestDataForModelClass ( ocanStaffForm3 ) ; ocanStaffForm3 . setFacilityId ( facilityId1 ) ; ocanStaffForm3 . setSubmissionId ( submissionId1 ) ; ocanStaffForm3 . setCreated ( created3 ) ; dao . persist ( ocanStaffForm3 ) ; java . util . List < org . oscarehr . common . model . OcanStaffForm > expectedResult = new java . util . ArrayList < org . oscarehr . common . model . OcanStaffForm > ( java . util . Arrays . asList ( ocanStaffForm3 , ocanStaffForm1 ) ) ; java . util . List < org . oscarehr . common . model . OcanStaffForm > result = dao . findBySubmissionId ( facilityId1 , submissionId1 ) ; org . apache . log4j . Logger logger = org . oscarehr . util . MiscUtils . getLogger ( ) ; if ( ( result . size ( ) ) != ( expectedResult . size ( ) ) ) { logger . warn ( "Array<sp>sizes<sp>do<sp>not<sp>match." ) ; org . junit . Assert . fail ( "Array<sp>sizes<sp>do<sp>not<sp>match." ) ; } for ( int i = 0 ; i < ( expectedResult . size ( ) ) ; i ++ ) { if ( ! ( expectedResult . get ( i ) . equals ( result . get ( i ) ) ) ) { logger . warn ( "Items<sp>do<sp>not<sp>match." ) ; org . junit . Assert . fail ( "Items<sp>do<sp>not<sp>match." ) ; } } "<AssertPlaceHolder>" ; } get ( java . lang . String ) { try { return terser . get ( path ) ; } catch ( ca . uhn . hl7v2 . HL7Exception e ) { oscar . oscarLab . ca . all . parsers . CLSHandler . logger . warn ( ( "Unable<sp>to<sp>get<sp>field<sp>at<sp>" + path ) , e ) ; return null ; } }
|
org . junit . Assert . assertTrue ( true )
|
test ( ) { doInJPA ( new com . vladmihalcea . hibernate . type . util . transaction . JPATransactionFunction < java . lang . Void > ( ) { @ com . vladmihalcea . hibernate . type . basic . Override public com . vladmihalcea . hibernate . type . basic . Void apply ( com . vladmihalcea . hibernate . type . basic . EntityManager entityManager ) { com . vladmihalcea . hibernate . type . basic . PostgreSQLEnumTest . Post post = new com . vladmihalcea . hibernate . type . basic . PostgreSQLEnumTest . Post ( ) ; post . setId ( 1L ) ; post . setTitle ( "High-Performance<sp>Java<sp>Persistence" ) ; post . setStatus ( com . vladmihalcea . hibernate . type . basic . PostgreSQLEnumTest . PostStatus . PENDING ) ; entityManager . persist ( post ) ; return null ; } } ) ; doInJPA ( new com . vladmihalcea . hibernate . type . util . transaction . JPATransactionFunction < java . lang . Void > ( ) { @ com . vladmihalcea . hibernate . type . basic . Override public com . vladmihalcea . hibernate . type . basic . Void apply ( com . vladmihalcea . hibernate . type . basic . EntityManager entityManager ) { com . vladmihalcea . hibernate . type . basic . PostgreSQLEnumTest . Post post = entityManager . find ( com . vladmihalcea . hibernate . type . basic . PostgreSQLEnumTest . Post . class , 1L ) ; "<AssertPlaceHolder>" ; return null ; } } ) ; } setTitle ( java . lang . String ) { this . title = title ; }
|
org . junit . Assert . assertEquals ( com . vladmihalcea . hibernate . type . basic . PostgreSQLEnumTest . PostStatus . PENDING , post . getStatus ( ) )
|
testMarshallXML ( ) { course . freedb . domain . Dictionary dictionary = generateFluentStructure ( ) ; java . lang . String outputXml = "" ; try { javax . xml . bind . JAXBContext context = javax . xml . bind . JAXBContext . newInstance ( "course.freedb.domain" ) ; javax . xml . bind . Marshaller marshaller = context . createMarshaller ( ) ; marshaller . setProperty ( Marshaller . JAXB_FORMATTED_OUTPUT , true ) ; java . io . Writer writer = new java . io . StringWriter ( ) ; course . freedb . domain . ObjectFactory objectFactory = new course . freedb . domain . ObjectFactory ( ) ; javax . xml . bind . JAXBElement < course . freedb . domain . Dictionary > element = objectFactory . createDictionary ( dictionary ) ; marshaller . marshal ( element , writer ) ; outputXml = writer . toString ( ) ; writer . close ( ) ; } catch ( javax . xml . bind . JAXBException e ) { e . printStackTrace ( ) ; } catch ( java . io . IOException e ) { e . printStackTrace ( ) ; org . junit . Assert . fail ( e . getMessage ( ) ) ; } System . out . println ( outputXml ) ; "<AssertPlaceHolder>" ; } getMessage ( ) { return message ; }
|
org . junit . Assert . assertTrue ( ( ( outputXml != null ) && ( ! ( outputXml . isEmpty ( ) ) ) ) )
|
testMatch ( ) { de . malkusch . whoisServerList . publicSuffixList . rule . RuleMatcher matcher = new de . malkusch . whoisServerList . publicSuffixList . rule . RuleMatcher ( rule ) ; "<AssertPlaceHolder>" ; } match ( java . lang . String ) { if ( domain == null ) { return null ; } java . lang . String match = matcher . match ( domain ) ; if ( match == null ) { return null ; } if ( ! ( isExceptionRule ( ) ) ) { return match ; } else { java . lang . String [ ] labels = de . malkusch . whoisServerList . publicSuffixList . util . DomainUtil . splitLabels ( match ) ; java . lang . String [ ] reducedLabels = java . util . Arrays . copyOfRange ( labels , 1 , labels . length ) ; return de . malkusch . whoisServerList . publicSuffixList . util . DomainUtil . joinLabels ( reducedLabels ) ; } }
|
org . junit . Assert . assertEquals ( match , matcher . match ( domain ) )
|
testAddedJson ( ) { net . nikr . eve . jeveasset . io . local . AssetAddedReaderTest . TestAssetAddedReader . load ( ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { java . lang . String sql = "SELECT<sp>*<sp>FROM<sp>assetadded" ; try ( java . sql . Connection connection = java . sql . DriverManager . getConnection ( net . nikr . eve . jeveasset . data . settings . AssetAddedData . CONNECTION_URL ) ; java . sql . PreparedStatement statement = connection . prepareStatement ( sql ) ; java . sql . ResultSet rs = statement . executeQuery ( ) ) { while ( rs . next ( ) ) { return false ; } } catch ( java . sql . SQLException ex ) { net . nikr . eve . jeveasset . data . settings . AssetAddedData . LOG . error ( ex . getMessage ( ) , ex ) ; } return true ; }
|
org . junit . Assert . assertTrue ( ( ! ( net . nikr . eve . jeveasset . data . settings . AssetAddedData . isEmpty ( ) ) ) )
|
testMapping ( ) { com . github . jmkgreen . morphia . mapping . ClassMappingTest . E e = new com . github . jmkgreen . morphia . mapping . ClassMappingTest . E ( ) ; e . testClass = java . util . LinkedList . class ; ds . save ( e ) ; e = ds . get ( e ) ; "<AssertPlaceHolder>" ; } get ( java . lang . Class ) { com . github . jmkgreen . morphia . logging . MorphiaLoggerFactory . init ( ) ; return com . github . jmkgreen . morphia . logging . MorphiaLoggerFactory . loggerFactory . get ( c ) ; }
|
org . junit . Assert . assertEquals ( java . util . LinkedList . class , e . testClass )
|
testOne ( ) { com . clearspring . analytics . stream . cardinality . HyperLogLogPlus one = new com . clearspring . analytics . stream . cardinality . HyperLogLogPlus ( 8 , 25 ) ; one . offer ( "a" ) ; "<AssertPlaceHolder>" ; } cardinality ( ) { double registerSum = 0 ; int count = registerSet . count ; double zeros = 0.0 ; for ( int j = 0 ; j < ( registerSet . count ) ; j ++ ) { int val = registerSet . get ( j ) ; registerSum += 1.0 / ( 1 << val ) ; if ( val == 0 ) { zeros ++ ; } } double estimate = ( alphaMM ) * ( 1 / registerSum ) ; if ( estimate <= ( ( 5.0 / 2.0 ) * count ) ) { return java . lang . Math . round ( com . clearspring . analytics . stream . cardinality . HyperLogLog . linearCounting ( count , zeros ) ) ; } else { return java . lang . Math . round ( estimate ) ; } }
|
org . junit . Assert . assertEquals ( 1 , one . cardinality ( ) )
|
whenGetDay_thenCorrectDay ( ) { int actualDayOfMonth = offsetDateTimeExtractYearMonthDayIntegerValues . getDay ( offsetDateTime ) ; "<AssertPlaceHolder>" ; } getDay ( java . time . LocalDate ) { return localDate . getDayOfMonth ( ) ; }
|
org . junit . Assert . assertThat ( actualDayOfMonth , org . hamcrest . CoreMatchers . is ( 3 ) )
|
whenDeletingARule_cacheContainsUpdatedEntry ( ) { try ( grakn . core . server . session . TransactionOLTP tx = grakn . core . graql . reasoner . cache . RuleCacheIT . ruleApplicabilitySession . transaction ( ) . write ( ) ) { tx . execute ( graql . lang . Graql . undefine ( graql . lang . Graql . type ( "rule-0" ) . sub ( "rule" ) ) ) ; tx . commit ( ) ; } try ( grakn . core . server . session . TransactionOLTP tx = grakn . core . graql . reasoner . cache . RuleCacheIT . ruleApplicabilitySession . transaction ( ) . write ( ) ) { grakn . core . concept . type . Type binary = tx . getType ( grakn . core . concept . Label . of ( "binary" ) ) ; java . util . Set < grakn . core . concept . type . Rule > rules = tx . ruleCache ( ) . getRulesWithType ( binary ) . collect ( java . util . stream . Collectors . toSet ( ) ) ; "<AssertPlaceHolder>" ; } } isEmpty ( ) { return ( ( ! ( isLookupExplanation ( ) ) ) && ( ! ( isRuleExplanation ( ) ) ) ) && ( getAnswers ( ) . isEmpty ( ) ) ; }
|
org . junit . Assert . assertTrue ( rules . isEmpty ( ) )
|
testMajorIssueConfigRoundTrip ( ) { java . lang . String majorIssue = "majorIssue" ; org . batfish . identifiers . IssueSettingsId issueSettingsId = new org . batfish . identifiers . IssueSettingsId ( "issueSettingsId" ) ; org . batfish . identifiers . NetworkId network = new org . batfish . identifiers . NetworkId ( "network" ) ; org . batfish . datamodel . answers . MinorIssueConfig minorIssueConfig = new org . batfish . datamodel . answers . MinorIssueConfig ( "minorIssue" , 100 , "www.google.com" ) ; org . batfish . datamodel . answers . MajorIssueConfig majorIssueConfig = new org . batfish . datamodel . answers . MajorIssueConfig ( majorIssue , com . google . common . collect . ImmutableList . of ( minorIssueConfig ) ) ; _storage . storeMajorIssueConfig ( network , issueSettingsId , majorIssueConfig ) ; "<AssertPlaceHolder>" ; } loadMajorIssueConfig ( org . batfish . identifiers . NetworkId , org . batfish . identifiers . IssueSettingsId ) { java . nio . file . Path path = _d . getMajorIssueConfigDir ( network , majorIssueType ) ; if ( ! ( java . nio . file . Files . exists ( path ) ) ) { return null ; } java . lang . String majorIssueFileText = org . batfish . common . util . CommonUtil . readFile ( path ) ; try { return org . batfish . common . util . BatfishObjectMapper . mapper ( ) . readValue ( majorIssueFileText , org . batfish . datamodel . answers . MajorIssueConfig . class ) ; } catch ( java . io . IOException e ) { _logger . errorf ( "ERROR:<sp>Could<sp>not<sp>cast<sp>file<sp>for<sp>major<sp>issue<sp>settings<sp>with<sp>ID<sp>%s<sp>in<sp>network<sp>%s<sp>to<sp>MajorIssueConfig:<sp>%s" , majorIssueType , network , com . google . common . base . Throwables . getStackTraceAsString ( e ) ) ; return null ; } }
|
org . junit . Assert . assertThat ( _storage . loadMajorIssueConfig ( network , issueSettingsId ) , org . hamcrest . Matchers . equalTo ( majorIssueConfig ) )
|
assertFindBlankStatisticTimes ( ) { for ( io . elasticjob . cloud . statistics . StatisticInterval each : io . elasticjob . cloud . statistics . StatisticInterval . values ( ) ) { int num = - 2 ; for ( java . util . Date eachTime : testStatisticJob . findBlankStatisticTimes ( io . elasticjob . cloud . scheduler . statistics . util . StatisticTimeUtils . getStatisticTime ( each , ( num - 1 ) ) , each ) ) { "<AssertPlaceHolder>" ; } } } getStatisticTime ( io . elasticjob . cloud . statistics . StatisticInterval , int ) { java . util . Calendar calendar = java . util . Calendar . getInstance ( ) ; calendar . set ( Calendar . MILLISECOND , 0 ) ; calendar . set ( Calendar . SECOND , 0 ) ; switch ( interval ) { case DAY : calendar . set ( Calendar . MINUTE , 0 ) ; calendar . set ( Calendar . HOUR_OF_DAY , 0 ) ; calendar . add ( Calendar . DATE , offset ) ; break ; case HOUR : calendar . set ( Calendar . MINUTE , 0 ) ; calendar . add ( Calendar . HOUR_OF_DAY , offset ) ; break ; case MINUTE : default : calendar . add ( Calendar . MINUTE , offset ) ; break ; } return calendar . getTime ( ) ; }
|
org . junit . Assert . assertThat ( eachTime . getTime ( ) , org . hamcrest . CoreMatchers . is ( io . elasticjob . cloud . scheduler . statistics . util . StatisticTimeUtils . getStatisticTime ( each , ( num ++ ) ) . getTime ( ) ) )
|
test2 ( ) { final java . io . File tempFile = java . io . File . createTempFile ( "ConcurrentWriteTest_test2" , ".bin" ) ; final long limit = 4000000L ; final long partition = limit / 4L ; final java . io . RandomAccessFile tmpRAF0 = new tlc2 . util . BufferedRandomAccessFile ( tempFile , "rw" ) ; tmpRAF0 . setLength ( ( limit * ( Long . BYTES ) ) ) ; final java . io . RandomAccessFile tmpRAF1 = new tlc2 . util . BufferedRandomAccessFile ( tempFile , "rw" ) ; tmpRAF1 . setLength ( ( limit * ( Long . BYTES ) ) ) ; tmpRAF1 . seek ( ( partition * ( Long . BYTES ) ) ) ; final java . io . RandomAccessFile tmpRAF2 = new tlc2 . util . BufferedRandomAccessFile ( tempFile , "rw" ) ; tmpRAF2 . setLength ( ( limit * ( Long . BYTES ) ) ) ; tmpRAF2 . seek ( ( ( 2L * partition ) * ( Long . BYTES ) ) ) ; final java . io . RandomAccessFile tmpRAF3 = new tlc2 . util . BufferedRandomAccessFile ( tempFile , "rw" ) ; tmpRAF3 . setLength ( ( limit * ( Long . BYTES ) ) ) ; tmpRAF3 . seek ( ( ( 3L * partition ) * ( Long . BYTES ) ) ) ; for ( long i = 0L ; i < ( 1L * partition ) ; i ++ ) { tmpRAF0 . writeLong ( i ) ; } for ( long i = 1L * partition ; i < ( 2L * partition ) ; i ++ ) { tmpRAF1 . writeLong ( i ) ; } for ( long i = 2L * partition ; i < ( 3L * partition ) ; i ++ ) { tmpRAF2 . writeLong ( i ) ; } for ( long i = 3L * partition ; i < ( 4L * partition ) ; i ++ ) { tmpRAF3 . writeLong ( i ) ; } tmpRAF0 . close ( ) ; tmpRAF1 . close ( ) ; tmpRAF2 . close ( ) ; tmpRAF3 . close ( ) ; final java . io . RandomAccessFile tmpRAF = new tlc2 . util . BufferedRandomAccessFile ( tempFile , "r" ) ; for ( long i = 0L ; i < limit ; i ++ ) { "<AssertPlaceHolder>" ; } tmpRAF . close ( ) ; } readLong ( ) { this . readFully ( this . temp , 0 , 8 ) ; long res = temp [ 0 ] ; res <<= 8 ; res |= ( temp [ 1 ] ) & 255 ; res <<= 8 ; res |= ( temp [ 2 ] ) & 255 ; res <<= 8 ; res |= ( temp [ 3 ] ) & 255 ; res <<= 8 ; res |= ( temp [ 4 ] ) & 255 ; res <<= 8 ; res |= ( temp [ 5 ] ) & 255 ; res <<= 8 ; res |= ( temp [ 6 ] ) & 255 ; res <<= 8 ; res |= ( temp [ 7 ] ) & 255 ; return res ; }
|
org . junit . Assert . assertEquals ( i , tmpRAF . readLong ( ) )
|
testCount ( ) { final com . spotify . heroic . aggregation . simple . QuantileBucket b = new com . spotify . heroic . aggregation . simple . QuantileBucket ( 0 , 0.5 , com . spotify . heroic . aggregation . simple . QuantileBucketTest . ERROR ) ; b . updatePoint ( com . spotify . heroic . aggregation . simple . QuantileBucketTest . TAGS , new com . spotify . heroic . metric . Point ( 0 , 1337.0 ) ) ; "<AssertPlaceHolder>" ; } value ( ) { return limits ; }
|
org . junit . Assert . assertEquals ( 1337.0 , b . value ( ) , 0.0 )
|
testKeyIterator ( ) { org . redisson . api . RMapRx < java . lang . Integer , java . lang . Integer > map = redisson . getMap ( "simple" ) ; sync ( map . put ( 1 , 0 ) ) ; sync ( map . put ( 3 , 5 ) ) ; sync ( map . put ( 4 , 6 ) ) ; sync ( map . put ( 7 , 8 ) ) ; java . util . List < java . lang . Integer > keys = new java . util . ArrayList < java . lang . Integer > ( java . util . Arrays . asList ( 1 , 3 , 4 , 7 ) ) ; for ( java . util . Iterator < java . lang . Integer > iterator = toIterator ( map . keyIterator ( ) ) ; iterator . hasNext ( ) ; ) { java . lang . Integer value = iterator . next ( ) ; if ( ! ( keys . remove ( value ) ) ) { org . junit . Assert . fail ( ) ; } } "<AssertPlaceHolder>" ; } size ( ) { return ( ( long ) ( cache . getNativeCache ( ) . size ( ) ) ) ; }
|
org . junit . Assert . assertEquals ( 0 , keys . size ( ) )
|
importProducts ( ) { int initial = jdbcTemplate . queryForInt ( "select<sp>count(1)<sp>from<sp>product" ) ; jobLauncher . run ( job , new org . springframework . batch . core . JobParametersBuilder ( ) . addString ( "inputResource" , "classpath:/input/products.zip" ) . addString ( "targetDirectory" , "./target/importproductsbatch/" ) . addString ( "targetFile" , "products.txt" ) . addLong ( "timestamp" , java . lang . System . currentTimeMillis ( ) ) . toJobParameters ( ) ) ; int nbOfNewProducts = 7 ; "<AssertPlaceHolder>" ; }
|
org . junit . Assert . assertEquals ( ( initial + nbOfNewProducts ) , jdbcTemplate . queryForInt ( "select<sp>count(1)<sp>from<sp>product" ) )
|
testPutQueueWithSlash ( ) { final java . lang . String queueWithSlash = "MY/QUEUE" ; final java . lang . String cmdLine = ( ( ( ( getConnectCommand ( ) ) + "-" ) + ( CMD_PUT ) ) + "<sp>\"test\"<sp>" ) + queueWithSlash ; System . out . println ( ( "Testing<sp>cmd:<sp>" + cmdLine ) ) ; a . run ( cmdLine . split ( "<sp>" ) ) ; javax . jms . MessageConsumer mc = session . createConsumer ( session . createQueue ( queueWithSlash ) ) ; javax . jms . TextMessage msg = ( ( javax . jms . TextMessage ) ( mc . receive ( co . nordlander . a . BaseTest . TEST_TIMEOUT ) ) ) ; "<AssertPlaceHolder>" ; } run ( java . lang . String [ ] ) { org . apache . commons . cli . Options opts = createOptions ( ) ; if ( ( args . length ) == 0 ) { org . apache . commons . cli . HelpFormatter helpFormatter = new org . apache . commons . cli . HelpFormatter ( ) ; helpFormatter . printHelp ( "java<sp>-jar<sp>a-<version>-with-dependencies.jar" , opts , true ) ; java . lang . System . exit ( 0 ) ; } org . apache . commons . cli . CommandLineParser cmdParser = new org . apache . commons . cli . PosixParser ( ) ; try { cmdLine = cmdParser . parse ( opts , args ) ; if ( cmdLine . hasOption ( co . nordlander . a . A . CMD_VERSION ) ) { executeShowVersion ( ) ; return ; } co . nordlander . a . A . Protocol protocol = co . nordlander . a . A . Protocol . OpenWire ; if ( cmdLine . hasOption ( co . nordlander . a . A . CMD_AMQP ) ) { protocol = co . nordlander . a . A . Protocol . AMQP ; } else if ( cmdLine . hasOption ( co . nordlander . a . A . CMD_ARTEMIS_CORE ) ) { protocol = co . nordlander . a . A . Protocol . ArtemisCore ; } connect ( cmdLine . getOptionValue ( co . nordlander . a . A . CMD_BROKER , "tcp://localhost:61616" ) , cmdLine . getOptionValue ( co . nordlander . a . A . CMD_USER ) , cmdLine . getOptionValue ( co . nordlander . a . A . CMD_PASS ) , protocol , cmdLine . getOptionValue ( co . nordlander . a . A . CMD_JNDI , "" ) , cmdLine . hasOption ( co . nordlander . a . A . CMD_NO_TRANSACTION_SUPPORT ) ) ; long startTime = java . lang . System . currentTimeMillis ( ) ; executeCommandLine ( cmdLine ) ; long stopTime = java . lang . System . currentTimeMillis ( ) ; long elapsedTime = stopTime - startTime ; output ( "Operation<sp>completed<sp>in<sp>" , java . lang . Long . toString ( elapsedTime ) , "ms<sp>(excluding<sp>connect)" ) ; } finally { try { if ( ( sess ) != null ) { sess . close ( ) ; } if ( ( conn ) != null ) { conn . close ( ) ; } } catch ( javax . jms . JMSException e2 ) { e2 . printStackTrace ( ) ; } } co . nordlander . a . A . logger . debug ( "Active<sp>threads<sp>{}" , java . lang . Thread . activeCount ( ) ) ; co . nordlander . a . A . logger . debug ( "At<sp>the<sp>end<sp>of<sp>the<sp>road" ) ; }
|
org . junit . Assert . assertEquals ( "test" , msg . getText ( ) )
|
testVerifyProfile1 ( ) { java . util . List < org . commonjava . maven . ext . common . model . Project > p = getProject ( ) ; org . commonjava . maven . ext . core . ManipulationManager m = new org . commonjava . maven . ext . core . ManipulationManager ( null , java . util . Collections . emptyMap ( ) , java . util . Collections . emptyMap ( ) , null ) ; org . commonjava . maven . ext . core . ManipulationSession ms = org . commonjava . maven . ext . core . fixture . TestUtils . createSession ( null ) ; m . init ( ms ) ; java . util . Set < java . lang . String > activeProfiles = ( ( java . util . Set < java . lang . String > ) ( org . commonjava . maven . ext . core . fixture . TestUtils . executeMethod ( m , "parseActiveProfiles" , new java . lang . Class [ ] { org . commonjava . maven . ext . core . ManipulationSession . class , java . util . List . class } , new java . lang . Object [ ] { ms , p } ) ) ) ; logger . info ( "Returning<sp>active<sp>profiles<sp>of<sp>{}<sp>" , activeProfiles ) ; "<AssertPlaceHolder>" ; } size ( ) { return ( ( ( list . size ( ) ) + ( size ) ) - 1 ) / ( size ) ; }
|
org . junit . Assert . assertEquals ( 2 , activeProfiles . size ( ) )
|
assertSkipHint ( ) { java . lang . String comment = "/*--xyz<sp>\n<sp>WHERE<sp>XX=1<sp>//xyz*/" ; java . lang . String sql = ( "SELECT<sp>*<sp>FROM<sp>XXX_TABLE<sp>" + comment ) + "WHERE<sp>YY>2" ; org . apache . shardingsphere . core . parse . old . lexer . analyzer . Tokenizer tokenizer = new org . apache . shardingsphere . core . parse . old . lexer . analyzer . Tokenizer ( sql , dictionary , sql . indexOf ( "/" ) ) ; int expected = ( sql . indexOf ( "/" ) ) + ( comment . length ( ) ) ; "<AssertPlaceHolder>" ; } skipHint ( ) { return untilCommentAndHintTerminateSign ( org . apache . shardingsphere . core . parse . old . lexer . analyzer . Tokenizer . HINT_BEGIN_SYMBOL_LENGTH ) ; }
|
org . junit . Assert . assertThat ( tokenizer . skipHint ( ) , org . hamcrest . CoreMatchers . is ( expected ) )
|
shouldSerializeGenericClass ( ) { java . lang . String expectedResult = "{\"genericWrapper\":<sp>{\"entityList\":<sp>[{\"name\":<sp>\"washington<sp>botelho\"},{\"name\":<sp>\"washington<sp>botelho\"}],\"total\":<sp>2}}" ; java . util . Collection < br . com . caelum . vraptor . serialization . xstream . XStreamJSONSerializationTest . Client > entityList = new java . util . ArrayList < br . com . caelum . vraptor . serialization . xstream . XStreamJSONSerializationTest . Client > ( ) ; entityList . add ( new br . com . caelum . vraptor . serialization . xstream . XStreamJSONSerializationTest . Client ( "washington<sp>botelho" ) ) ; entityList . add ( new br . com . caelum . vraptor . serialization . xstream . XStreamJSONSerializationTest . Client ( "washington<sp>botelho" ) ) ; br . com . caelum . vraptor . serialization . xstream . XStreamJSONSerializationTest . GenericWrapper < br . com . caelum . vraptor . serialization . xstream . XStreamJSONSerializationTest . Client > wrapper = new br . com . caelum . vraptor . serialization . xstream . XStreamJSONSerializationTest . GenericWrapper < br . com . caelum . vraptor . serialization . xstream . XStreamJSONSerializationTest . Client > ( entityList , entityList . size ( ) ) ; serialization . from ( wrapper ) . include ( "entityList" ) . serialize ( ) ; "<AssertPlaceHolder>" ; } result ( ) { return new java . lang . String ( stream . toByteArray ( ) ) ; }
|
org . junit . Assert . assertThat ( result ( ) , org . hamcrest . Matchers . is ( org . hamcrest . Matchers . equalTo ( expectedResult ) ) )
|
testAddNotification_mulitpleFinalRecipAdded_assertTxAdded ( ) { org . nhindirect . monitor . processor . impl . DefaultDuplicateNotificationStateManager mgr = new org . nhindirect . monitor . processor . impl . DefaultDuplicateNotificationStateManager ( ) ; mgr . setDao ( notifDao ) ; org . nhindirect . common . tx . model . Tx tx = org . nhindirect . monitor . util . TestUtils . makeMessage ( TxMessageType . DSN , "1234" , "5678" , "" , "" , "gm2552@cerner.com,ah4626@cerner.com" ) ; mgr . addNotification ( tx ) ; java . util . Set < java . lang . String > addedAddr = notifDao . getReceivedAddresses ( "5678" , java . util . Arrays . asList ( "gm2552@cerner.com" , "ah4626@cerner.com" ) ) ; "<AssertPlaceHolder>" ; }
|
org . junit . Assert . assertEquals ( 2 , addedAddr . size ( ) )
|
testToXml ( ) { addElevationModel ( "test1" , "file://sun-web-common/sandpit/symbolic-links/world-wind/current/dataset/standard/layers/earth_elevation_model.xml" ) ; addExaggerator ( 1.0 , 0.0 ) ; addExaggerator ( 2.0 , 100.0 ) ; addExaggerator ( 3.0 , 200.0 ) ; addKeyFrame ( 0 , 1.2 , getParameter ( 0 ) ) ; addKeyFrame ( 10 , 2.5 , getParameter ( 0 ) ) ; addKeyFrame ( 20 , 3.5 , getParameter ( 2 ) ) ; org . w3c . dom . Document xmlDocument = writeClassToBeTestedToXml ( ) ; java . io . ByteArrayOutputStream resultStream = writeDocumentToStream ( xmlDocument ) ; java . lang . String result = normalise ( new java . lang . String ( resultStream . toByteArray ( ) ) ) ; java . lang . String expected = normalise ( au . gov . ga . worldwind . test . util . TestUtils . readStreamToString ( getClass ( ) . getResourceAsStream ( "animatableElevationXmlSnippet.xml" ) ) ) ; "<AssertPlaceHolder>" ; } readStreamToString ( java . io . InputStream ) { java . io . BufferedReader reader = new java . io . BufferedReader ( new java . io . InputStreamReader ( stream ) ) ; java . lang . StringBuffer result = new java . lang . StringBuffer ( ) ; java . lang . String readLine = null ; while ( ( readLine = reader . readLine ( ) ) != null ) { if ( ( result . length ( ) ) > 0 ) { result . append ( '\n' ) ; } result . append ( readLine ) ; } reader . close ( ) ; return result . toString ( ) ; }
|
org . junit . Assert . assertEquals ( expected , result )
|
testAnotherPatterns1 ( ) { org . apache . storm . utils . Utils prevUtils = null ; try { org . apache . storm . utils . Utils mockedUtil = mock ( org . apache . storm . utils . Utils . class ) ; prevUtils = org . apache . storm . utils . Utils . setInstance ( mockedUtil ) ; when ( mockedUtil . hostname ( ) ) . thenReturn ( expectedHost ) ; final java . io . File file = new java . io . File ( java . lang . String . join ( File . separator , "src" , "test" , "X" 3 ) , "test-worker.log.test" ) ; java . lang . String pattern = org . jooq . lambda . Seq . range ( 0 , 1024 ) . map ( ( x ) -> "X" ) . collect ( java . util . stream . Collectors . joining ( ) ) ; java . util . Map < java . lang . String , java . lang . Object > expected = new java . util . HashMap ( ) ; expected . put ( "X" 6 , "no" ) ; expected . put ( "X" 0 , pattern ) ; expected . put ( "X" 5 , 0 ) ; expected . put ( "nextByteOffset" , 6183 ) ; java . util . List < java . util . Map < java . lang . String , java . lang . Object > > matches = new java . util . ArrayList ( ) ; matches . add ( buildMatchData ( 4075 , "X" 1 , "\nThe<sp>following<sp>max-size<sp>match<sp>straddles<sp>a<sp>1024<sp>byte<sp>buffer.\nXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX" , pattern , ( ( ( ( ( "/api/v1/log?file=test" + ( encodedFileSeparator ( ) ) ) + "X" 3 ) + ( encodedFileSeparator ( ) ) ) + ( file . getName ( ) ) ) + "X" 2 ) ) ) ; matches . add ( buildMatchData ( 5159 , "XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX\nThe<sp>following<sp>max-size<sp>match<sp>straddles<sp>a<sp>1024<sp>byte<sp>buffer.\n" , "\n\nHere<sp>are<sp>four<sp>non-ascii<sp>1-byte<sp>UTF-8<sp>characters:<sp>αβγδε\n\nneedle\n\nHere<sp>are<sp>four<sp>printable<sp>2-byte<sp>UTF-8<sp>characters:<sp>¡¢£¤" , pattern , ( ( ( ( ( "/api/v1/log?file=test" + ( encodedFileSeparator ( ) ) ) + "X" 3 ) + ( encodedFileSeparator ( ) ) ) + ( file . getName ( ) ) ) + "X" 2 ) ) ) ; expected . put ( "X" 4 , matches ) ; org . apache . storm . daemon . logviewer . handler . LogviewerLogSearchHandler handler = org . apache . storm . daemon . logviewer . handler . LogviewerLogSearchHandlerTest . getSearchHandlerWithPort ( expectedPort ) ; java . util . Map < java . lang . String , java . lang . Object > searchResult = handler . substringSearch ( file . toPath ( ) , pattern , 2 ) ; "<AssertPlaceHolder>" ; } finally { org . apache . storm . utils . Utils . setInstance ( prevUtils ) ; } } substringSearch ( java . nio . file . Path , java . lang . String , int ) { return substringSearch ( file , searchString , false , numMatches , 0 ) ; }
|
org . junit . Assert . assertEquals ( expected , searchResult )
|
autoCreateLocalDirCanBeDisabled ( ) { org . springframework . context . annotation . AnnotationConfigApplicationContext context = new org . springframework . context . annotation . AnnotationConfigApplicationContext ( ) ; org . springframework . boot . test . EnvironmentTestUtils . addEnvironment ( context , "sftp.autoCreateLocalDir:false" ) ; context . register ( org . springframework . cloud . stream . app . sftp . source . SftpSourcePropertiesTests . Conf . class ) ; context . refresh ( ) ; org . springframework . cloud . stream . app . sftp . source . SftpSourceProperties properties = context . getBean ( org . springframework . cloud . stream . app . sftp . source . SftpSourceProperties . class ) ; "<AssertPlaceHolder>" ; } isAutoCreateLocalDir ( ) { return autoCreateLocalDir ; }
|
org . junit . Assert . assertTrue ( ( ! ( properties . isAutoCreateLocalDir ( ) ) ) )
|
shouldPopulateDtoWithActiveBranches ( ) { final java . lang . Short branchId = userContext . getBranchId ( ) ; final java . lang . Short branchId2 = java . lang . Short . valueOf ( "2" ) ; final java . lang . Short levelId = OfficeLevel . BRANCHOFFICE . getValue ( ) ; org . mifos . dto . domain . OfficeDetailsDto officeStub1 = new org . mifos . dto . domain . OfficeDetailsDto ( branchId , "branchName1" , levelId , 1 ) ; org . mifos . dto . domain . OfficeDetailsDto officeStub2 = new org . mifos . dto . domain . OfficeDetailsDto ( branchId2 , "branchName2" , levelId , 1 ) ; java . util . List < org . mifos . dto . domain . OfficeDetailsDto > activeOffices = java . util . Arrays . asList ( officeStub1 , officeStub2 ) ; when ( officePersistence . getActiveOffices ( branchId ) ) . thenReturn ( activeOffices ) ; org . mifos . application . servicefacade . CollectionSheetEntryFormDto formDto = collectionSheetServiceFacadeWebTier . loadAllActiveBranchesAndSubsequentDataIfApplicable ( userContext ) ; "<AssertPlaceHolder>" ; } getActiveBranchesList ( ) { return this . activeBranchesList ; }
|
org . junit . Assert . assertThat ( formDto . getActiveBranchesList ( ) , org . hamcrest . CoreMatchers . is ( activeOffices ) )
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.