input stringlengths 28 18.7k | output stringlengths 39 1.69k |
|---|---|
testGetStorageAttributeIntegerValueByNameWithDefault_3 ( ) { java . lang . String attributeName = "test" ; java . lang . String attributeValue = "1234" ; org . finra . herd . model . jpa . StorageEntity storageEntity = storageDaoTestHelper . createStorageEntityWithAttributes ( attributeName , attributeValue ) ; java . lang . Integer defaultValue = 2345 ; java . lang . Integer value = storageHelper . getStorageAttributeIntegerValueByName ( attributeName , storageEntity , defaultValue ) ; "<AssertPlaceHolder>" ; } getStorageAttributeIntegerValueByName ( java . lang . String , org . finra . herd . model . jpa . StorageEntity , java . lang . Integer ) { java . lang . Integer value = getStorageAttributeIntegerValueByName ( attributeName , storageEntity , false , false ) ; if ( value == null ) { value = defaultValue ; } return value ; } | org . junit . Assert . assertEquals ( new java . lang . Integer ( attributeValue ) , value ) |
getNumberOfProcessDeploymentInfosWithAssignedOrPendingHumanTasksSupervisedBy ( ) { final org . bonitasoft . engine . persistence . QueryOptions options = new org . bonitasoft . engine . persistence . QueryOptions ( 0 , 10 ) ; final long userId = 9 ; when ( persistenceService . getNumberOfEntities ( org . bonitasoft . engine . core . process . definition . model . SProcessDefinitionDeployInfo . class , "WithAssignedOrPendingHumanTasksSupervisedBy" , options , java . util . Collections . singletonMap ( "userId" , ( ( java . lang . Object ) ( userId ) ) ) ) ) . thenReturn ( 1L ) ; final long result = processDefinitionServiceImpl . getNumberOfProcessDeploymentInfosWithAssignedOrPendingHumanTasksSupervisedBy ( userId , options ) ; "<AssertPlaceHolder>" ; } getNumberOfProcessDeploymentInfosWithAssignedOrPendingHumanTasksSupervisedBy ( long , org . bonitasoft . engine . persistence . QueryOptions ) { final java . util . Map < java . lang . String , java . lang . Object > parameters = java . util . Collections . singletonMap ( org . bonitasoft . engine . core . process . definition . USER_ID , ( ( java . lang . Object ) ( userId ) ) ) ; return persistenceService . getNumberOfEntities ( org . bonitasoft . engine . core . process . definition . model . SProcessDefinitionDeployInfo . class , "WithAssignedOrPendingHumanTasksSupervisedBy" , queryOptions , parameters ) ; } | org . junit . Assert . assertEquals ( 1L , result ) |
checkRadioButtonGroup ( ) { com . vaadin . testbench . elements . RadioButtonGroupElement elem = $ ( com . vaadin . testbench . elements . RadioButtonGroupElement . class ) . get ( 0 ) ; java . lang . String expected = ComponentElementGetValue . TEST_STRING_VALUE ; java . lang . String actual = elem . getValue ( ) ; "<AssertPlaceHolder>" ; } getValue ( ) { java . util . List < org . openqa . selenium . WebElement > selectedElements = findElements ( org . openqa . selenium . By . className ( "v-tree-node-selected" ) ) ; if ( selectedElements . isEmpty ( ) ) { return "" ; } else { return selectedElements . get ( 0 ) . getText ( ) ; } } | org . junit . Assert . assertEquals ( expected , actual ) |
testDoGenerateEdgesValidity ( ) { int meanDegree = 4 ; int numberOfNodes = 10 ; double betaCoefficient = 0.5 ; com . graphaware . module . algo . generator . relationship . WattsStrogatzRelationshipGenerator generator = new com . graphaware . module . algo . generator . relationship . WattsStrogatzRelationshipGenerator ( new com . graphaware . module . algo . generator . config . WattsStrogatzConfig ( numberOfNodes , meanDegree , betaCoefficient ) ) ; "<AssertPlaceHolder>" ; } doGenerateEdges ( ) { java . util . List < com . graphaware . common . util . SameTypePair < java . lang . Integer > > edges = new java . util . ArrayList ( ) ; com . graphaware . module . algo . generator . distribution . MutableDegreeDistribution distribution = new com . graphaware . module . algo . generator . distribution . MutableSimpleDegreeDistribution ( getConfiguration ( ) . getDegrees ( ) ) ; while ( ! ( distribution . isZeroList ( ) ) ) { int index = 0 ; int min = Integer . MAX_VALUE ; for ( int i = 0 ; i < ( distribution . size ( ) ) ; ++ i ) { int elem = distribution . get ( i ) ; if ( ( elem != 0 ) && ( elem < min ) ) { min = elem ; index = i ; } } com . graphaware . module . algo . generator . utils . WeightedReservoirSampler sampler = new com . graphaware . module . algo . generator . utils . WeightedReservoirSampler ( ) ; while ( true ) { com . graphaware . module . algo . generator . distribution . MutableDegreeDistribution temp = new com . graphaware . module . algo . generator . distribution . MutableSimpleDegreeDistribution ( distribution . getDegrees ( ) ) ; int candidateIndex = sampler . randomIndexChoice ( temp . getDegrees ( ) , index ) ; com . graphaware . common . util . SameTypePair < java . lang . Integer > edgeCandidate = new com . graphaware . common . util . UnorderedPair ( candidateIndex , index ) ; boolean skip = false ; for ( com . graphaware . common . util . SameTypePair < java . lang . Integer > edge : edges ) { if ( edge . equals ( edgeCandidate ) ) { skip = true ; break ; } } if ( skip ) { continue ; } temp . decrease ( index ) ; temp . decrease ( candidateIndex ) ; if ( temp . isValid ( ) ) { distribution = temp ; edges . add ( edgeCandidate ) ; break ; } } } return edges ; } | org . junit . Assert . assertEquals ( ( ( int ) ( ( meanDegree * numberOfNodes ) * 0.5 ) ) , generator . doGenerateEdges ( ) . size ( ) ) |
shouldCoerceAShortTargetWithASingleFlavor ( ) { com . facebook . buck . core . model . BuildTarget seen = new com . facebook . buck . rules . coercer . BuildTargetTypeCoercer ( unconfiguredBuildTargetTypeCoercer ) . coerce ( createCellRoots ( filesystem ) , filesystem , basePath , EmptyTargetConfiguration . INSTANCE , ":bar#baz" ) ; com . facebook . buck . core . model . BuildTarget expected = com . facebook . buck . core . model . BuildTargetFactory . newInstance ( "//java/com/facebook/buck/example:bar#baz" ) ; "<AssertPlaceHolder>" ; } newInstance ( java . lang . String ) { return com . facebook . buck . core . model . BuildTargetFactory . newInstance ( ( ( java . nio . file . Path ) ( null ) ) , fullyQualifiedName ) ; } | org . junit . Assert . assertEquals ( expected , seen ) |
testCloseCaseInstance ( ) { final org . jbpm . workbench . cm . model . CaseInstanceSummary cis = setupCaseInstance ( ) ; presenter . closeCaseInstance ( ) ; verify ( caseManagementService ) . closeCaseInstance ( cis . getContainerId ( ) , cis . getCaseId ( ) , null ) ; final org . mockito . ArgumentCaptor < org . jbpm . workbench . cm . client . events . CaseClosedEvent > captor = org . mockito . ArgumentCaptor . forClass ( org . jbpm . workbench . cm . client . events . CaseClosedEvent . class ) ; verify ( caseDestroyEvent ) . fire ( captor . capture ( ) ) ; "<AssertPlaceHolder>" ; } getCaseId ( ) { return caseId ; } | org . junit . Assert . assertEquals ( cis . getCaseId ( ) , captor . getValue ( ) . getCaseId ( ) ) |
testAdd ( ) { com . corefire . model . AlipayAnnal pay = new com . corefire . model . AlipayAnnal ( ) ; pay . setAppid ( "alipay-appid" ) ; pay . setValue ( new java . math . BigDecimal ( "999.99" ) ) ; pay . setCharge ( new java . math . BigDecimal ( "0.99" ) ) ; java . text . SimpleDateFormat sdf = new java . text . SimpleDateFormat ( "yyyy-MM-dd" ) ; try { pay . setAdd_time ( sdf . parse ( "2016-12-12" ) ) ; pay . setModify_time ( sdf . parse ( "2016-12-31" ) ) ; } catch ( java . text . ParseException e ) { e . printStackTrace ( ) ; } com . corefire . dao . AlipayAnnalDao payDao = new com . corefire . dao . AlipayAnnalDao ( ) ; int res = payDao . add ( pay ) ; System . out . println ( ( "AlipayDao.add" + ( payDao . add ( pay ) ) ) ) ; "<AssertPlaceHolder>" ; } add ( model . Goodess ) { dao . GoodessDao dao = new dao . GoodessDao ( ) ; dao . addGoddess ( goodess ) ; } | org . junit . Assert . assertEquals ( 1 , res ) |
testValueFromPro ( ) { "<AssertPlaceHolder>" ; System . out . println ( app . ioc ( ) . getBean ( "com.blade.model.ValueBean" ) ) ; } ioc ( ) { return this . ioc ; } | org . junit . Assert . assertNotNull ( app . ioc ( ) . getBean ( "com.blade.model.ValueBean" ) ) |
getEmptySample ( ) { runtime . initialize ( null , org . talend . components . pubsub . runtime . PubSubTestConstants . addSubscriptionForDataset ( org . talend . components . pubsub . runtime . PubSubTestConstants . createDatasetFromAvro ( org . talend . components . pubsub . runtime . PubSubTestConstants . createDatastore ( ) , org . talend . components . pubsub . runtime . PubSubDatasetRuntimeTestIT . topics . get ( 3 ) , Person . schema . toString ( ) ) , org . talend . components . pubsub . runtime . PubSubDatasetRuntimeTestIT . subForTP4 ) ) ; final java . util . List < java . lang . String > actual = new java . util . ArrayList ( ) ; runtime . getSample ( 10 , new org . talend . daikon . java8 . Consumer < org . apache . avro . generic . IndexedRecord > ( ) { @ org . talend . components . pubsub . runtime . Override public void accept ( org . apache . avro . generic . IndexedRecord indexedRecord ) { actual . add ( indexedRecord . toString ( ) ) ; } } ) ; "<AssertPlaceHolder>" ; } size ( ) { if ( ( tokenName . getValue ( ) ) == null ) { return 0 ; } return tokenName . getValue ( ) . size ( ) ; } | org . junit . Assert . assertThat ( actual . size ( ) , org . hamcrest . core . Is . is ( 0 ) ) |
testSubclassRemovesObserverAnnotationByOverriding ( ) { java . util . Collection < java . lang . Class < ? > > beanClasses = new java . util . ArrayList < java . lang . Class < ? > > ( ) ; beanClasses . add ( org . apache . webbeans . test . events . observer . Superclass . class ) ; beanClasses . add ( org . apache . webbeans . test . events . observer . BeanB . class ) ; startContainer ( beanClasses , null ) ; org . apache . webbeans . test . events . observer . TestEvent testEvent = new org . apache . webbeans . test . events . observer . TestEvent ( ) ; getBeanManager ( ) . fireEvent ( testEvent ) ; "<AssertPlaceHolder>" ; shutDownContainer ( ) ; } getCalledObservers ( ) { return calledObserverNames ; } | org . junit . Assert . assertEquals ( 0 , testEvent . getCalledObservers ( ) . size ( ) ) |
testNoZk ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . conf . Configuration ( ) ; org . apache . sentry . provider . db . service . persistent . LeaderStatusMonitor monitor = new org . apache . sentry . provider . db . service . persistent . LeaderStatusMonitor ( conf ) ; "<AssertPlaceHolder>" ; } isLeader ( ) { if ( isSingleNodeMode ) { return true ; } lock . lock ( ) ; @ org . apache . sentry . provider . db . service . persistent . SuppressWarnings ( "FieldAccessNotGuarded" ) boolean leader = isLeader ; lock . unlock ( ) ; return leader ; } | org . junit . Assert . assertTrue ( monitor . isLeader ( ) ) |
shouldNotThrowExceptionWhenFetchRangeIsExpired ( ) { windowStore = createInMemoryWindowStore ( context , false ) ; windowStore . put ( 1 , "one" , 0L ) ; windowStore . put ( 1 , "two" , retentionPeriod ) ; final org . apache . kafka . streams . state . WindowStoreIterator < java . lang . String > iterator = windowStore . fetch ( 1 , 0L , 10L ) ; "<AssertPlaceHolder>" ; } hasNext ( ) { return false ; } | org . junit . Assert . assertFalse ( iterator . hasNext ( ) ) |
testValidate ( ) { org . meridor . perspective . shell . common . validator . PatternValidator patternValidator = new org . meridor . perspective . shell . common . validator . PatternValidator ( ) ; java . lang . Object instance = new java . lang . Object ( ) ; org . meridor . perspective . shell . common . validator . annotation . Pattern annotation = new org . meridor . perspective . shell . common . validator . annotation . Pattern ( ) { @ org . meridor . perspective . shell . common . validator . Override public java . lang . Class < ? extends java . lang . annotation . Annotation > annotationType ( ) { return org . meridor . perspective . shell . common . validator . annotation . Pattern . class ; } } ; "<AssertPlaceHolder>" ; } validate ( java . lang . Object , java . lang . annotation . Annotation , java . lang . Object ) { java . util . List < java . lang . String > values = getValues ( annotation ) ; return ( value == null ) || ( values . contains ( value . toString ( ) ) ) ; } | org . junit . Assert . assertThat ( patternValidator . validate ( instance , annotation , pattern ) , org . hamcrest . Matchers . is ( result ) ) |
testGlyph ( ) { "<AssertPlaceHolder>" ; tested . getGlyph ( canvasHandler , org . kie . workbench . common . stunner . core . client . components . toolbox . actions . MorphNodeToolboxActionTest . E_UUID ) ; verify ( shapeFactory ) . getGlyph ( org . kie . workbench . common . stunner . core . client . components . toolbox . actions . MorphNodeToolboxActionTest . MORPH_TARGET_ID , AbstractToolboxAction . ToolboxGlyphConsumer . class ) ; } getGlyphId ( org . kie . workbench . common . stunner . core . client . canvas . AbstractCanvasHandler , java . lang . String ) { return targetDefinitionId ; } | org . junit . Assert . assertEquals ( org . kie . workbench . common . stunner . core . client . components . toolbox . actions . MorphNodeToolboxActionTest . MORPH_TARGET_ID , tested . getGlyphId ( canvasHandler , org . kie . workbench . common . stunner . core . client . components . toolbox . actions . MorphNodeToolboxActionTest . E_UUID ) ) |
testGetPostsWithGreatComments ( ) { java . util . List < org . blog . core . domain . BlogPost > found = postService . getPostsWithGreatComments ( getServiceContext ( ) ) ; "<AssertPlaceHolder>" ; } getServiceContext ( ) { return org . sculptor . framework . test . AppEngineTestHelper . getServiceContext ( ) ; } | org . junit . Assert . assertEquals ( 1 , found . size ( ) ) |
testDispose ( ) { org . eclipse . swt . graphics . Cursor cursor = new org . eclipse . swt . graphics . Cursor ( device , org . eclipse . swt . SWT . CURSOR_ARROW ) ; cursor . dispose ( ) ; "<AssertPlaceHolder>" ; } isDisposed ( ) { org . eclipse . swt . widgets . Control c = getMessageControl ( ) ; return ( c != null ) && ( c . isDisposed ( ) ) ; } | org . junit . Assert . assertTrue ( cursor . isDisposed ( ) ) |
testGetSourceNode ( ) { com . fasterxml . jackson . databind . JsonNode expResult = com . fasterxml . jackson . databind . node . NullNode . getInstance ( ) ; com . redhat . lightblue . util . JsonObject instance = new com . redhat . lightblue . util . JsonObjectTest . JsonObjectImpl ( expResult ) ; com . fasterxml . jackson . databind . JsonNode result = instance . getSourceNode ( ) ; "<AssertPlaceHolder>" ; } getSourceNode ( ) { return sourceNode ; } | org . junit . Assert . assertEquals ( expResult , result ) |
testWithBorderBlocking ( ) { com . graphhopper . routing . AbstractRoutingAlgorithmTester . initBiGraph ( ghStorage ) ; com . graphhopper . routing . lm . LandmarkStorage storage = new com . graphhopper . routing . lm . LandmarkStorage ( ghStorage , new com . graphhopper . routing . lm . RAMDirectory ( ) , new com . graphhopper . routing . weighting . FastestWeighting ( encoder ) , 2 ) ; final com . graphhopper . routing . util . spatialrules . SpatialRule ruleRight = new com . graphhopper . routing . util . spatialrules . DefaultSpatialRule ( ) { @ com . graphhopper . routing . lm . Override public java . lang . String getId ( ) { return "right" ; } } ; final com . graphhopper . routing . util . spatialrules . SpatialRule ruleLeft = new com . graphhopper . routing . util . spatialrules . DefaultSpatialRule ( ) { @ com . graphhopper . routing . lm . Override public java . lang . String getId ( ) { return "left" ; } } ; final com . graphhopper . routing . util . spatialrules . SpatialRuleLookup lookup = new com . graphhopper . routing . util . spatialrules . SpatialRuleLookup ( ) { @ com . graphhopper . routing . lm . Override public com . graphhopper . routing . util . spatialrules . SpatialRule lookupRule ( double lat , double lon ) { if ( lon > 0.00105 ) return ruleRight ; return ruleLeft ; } @ com . graphhopper . routing . lm . Override public com . graphhopper . routing . util . spatialrules . SpatialRule lookupRule ( com . graphhopper . util . shapes . GHPoint point ) { return lookupRule ( point . lat , point . lon ) ; } @ com . graphhopper . routing . lm . Override public int getSpatialId ( com . graphhopper . routing . util . spatialrules . SpatialRule rule ) { throw new java . lang . IllegalStateException ( ) ; } @ com . graphhopper . routing . lm . Override public com . graphhopper . routing . util . spatialrules . SpatialRule getSpatialRule ( int spatialId ) { throw new java . lang . IllegalStateException ( ) ; } @ com . graphhopper . routing . lm . Override public int size ( ) { return 2 ; } @ com . graphhopper . routing . lm . Override public com . graphhopper . util . shapes . BBox getBounds ( ) { return new com . graphhopper . util . shapes . BBox ( ( - 180 ) , 180 , ( - 90 ) , 90 ) ; } } ; storage . setSpatialRuleLookup ( lookup ) ; storage . setMinimumNodes ( 2 ) ; storage . createLandmarks ( ) ; "<AssertPlaceHolder>" ; } getSubnetworksWithLandmarks ( ) { return lms . getSubnetworksWithLandmarks ( ) ; } | org . junit . Assert . assertEquals ( 3 , storage . getSubnetworksWithLandmarks ( ) ) |
testLicensesProperties ( ) { "<AssertPlaceHolder>" ; } licenseProperties ( ) { if ( ( org . gbif . ipt . action . manage . MetadataAction . licenseProperties ) == null ) { java . util . Properties p = new java . util . Properties ( ) ; try { p . load ( org . gbif . ipt . action . manage . MetadataAction . class . getResourceAsStream ( org . gbif . ipt . action . manage . MetadataAction . LICENSES_PROPFILE_PATH ) ) ; org . gbif . ipt . action . manage . MetadataAction . LOG . debug ( ( "Loaded<sp>licenses<sp>from<sp>" + ( org . gbif . ipt . action . manage . MetadataAction . LICENSES_PROPFILE_PATH ) ) ) ; } catch ( java . io . IOException e ) { throw new org . gbif . ipt . service . InvalidConfigException ( InvalidConfigException . TYPE . INVALID_PROPERTIES_FILE , ( "Failed<sp>to<sp>load<sp>licenses<sp>from<sp>" + ( org . gbif . ipt . action . manage . MetadataAction . LICENSES_PROPFILE_PATH ) ) ) ; } finally { org . gbif . ipt . action . manage . MetadataAction . licenseProperties = p ; } } return org . gbif . ipt . action . manage . MetadataAction . licenseProperties ; } | org . junit . Assert . assertEquals ( 6 , org . gbif . ipt . action . manage . MetadataAction . licenseProperties ( ) . size ( ) ) |
apply_returning_tsv ( ) { final com . groupon . lex . metrics . expression . LiteralGroupExpression expr = new com . groupon . lex . metrics . expression . LiteralGroupExpression ( resolver ) ; final com . groupon . lex . metrics . timeseries . TimeSeriesValueSet expect = new com . groupon . lex . metrics . timeseries . TimeSeriesValueSet ( java . util . stream . Stream . of ( tsv ) ) ; when ( resolver . apply ( ctx ) ) . thenReturn ( java . util . Optional . of ( com . groupon . lex . metrics . SimpleGroupPath . valueOf ( "foobar" ) ) ) ; when ( ctx . getTSData ( ) ) . thenReturn ( ts_data ) ; when ( ts_data . getTSValue ( com . groupon . lex . metrics . SimpleGroupPath . valueOf ( "foobar" ) ) ) . thenReturn ( expect ) ; "<AssertPlaceHolder>" ; } getTSDelta ( com . groupon . lex . metrics . timeseries . expression . Context ) { throw new java . lang . UnsupportedOperationException ( "Not<sp>supported<sp>yet." ) ; } | org . junit . Assert . assertEquals ( expect , expr . getTSDelta ( ctx ) ) |
testGetNoDataColor ( ) { "<AssertPlaceHolder>" ; } getNoDataColor ( ) { org . esa . s2tbx . dataio . deimos . dimap . Color color ; try { int red = ( ( int ) ( ( DeimosConstants . MAX_LEVEL ) * ( java . lang . Double . parseDouble ( getAttributeSiblingValue ( DeimosConstants . PATH_SPECIAL_VALUE_TEXT , DeimosConstants . NODATA_VALUE , DeimosConstants . PATH_SPECIAL_VALUE_COLOR_RED_LEVEL , DeimosConstants . STRING_ZERO ) ) ) ) ) ; int green = ( ( int ) ( ( DeimosConstants . MAX_LEVEL ) * ( java . lang . Double . parseDouble ( getAttributeSiblingValue ( DeimosConstants . PATH_SPECIAL_VALUE_TEXT , DeimosConstants . NODATA_VALUE , DeimosConstants . PATH_SPECIAL_VALUE_COLOR_GREEN_LEVEL , DeimosConstants . STRING_ZERO ) ) ) ) ) ; int blue = ( ( int ) ( ( DeimosConstants . MAX_LEVEL ) * ( java . lang . Double . parseDouble ( getAttributeSiblingValue ( DeimosConstants . PATH_SPECIAL_VALUE_TEXT , DeimosConstants . NODATA_VALUE , DeimosConstants . PATH_SPECIAL_VALUE_COLOR_BLUE_LEVEL , DeimosConstants . STRING_ZERO ) ) ) ) ) ; color = new org . esa . s2tbx . dataio . deimos . dimap . Color ( red , green , blue ) ; } catch ( java . lang . NumberFormatException e ) { color = Color . BLACK ; } return color ; } | org . junit . Assert . assertEquals ( Color . BLACK , metadata . getNoDataColor ( ) ) |
testReload ( ) { final com . orientechnologies . orient . core . db . document . ODatabaseDocumentTx databaseDocumentTx = new com . orientechnologies . orient . core . db . document . ODatabaseDocumentTx ( "memory:osimplekeyindexdefinitiontest" ) ; databaseDocumentTx . create ( ) ; final com . orientechnologies . orient . core . record . impl . ODocument storeDocument = simpleKeyIndexDefinition . toStream ( ) ; storeDocument . save ( databaseDocumentTx . getClusterNameById ( databaseDocumentTx . getDefaultClusterId ( ) ) ) ; final com . orientechnologies . orient . core . record . impl . ODocument loadDocument = databaseDocumentTx . load ( storeDocument . getIdentity ( ) ) ; final com . orientechnologies . orient . core . index . OSimpleKeyIndexDefinition loadedKeyIndexDefinition = new com . orientechnologies . orient . core . index . OSimpleKeyIndexDefinition ( ) ; loadedKeyIndexDefinition . fromStream ( loadDocument ) ; databaseDocumentTx . drop ( ) ; "<AssertPlaceHolder>" ; } drop ( ) { graph . drop ( ) ; } | org . junit . Assert . assertEquals ( loadedKeyIndexDefinition , simpleKeyIndexDefinition ) |
shouldNotHaveNextIfNoIteratorsAreAdded ( ) { final java . util . Iterator < java . lang . String > itty = new org . apache . tinkerpop . gremlin . util . iterator . MultiIterator ( ) ; "<AssertPlaceHolder>" ; } hasNext ( ) { if ( ! ( this . queue . isEmpty ( ) ) ) return true ; else if ( ! ( this . inputIterator . hasNext ( ) ) ) { this . mapReduce . workerEnd ( MapReduce . Stage . REDUCE ) ; return false ; } else { this . processNext ( ) ; return this . hasNext ( ) ; } } | org . junit . Assert . assertFalse ( itty . hasNext ( ) ) |
testCombine ( ) { org . apache . accumulo . core . client . summary . SummarizerConfiguration sc = org . apache . accumulo . core . client . summary . SummarizerConfiguration . builder ( org . apache . accumulo . core . client . summary . summarizers . EntryLengthSummarizer . class ) . build ( ) ; org . apache . accumulo . core . client . summary . summarizers . EntryLengthSummarizer entrySum = new org . apache . accumulo . core . client . summary . summarizers . EntryLengthSummarizer ( ) ; org . apache . accumulo . core . client . summary . Summarizer . Collector collector1 = entrySum . collector ( sc ) ; collector1 . accept ( new org . apache . accumulo . core . data . Key ( "row.logHist.0" 2 , "qualifier.logHist.1" 7 , "row.logHist.0" 1 ) , new org . apache . accumulo . core . data . Value ( "qualifier.logHist.3" 7 ) ) ; collector1 . accept ( new org . apache . accumulo . core . data . Key ( "qualifier.logHist.1" 2 , "qualifier.logHist.1" 7 , "row.logHist.0" 1 ) , new org . apache . accumulo . core . data . Value ( "qualifier.logHist.1" 1 ) ) ; collector1 . accept ( new org . apache . accumulo . core . data . Key ( "row.logHist.0" 0 , "qualifier.logHist.1" 7 , "row.logHist.0" 1 ) , new org . apache . accumulo . core . data . Value ( "12" 7 ) ) ; java . util . HashMap < java . lang . String , java . lang . Long > stats1 = new java . util . HashMap ( ) ; collector1 . summarize ( stats1 :: put ) ; org . apache . accumulo . core . client . summary . Summarizer . Collector collector2 = entrySum . collector ( sc ) ; collector2 . accept ( new org . apache . accumulo . core . data . Key ( "12" 3 , "row.logHist.0" 8 , "qualifier.logHist.3" 4 ) , new org . apache . accumulo . core . data . Value ( "2" ) ) ; collector2 . accept ( new org . apache . accumulo . core . data . Key ( "12" , "row.logHist.0" 8 , "row.logHist.0" 9 ) , new org . apache . accumulo . core . data . Value ( "12" ) ) ; collector2 . accept ( new org . apache . accumulo . core . data . Key ( "12" , "row.logHist.0" 8 , "row.logHist.0" 7 ) , new org . apache . accumulo . core . data . Value ( "qualifier.logHist.1" 3 ) ) ; java . util . HashMap < java . lang . String , java . lang . Long > stats2 = new java . util . HashMap ( ) ; collector2 . summarize ( stats2 :: put ) ; org . apache . accumulo . core . client . summary . Summarizer . Combiner combiner = entrySum . combiner ( sc ) ; combiner . merge ( stats1 , stats2 ) ; java . util . HashMap < java . lang . String , java . lang . Long > expected = new java . util . HashMap ( ) ; expected . put ( "key.min" , 5L ) ; expected . put ( "qualifier.logHist.1" 0 , 14L ) ; expected . put ( "qualifier.logHist.3" 3 , 59L ) ; expected . put ( "row.logHist.0" 3 , 1L ) ; expected . put ( "value.min" 0 , 3L ) ; expected . put ( "qualifier.logHist.3" 9 , 2L ) ; expected . put ( "qualifier.logHist.1" 4 , 1L ) ; expected . put ( "qualifier.logHist.3" 2 , 8L ) ; expected . put ( "row.sum" , 21L ) ; expected . put ( "row.logHist.0" , 1L ) ; expected . put ( "12" 1 , 2L ) ; expected . put ( "row.logHist.0" 4 , 2L ) ; expected . put ( "qualifier.logHist.1" 5 , 1L ) ; expected . put ( "12" 9 , 2L ) ; expected . put ( "qualifier.logHist.3" 5 , 3L ) ; expected . put ( "family.sum" , 15L ) ; expected . put ( "qualifier.logHist.3" 0 , 3L ) ; expected . put ( "row.logHist.0" 6 , 3L ) ; expected . put ( "qualifier.min" , 2L ) ; expected . put ( "qualifier.logHist.3" 1 , 9L ) ; expected . put ( "qualifier.logHist.3" 8 , 23L ) ; expected . put ( "qualifier.logHist.1" , 3L ) ; expected . put ( "12" 0 , 2L ) ; expected . put ( "qualifier.logHist.3" , 1L ) ; expected . put ( "value.min" 1 , 0L ) ; expected . put ( "qualifier.logHist.1" 8 , 0L ) ; expected . put ( "qualifier.logHist.1" 9 , 0L ) ; expected . put ( "12" 4 , 6L ) ; expected . put ( "value.min" , 1L ) ; expected . put ( "qualifier.logHist.3" 6 , 7L ) ; expected . put ( "12" 8 , 20L ) ; expected . put ( "12" 2 , 1L ) ; expected . put ( "row.logHist.0" 5 , 2L ) ; expected . put ( "12" 5 , 2L ) ; expected . put ( "qualifier.logHist.1" 6 , 1L ) ; expected . put ( "12" 6 , 6L ) ; "<AssertPlaceHolder>" ; } put ( java . lang . Long , org . apache . accumulo . server . tabletserver . LargestFirstMemoryManager$TabletInfo ) { if ( ( map . size ( ) ) == ( max ) ) { if ( ( key . compareTo ( map . firstKey ( ) ) ) < 0 ) return false ; try { add ( key , value ) ; return true ; } finally { map . remove ( map . firstKey ( ) ) ; } } else { add ( key , value ) ; return true ; } } | org . junit . Assert . assertEquals ( expected , stats1 ) |
testGetActionUrl_relativePath ( ) { java . lang . String contextPath = "/jenkins" ; java . lang . String itUrl = "iturl/" ; org . kohsuke . stapler . StaplerRequest req = hudson . FunctionsTest . createMockRequest ( contextPath ) ; java . lang . String [ ] paths = new java . lang . String [ ] { "foo/bar" , "./foo/bar" , "../foo/bar" } ; mockStatic ( org . kohsuke . stapler . Stapler . class ) ; when ( org . kohsuke . stapler . Stapler . getCurrentRequest ( ) ) . thenReturn ( req ) ; for ( java . lang . String path : paths ) { java . lang . String result = hudson . Functions . getActionUrl ( itUrl , hudson . FunctionsTest . createMockAction ( path ) ) ; "<AssertPlaceHolder>" ; } } createMockAction ( java . lang . String ) { hudson . model . Action action = mock ( hudson . model . Action . class ) ; when ( action . getUrlName ( ) ) . thenReturn ( uri ) ; return action ; } | org . junit . Assert . assertEquals ( ( ( ( contextPath + "/" ) + itUrl ) + path ) , result ) |
getDelayTillHourBeforeHourTest ( ) { final int minute = 10 ; final int currentHour = 16 ; final int toHour = 20 ; long now = org . threadly . concurrent . SchedulingUtilsTest . getNowToDay ( ) ; now += TimeUnit . HOURS . toMillis ( currentHour ) ; now += TimeUnit . MINUTES . toMillis ( minute ) ; long result = org . threadly . concurrent . SchedulingUtils . getDelayTillHour ( now , toHour , minute ) ; "<AssertPlaceHolder>" ; } getDelayTillHour ( long , int , int ) { long delayInMillis = TimeUnit . MINUTES . toMillis ( minute ) ; long currentHour = TimeUnit . MILLISECONDS . toHours ( ( now % ( TimeUnit . DAYS . toMillis ( 1 ) ) ) ) ; if ( hour > currentHour ) { delayInMillis += TimeUnit . HOURS . toMillis ( ( hour - currentHour ) ) ; } else if ( hour < currentHour ) { delayInMillis += TimeUnit . HOURS . toMillis ( ( ( ( TimeUnit . DAYS . toHours ( 1 ) ) - currentHour ) + hour ) ) ; } else { long result = org . threadly . concurrent . SchedulingUtils . getDelayTillMinute ( org . threadly . util . Clock . lastKnownTimeMillis ( ) , minute ) ; if ( ( TimeUnit . MILLISECONDS . toMinutes ( result ) ) <= minute ) { return result ; } else { return result + ( TimeUnit . HOURS . toMillis ( ( ( TimeUnit . DAYS . toHours ( 1 ) ) - 1 ) ) ) ; } } long offset = now % ( TimeUnit . HOURS . toMillis ( 1 ) ) ; return delayInMillis - offset ; } | org . junit . Assert . assertEquals ( TimeUnit . HOURS . toMillis ( ( toHour - currentHour ) ) , result ) |
shouldReturnTrueOnEvaluatingSpringExpressionGivenValidSpringExpression ( ) { org . flips . model . FeatureContext featureContext = mock ( org . flips . model . FeatureContext . class ) ; org . flips . model . FlipAnnotationAttributes flipAnnotationAttributes = mock ( org . flips . model . FlipAnnotationAttributes . class ) ; java . lang . String inputExpression = "@bean.property.equals('test')" ; org . springframework . expression . ExpressionParser expressionParser = mock ( org . springframework . expression . ExpressionParser . class ) ; org . springframework . expression . spel . support . StandardEvaluationContext context = mock ( org . springframework . expression . spel . support . StandardEvaluationContext . class ) ; org . springframework . expression . Expression expression = mock ( org . springframework . expression . Expression . class ) ; when ( flipAnnotationAttributes . getAttributeValue ( "expression" , "" ) ) . thenReturn ( inputExpression ) ; when ( featureContext . getExpressionParser ( ) ) . thenReturn ( expressionParser ) ; when ( featureContext . getEvaluationContext ( ) ) . thenReturn ( context ) ; when ( expressionParser . parseExpression ( inputExpression ) ) . thenReturn ( expression ) ; when ( expression . getValue ( context ) ) . thenReturn ( true ) ; org . flips . condition . SpringExpressionFlipCondition condition = new org . flips . condition . SpringExpressionFlipCondition ( ) ; boolean result = condition . evaluateCondition ( featureContext , flipAnnotationAttributes ) ; "<AssertPlaceHolder>" ; verify ( flipAnnotationAttributes ) . getAttributeValue ( "expression" , "" ) ; verify ( featureContext ) . getExpressionParser ( ) ; verify ( featureContext ) . getEvaluationContext ( ) ; verify ( expressionParser ) . parseExpression ( inputExpression ) ; verify ( expression ) . getValue ( context ) ; } evaluateCondition ( org . flips . model . FeatureContext , org . flips . model . FlipAnnotationAttributes ) { java . lang . String expression = flipAnnotationAttributes . getAttributeValue ( "expression" , "" ) ; org . flips . utils . ValidationUtils . requireNonEmpty ( expression , "expression<sp>element<sp>can<sp>not<sp>be<sp>NULL<sp>or<sp>EMPTY<sp>when<sp>using<sp>@FlipOnSpringExpression" ) ; return evaluateExpression ( featureContext , expression ) ; } | org . junit . Assert . assertEquals ( true , result ) |
deveObterAmbienteComoFoiSetado ( ) { final com . fincatto . documentofiscal . nfe400 . classes . evento . cancelamento . NFInfoEventoCancelamento infoEventoCancelamento = new com . fincatto . documentofiscal . nfe400 . classes . evento . cancelamento . NFInfoEventoCancelamento ( ) ; final com . fincatto . documentofiscal . DFAmbiente ambiente = com . fincatto . documentofiscal . DFAmbiente . PRODUCAO ; infoEventoCancelamento . setAmbiente ( ambiente ) ; "<AssertPlaceHolder>" ; } getAmbiente ( ) { return this . ambiente ; } | org . junit . Assert . assertEquals ( ambiente , infoEventoCancelamento . getAmbiente ( ) ) |
createDynamicModule ( ) { de . devsurf . injection . guice . scanner . StartupModule startup = de . devsurf . injection . guice . scanner . StartupModule . create ( de . devsurf . injection . guice . scanner . asm . ASMClasspathScanner . class , de . devsurf . injection . guice . scanner . PackageFilter . create ( de . devsurf . injection . guice . test . configuration . failure . FailureConfigTests . class ) ) ; startup . addFeature ( de . devsurf . injection . guice . configuration . features . ConfigurationFeature . class ) ; com . google . inject . Injector injector = com . google . inject . Guice . createInjector ( startup ) ; "<AssertPlaceHolder>" ; } addFeature ( de . devsurf . injection . guice . scanner . features . ScannerFeature ) { _collector . addScannerFeature ( listener ) ; } | org . junit . Assert . assertNotNull ( injector ) |
testResourceSizeStringNoneSet ( ) { ddf . catalog . data . impl . MetacardImpl metacard = new ddf . catalog . data . impl . MetacardImpl ( ) ; java . lang . String result = org . codice . ddf . spatial . kml . transformer . TestDescriptionTemplateHelper . helper . resourceSizeString ( metacard ) ; "<AssertPlaceHolder>" ; } resourceSizeString ( ddf . catalog . data . Metacard ) { java . lang . String resourceSize = context . getResourceSize ( ) ; java . lang . String sizePrefixes = "<sp>KMGTPEZYXWVU" ; if ( ( ( resourceSize == null ) || ( ( resourceSize . trim ( ) . length ( ) ) == 0 ) ) || ( ( resourceSize . toLowerCase ( ) . indexOf ( "n/a" ) ) >= 0 ) ) { return null ; } long size = 0 ; try { size = java . lang . Long . parseLong ( resourceSize ) ; } catch ( java . lang . NumberFormatException nfe ) { org . codice . ddf . spatial . kml . transformer . DescriptionTemplateHelper . LOGGER . debug ( "Failed<sp>to<sp>parse<sp>resourceSize<sp>({}),<sp>assuming<sp>already<sp>formatted." , org . apache . commons . lang . StringUtils . trim ( resourceSize ) ) ; return resourceSize ; } if ( size <= 0 ) { return "0" ; } int t2 = ( ( int ) ( java . lang . Math . min ( java . lang . Math . floor ( ( ( java . lang . Math . log ( size ) ) / ( java . lang . Math . log ( 1024 ) ) ) ) , 12 ) ) ) ; char c = sizePrefixes . charAt ( t2 ) ; return ( ( ( ( java . lang . Math . round ( ( ( size * 100 ) / ( java . lang . Math . pow ( 1024 , t2 ) ) ) ) ) / 100 ) + "<sp>" ) + ( c == '<sp>' ? "" : c ) ) + "B" ; } | org . junit . Assert . assertNull ( result ) |
getArchivedFlowNodeInstance ( ) { logoutOnTenant ( ) ; loginOnDefaultTenantWith ( org . bonitasoft . engine . test . USERNAME , "bpm" ) ; final org . bonitasoft . engine . bpm . process . impl . ProcessDefinitionBuilder processDefinitionBuilder = new org . bonitasoft . engine . bpm . process . impl . ProcessDefinitionBuilder ( ) . createNewInstance ( "ProcessToDelete" , "1.0" ) ; processDefinitionBuilder . addActor ( "actor" ) ; processDefinitionBuilder . addUserTask ( "step1" , "actor" ) . addDescription ( "My<sp>Description" ) . addDisplayName ( new org . bonitasoft . engine . expression . ExpressionBuilder ( ) . createConstantStringExpression ( "My<sp>Display<sp>Name" ) ) . addDisplayDescriptionAfterCompletion ( new org . bonitasoft . engine . expression . ExpressionBuilder ( ) . createConstantStringExpression ( "My<sp>Display<sp>Description" ) ) ; final org . bonitasoft . engine . bpm . process . DesignProcessDefinition designProcessDefinition = processDefinitionBuilder . getProcess ( ) ; final org . bonitasoft . engine . bpm . process . ProcessDefinition processDefinition = deployAndEnableProcessWithActor ( designProcessDefinition , "actor" , john ) ; final org . bonitasoft . engine . bpm . process . ProcessInstance p1 = getProcessAPI ( ) . startProcess ( processDefinition . getId ( ) ) ; final long userTaskId = waitForUserTaskAndExecuteIt ( p1 , "step1" , john ) ; waitForProcessToFinish ( p1 ) ; final org . bonitasoft . engine . bpm . flownode . ArchivedActivityInstance archivedUserTask = getProcessAPI ( ) . getArchivedActivityInstance ( userTaskId ) ; "<AssertPlaceHolder>" ; disableAndDeleteProcess ( processDefinition ) ; } getProcessAPI ( ) { return getApiClient ( ) . getProcessAPI ( ) ; } | org . junit . Assert . assertEquals ( archivedUserTask , getProcessAPI ( ) . getArchivedFlowNodeInstance ( archivedUserTask . getId ( ) ) ) |
run ( ) { final java . io . File tempWar = new java . io . File ( "target/AppDataSourceTest" ) ; tempWar . mkdirs ( ) ; new java . io . File ( tempWar , "WEB-INF" ) . mkdirs ( ) ; org . apache . openejb . loader . IO . writeString ( new java . io . File ( tempWar , ">\n" 0 ) , ( "<resources>\n" + ( ( ( "<Resource<sp>id=\"java:app/gace/MyDS\"<sp>type=\"DataSource\">\n" + "DataSourceCreator=dbcp\n" ) + "</Resource>\n" ) + "</resources>\n" ) ) ) ; final java . util . Collection < java . util . logging . LogRecord > records = new java . util . ArrayList ( ) ; try ( final org . apache . tomee . embedded . Container c = new org . apache . tomee . embedded . Container ( new org . apache . tomee . embedded . Configuration ( ) . randomHttpPort ( ) ) ) { org . apache . commons . logging . impl . Jdk14Logger . class . cast ( org . apache . commons . logging . LogFactory . getLog ( org . apache . commons . dbcp2 . BasicDataSource . class ) ) . getLogger ( ) . addHandler ( new java . util . logging . Handler ( ) { @ org . apache . tomee . embedded . Override public void publish ( final java . util . logging . LogRecord record ) { if ( ( ( record . getLevel ( ) ) == ( java . util . logging . Level . SEVERE ) ) || ( ( record . getLevel ( ) ) == ( java . util . logging . Level . WARNING ) ) ) { records . add ( record ) ; } } @ org . apache . tomee . embedded . Override public void flush ( ) { } @ org . apache . tomee . embedded . Override public void close ( ) throws org . apache . tomee . embedded . SecurityException { } } ) ; c . deploy ( null , tempWar ) ; } "<AssertPlaceHolder>" ; } isEmpty ( ) { return map . isEmpty ( ) ; } | org . junit . Assert . assertTrue ( records . isEmpty ( ) ) |
shouldCloseExhaustedCursors ( ) { org . neo4j . cursor . RawCursor < org . neo4j . index . internal . gbptree . Hit < org . neo4j . kernel . impl . index . labelscan . LabelScanKey , org . neo4j . kernel . impl . index . labelscan . LabelScanValue > , java . io . IOException > cursor = mock ( org . neo4j . cursor . RawCursor . class ) ; when ( cursor . next ( ) ) . thenReturn ( false ) ; java . util . Collection < org . neo4j . cursor . RawCursor < org . neo4j . index . internal . gbptree . Hit < org . neo4j . kernel . impl . index . labelscan . LabelScanKey , org . neo4j . kernel . impl . index . labelscan . LabelScanValue > , java . io . IOException > > toRemoveFrom = new java . util . HashSet ( ) ; org . neo4j . kernel . impl . index . labelscan . LabelScanValueIndexProgressor iterator = new org . neo4j . kernel . impl . index . labelscan . LabelScanValueIndexProgressor ( cursor , toRemoveFrom , mock ( IndexProgressor . NodeLabelClient . class ) ) ; verify ( cursor , never ( ) ) . close ( ) ; exhaust ( iterator ) ; verify ( cursor , times ( 1 ) ) . close ( ) ; iterator . next ( ) ; verify ( cursor , times ( 1 ) ) . close ( ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { try ( org . neo4j . cursor . RawCursor < org . neo4j . index . internal . gbptree . Hit < org . neo4j . kernel . impl . index . labelscan . LabelScanKey , org . neo4j . kernel . impl . index . labelscan . LabelScanValue > , java . io . IOException > cursor = index . seek ( new org . neo4j . kernel . impl . index . labelscan . LabelScanKey ( 0 , 0 ) , new org . neo4j . kernel . impl . index . labelscan . LabelScanKey ( Integer . MAX_VALUE , Long . MAX_VALUE ) ) ) { return ! ( cursor . next ( ) ) ; } } | org . junit . Assert . assertTrue ( toRemoveFrom . isEmpty ( ) ) |
detectInvalidSchemaId ( ) { java . io . ByteArrayOutputStream out = new java . io . ByteArrayOutputStream ( ) ; out . write ( AvroSchemaHelper . MAGIC_BYTE ) ; com . streamsets . pipeline . lib . util . AvroSchemaHelper helper = new com . streamsets . pipeline . lib . util . AvroSchemaHelper ( getSettings ( null , OriginAvroSchemaSource . SOURCE , false ) ) ; java . util . Optional < java . lang . Integer > schemaId = helper . detectSchemaId ( out . toByteArray ( ) ) ; "<AssertPlaceHolder>" ; } empty ( ) { return new com . streamsets . datacollector . event . handler . remote . RemoteDataCollectorResult ( null , null , false , null ) ; } | org . junit . Assert . assertEquals ( java . util . Optional . empty ( ) , schemaId ) |
testSetTaxReturnsOrder ( ) { com . eclipsesource . tabris . tracking . Order order = new com . eclipsesource . tabris . tracking . Order ( "foo" , java . math . BigDecimal . ONE ) ; com . eclipsesource . tabris . tracking . Order actualOrder = order . setTax ( java . math . BigDecimal . valueOf ( 3 ) ) ; "<AssertPlaceHolder>" ; } setTax ( double ) { addParameter ( getRequestKey ( com . eclipsesource . tabris . tracking . internal . analytics . model . hit . TRANSACTION_TAX ) , java . lang . Double . valueOf ( tax ) ) ; return this ; } | org . junit . Assert . assertSame ( order , actualOrder ) |
getAllUsersClient ( ) { org . springframework . web . reactive . function . client . WebClient webClient = org . springframework . web . reactive . function . client . WebClient . create ( ( "http://localhost:" + ( port ) ) ) ; java . util . List < com . apress . webfluxdemo . User > users = webClient . get ( ) . uri ( "/api/users/" ) . accept ( com . apress . webfluxdemo . APPLICATION_JSON ) . exchange ( ) . flatMap ( ( response ) -> response . bodyToFlux ( . class ) . collectList ( ) ) . block ( java . time . Duration . ofSeconds ( 100 ) ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertNotNull ( users ) |
verifyGetMachinesNameDoesNotReturnExistingMachines ( ) { oracle . kubernetes . operator . wlsconfig . WlsMachineConfig machine1 = new oracle . kubernetes . operator . wlsconfig . WlsMachineConfig ( "domain1-machine1" , 5556 , "localhost" , "SSL" ) ; oracle . kubernetes . operator . wlsconfig . WlsMachineConfig machine2 = new oracle . kubernetes . operator . wlsconfig . WlsMachineConfig ( "domain1-machine2" , 5556 , "localhost" , "SSL" ) ; oracle . kubernetes . operator . wlsconfig . WlsMachineConfig machine3 = new oracle . kubernetes . operator . wlsconfig . WlsMachineConfig ( "domain1-machine3" , 5556 , "localhost" , "SSL" ) ; java . util . Map < java . lang . String , oracle . kubernetes . operator . wlsconfig . WlsMachineConfig > machines = new java . util . HashMap ( ) ; machines . put ( machine1 . getName ( ) , machine1 ) ; machines . put ( machine2 . getName ( ) , machine2 ) ; machines . put ( machine3 . getName ( ) , machine3 ) ; oracle . kubernetes . operator . wlsconfig . WlsClusterConfig wlsClusterConfig = new oracle . kubernetes . operator . wlsconfig . WlsClusterConfig ( "cluster1" , oracle . kubernetes . operator . wlsconfig . WlsClusterConfigTest . createDynamicServersConfig ( 1 , 5 , "ms-" , "cluster1" ) ) ; java . util . Map < java . lang . String , oracle . kubernetes . operator . wlsconfig . WlsClusterConfig > clusters = new java . util . HashMap ( ) ; clusters . put ( wlsClusterConfig . getClusterName ( ) , wlsClusterConfig ) ; oracle . kubernetes . operator . wlsconfig . WlsDomainConfig wlsDomainConfig = new oracle . kubernetes . operator . wlsconfig . WlsDomainConfig ( "base_domain" , oracle . kubernetes . operator . wlsconfig . WlsClusterConfigTest . ADMIN_SERVER , clusters , null , null , machines ) ; wlsClusterConfig . setWlsDomainConfig ( wlsDomainConfig ) ; java . lang . String [ ] names = wlsClusterConfig . getMachineNamesForDynamicServers ( "domain1-machine" , 3 ) ; "<AssertPlaceHolder>" ; } getMachineNamesForDynamicServers ( java . lang . String , int ) { if ( ( ( targetClusterSize < 1 ) || ( ! ( hasDynamicServers ( ) ) ) ) || ( ( wlsDomainConfig ) == null ) ) { return new java . lang . String [ 0 ] ; } java . util . ArrayList < java . lang . String > names = new java . util . ArrayList ( ) ; for ( int suffix = 1 ; suffix <= targetClusterSize ; suffix ++ ) { java . lang . String newMachineName = ( machineNamePrefix == null ) ? "" + suffix : machineNamePrefix + suffix ; if ( ( wlsDomainConfig . getMachineConfig ( newMachineName ) ) == null ) { names . add ( newMachineName ) ; } } java . lang . String [ ] machineNameArray = new java . lang . String [ names . size ( ) ] ; names . toArray ( machineNameArray ) ; return machineNameArray ; } | org . junit . Assert . assertEquals ( 0 , names . length ) |
should_get_empty_string ( ) { com . moandjiezana . toml . Toml toml = new com . moandjiezana . toml . Toml ( ) . read ( "a<sp>=<sp>\"\"" ) ; "<AssertPlaceHolder>" ; } getString ( java . lang . String ) { return ( ( java . lang . String ) ( get ( key ) ) ) ; } | org . junit . Assert . assertEquals ( "" , toml . getString ( "a" ) ) |
testPastDayOK ( ) { org . ff4j . core . Feature f = ff4j . getFeature ( "past1" ) ; org . ff4j . strategy . time . ReleaseDateFlipStrategy rds = ( ( org . ff4j . strategy . time . ReleaseDateFlipStrategy ) ( f . getFlippingStrategy ( ) ) ) ; "<AssertPlaceHolder>" ; } evaluate ( java . lang . String , org . ff4j . core . FeatureStore , org . ff4j . core . FlippingExecutionContext ) { if ( ( null == executionContext ) || ( ! ( executionContext . containsKey ( org . ff4j . strategy . ClientFilterStrategy . CLIENT_HOSTNAME ) ) ) ) { throw new java . lang . IllegalArgumentException ( ( ( ( ( "To<sp>work<sp>with<sp>" + ( getClass ( ) . getName ( ) ) ) + "<sp>you<sp>must<sp>provide<sp>'" ) + ( org . ff4j . strategy . ClientFilterStrategy . CLIENT_HOSTNAME ) ) + "'<sp>parameter<sp>in<sp>execution<sp>context" ) ) ; } return setOfGrantedClient . contains ( executionContext . getString ( org . ff4j . strategy . ClientFilterStrategy . CLIENT_HOSTNAME ) ) ; } | org . junit . Assert . assertTrue ( rds . evaluate ( "past1" , null , null ) ) |
g_RowMatrix ( ) { mikera . matrixx . impl . RowMatrix rm = new mikera . matrixx . impl . RowMatrix ( mikera . vectorz . Vector . of ( 1 , 2 , 3 , 4 , 7 ) ) ; "<AssertPlaceHolder>" ; doGenericTests ( rm ) ; doGenericTests ( new mikera . matrixx . impl . RowMatrix ( mikera . vectorz . Vector3 . of ( 1 , 2 , 3 ) ) ) ; } elementCount ( ) { return data . length ; } | org . junit . Assert . assertEquals ( 5 , rm . elementCount ( ) ) |
testRandomByteArrayTransfer1 ( ) { byte [ ] value = new byte [ io . netty . buffer . AbstractByteBufTest . BLOCK_SIZE ] ; for ( int i = 0 ; i < ( ( ( buffer . capacity ( ) ) - ( io . netty . buffer . AbstractByteBufTest . BLOCK_SIZE ) ) + 1 ) ; i += io . netty . buffer . AbstractByteBufTest . BLOCK_SIZE ) { random . nextBytes ( value ) ; buffer . setBytes ( i , value ) ; } random . setSeed ( seed ) ; byte [ ] expectedValueContent = new byte [ io . netty . buffer . AbstractByteBufTest . BLOCK_SIZE ] ; io . netty . buffer . ByteBuf expectedValue = io . netty . buffer . Unpooled . wrappedBuffer ( expectedValueContent ) ; for ( int i = 0 ; i < ( ( ( buffer . capacity ( ) ) - ( io . netty . buffer . AbstractByteBufTest . BLOCK_SIZE ) ) + 1 ) ; i += io . netty . buffer . AbstractByteBufTest . BLOCK_SIZE ) { random . nextBytes ( expectedValueContent ) ; buffer . getBytes ( i , value ) ; for ( int j = 0 ; j < ( io . netty . buffer . AbstractByteBufTest . BLOCK_SIZE ) ; j ++ ) { "<AssertPlaceHolder>" ; } } } getByte ( int ) { return buf . getByte ( index ) ; } | org . junit . Assert . assertEquals ( expectedValue . getByte ( j ) , value [ j ] ) |
testGetNormalizedPolarizability ( ) { org . openscience . cdk . tools . AtomicProperties props = org . openscience . cdk . tools . AtomicProperties . getInstance ( ) ; double polar = props . getNormalizedPolarizability ( "C" ) ; "<AssertPlaceHolder>" ; } getNormalizedPolarizability ( java . lang . String ) { return ( this . getPolarizability ( symbol ) ) / ( this . getPolarizability ( "C" ) ) ; } | org . junit . Assert . assertTrue ( ( polar > 0 ) ) |
testClientID5 ( ) { org . skyscreamer . nevado . jms . facilities . Connection conn = createConnection ( getConnectionFactory ( ) ) ; "<AssertPlaceHolder>" ; conn . createSession ( false , Session . AUTO_ACKNOWLEDGE ) ; conn . setClientID ( org . skyscreamer . nevado . jms . facilities . ClientIDTest . TEST_CLIENT_ID ) ; } getClientID ( ) { return _clientID ; } | org . junit . Assert . assertNull ( conn . getClientID ( ) ) |
lotsOfInsertsTest ( ) { io . teknek . nibiru . Keyspace ks1 = io . teknek . nibiru . engine . AbstractMemtableTest . keyspaceWithNaturalPartitioner ( testFolder ) ; ks1 . createStore ( "abc" , new io . teknek . nibiru . transport . Response ( ) . withProperty ( StoreMetaData . IMPLEMENTING_CLASS , io . teknek . nibiru . engine . DefaultColumnFamily . class . getName ( ) ) ) ; io . teknek . nibiru . engine . AbstractMemtable m = makeMemtable ( ks1 ) ; { long start = java . lang . System . currentTimeMillis ( ) ; for ( int i = 0 ; i < 1000000 ; i ++ ) { m . put ( ks1 . getKeyspaceMetaData ( ) . getPartitioner ( ) . partition ( java . lang . String . valueOf ( i ) ) , "column2" , "c" , 1 , 0L ) ; } System . out . println ( ( "10000<sp>inserts" + ( ( java . lang . System . currentTimeMillis ( ) ) - start ) ) ) ; } { long start = java . lang . System . currentTimeMillis ( ) ; for ( int i = 0 ; i < 1000000 ; i ++ ) { io . teknek . nibiru . engine . atom . AtomValue o = m . get ( ks1 . getKeyspaceMetaData ( ) . getPartitioner ( ) . partition ( java . lang . String . valueOf ( i ) ) , "column2" ) ; "<AssertPlaceHolder>" ; } System . out . println ( ( "10000<sp>gets" + ( ( java . lang . System . currentTimeMillis ( ) ) - start ) ) ) ; } } partition ( java . lang . String ) { io . teknek . nibiru . Token t = new io . teknek . nibiru . Token ( ) ; t . setRowkey ( in ) ; t . setToken ( MD5 ( in ) ) ; return t ; } | org . junit . Assert . assertNotNull ( o ) |
onCreateShouldNotRegisterKeyListenerWhenUnknownUser ( ) { activityController . withIntent ( null ) ; activityController . create ( ) ; final android . widget . EditText privateChatInput = ( ( android . widget . EditText ) ( controller . findViewById ( R . id . privateChatInput ) ) ) ; privateChatInput . setText ( "Hello" ) ; privateChatInput . dispatchKeyEvent ( new android . view . KeyEvent ( android . view . KeyEvent . ACTION_UP , android . view . KeyEvent . KEYCODE_ENTER ) ) ; verify ( ui , never ( ) ) . sendPrivateMessage ( anyString ( ) , any ( net . usikkert . kouchat . misc . User . class ) ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( ( "Line{" + "lineNumber=" ) + ( lineNumber ) ) + ",<sp>lineText=" ) + ( lineText ) ) + '}' ; } | org . junit . Assert . assertEquals ( "Hello" , privateChatInput . getText ( ) . toString ( ) ) |
returnsTrueIfCpuDemandLow ( ) { com . microsoft . dhalion . conf . PolicyConfig policyConf = new com . microsoft . dhalion . conf . PolicyConfig ( "policy" , java . util . Collections . emptyMap ( ) ) ; com . microsoft . dhalion . detectors . ExcessCpuDetector detector = new com . microsoft . dhalion . detectors . ExcessCpuDetector ( policyConf ) ; "<AssertPlaceHolder>" ; } evaluate ( java . lang . String , double , double ) { if ( demand <= 0 ) { return false ; } if ( free <= 0 ) { return true ; } return ( demand / free ) >= ( thresholdRatio ) ; } | org . junit . Assert . assertTrue ( detector . evaluate ( "" , 10 , 5 ) ) |
testParseWKT ( ) { java . lang . String wkt = dq ( ( "GEOGCS['GCS_WGS_1984'," + ( ( ( "DATUM['WGS_1984',<sp>" + "SPHEROID['WGS_1984',6378137,298.257223563]]," ) + "PRIMEM['Greenwich',0]," ) + "UNIT['Degree',0.017453292519943295]]" ) ) ) ; org . osgeo . proj4j . CoordinateReferenceSystem crs1 = io . jeo . proj . Proj . fromWKT ( wkt ) ; org . osgeo . proj4j . CoordinateReferenceSystem crs2 = io . jeo . proj . Proj . crs ( "epsg:4326" ) ; org . osgeo . proj4j . CoordinateReferenceSystem crs3 = io . jeo . proj . Proj . crs ( "epsg:3157" ) ; com . vividsolutions . jts . geom . Point p1 = new io . jeo . geom . GeomBuilder ( ) . point ( ( - 117 ) , 63.15 ) . toPoint ( ) ; com . vividsolutions . jts . geom . Point p2 = new io . jeo . geom . GeomBuilder ( ) . point ( ( - 117 ) , 63.15 ) . toPoint ( ) ; p1 = io . jeo . proj . Proj . reproject ( p1 , crs1 , crs3 ) ; p2 = io . jeo . proj . Proj . reproject ( p2 , crs2 , crs3 ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ( this ) == obj ) return true ; if ( obj == null ) return false ; if ( ( getClass ( ) ) != ( obj . getClass ( ) ) ) return false ; io . jeo . vector . Field other = ( ( io . jeo . vector . Field ) ( obj ) ) ; if ( ( name ) == null ) { if ( ( other . name ) != null ) return false ; } else if ( ! ( name . equals ( other . name ) ) ) return false ; if ( ( type ) == null ) { if ( ( other . type ) != null ) return false ; } else if ( ! ( type . equals ( other . type ) ) ) return false ; if ( ( crs ) == null ) { if ( ( other . crs ) != null ) return false ; } else if ( ! ( crs . equals ( other . crs ) ) ) return false ; return true ; } | org . junit . Assert . assertTrue ( p1 . equals ( p2 ) ) |
testError ( ) { org . mockito . ArgumentCaptor < org . pentaho . di . core . logging . Log4jFileAppender > captor = org . mockito . ArgumentCaptor . forClass ( org . pentaho . di . core . logging . Log4jFileAppender . class ) ; logWriter = mock ( org . pentaho . di . core . logging . LogWriter . class ) ; doThrow ( new java . lang . RuntimeException ( ) ) . when ( logWriter ) . addAppender ( isA ( org . pentaho . di . core . logging . Log4jFileAppender . class ) ) ; writerAppenderManager = new org . pentaho . big . data . impl . shim . pig . WriterAppenderManager ( logChannelInterface , logLevel , testName , logWriter ) ; verify ( logWriter ) . addAppender ( captor . capture ( ) ) ; "<AssertPlaceHolder>" ; writerAppenderManager . close ( ) ; verify ( logWriter ) . removeAppender ( captor . getValue ( ) ) ; } getFile ( ) { return ( appender ) == null ? null : appender . getFile ( ) ; } | org . junit . Assert . assertNotNull ( writerAppenderManager . getFile ( ) ) |
disposeInvalid ( ) { list . remove ( 1 ) ; "<AssertPlaceHolder>" ; p1 . dispose ( ) ; exception . expect ( jetbrains . jetpad . model . property . IllegalStateException . class ) ; p1 . dispose ( ) ; } isValid ( ) { return ( getIndex ( ) . get ( ) ) != null ; } | org . junit . Assert . assertFalse ( p1 . isValid ( ) ) |
testCapitalizeWords11 ( ) { java . lang . Object s = "" ; java . lang . String expResult = "" ; java . lang . String result = org . thymeleaf . util . StringUtils . capitalizeWords ( s , "<sp>." ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertEquals ( expResult , result ) |
test ( ) { cn . bc . spider . callable . TextCallable c = new cn . bc . spider . callable . TextCallable ( ) ; c . setMethod ( "get" ) ; c . setGroup ( "baidu" ) ; c . setUrl ( "http://www.baidu.com" ) ; c . setSuccessExpression ( "length()<sp>><sp>0" ) ; cn . bc . spider . Result < java . lang . String > result = cn . bc . spider . http . TaskExecutor . get ( c ) ; "<AssertPlaceHolder>" ; if ( cn . bc . spider . callable . TextCallableTest . logger . isDebugEnabled ( ) ) cn . bc . spider . callable . TextCallableTest . logger . debug ( result . getData ( ) ) ; } isSuccess ( ) { return success ; } | org . junit . Assert . assertTrue ( result . isSuccess ( ) ) |
insertLastNotEmpty ( ) { lijst . insertLast ( 2 ) ; "<AssertPlaceHolder>" ; } getLast ( ) { return last . data ; } | org . junit . Assert . assertEquals ( 2 , lijst . getLast ( ) ) |
testGetGroupFaultyHttpResponse ( ) { super . setResponseResource ( new org . springframework . core . io . ByteArrayResource ( "not-even-valid-json" . getBytes ( ) ) ) ; super . setStatus ( HttpServletResponse . SC_BAD_REQUEST ) ; nl . surfnet . coin . api . client . domain . Group20 group20 = groupService . getGroup20 ( provider , "personId" , "whatever" ) ; "<AssertPlaceHolder>" ; } getGroup20 ( nl . surfnet . coin . teams . domain . GroupProvider , java . lang . String , java . lang . String ) { java . lang . String strippedPersonID = convertToExternalPersonId ( personId , groupProvider ) ; java . lang . String strippedGroupID = convertToExternalGroupId ( groupId , groupProvider ) ; com . sun . jersey . api . client . Client client = getClient ( groupProvider ) ; java . lang . String url = java . lang . String . format ( "%s/groups/%s/%s" , groupProvider . getAllowedOptionAsString ( GroupProviderOptionParameters . URL ) , strippedPersonID , strippedGroupID ) ; nl . surfnet . coin . teams . service . impl . GroupServiceBasicAuthentication . LOG . info ( "Getting<sp>group<sp>information<sp>for<sp>person<sp>{}<sp>and<sp>group<sp>{}<sp>at<sp>groupProvider<sp>{},<sp>using<sp>URL:<sp>{}" , personId , groupId , groupProvider . getIdentifier ( ) , url ) ; com . sun . jersey . api . client . WebResource webResource = client . resource ( url ) ; nl . surfnet . coin . api . client . domain . Group20Entry entry = getGroup20Entry ( groupProvider , webResource ) ; nl . surfnet . coin . teams . service . impl . GroupServiceBasicAuthentication . LOG . debug ( "Got<sp>group<sp>information:<sp>{}" , entry ) ; if ( ( ( entry == null ) || ( ( entry . getEntry ( ) ) == null ) ) || ( entry . getEntry ( ) . isEmpty ( ) ) ) { return null ; } final java . util . List < nl . surfnet . coin . api . client . domain . Group20 > group20s = entry . getEntry ( ) ; if ( ( group20s != null ) && ( ( group20s . size ( ) ) == 1 ) ) { return group20s . get ( 0 ) ; } throw new java . lang . RuntimeException ( java . lang . String . format ( "Received<sp>%s<sp>groups<sp>for<sp>groupid<sp>%s" , group20s . size ( ) , groupId ) ) ; } | org . junit . Assert . assertNull ( group20 ) |
test1ValidateForCreate ( ) { org . apache . ranger . plugin . model . RangerPolicy rangerPolicy = rangerPolicy ( ) ; policyService . validateForCreate ( rangerPolicy ) ; "<AssertPlaceHolder>" ; } validateForCreate ( org . apache . ranger . plugin . model . RangerSecurityZone ) { } | org . junit . Assert . assertNotNull ( rangerPolicy ) |
parallel_001 ( ) { final java . lang . String rulebase = "rules/reloaded/parallel.prova" ; java . util . concurrent . atomic . AtomicInteger count = new java . util . concurrent . atomic . AtomicInteger ( ) ; java . util . Map < java . lang . String , java . lang . Object > globals = new java . util . HashMap < java . lang . String , java . lang . Object > ( ) ; globals . put ( "$Count" , count ) ; prova = new ws . prova . api2 . ProvaCommunicatorImpl ( test . ws . prova . test2 . ProvaMessagingTest . kAgent , test . ws . prova . test2 . ProvaMessagingTest . kPort , rulebase , ws . prova . api2 . ProvaCommunicatorImpl . SYNC , globals ) ; try { synchronized ( this ) { wait ( 200 ) ; "<AssertPlaceHolder>" ; } } catch ( java . lang . Exception e ) { } } get ( ) { return count ; } | org . junit . Assert . assertEquals ( 1 , count . get ( ) ) |
testSupportRegexp ( ) { try { org . talend . dq . dbms . DbmsLanguage dbms = getMysqlDbmsLanguage ( ) ; "<AssertPlaceHolder>" ; } catch ( java . lang . Exception e ) { org . junit . Assert . fail ( e . getMessage ( ) ) ; } } supportRegexp ( ) { org . talend . utils . ProductVersion dbVersion = getDbVersion ( ) ; if ( dbVersion != null ) { return ( dbVersion . getMajor ( ) ) >= 5 ; } return false ; } | org . junit . Assert . assertTrue ( dbms . supportRegexp ( ) ) |
placeItem ( ) { com . jadventure . game . navigation . Location cave = createLocation ( ) ; com . jadventure . game . items . Item bottle = new com . jadventure . game . items . Item ( "bottle" , "tool" , "bottle" , "old<sp>bottle" , 1 , null ) ; cave . addItem ( bottle ) ; com . jadventure . game . items . Item found = cave . removeItem ( bottle ) ; "<AssertPlaceHolder>" ; } getId ( ) { return id ; } | org . junit . Assert . assertEquals ( bottle . getId ( ) , found . getId ( ) ) |
executeRulesInCorrectOrder ( ) { org . junit . experimental . results . PrintableResult . testResult ( org . junit . tests . experimental . rules . RuleChainTest . UseRuleChain . class ) ; java . util . List < java . lang . String > expectedLog = java . util . Arrays . asList ( "starting<sp>outer<sp>rule" , "starting<sp>middle<sp>rule" , "starting<sp>inner<sp>rule" , "finished<sp>inner<sp>rule" , "finished<sp>middle<sp>rule" , "finished<sp>outer<sp>rule" ) ; "<AssertPlaceHolder>" ; } testResult ( java . lang . Class ) { return org . junit . experimental . results . PrintableResult . testResult ( org . junit . runner . Request . aClass ( type ) ) ; } | org . junit . Assert . assertEquals ( expectedLog , org . junit . tests . experimental . rules . RuleChainTest . LOG ) |
testEmptyNumTokens ( ) { refillStrategy . addTokens ( org . isomorphism . util . TokenBucketImplTest . CAPACITY ) ; bucket . consume ( org . isomorphism . util . TokenBucketImplTest . CAPACITY ) ; "<AssertPlaceHolder>" ; } getNumTokens ( ) { refill ( refillStrategy . refill ( ) ) ; return size ; } | org . junit . Assert . assertEquals ( 0 , bucket . getNumTokens ( ) ) |
peerTypeExtractionFromConfiguration ( ) { org . apache . accumulo . server . fs . VolumeManager fs = org . easymock . EasyMock . createMock ( org . apache . accumulo . server . fs . VolumeManager . class ) ; org . apache . accumulo . server . ServerContext context = org . easymock . EasyMock . createMock ( org . apache . accumulo . server . ServerContext . class ) ; java . util . Map < java . lang . String , java . lang . String > data = new java . util . HashMap ( ) ; java . lang . String peerName = "peer" ; java . lang . String configuration = "java.lang.String,foo" ; data . put ( ( ( org . apache . accumulo . core . conf . Property . REPLICATION_PEERS ) + peerName ) , configuration ) ; org . apache . accumulo . core . conf . ConfigurationCopy conf = new org . apache . accumulo . core . conf . ConfigurationCopy ( data ) ; org . apache . accumulo . tserver . replication . ReplicationProcessor proc = new org . apache . accumulo . tserver . replication . ReplicationProcessor ( context , conf , fs ) ; "<AssertPlaceHolder>" ; } getPeerType ( java . lang . String ) { java . util . Map < java . lang . String , java . lang . String > configuredPeers = conf . getAllPropertiesWithPrefix ( Property . REPLICATION_PEERS ) ; java . lang . String peerType = configuredPeers . get ( ( ( Property . REPLICATION_PEERS . getKey ( ) ) + peerName ) ) ; if ( peerType == null ) { java . lang . String msg = "Cannot<sp>process<sp>replication<sp>for<sp>unknown<sp>peer:<sp>" + peerName ; org . apache . accumulo . tserver . replication . ReplicationProcessor . log . warn ( msg ) ; throw new java . lang . IllegalArgumentException ( msg ) ; } return peerType ; } | org . junit . Assert . assertEquals ( configuration , proc . getPeerType ( peerName ) ) |
fetchAllEventsOfSeveralUnexistingAlmanachs ( ) { java . util . Collection < com . stratelia . webactiv . almanach . model . EventDetail > allEvents = eventDAO . findAllEvents ( "almanach1" , "almanach2" , "almanach3" ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return ( this . size ( ) ) == 0 ; } | org . junit . Assert . assertThat ( allEvents . isEmpty ( ) , is ( true ) ) |
testOKFullRun ( ) { com . streamsets . datacollector . main . DataCollectorMain main = new com . streamsets . datacollector . main . TestMain . TMain ( ) ; org . mockito . Mockito . verifyZeroInteractions ( com . streamsets . datacollector . main . TestMain . runtime ) ; org . mockito . Mockito . verifyZeroInteractions ( com . streamsets . datacollector . main . TestMain . logConfigurator ) ; org . mockito . Mockito . verifyZeroInteractions ( com . streamsets . datacollector . main . TestMain . buildInfo ) ; org . mockito . Mockito . verifyZeroInteractions ( com . streamsets . datacollector . main . TestMain . runtimeInfo ) ; org . mockito . Mockito . verifyZeroInteractions ( com . streamsets . datacollector . main . TestMain . task ) ; "<AssertPlaceHolder>" ; org . mockito . Mockito . verify ( com . streamsets . datacollector . main . TestMain . logConfigurator , org . mockito . Mockito . times ( 1 ) ) . configure ( ) ; org . mockito . Mockito . verify ( com . streamsets . datacollector . main . TestMain . buildInfo , org . mockito . Mockito . times ( 1 ) ) . log ( org . mockito . Mockito . any ( org . slf4j . Logger . class ) ) ; org . mockito . Mockito . verify ( com . streamsets . datacollector . main . TestMain . runtimeInfo , org . mockito . Mockito . times ( 1 ) ) . log ( org . mockito . Mockito . any ( org . slf4j . Logger . class ) ) ; org . mockito . Mockito . verify ( com . streamsets . datacollector . main . TestMain . task , org . mockito . Mockito . times ( 1 ) ) . init ( ) ; org . mockito . Mockito . verify ( com . streamsets . datacollector . main . TestMain . runtime , org . mockito . Mockito . times ( 1 ) ) . addShutdownHook ( org . mockito . Mockito . any ( com . streamsets . datacollector . main . Thread . class ) ) ; org . mockito . Mockito . verify ( com . streamsets . datacollector . main . TestMain . task , org . mockito . Mockito . times ( 1 ) ) . run ( ) ; org . mockito . Mockito . verify ( com . streamsets . datacollector . main . TestMain . runtime , org . mockito . Mockito . times ( 1 ) ) . removeShutdownHook ( org . mockito . Mockito . any ( com . streamsets . datacollector . main . Thread . class ) ) ; } doMain ( ) { com . streamsets . datacollector . security . SecurityContext securityContext ; org . slf4j . Logger log = null ; try { final com . streamsets . datacollector . task . Task task = this . task ; dagger . get ( com . streamsets . datacollector . main . LogConfigurator . class ) . configure ( ) ; log = org . slf4j . LoggerFactory . getLogger ( com . streamsets . datacollector . main . Main . class ) ; log . info ( "<sp>Security<sp>Manager<sp>:<sp>DISABLED" 1 ) ; dagger . get ( com . streamsets . datacollector . main . BuildInfo . class ) . log ( log ) ; log . info ( "<sp>Security<sp>Manager<sp>:<sp>DISABLED" 1 ) ; com . streamsets . datacollector . main . RuntimeInfo runtimeInfo = dagger . get ( com . streamsets . datacollector . main . RuntimeInfo . class ) ; runtimeInfo . log ( log ) ; log . info ( "<sp>Security<sp>Manager<sp>:<sp>DISABLED" 1 ) ; com . streamsets . datacollector . util . Configuration configuration = dagger . get ( com . streamsets . datacollector . util . Configuration . class ) ; if ( ( java . lang . System . getSecurityManager ( ) ) != null ) { if ( configuration . get ( PROPERTY_USE_SDC_SECURITY_MANAGER , DEFAULT_USE_SDC_SECURITY_MANAGER ) ) { java . lang . System . setSecurityManager ( new com . streamsets . datacollector . security . SdcSecurityManager ( runtimeInfo , configuration ) ) ; } log . info ( "<sp>Security<sp>Manager<sp>:<sp>ENABLED,<sp>policy<sp>file:<sp>{},<sp>implementation:<sp>{}" , java . lang . System . getProperty ( "<sp>Security<sp>Manager<sp>:<sp>DISABLED" 8 ) , java . lang . System . getSecurityManager ( ) . getClass ( ) . getName ( ) ) ; } else { log . warn ( "<sp>Security<sp>Manager<sp>:<sp>DISABLED" ) ; } log . info ( "<sp>Security<sp>Manager<sp>:<sp>DISABLED" 1 ) ; log . info ( "Starting<sp>..." ) ; java . net . Authenticator . setDefault ( new com . streamsets . datacollector . main . UserPasswordAuthenticator ( ) ) ; securityContext = new com . streamsets . datacollector . security . SecurityContext ( dagger . get ( com . streamsets . datacollector . main . RuntimeInfo . class ) , configuration ) ; securityContext . login ( ) ; log . info ( "<sp>Security<sp>Manager<sp>:<sp>DISABLED" 1 ) ; log . info ( "<sp>Security<sp>Manager<sp>:<sp>DISABLED" 7 , securityContext . getSecurityConfiguration ( ) . isKerberosEnabled ( ) ) ; if ( securityContext . getSecurityConfiguration ( ) . isKerberosEnabled ( ) ) { log . info ( "<sp>Security<sp>Manager<sp>:<sp>DISABLED" 2 , securityContext . getSecurityConfiguration ( ) . getKerberosPrincipal ( ) ) ; log . info ( "<sp>Security<sp>Manager<sp>:<sp>DISABLED" 3 , securityContext . getSecurityConfiguration ( ) . getKerberosKeytab ( ) ) ; } try { boolean unlimited = ( javax . crypto . Cipher . getMaxAllowedKeyLength ( "RC5" ) ) >= 256 ; log . info ( "<sp>Unlimited<sp>cryptography<sp>enabled:<sp>{}" , unlimited ) ; } catch ( java . security . NoSuchAlgorithmException ex ) { log . info ( "<sp>Unlimited<sp>cryptography<sp>check:<sp>algorithm<sp>RC5<sp>not<sp>found." ) ; } log . info ( "<sp>Security<sp>Manager<sp>:<sp>DISABLED" 1 ) ; if ( configuration . get ( "monitor.memory" , false ) ) { log . warn ( "Memory<sp>monitoring<sp>(monitor.memory=true)<sp>is<sp>no<sp>longer<sp>supported." ) ; } log . info ( "Starting<sp>..." ) ; final org . slf4j . | org . junit . Assert . assertEquals ( 0 , main . doMain ( ) ) |
testJobStatusContainsOnlyOneFinished ( ) { java . util . concurrent . atomic . AtomicBoolean jobStarted = new java . util . concurrent . atomic . AtomicBoolean ( true ) ; int countOfCheckers = 100 ; java . util . HashMap < java . lang . Integer , java . util . concurrent . Future < java . lang . StringBuilder > > checkers = new java . util . HashMap ( ) ; java . lang . String testKJBFullPath = this . getClass ( ) . getResource ( "testJob.kjb" ) . getFile ( ) ; org . pentaho . di . job . JobMeta jm = new org . pentaho . di . job . JobMeta ( testKJBFullPath , null ) ; final org . pentaho . di . job . Job job = new org . pentaho . di . job . Job ( null , jm ) ; java . util . concurrent . ExecutorService execServ = java . util . concurrent . Executors . newFixedThreadPool ( countOfCheckers ) ; for ( int i = 0 ; i < countOfCheckers ; i ++ ) { java . util . concurrent . Future < java . lang . StringBuilder > future = execServ . submit ( new java . util . concurrent . Callable < java . lang . StringBuilder > ( ) { private org . pentaho . di . job . StringBuilder statusCollector = new java . lang . StringBuilder ( ) ; @ org . pentaho . di . job . Override public org . pentaho . di . job . StringBuilder call ( ) throws org . pentaho . di . job . Exception { while ( jobStarted . get ( ) ) { statusCollector . append ( job . getStatus ( ) ) ; java . lang . Thread . sleep ( 10 ) ; } return statusCollector ; } } ) ; checkers . put ( i , future ) ; } job . start ( ) ; job . waitUntilFinished ( ) ; jobStarted . set ( false ) ; for ( java . util . concurrent . Future < java . lang . StringBuilder > checker : checkers . values ( ) ) { "<AssertPlaceHolder>" ; } execServ . shutdown ( ) ; } checkOrderStatus ( java . lang . String ) { java . lang . String [ ] tokens = new java . lang . String [ ] { "Waiting" , "Running" , "Finished" } ; int offset = 0 ; for ( java . lang . String t : tokens ) { while ( status . startsWith ( t , offset ) ) offset += t . length ( ) ; } return offset == ( status . length ( ) ) ; } | org . junit . Assert . assertTrue ( checkOrderStatus ( checker . get ( ) . toString ( ) ) ) |
testToGradleArtifactFileName_withClassifier ( ) { java . lang . String group = "org.wildfly.swarm" ; java . lang . String packaging = "jar" ; java . lang . String artifact = "test" ; java . lang . String version = "1.0" ; java . lang . String classifier = "sources" ; org . jboss . modules . maven . ArtifactCoordinates artifactCoordinates = new org . jboss . modules . maven . ArtifactCoordinates ( group , artifact , version , classifier ) ; org . wildfly . swarm . bootstrap . modules . GradleResolver resolver = new org . wildfly . swarm . bootstrap . modules . GradleResolver ( null ) ; java . lang . String artifactFileName = resolver . toGradleArtifactFileName ( artifactCoordinates , packaging ) ; "<AssertPlaceHolder>" ; } toGradleArtifactFileName ( org . jboss . modules . maven . ArtifactCoordinates , java . lang . String ) { java . lang . StringBuilder sbFileFilter = new java . lang . StringBuilder ( ) ; sbFileFilter . append ( artifactCoordinates . getArtifactId ( ) ) . append ( "-" ) . append ( artifactCoordinates . getVersion ( ) ) ; if ( ( ( artifactCoordinates . getClassifier ( ) ) != null ) && ( ( artifactCoordinates . getClassifier ( ) . length ( ) ) > 0 ) ) { sbFileFilter . append ( "-" ) . append ( artifactCoordinates . getClassifier ( ) ) ; } sbFileFilter . append ( "." ) . append ( packaging ) ; return sbFileFilter . toString ( ) ; } | org . junit . Assert . assertEquals ( ( ( ( ( ( ( artifact + "-" ) + version ) + "-" ) + classifier ) + "." ) + packaging ) , artifactFileName ) |
isIdentifierInUseByAnotherPatient_shouldReturnFalseWhenPatientIdentifierDoesNotContainAPatientAndNoPatientHasThisId ( ) { org . openmrs . PatientIdentifierType pit = org . openmrs . api . PatientServiceTest . patientService . getPatientIdentifierType ( 1 ) ; org . openmrs . PatientIdentifier patientIdentifier = new org . openmrs . PatientIdentifier ( "Nobody<sp>could<sp>possibly<sp>have<sp>this<sp>identifier" , pit , null ) ; "<AssertPlaceHolder>" ; } isIdentifierInUseByAnotherPatient ( org . openmrs . PatientIdentifier ) { boolean checkPatient = ( ( patientIdentifier . getPatient ( ) ) != null ) && ( ( patientIdentifier . getPatient ( ) . getPatientId ( ) ) != null ) ; boolean checkLocation = ( ( patientIdentifier . getLocation ( ) ) != null ) && ( ( patientIdentifier . getIdentifierType ( ) . getUniquenessBehavior ( ) ) == ( org . openmrs . PatientIdentifierType . UniquenessBehavior . LOCATION ) ) ; java . lang . String hql = "select<sp>count(*)<sp>from<sp>PatientIdentifier<sp>pi,<sp>Patient<sp>p<sp>where<sp>pi.patient.patientId<sp>=<sp>p.patient.patientId<sp>" + "and<sp>p.voided<sp>=<sp>false<sp>and<sp>pi.voided<sp>=<sp>false<sp>and<sp>pi.identifier<sp>=<sp>:identifier<sp>and<sp>pi.identifierType<sp>=<sp>:idType" ; if ( checkPatient ) { hql += "<sp>and<sp>p.patientId<sp>!=<sp>:ptId" ; } if ( checkLocation ) { hql += "<sp>and<sp>pi.location<sp>=<sp>:locationId" ; } org . hibernate . Query query = sessionFactory . getCurrentSession ( ) . createQuery ( hql ) ; query . setString ( "identifier" , patientIdentifier . getIdentifier ( ) ) ; query . setInteger ( "idType" , patientIdentifier . getIdentifierType ( ) . getPatientIdentifierTypeId ( ) ) ; if ( checkPatient ) { query . setInteger ( "ptId" , patientIdentifier . getPatient ( ) . getPatientId ( ) ) ; } if ( checkLocation ) { query . setInteger ( "locationId" , patientIdentifier . getLocation ( ) . getLocationId ( ) ) ; } return ! ( "0" . equals ( query . uniqueResult ( ) . toString ( ) ) ) ; } | org . junit . Assert . assertFalse ( org . openmrs . api . PatientServiceTest . patientService . isIdentifierInUseByAnotherPatient ( patientIdentifier ) ) |
testAnonymousAccess ( ) { java . io . File temp = java . io . File . createTempFile ( "temp" , java . lang . Long . toString ( java . lang . System . nanoTime ( ) ) ) ; temp . mkdirs ( ) ; io . findify . s3mock . S3Mock mockS3 = new io . findify . s3mock . S3Mock . Builder ( ) . withPort ( 8001 ) . withFileBackend ( temp . getAbsolutePath ( ) ) . withInMemoryBackend ( ) . build ( ) ; mockS3 . start ( ) ; org . locationtech . geowave . core . ingest . IngestUtils . setURLStreamHandlerFactory ( URLTYPE . S3 ) ; org . locationtech . geowave . core . ingest . spark . SparkIngestDriver sparkDriver = new org . locationtech . geowave . core . ingest . spark . SparkIngestDriver ( ) ; com . upplication . s3fs . S3FileSystem s3 = sparkDriver . initializeS3FS ( "s3://s3.amazonaws.com" ) ; s3 . getClient ( ) . setEndpoint ( "http://127.0.0.1:8001" ) ; s3 . getClient ( ) . createBucket ( "testbucket" ) ; s3 . getClient ( ) . putObject ( "testbucket" , "test" , "content" ) ; try ( java . util . stream . Stream < java . nio . file . Path > s = java . nio . file . Files . list ( org . locationtech . geowave . core . ingest . IngestUtils . setupS3FileSystem ( "s3://testbucket/" , "s3://s3.amazonaws.com" ) ) ) { "<AssertPlaceHolder>" ; } mockS3 . shutdown ( ) ; } count ( ) { return CountDataStatistics . STATS_TYPE . newBuilder ( ) ; } | org . junit . Assert . assertEquals ( 1 , s . count ( ) ) |
testStateStoreLazyEval ( ) { final org . apache . kafka . streams . StreamsBuilder builder = new org . apache . kafka . streams . StreamsBuilder ( ) ; final java . lang . String topic1 = "topic1" ; final java . lang . String topic2 = "topic2" ; final org . apache . kafka . streams . kstream . internals . KTableImpl < java . lang . String , java . lang . String , java . lang . String > table1 = ( ( org . apache . kafka . streams . kstream . internals . KTableImpl < java . lang . String , java . lang . String , java . lang . String > ) ( builder . table ( topic1 , consumed ) ) ) ; builder . table ( topic2 , consumed ) ; final org . apache . kafka . streams . kstream . internals . KTableImpl < java . lang . String , java . lang . String , java . lang . Integer > table1Mapped = ( ( org . apache . kafka . streams . kstream . internals . KTableImpl < java . lang . String , java . lang . String , java . lang . Integer > ) ( table1 . mapValues ( Integer :: new ) ) ) ; table1Mapped . filter ( ( key , value ) -> ( value % 2 ) == 0 ) ; try ( final org . apache . kafka . streams . TopologyTestDriver driver = new org . apache . kafka . streams . TopologyTestDriver ( builder . build ( ) , props ) ) { "<AssertPlaceHolder>" ; } } getAllStateStores ( ) { final java . util . Map < java . lang . String , org . apache . kafka . streams . processor . StateStore > allStores = new java . util . HashMap ( ) ; for ( final java . lang . String storeName : internalTopologyBuilder . allStateStoreName ( ) ) { allStores . put ( storeName , getStateStore ( storeName ) ) ; } return allStores ; } | org . junit . Assert . assertEquals ( 0 , driver . getAllStateStores ( ) . size ( ) ) |
testInputRepartitionColumnsAndPartitionCount ( ) { java . util . Map < java . lang . String , java . lang . Object > configMap = com . google . common . collect . Maps . newHashMap ( ) ; configMap . put ( BatchStep . REPARTITION_COLUMNS_PROPERTY , com . google . common . collect . Lists . newArrayList ( "modulo" ) ) ; configMap . put ( BatchStep . REPARTITION_NUM_PARTITIONS_PROPERTY , 5 ) ; configMap . put ( ( ( ( DataStep . INPUT_TYPE ) + "." ) + ( com . cloudera . labs . envelope . input . InputFactory . TYPE_CONFIG_NAME ) ) , com . cloudera . labs . envelope . run . DummyInput . class . getName ( ) ) ; configMap . put ( ( ( ( DataStep . INPUT_TYPE ) + "." ) + "starting.partitions" ) , 10 ) ; com . typesafe . config . Config config = com . typesafe . config . ConfigFactory . parseMap ( configMap ) ; com . cloudera . labs . envelope . run . BatchStep batchStep = new com . cloudera . labs . envelope . run . BatchStep ( "test" ) ; batchStep . configure ( config ) ; batchStep . submit ( com . google . common . collect . Sets . < com . cloudera . labs . envelope . run . Step > newHashSet ( ) ) ; org . apache . spark . sql . Dataset < org . apache . spark . sql . Row > df = batchStep . getData ( ) ; int numPartitions = df . javaRDD ( ) . getNumPartitions ( ) ; "<AssertPlaceHolder>" ; } getData ( ) { return data ; } | org . junit . Assert . assertEquals ( 5 , numPartitions ) |
givenUserService_whenSearchForUserByEmailEndsWith_thenFound ( ) { expect ( userService . findByEmail ( endsWith ( ".com" ) ) ) . andReturn ( java . util . Collections . emptyList ( ) ) ; replay ( userService ) ; java . util . List < com . baeldung . easymock . User > result = userService . findByEmail ( "test@example.com" ) ; verify ( userService ) ; "<AssertPlaceHolder>" ; } size ( ) { return elements . size ( ) ; } | org . junit . Assert . assertEquals ( 0 , result . size ( ) ) |
testDeserializationException ( ) { java . lang . String json = "{\"k1\":[\"v1\",\"v2\"],\"k3\":\"v3\",\"k4\":500,000}" ; boolean ex = false ; try { org . apache . tika . metadata . Metadata deserialized = org . apache . tika . metadata . serialization . JsonMetadata . fromJson ( new java . io . StringReader ( json ) ) ; } catch ( org . apache . tika . exception . TikaException e ) { ex = true ; } "<AssertPlaceHolder>" ; } fromJson ( java . io . Reader ) { java . util . List < org . apache . tika . metadata . Metadata > ms = null ; if ( reader == null ) { return ms ; } try { ms = org . apache . tika . metadata . serialization . JsonMetadataList . GSON . fromJson ( reader , org . apache . tika . metadata . serialization . JsonMetadataList . listType ) ; } catch ( org . apache . tika . metadata . serialization . com . google e ) { throw new org . apache . tika . exception . TikaException ( e . getMessage ( ) ) ; } if ( ms == null ) { return null ; } if ( ( ms . size ( ) ) > 1 ) { org . apache . tika . metadata . Metadata last = ms . get ( ( ( ms . size ( ) ) - 1 ) ) ; java . lang . String embResourcePath = last . get ( RecursiveParserWrapperHandler . EMBEDDED_RESOURCE_PATH ) ; if ( ( embResourcePath == null ) && ( ( ms . get ( 0 ) . get ( RecursiveParserWrapperHandler . EMBEDDED_RESOURCE_PATH ) ) != null ) ) { ms . add ( 0 , ms . remove ( ( ( ms . size ( ) ) - 1 ) ) ) ; } } return ms ; } | org . junit . Assert . assertTrue ( ex ) |
testRunBackup ( ) { final java . util . Collection < java . lang . Throwable > problems = new java . util . ArrayList ( ) ; when ( mockService . backupRepository ( any ( org . fcrepo . kernel . api . FedoraSession . class ) , any ( java . io . File . class ) ) ) . thenReturn ( problems ) ; final java . lang . String backupPath = ( ( java . lang . String ) ( repoBackup . runBackup ( null ) . getEntity ( ) ) ) ; "<AssertPlaceHolder>" ; } runBackup ( java . io . InputStream ) { java . io . File backupDirectory ; if ( null != bodyStream ) { final java . lang . String body = org . apache . commons . io . IOUtils . toString ( bodyStream ) . trim ( ) ; backupDirectory = new java . io . File ( body . trim ( ) ) ; if ( body . isEmpty ( ) ) { backupDirectory = createTempDir ( ) ; } else if ( ( ! ( backupDirectory . exists ( ) ) ) || ( ! ( backupDirectory . canWrite ( ) ) ) ) { throw new javax . ws . rs . WebApplicationException ( serverError ( ) . entity ( ( "Backup<sp>directory<sp>does<sp>not<sp>exist<sp>or<sp>is<sp>not<sp>writable:<sp>" + ( backupDirectory . getAbsolutePath ( ) ) ) ) . build ( ) ) ; } } else { backupDirectory = createTempDir ( ) ; } org . fcrepo . http . api . repository . FedoraRepositoryBackup . LOGGER . debug ( "Backing<sp>up<sp>to:<sp>{}" , backupDirectory . getAbsolutePath ( ) ) ; final java . util . Collection < java . lang . Throwable > problems = repositoryService . backupRepository ( session . getFedoraSession ( ) , backupDirectory ) ; if ( ! ( problems . isEmpty ( ) ) ) { org . fcrepo . http . api . repository . FedoraRepositoryBackup . LOGGER . error ( "Problems<sp>backing<sp>up<sp>the<sp>repository:" ) ; final java . lang . String output = problems . stream ( ) . map ( Throwable :: getMessage ) . peek ( org . fcrepo . http . api . repository . FedoraRepositoryBackup . LOGGER :: error ) . collect ( java . util . stream . Collectors . joining ( "\n" ) ) ; throw new javax . ws . rs . WebApplicationException ( serverError ( ) . entity ( output ) . build ( ) ) ; } return ok ( ) . header ( "Warning" , "This<sp>endpoint<sp>will<sp>be<sp>moving<sp>to<sp>an<sp>extension<sp>module<sp>in<sp>a<sp>future<sp>release<sp>of<sp>Fedora" ) . entity ( backupDirectory . getCanonicalPath ( ) ) . build ( ) ; } | org . junit . Assert . assertNotNull ( backupPath ) |
whenLoggerEmitsLoggingEvent_thenAppenderReceivesEvent ( ) { logger . info ( "Test<sp>from<sp>{}" , this . getClass ( ) . getSimpleName ( ) ) ; org . apache . logging . log4j . core . LoggerContext context = org . apache . logging . log4j . core . LoggerContext . getContext ( false ) ; org . apache . logging . log4j . core . config . Configuration config = context . getConfiguration ( ) ; com . baeldung . logging . log4j2 . appender . MapAppender appender = config . getAppender ( "MapAppender" ) ; "<AssertPlaceHolder>" ; } getEventMap ( ) { return eventMap ; } | org . junit . Assert . assertEquals ( appender . getEventMap ( ) . size ( ) , 1 ) |
testGetServerContext ( ) { javax . net . ssl . SSLContext context = sslContextFactory . getServerContext ( ) ; "<AssertPlaceHolder>" ; } getServerContext ( ) { java . lang . String algorithm = java . security . Security . getProperty ( "ssl.KeyManagerFactory.algorithm" ) ; if ( algorithm == null ) { algorithm = "SunX509" ; } javax . net . ssl . SSLContext serverContext = null ; try { java . security . KeyStore ks = java . security . KeyStore . getInstance ( tlsConfig . getTlsKeystoreType ( ) . name ( ) ) ; ks . load ( org . opendaylight . openflowjava . protocol . impl . core . SslKeyStore . asInputStream ( tlsConfig . getTlsKeystore ( ) , tlsConfig . getTlsKeystorePathType ( ) ) , tlsConfig . getKeystorePassword ( ) . toCharArray ( ) ) ; javax . net . ssl . KeyManagerFactory kmf = javax . net . ssl . KeyManagerFactory . getInstance ( algorithm ) ; kmf . init ( ks , tlsConfig . getCertificatePassword ( ) . toCharArray ( ) ) ; java . security . KeyStore ts = java . security . KeyStore . getInstance ( tlsConfig . getTlsTruststoreType ( ) . name ( ) ) ; ts . load ( org . opendaylight . openflowjava . protocol . impl . core . SslKeyStore . asInputStream ( tlsConfig . getTlsTruststore ( ) , tlsConfig . getTlsTruststorePathType ( ) ) , tlsConfig . getTruststorePassword ( ) . toCharArray ( ) ) ; javax . net . ssl . TrustManagerFactory tmf = javax . net . ssl . TrustManagerFactory . getInstance ( algorithm ) ; tmf . init ( ts ) ; serverContext = javax . net . ssl . SSLContext . getInstance ( org . opendaylight . openflowjava . protocol . impl . core . SslContextFactory . PROTOCOL ) ; serverContext . init ( kmf . getKeyManagers ( ) , tmf . getTrustManagers ( ) , null ) ; } catch ( java . io . IOException e ) { org . opendaylight . openflowjava . protocol . impl . core . SslContextFactory . LOG . warn ( ( "IOException<sp>-<sp>Failed<sp>to<sp>load<sp>keystore<sp>/<sp>truststore." + "<sp>Failed<sp>to<sp>initialize<sp>the<sp>server-side<sp>SSLContext" ) , e ) ; } catch ( java . security . NoSuchAlgorithmException e ) { org . opendaylight . openflowjava . protocol . impl . core . SslContextFactory . LOG . warn ( ( "NoSuchAlgorithmException<sp>-<sp>Unsupported<sp>algorithm." + "<sp>Failed<sp>to<sp>initialize<sp>the<sp>server-side<sp>SSLContext" ) , e ) ; } catch ( java . security . cert . CertificateException e ) { org . opendaylight . openflowjava . protocol . impl . core . SslContextFactory . LOG . warn ( ( "CertificateException<sp>-<sp>Unable<sp>to<sp>access<sp>certificate<sp>(check<sp>password)." + "<sp>Failed<sp>to<sp>initialize<sp>the<sp>server-side<sp>SSLContext" ) , e ) ; } catch ( java . security . KeyManagementException | java . security . KeyStoreException | java . security . UnrecoverableKeyException e ) { org . opendaylight . openflowjava . protocol . impl . core . SslContextFactory . LOG . warn ( "Exception<sp>-<sp>Failed<sp>to<sp>initialize<sp>the<sp>server-side<sp>SSLContext" , e ) ; } return serverContext ; } | org . junit . Assert . assertNotNull ( context ) |
testPeriodBarriersCrossedWhenLeavingDaylightSaving ( ) { ch . qos . logback . core . rolling . helper . RollingCalendar rc = new ch . qos . logback . core . rolling . helper . RollingCalendar ( dailyPattern , java . util . TimeZone . getTimeZone ( "CET" ) , java . util . Locale . US ) ; long start = 1509228123333L ; long end = 1509228123333L + ( 25 * ( ch . qos . logback . core . CoreConstants . MILLIS_IN_ONE_HOUR ) ) ; "<AssertPlaceHolder>" ; } periodBarriersCrossed ( long , long ) { if ( start > end ) throw new java . lang . IllegalArgumentException ( "Start<sp>cannot<sp>come<sp>before<sp>end" ) ; long startFloored = getStartOfCurrentPeriodWithGMTOffsetCorrection ( start , getTimeZone ( ) ) ; long endFloored = getStartOfCurrentPeriodWithGMTOffsetCorrection ( end , getTimeZone ( ) ) ; long diff = endFloored - startFloored ; switch ( periodicityType ) { case TOP_OF_MILLISECOND : return diff ; case TOP_OF_SECOND : return diff / ( ch . qos . logback . core . CoreConstants . MILLIS_IN_ONE_SECOND ) ; case TOP_OF_MINUTE : return diff / ( ch . qos . logback . core . CoreConstants . MILLIS_IN_ONE_MINUTE ) ; case TOP_OF_HOUR : return ( ( int ) ( diff ) ) / ( ch . qos . logback . core . CoreConstants . MILLIS_IN_ONE_HOUR ) ; case TOP_OF_DAY : return diff / ( ch . qos . logback . core . CoreConstants . MILLIS_IN_ONE_DAY ) ; case TOP_OF_WEEK : return diff / ( ch . qos . logback . core . CoreConstants . MILLIS_IN_ONE_WEEK ) ; case TOP_OF_MONTH : return ch . qos . logback . core . rolling . helper . RollingCalendar . diffInMonths ( start , end ) ; default : throw new java . lang . IllegalStateException ( "Unknown<sp>periodicity<sp>type." ) ; } } | org . junit . Assert . assertEquals ( 1 , rc . periodBarriersCrossed ( start , end ) ) |
retrieveTopicName ( ) { cern . c2mon . client . core . tag . TagController tagController = new cern . c2mon . client . core . tag . TagController ( 1234L ) ; tagController . onUpdate ( createValidTransferTag ( 1234L ) ) ; tagController . getTagImpl ( ) . setTopicName ( "test" ) ; "<AssertPlaceHolder>" ; } getTopicName ( ) { return "c2mon.JmsProxy.test.topic.registration" ; } | org . junit . Assert . assertTrue ( tagController . getTagImpl ( ) . getTopicName ( ) . equals ( "test" ) ) |
testBuildWithEmptyContext ( ) { java . util . List < java . lang . Long > context = java . util . Collections . emptyList ( ) ; java . lang . String expectedQuery = "SELECT<sp>e<sp>FROM<sp>UserGroupView<sp>e<sp>WHERE<sp>e.status=:status<sp>AND<sp>e.crtUserGroup<sp>IN<sp>(:userGroups)<sp>" ; org . lnu . is . pagination . MultiplePagedSearch < java . util . List < java . lang . Long > > pagedSearch = new org . lnu . is . pagination . MultiplePagedSearch ( ) ; pagedSearch . setEntity ( context ) ; java . lang . String actualQuery = unit . build ( pagedSearch ) ; "<AssertPlaceHolder>" ; } setEntity ( T ) { this . entity = entity ; } | org . junit . Assert . assertEquals ( expectedQuery , actualQuery ) |
testNullTaskObject ( ) { java . util . List l = interceptor . getHandlers ( null ) ; "<AssertPlaceHolder>" ; } getHandlers ( java . lang . Object ) { if ( o == null ) return org . cytoscape . work . Collections . emptyList ( ) ; return loadTunables ( o , 0.0 ) ; } | org . junit . Assert . assertTrue ( l . isEmpty ( ) ) |
reportIdsStartWithRepositoryIdTest ( ) { java . util . List < org . opennms . features . reporting . model . basicreport . BasicReportDefinition > reports = m_defaultRemoteRepository . getReports ( ) ; logger . debug ( "reportIdsStartWithRepositoryIdTest" ) ; for ( org . opennms . features . reporting . model . basicreport . BasicReportDefinition report : reports ) { "<AssertPlaceHolder>" ; logger . debug ( report . getId ( ) ) ; logger . debug ( "'{}'" , report . getRepositoryId ( ) ) ; } } getId ( ) { return m_id ; } | org . junit . Assert . assertTrue ( report . getId ( ) . startsWith ( m_defaultRemoteRepository . getRepositoryId ( ) ) ) |
testReSeed ( ) { empiricalDistribution . load ( url ) ; empiricalDistribution . reSeed ( 100 ) ; final double [ ] values = new double [ 10 ] ; for ( int i = 0 ; i < 10 ; i ++ ) { values [ i ] = empiricalDistribution . getNextValue ( ) ; } empiricalDistribution . reSeed ( 100 ) ; for ( int i = 0 ; i < 10 ; i ++ ) { "<AssertPlaceHolder>" ; } } load ( java . io . File ) { org . hipparchus . util . MathUtils . checkNotNull ( file ) ; java . nio . charset . Charset charset = java . nio . charset . Charset . forName ( org . hipparchus . stat . fitting . EmpiricalDistribution . FILE_CHARSET ) ; try ( java . io . InputStream is1 = java . nio . file . Files . newInputStream ( file . toPath ( ) ) ; java . io . BufferedReader in1 = new java . io . BufferedReader ( new java . io . InputStreamReader ( is1 , charset ) ) ) { org . hipparchus . stat . fitting . EmpiricalDistribution . DataAdapter da = new org . hipparchus . stat . fitting . EmpiricalDistribution . StreamDataAdapter ( in1 ) ; da . computeStats ( ) ; try ( java . io . InputStream is2 = java . nio . file . Files . newInputStream ( file . toPath ( ) ) ; java . io . BufferedReader in2 = new java . io . BufferedReader ( new java . io . InputStreamReader ( is2 , charset ) ) ) { fillBinStats ( new org . hipparchus . stat . fitting . EmpiricalDistribution . StreamDataAdapter ( in2 ) ) ; } loaded = true ; } } | org . junit . Assert . assertEquals ( values [ i ] , empiricalDistribution . getNextValue ( ) , 0.0 ) |
shouldOnlyBindCookieWhenNoBindingAlreadyMade ( ) { com . threewks . thundr . introspection . ParameterDescription varParam = new com . threewks . thundr . introspection . ParameterDescription ( "var" , java . lang . String . class ) ; bindings . put ( varParam , "original" ) ; req . withCookie ( "var" , "overridden" ) ; binder . bindAll ( bindings , req , resp ) ; "<AssertPlaceHolder>" ; } get ( java . lang . String ) { return delegate . get ( java . util . Collections . singletonList ( arg0 ) ) ; } | org . junit . Assert . assertThat ( bindings . get ( varParam ) , is ( ( ( java . lang . Object ) ( "original" ) ) ) ) |
testGetModifiedDate ( ) { net . holmes . core . business . media . model . AbstractNodeTest . AbstractNodeTester node = buildAbstractNodeTester ( "" ) ; "<AssertPlaceHolder>" ; } getModifiedDate ( ) { return modifiedDate ; } | org . junit . Assert . assertNull ( node . getModifiedDate ( ) ) |
testToStringOther ( ) { org . eclipse . kura . core . net . NetworkConfiguration config = new org . eclipse . kura . core . net . NetworkConfiguration ( ) ; org . eclipse . kura . core . net . EthernetInterfaceConfigImpl interfaceConfig = new org . eclipse . kura . core . net . EthernetInterfaceConfigImpl ( "if1" ) ; java . util . List < org . eclipse . kura . net . NetInterfaceAddressConfig > interfaceAddresses = new java . util . ArrayList ( ) ; org . eclipse . kura . core . net . NetInterfaceAddressConfigImpl addressConfig = new org . eclipse . kura . core . net . NetInterfaceAddressConfigImpl ( ) ; java . util . List < org . eclipse . kura . net . NetConfig > netConfigs = new java . util . ArrayList ( ) ; try { org . eclipse . kura . net . dhcp . DhcpServerCfg dhcpServerCfg = new org . eclipse . kura . net . dhcp . DhcpServerCfg ( "eth0" , true , 7200 , 7200 , false ) ; org . eclipse . kura . net . dhcp . DhcpServerCfgIP4 dhcpServerCfgIP4 = new org . eclipse . kura . net . dhcp . DhcpServerCfgIP4 ( ( ( org . eclipse . kura . net . IP4Address ) ( org . eclipse . kura . net . IPAddress . parseHostAddress ( "192.168.1.0" ) ) ) , ( ( org . eclipse . kura . net . IP4Address ) ( org . eclipse . kura . net . IPAddress . parseHostAddress ( "255.255.255.0" ) ) ) , ( ( short ) ( 24 ) ) , ( ( org . eclipse . kura . net . IP4Address ) ( org . eclipse . kura . net . IPAddress . parseHostAddress ( "failed:<sp>" 1 ) ) ) , ( ( org . eclipse . kura . net . IP4Address ) ( org . eclipse . kura . net . IPAddress . parseHostAddress ( "192.168.1.100" ) ) ) , ( ( org . eclipse . kura . net . IP4Address ) ( org . eclipse . kura . net . IPAddress . parseHostAddress ( "192.168.1.254" ) ) ) , null ) ; netConfigs . add ( new org . eclipse . kura . net . dhcp . DhcpServerConfigIP4 ( dhcpServerCfg , dhcpServerCfgIP4 ) ) ; } catch ( java . lang . Exception e ) { org . junit . Assert . fail ( ( "failed:<sp>" + e ) ) ; } netConfigs . add ( new org . eclipse . kura . net . firewall . FirewallAutoNatConfig ( ) ) ; netConfigs . add ( null ) ; netConfigs . add ( new org . eclipse . kura . core . net . MockConfig ( ) ) ; addressConfig . setNetConfigs ( netConfigs ) ; interfaceAddresses . add ( addressConfig ) ; interfaceConfig . setNetInterfaceAddresses ( interfaceAddresses ) ; config . addNetInterfaceConfig ( interfaceConfig ) ; java . lang . String expected = "\nname:<sp>if1<sp>::<sp>Loopback?<sp>false<sp>::<sp>Point<sp>to<sp>Point?<sp>false<sp>::<sp>Up?<sp>false<sp>::<sp>Virtual?<sp>false" + ( ( "<sp>::<sp>Driver:<sp>null<sp>::<sp>Driver<sp>Version:<sp>null<sp>::<sp>Firmware<sp>Version:<sp>null<sp>::<sp>MTU:<sp>0<sp>::<sp>State:<sp>null" + "<sp>::<sp>Type:<sp>ETHERNET<sp>::<sp>Usb<sp>Device:<sp>null<sp>::<sp>Prefix:<sp>0\n\tDhcpServerConfig<sp>\n\tFirewallAutoNatConfig<sp>" ) + "failed:<sp>" 0 ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( ( ( "ComponentConfigurationImpl<sp>[pid=" + ( pid ) ) + ",<sp>definition=" ) + ( definition ) ) + ",<sp>properties=" ) + ( properties ) ) + "]" ; } | org . junit . Assert . assertEquals ( expected , config . toString ( ) ) |
testBigDecimal ( ) { final java . lang . String drl = ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( "end" 0 + ( "\n" + "end" 3 ) ) + ( org . drools . testcoverage . common . model . Cheese . class . getCanonicalName ( ) ) ) + "end" 1 ) + "end" 3 ) + ( org . drools . testcoverage . common . model . Primitives . class . getCanonicalName ( ) ) ) + "end" 1 ) + "end" 2 ) + "\n" ) + "global<sp>java.util.List<sp>list;\n" ) + "\n" ) + "rule<sp>\"BigDec\"\n" ) + "\n" ) + "<sp>when\n" ) + "<sp>Cheese($price<sp>:<sp>price)\n" ) + "<sp>p<sp>:<sp>Primitives(bigDecimal<sp><<sp>$price)\n" ) + "<sp>then\n" ) + "end" 6 ) + "end" ; final org . kie . api . KieBase kbase = org . drools . testcoverage . common . util . KieBaseUtil . getKieBaseFromKieModuleFromDrl ( "end" 4 , kieBaseTestConfiguration , drl ) ; final org . kie . api . runtime . KieSession session = kbase . newKieSession ( ) ; try { final java . util . List list = new java . util . ArrayList ( ) ; session . setGlobal ( "end" 5 , list ) ; final org . drools . testcoverage . common . model . Primitives bill = new org . drools . testcoverage . common . model . Primitives ( ) ; bill . setBigDecimal ( new java . math . BigDecimal ( "42" ) ) ; final org . drools . testcoverage . common . model . Primitives ben = new org . drools . testcoverage . common . model . Primitives ( ) ; ben . setBigDecimal ( new java . math . BigDecimal ( "end" 7 ) ) ; session . insert ( bill ) ; session . insert ( new org . drools . testcoverage . common . model . Cheese ( "gorgonzola" , 43 ) ) ; session . insert ( ben ) ; session . fireAllRules ( ) ; "<AssertPlaceHolder>" ; } finally { session . dispose ( ) ; } } getGlobal ( java . lang . String ) { return null ; } | org . junit . Assert . assertEquals ( 1 , ( ( java . util . List ) ( session . getGlobal ( "end" 5 ) ) ) . size ( ) ) |
testCountingBloomFilter2 ( ) { final int countingSize = 20 ; final int items = 100 ; final int error = 8 ; int [ ] counting = new int [ countingSize ] ; net . tomp2p . rpc . CountingBloomFilter < java . lang . String > cbs = new net . tomp2p . rpc . CountingBloomFilter < java . lang . String > ( bfSizeLarge , counting ) ; System . out . println ( cbs . expectedFalsePositiveProbability ( ) ) ; for ( int i = 0 ; i < items ; i ++ ) { cbs . add ( "abc" ) ; } for ( int i = 0 ; i < items ; i ++ ) { cbs . add ( ( "abc" + i ) ) ; } "<AssertPlaceHolder>" ; } approximateCount ( E ) { int retVal = Integer . MAX_VALUE ; java . util . Random r = new java . util . Random ( key . hashCode ( ) ) ; for ( int x = 0 ; x < ( k ) ; x ++ ) { retVal = java . lang . Math . min ( retVal , intSet [ r . nextInt ( intArraySize ) ] ) ; } return retVal ; } | org . junit . Assert . assertEquals ( ( items + error ) , cbs . approximateCount ( "abc" ) ) |
scryptViaLambdaWorksTest ( ) { java . lang . String encodedValue = com . jedivision . Hash . scryptViaLambdaWorks ( com . jedivision . HashTest . GLOBAL_VALUE ) ; boolean matched = com . lambdaworks . crypto . SCryptUtil . check ( com . jedivision . HashTest . GLOBAL_VALUE , encodedValue ) ; "<AssertPlaceHolder>" ; } scryptViaLambdaWorks ( java . lang . String ) { com . jedivision . Hash . LOGGER . debug ( "Hashing<sp>{}<sp>with<sp>scrypt<sp>hashing<sp>algorithm" , value ) ; return com . lambdaworks . crypto . SCryptUtil . scrypt ( value , 16 , 16 , 16 ) ; } | org . junit . Assert . assertTrue ( matched ) |
testConstructorWithBeginningOfQuery ( ) { io . quantumdb . core . utils . QueryBuilder builder = new io . quantumdb . core . utils . QueryBuilder ( "SELECT" ) ; "<AssertPlaceHolder>" ; } toString ( ) { return io . quantumdb . core . schema . definitions . PrettyPrinter . prettyPrint ( this ) ; } | org . junit . Assert . assertEquals ( "SELECT" , builder . toString ( ) ) |
i02a ( ) { boolean r = proxy . httpStatusReturnBool200 ( ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertEquals ( true , r ) |
testWithOneFileWithWrongOrder ( ) { uk . gov . nationalarchives . droid . core . interfaces . archive . AsynchDroid droid = mock ( uk . gov . nationalarchives . droid . core . interfaces . archive . AsynchDroid . class ) ; uk . gov . nationalarchives . droid . core . interfaces . archive . IdentificationRequestFactory < java . io . InputStream > factory = mock ( uk . gov . nationalarchives . droid . core . interfaces . archive . IdentificationRequestFactory . class ) ; when ( factory . newRequest ( any ( uk . gov . nationalarchives . droid . core . interfaces . resource . RequestMetaData . class ) , any ( uk . gov . nationalarchives . droid . core . interfaces . archive . RequestIdentifier . class ) ) ) . thenReturn ( mock ( uk . gov . nationalarchives . droid . core . interfaces . archive . IdentificationRequest . class ) ) ; uk . gov . nationalarchives . droid . core . interfaces . archive . ResultHandler resultHandler = mock ( uk . gov . nationalarchives . droid . core . interfaces . archive . ResultHandler . class ) ; when ( resultHandler . handleDirectory ( any ( uk . gov . nationalarchives . droid . core . interfaces . archive . IdentificationResult . class ) , any ( uk . gov . nationalarchives . droid . core . interfaces . archive . ResourceId . class ) , anyBoolean ( ) ) ) . thenReturn ( mock ( uk . gov . nationalarchives . droid . core . interfaces . archive . ResourceId . class ) ) ; com . github . stephenc . javaisotools . loopfs . iso9660 . Iso9660FileSystem filesystem = mock ( com . github . stephenc . javaisotools . loopfs . iso9660 . Iso9660FileSystem . class ) ; uk . gov . nationalarchives . droid . core . interfaces . archive . RequestIdentifier requestIdentifier = new uk . gov . nationalarchives . droid . core . interfaces . archive . RequestIdentifier ( new java . net . URI ( "mock://some/path/to/iso" ) ) ; requestIdentifier . setNodeId ( 10L ) ; uk . gov . nationalarchives . droid . core . interfaces . archive . ISOImageArchiveHandler . ISOImageArchiveWalker walker = new uk . gov . nationalarchives . droid . core . interfaces . archive . ISOImageArchiveHandler . ISOImageArchiveWalker ( droid , factory , resultHandler , filesystem , requestIdentifier ) ; com . github . stephenc . javaisotools . loopfs . iso9660 . Iso9660FileEntry rootEntry = mock ( com . github . stephenc . javaisotools . loopfs . iso9660 . Iso9660FileEntry . class ) ; when ( rootEntry . getPath ( ) ) . thenReturn ( "" ) ; when ( rootEntry . isDirectory ( ) ) . thenReturn ( true ) ; com . github . stephenc . javaisotools . loopfs . iso9660 . Iso9660FileEntry contentDir = mock ( com . github . stephenc . javaisotools . loopfs . iso9660 . Iso9660FileEntry . class ) ; when ( contentDir . getPath ( ) ) . thenReturn ( "content/" ) ; when ( contentDir . isDirectory ( ) ) . thenReturn ( true ) ; com . github . stephenc . javaisotools . loopfs . iso9660 . Iso9660FileEntry testFile = mock ( com . github . stephenc . javaisotools . loopfs . iso9660 . Iso9660FileEntry . class ) ; when ( testFile . getName ( ) ) . thenReturn ( "test.txt" ) ; when ( testFile . getPath ( ) ) . thenReturn ( "content/test.txt" ) ; when ( testFile . isDirectory ( ) ) . thenReturn ( false ) ; java . util . List < com . github . stephenc . javaisotools . loopfs . iso9660 . Iso9660FileEntry > entryList = new java . util . ArrayList ( ) ; entryList . add ( rootEntry ) ; entryList . add ( testFile ) ; entryList . add ( contentDir ) ; walker . walk ( entryList ) ; org . mockito . InOrder inOrder = inOrder ( resultHandler , droid ) ; inOrder . verify ( resultHandler , times ( 1 ) ) . handleDirectory ( any ( uk . gov . nationalarchives . droid . core . interfaces . archive . IdentificationResult . class ) , any ( uk . gov . nationalarchives . droid . core . interfaces . archive . ResourceId . class ) , anyBoolean ( ) ) ; inOrder . verify ( droid , times ( 1 ) ) . submit ( any ( uk . gov . nationalarchives . droid . core . interfaces . archive . IdentificationRequest . class ) ) ; "<AssertPlaceHolder>" ; } submit ( uk . gov . nationalarchives . droid . core . interfaces . IdentificationRequest ) { jobCounter . increment ( ) ; requests . add ( request ) ; java . util . concurrent . Callable < uk . gov . nationalarchives . droid . core . interfaces . IdentificationResultCollection > callable = new java . util . concurrent . Callable < uk . gov . nationalarchives . droid . core . interfaces . IdentificationResultCollection > ( ) { @ uk . gov . nationalarchives . droid . submitter . Override public uk . gov . nationalarchives . droid . core . interfaces . IdentificationResultCollection call ( ) throws java . io . IOException { droidCore . setMaxBytesToScan ( maxBytesToScan ) ; uk . gov . nationalarchives . droid . core . interfaces . IdentificationResultCollection results = droidCore . matchBinarySignatures ( request ) ; return results ; } } ; java . util . concurrent . FutureTask < uk . gov . nationalarchives . droid . core . interfaces . IdentificationResultCollection > task = new uk . gov . nationalarchives . droid . submitter . SubmissionGateway . SubmissionFutureTask ( callable , request ) ; executorService . submit ( task ) ; return task ; } | org . junit . Assert . assertTrue ( true ) |
testLoadWithNullBundleContext ( ) { com . codeaffine . osgi . testuite . internal . DevPropertiesLoader loader = new com . codeaffine . osgi . testuite . internal . DevPropertiesLoader ( null ) ; java . util . Properties properties = loader . load ( ) ; "<AssertPlaceHolder>" ; } load ( ) { if ( ( bundleContext ) != null ) { java . lang . String osgiDevProperty = bundleContext . getProperty ( "osgi.dev" ) ; if ( ( osgiDevProperty != null ) && ( ( osgiDevProperty . length ( ) ) > 0 ) ) { java . net . URL url = com . codeaffine . osgi . testuite . internal . DevPropertiesLoader . toUrl ( osgiDevProperty ) ; load ( url ) ; } } return properties ; } | org . junit . Assert . assertEquals ( 0 , properties . size ( ) ) |
testRewriteCorrelatedSubqueryInHaving ( ) { java . lang . String sql = "select<sp>pm1.g1.e1<sp>from<sp>pm1.g1<sp>group<sp>by<sp>pm1.g1.e1<sp>having<sp>pm1.g1.e1<sp>in<sp>(select<sp>pm1.g1.e1<sp>from<sp>pm1.g2)" ; java . lang . String expected = "SELECT<sp>pm1.g1.e1<sp>FROM<sp>pm1.g1<sp>GROUP<sp>BY<sp>pm1.g1.e1<sp>HAVING<sp>pm1.g1.e1<sp>IN<sp>(SELECT<sp>pm1.g1.e1<sp>FROM<sp>pm1.g2)" ; org . teiid . query . rewriter . Query query = ( ( org . teiid . query . rewriter . Query ) ( org . teiid . query . rewriter . TestQueryRewriter . helpTestRewriteCommand ( sql , expected ) ) ) ; java . util . List < org . teiid . query . sql . symbol . Reference > refs = new java . util . LinkedList < org . teiid . query . sql . symbol . Reference > ( ) ; org . teiid . query . sql . symbol . GroupSymbol gs = new org . teiid . query . sql . symbol . GroupSymbol ( "pm1.g1" ) ; org . teiid . query . resolver . util . ResolverUtil . resolveGroup ( gs , org . teiid . query . unittest . RealMetadataFactory . example1Cached ( ) ) ; org . teiid . query . sql . visitor . CorrelatedReferenceCollectorVisitor . collectReferences ( query . getHaving ( ) , java . util . Arrays . asList ( gs ) , refs , org . teiid . query . unittest . RealMetadataFactory . example1Cached ( ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return size ; } | org . junit . Assert . assertEquals ( 1 , refs . size ( ) ) |
aBaseCalendarMustBeActive ( ) { org . libreplan . business . calendars . entities . BaseCalendar calendar = org . libreplan . business . test . calendars . entities . BaseCalendarTest . createBasicCalendar ( ) ; "<AssertPlaceHolder>" ; } isActive ( org . joda . time . LocalDate ) { return ( ( startDate . compareTo ( date ) ) <= 0 ) && ( ! ( ( ( endDate ) != null ) && ( ( endDate . compareTo ( date ) ) < 0 ) ) ) ; } | org . junit . Assert . assertTrue ( calendar . isActive ( new org . joda . time . LocalDate ( ) ) ) |
testAddLog ( ) { service . log ( "JUNIT<sp>TEST<sp>ADD<sp>LOG" ) ; "<AssertPlaceHolder>" ; } logEntries ( ) { return logEntryConverter . convert ( dao . findAll ( ) ) ; } | org . junit . Assert . assertEquals ( "JUNIT<sp>TEST<sp>ADD<sp>LOG" , service . logEntries ( ) . get ( 0 ) . getMessage ( ) ) |
testGivenMOKBuyWith2LOSellWhenMOKQuantityIsBigger ( ) { vn . com . vndirect . exchangesimulator . model . NewOrderSingle order = vn . com . vndirect . exchangesimulator . matching . OrderFactory . createLOSell ( "VND" , 1000 , 13000 ) ; sm . push ( order ) ; vn . com . vndirect . exchangesimulator . model . NewOrderSingle order2 = vn . com . vndirect . exchangesimulator . matching . OrderFactory . createLOSell ( "VND" , 2000 , 13500 ) ; sm . push ( order2 ) ; vn . com . vndirect . exchangesimulator . model . NewOrderSingle mokOrder = vn . com . vndirect . exchangesimulator . matching . OrderFactory . createMOKBuy ( "VND" , 7000 ) ; sm . push ( mokOrder ) ; java . util . List < vn . com . vndirect . exchangesimulator . model . ExecutionReport > reports = sm . getLastMatches ( ) ; "<AssertPlaceHolder>" ; verifyExpiredReport ( reports . get ( 0 ) , 7000 ) ; } size ( ) { return queue . size ( ) ; } | org . junit . Assert . assertEquals ( 1 , reports . size ( ) ) |
testGoodDesc ( ) { org . apache . kylin . cube . model . validation . rule . DictionaryRule rule = new org . apache . kylin . cube . model . validation . rule . DictionaryRule ( ) ; for ( java . io . File f : new java . io . File ( ( ( org . apache . kylin . common . util . LocalFileMetadataTestCase . LOCALMETA_TEST_DATA ) + "/cube_desc/" ) ) . listFiles ( ) ) { if ( ! ( f . getName ( ) . endsWith ( "json" ) ) ) { continue ; } org . apache . kylin . cube . model . CubeDesc desc = org . apache . kylin . common . util . JsonUtil . readValue ( new java . io . FileInputStream ( f ) , org . apache . kylin . cube . model . CubeDesc . class ) ; desc . init ( org . apache . kylin . cube . model . validation . rule . DictionaryRuleTest . config ) ; org . apache . kylin . cube . model . validation . ValidateContext vContext = new org . apache . kylin . cube . model . validation . ValidateContext ( ) ; rule . validate ( desc , vContext ) ; "<AssertPlaceHolder>" ; } } getResults ( ) { return results ; } | org . junit . Assert . assertTrue ( ( ( vContext . getResults ( ) . length ) == 0 ) ) |
marshall_BrokerRevenueShareResult ( ) { org . oscm . billingservice . business . model . brokershare . BrokerRevenueShareResult brokerRevenueShareResult = createBrokerRevenueShareResult ( ) ; byte [ ] serialized = sharesCalculator . marshallRevenueShareResults ( brokerRevenueShareResult ) ; "<AssertPlaceHolder>" ; } marshallRevenueShareResults ( java . lang . Object ) { javax . xml . bind . Marshaller marshaller = createMarshaller ( obj ) ; marshaller . setProperty ( Marshaller . JAXB_ENCODING , "UTF-8" ) ; marshaller . setProperty ( Marshaller . JAXB_FORMATTED_OUTPUT , Boolean . TRUE ) ; marshaller . setProperty ( Marshaller . JAXB_FRAGMENT , Boolean . TRUE ) ; java . io . ByteArrayOutputStream bos = new java . io . ByteArrayOutputStream ( ) ; marshaller . marshal ( obj , bos ) ; return bos . toByteArray ( ) ; } | org . junit . Assert . assertNotNull ( serialized ) |
testCopyFieldsValues ( ) { org . apache . avro . Schema intSchema = org . apache . avro . SchemaBuilder . record ( "intSchema" ) . fields ( ) . name ( "a" ) . type ( ) . intType ( ) . noDefault ( ) . endRecord ( ) ; org . apache . avro . generic . GenericRecord intRecord = new org . apache . avro . generic . GenericRecordBuilder ( intSchema ) . set ( "a" , 1 ) . build ( ) ; org . apache . avro . Schema stringSchema = org . apache . avro . SchemaBuilder . record ( "intSchema" ) . fields ( ) . name ( "a" ) . type ( ) . stringType ( ) . noDefault ( ) . endRecord ( ) ; org . apache . avro . generic . GenericRecordBuilder stringRecordBuilder = new org . apache . avro . generic . GenericRecordBuilder ( stringSchema ) . set ( "a" , "s" ) ; org . talend . components . processing . runtime . typeconverter . TypeConverterUtils . copyFieldsValues ( intRecord , stringRecordBuilder ) ; org . apache . avro . generic . GenericRecord stringRecord = stringRecordBuilder . build ( ) ; "<AssertPlaceHolder>" ; } get ( int ) { return values [ i ] ; } | org . junit . Assert . assertEquals ( intRecord . get ( "a" ) , stringRecord . get ( "a" ) ) |
SimpleExecute ( ) { java . lang . String driver = "org.h2.Driver" ; java . lang . Class . forName ( driver ) . newInstance ( ) ; java . lang . String url = "jdbc:h2:~/jell" ; java . lang . String username = "sa" ; java . lang . String password = "" ; java . sql . Connection conn = java . sql . DriverManager . getConnection ( url , username , password ) ; com . noelherrick . jell . Jell jell = new com . noelherrick . jell . Jell ( conn ) ; jell . execute ( "create<sp>table<sp>pojos<sp>(number<sp>int<sp>primary<sp>key,<sp>string<sp>varchar(50)<sp>not<sp>null)" ) ; int count = jell . execute ( "insert<sp>into<sp>pojos<sp>values<sp>(@number,<sp>@string)" , new com . noelherrick . jell . SillyPojo ( "a" , 2 ) ) ; "<AssertPlaceHolder>" ; jell . execute ( "drop<sp>table<sp>pojos" ) ; conn . close ( ) ; } execute ( java . lang . String ) { com . noelherrick . jell . Statement statement = connection . createStatement ( ) ; return statement . executeUpdate ( sql ) ; } | org . junit . Assert . assertEquals ( 1 , count ) |
testIntersectVersionRange_Invalid1 ( ) { org . apache . aries . util . VersionRange v1 = new org . apache . aries . util . VersionRange ( "[1.0.0,2.0.0]" ) ; org . apache . aries . util . VersionRange v2 = new org . apache . aries . util . VersionRange ( "(2.0.0,3.0.0]" ) ; org . apache . aries . util . VersionRange result = v1 . intersect ( v2 ) ; "<AssertPlaceHolder>" ; } intersect ( org . apache . aries . util . VersionRange ) { final org . osgi . framework . Version newMinimumVersion ; final boolean newMinimumExclusive ; int minCompare = minimumVersion . compareTo ( r . getMinimumVersion ( ) ) ; if ( minCompare > 0 ) { newMinimumVersion = minimumVersion ; newMinimumExclusive = minimumExclusive ; } else if ( minCompare < 0 ) { newMinimumVersion = r . getMinimumVersion ( ) ; newMinimumExclusive = r . isMinimumExclusive ( ) ; } else { newMinimumVersion = minimumVersion ; newMinimumExclusive = ( minimumExclusive ) || ( r . isMinimumExclusive ( ) ) ; } final org . osgi . framework . Version newMaximumVersion ; final boolean newMaximumExclusive ; if ( ( maximumVersion ) == null ) { newMaximumVersion = r . getMaximumVersion ( ) ; newMaximumExclusive = r . isMaximumExclusive ( ) ; } else if ( ( r . getMaximumVersion ( ) ) == null ) { newMaximumVersion = maximumVersion ; newMaximumExclusive = maximumExclusive ; } else { int maxCompare = maximumVersion . compareTo ( r . getMaximumVersion ( ) ) ; if ( maxCompare < 0 ) { newMaximumVersion = maximumVersion ; newMaximumExclusive = maximumExclusive ; } else if ( maxCompare > 0 ) { newMaximumVersion = r . getMaximumVersion ( ) ; newMaximumExclusive = r . isMaximumExclusive ( ) ; } else { newMaximumVersion = maximumVersion ; newMaximumExclusive = ( maximumExclusive ) || ( r . isMaximumExclusive ( ) ) ; } } org . apache . aries . util . VersionRange result ; if ( isRangeValid ( newMinimumVersion , newMinimumExclusive , newMaximumVersion , newMaximumExclusive ) ) { result = new org . apache . aries . util . VersionRange ( newMaximumVersion , newMaximumExclusive , newMinimumVersion , newMinimumExclusive ) ; } else { result = null ; } return result ; } | org . junit . Assert . assertNull ( result ) |
testBasicKeyboardInput ( ) { driver . get ( pages . javascriptPage ) ; org . openqa . selenium . qtwebkit . WebElement keyReporter = driver . findElement ( org . openqa . selenium . qtwebkit . By . id ( "keyReporter" ) ) ; org . openqa . selenium . interactions . Action sendLowercase = getBuilder ( driver ) . sendKeys ( keyReporter , "abc<sp>def" ) . build ( ) ; sendLowercase . perform ( ) ; "<AssertPlaceHolder>" ; } getAttribute ( java . lang . String ) { return commandProcessor . getString ( "getAttribute" , new java . lang . String [ ] { attributeLocator } ) ; } | org . junit . Assert . assertThat ( keyReporter . getAttribute ( "value" ) , org . hamcrest . Matchers . is ( "abc<sp>def" ) ) |
testAddChildAfter_withNullNewChild ( ) { final com . valkryst . VTerminal . TileGrid gridA = new com . valkryst . VTerminal . TileGrid ( new java . awt . Dimension ( 1 , 1 ) ) ; parentGrid . addChild ( gridA ) ; parentGrid . addChildAfter ( null , gridA ) ; "<AssertPlaceHolder>" ; } containsChild ( com . valkryst . VTerminal . TileGrid ) { if ( child == null ) { return false ; } childLock . readLock ( ) . lock ( ) ; final boolean containsChild = childGrids . contains ( child ) ; childLock . readLock ( ) . unlock ( ) ; return containsChild ; } | org . junit . Assert . assertTrue ( parentGrid . containsChild ( gridA ) ) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.