input stringlengths 28 18.7k | output stringlengths 39 1.69k |
|---|---|
testInnerProduct ( ) { mikera . matrixx . Matrix m = mikera . matrixx . Matrixx . createYAxisRotationMatrix ( 2.0 ) . toMatrix ( ) ; mikera . arrayz . NDArray a = mikera . arrayz . NDArray . wrap ( m ) ; "<AssertPlaceHolder>" ; } innerProduct ( mikera . matrixx . AMatrix ) { if ( ( a . rowCount ( ) ) != ( rows ) ) throw new java . lang . IllegalArgumentException ( mikera . vectorz . util . ErrorMessages . mismatch ( this , a ) ) ; int cc = a . columnCount ( ) ; mikera . matrixx . Matrix result = mikera . matrixx . Matrix . create ( rows , cc ) ; for ( int i = 0 ; i < ( rows ) ; i ++ ) { int dstIndex = i * cc ; int srcRow = perm . get ( i ) ; a . copyRowTo ( srcRow , result . data , dstIndex ) ; } return result ; } | org . junit . Assert . assertEquals ( a . innerProduct ( a ) , m . innerProduct ( m ) ) |
failedJobPartitionRequestTest ( ) { final org . apache . hyracks . control . nc . NodeControllerService nc1 = org . apache . asterix . runtime . PartitionManagerTest . integrationUtil . ncs [ 0 ] ; final org . apache . hyracks . control . nc . NodeControllerService nc2 = org . apache . asterix . runtime . PartitionManagerTest . integrationUtil . ncs [ 1 ] ; final org . apache . hyracks . api . job . JobId failedJob = new org . apache . hyracks . api . job . JobId ( ( - 1 ) ) ; nc2 . getPartitionManager ( ) . jobCompleted ( failedJob , JobStatus . FAILURE ) ; final org . apache . hyracks . api . comm . NetworkAddress localNetworkAddress = nc2 . getNetworkManager ( ) . getPublicNetworkAddress ( ) ; final java . net . InetSocketAddress nc2Address = new java . net . InetSocketAddress ( localNetworkAddress . getAddress ( ) , localNetworkAddress . getPort ( ) ) ; org . apache . hyracks . api . partitions . PartitionId id = new org . apache . hyracks . api . partitions . PartitionId ( failedJob , new org . apache . hyracks . api . dataflow . ConnectorDescriptorId ( 1 ) , 0 , 1 ) ; org . apache . hyracks . comm . channels . NetworkInputChannel inputChannel = new org . apache . hyracks . comm . channels . NetworkInputChannel ( nc1 . getNetworkManager ( ) , nc2Address , id , 1 ) ; org . apache . hyracks . dataflow . std . collectors . InputChannelFrameReader frameReader = new org . apache . hyracks . dataflow . std . collectors . InputChannelFrameReader ( inputChannel ) ; inputChannel . registerMonitor ( frameReader ) ; java . util . concurrent . atomic . AtomicBoolean failed = new java . util . concurrent . atomic . AtomicBoolean ( false ) ; java . lang . Thread reader = new java . lang . Thread ( ( ) -> { try { failed . set ( ( ! ( frameReader . nextFrame ( new org . apache . hyracks . api . comm . FixedSizeFrame ( ) ) ) ) ) ; } catch ( e ) { org . apache . asterix . runtime . e . printStackTrace ( ) ; } } ) ; reader . start ( ) ; final org . apache . hyracks . api . context . IHyracksCommonContext context = org . mockito . Mockito . mock ( org . apache . hyracks . api . context . IHyracksCommonContext . class ) ; org . mockito . Mockito . when ( context . getInitialFrameSize ( ) ) . thenReturn ( 2000 ) ; inputChannel . open ( context ) ; reader . join ( 5000 ) ; "<AssertPlaceHolder>" ; } get ( ) { return counter ; } | org . junit . Assert . assertTrue ( failed . get ( ) ) |
testACLPublic ( ) { org . geowebcache . s3 . S3BlobStoreInfo config = new org . geowebcache . s3 . S3BlobStoreInfo ( ) ; config . setAccess ( Access . PUBLIC ) ; "<AssertPlaceHolder>" ; } getAccessControlList ( ) { com . amazonaws . services . s3 . model . CannedAccessControlList accessControlList ; if ( ( access ) == ( Access . PRIVATE ) ) { accessControlList = com . amazonaws . services . s3 . model . CannedAccessControlList . BucketOwnerFullControl ; } else { accessControlList = com . amazonaws . services . s3 . model . CannedAccessControlList . PublicRead ; } return accessControlList ; } | org . junit . Assert . assertEquals ( CannedAccessControlList . PublicRead , config . getAccessControlList ( ) ) |
testAddSpecificQuery ( ) { java . lang . String query = "select<sp>count(data_id)<sp>from<sp>r_data_main" ; java . lang . String alias = "get_dataobject_ids" ; org . irods . jargon . core . pub . domain . SpecificQueryDefinition specificQuery = new org . irods . jargon . core . pub . domain . SpecificQueryDefinition ( query , alias ) ; org . irods . jargon . core . packinstr . GeneralAdminInpForSQ pi = org . irods . jargon . core . packinstr . GeneralAdminInpForSQ . instanceForAddSpecificQuery ( specificQuery ) ; "<AssertPlaceHolder>" ; } instanceForAddSpecificQuery ( org . irods . jargon . core . pub . domain . SpecificQueryDefinition ) { if ( specificQuery == null ) { throw new java . lang . IllegalArgumentException ( "null<sp>SpecificQueryDefinition<sp>object" ) ; } java . lang . String query = specificQuery . getSql ( ) ; java . lang . String alias = specificQuery . getAlias ( ) ; org . irods . jargon . core . packinstr . GeneralAdminInpForSQ generalAdminInpForSQ = new org . irods . jargon . core . packinstr . GeneralAdminInpForSQ ( query , alias , org . irods . jargon . core . packinstr . GeneralAdminInpForSQ . SQ_ADD ) ; return generalAdminInpForSQ ; } | org . junit . Assert . assertNotNull ( pi ) |
testEmptyContainer ( ) { org . openscience . cdk . interfaces . IAtomContainer emptyContainer = super . builder . newInstance ( org . openscience . cdk . interfaces . IAtomContainer . class ) ; org . openscience . cdk . renderer . elements . IRenderingElement root = generator . generate ( emptyContainer , model ) ; java . util . List < org . openscience . cdk . renderer . elements . IRenderingElement > elements = elementUtil . getAllSimpleElements ( root ) ; "<AssertPlaceHolder>" ; } size ( ) { return this . cells . size ( ) ; } | org . junit . Assert . assertEquals ( 0 , elements . size ( ) ) |
testGetProducts ( ) { com . welflex . order . dto . ProductListDto products = com . welflex . order . IntegrationTest . PRODUCT_CLIENT . getProducts ( ) ; "<AssertPlaceHolder>" ; products = com . welflex . order . IntegrationTest . PRODUCT_CLIENT . getProducts ( ) ; System . out . println ( "Sleeping<sp>for<sp>5<sp>seconds<sp>so<sp>that<sp>product<sp>list<sp>expires..." ) ; java . lang . Thread . sleep ( 5000 ) ; products = com . welflex . order . IntegrationTest . PRODUCT_CLIENT . getProducts ( ) ; } getProducts ( ) { java . lang . String path = resource . getURI ( ) . getPath ( ) ; com . welflex . client . CacheEntry entry = com . welflex . client . CacheManager . get ( path ) ; com . welflex . order . dto . ProductListDto productList = null ; if ( entry != null ) { com . welflex . client . ProductClientImpl . LOG . debug ( "Product<sp>Entry<sp>in<sp>cache<sp>is<sp>not<sp>null...checking<sp>expiration<sp>date.." ) ; java . util . Date cacheTillDate = entry . getCacheTillDate ( ) ; java . util . Date now = new java . util . Date ( ) ; if ( now . before ( cacheTillDate ) ) { com . welflex . client . ProductClientImpl . LOG . debug ( "Product<sp>List<sp>is<sp>not<sp>stale..using<sp>cached<sp>value" ) ; productList = ( ( com . welflex . order . dto . ProductListDto ) ( entry . getObject ( ) ) ) ; } else { com . welflex . client . ProductClientImpl . LOG . debug ( "Product<sp>List<sp>is<sp>stale..will<sp>request<sp>server<sp>for<sp>new<sp>Product<sp>List.." ) ; } } if ( productList == null ) { com . welflex . client . ProductClientImpl . LOG . debug ( "Fetching<sp>Product<sp>List<sp>from<sp>Service..." ) ; com . sun . jersey . api . client . ClientResponse response = resource . accept ( MediaType . APPLICATION_JSON ) . get ( com . sun . jersey . api . client . ClientResponse . class ) ; if ( response . getResponseStatus ( ) . equals ( Status . OK ) ) { productList = response . getEntity ( com . welflex . order . dto . ProductListDto . class ) ; java . lang . String cacheDate = response . getMetadata ( ) . getFirst ( "Expires" ) ; if ( cacheDate != null ) { java . util . Date ccDate ; try { ccDate = com . welflex . client . ProductClientImpl . DATE_FORMAT . parse ( cacheDate ) ; entry = new com . welflex . client . CacheEntry ( productList , null , ccDate ) ; com . welflex . client . CacheManager . cache ( path , entry ) ; } catch ( java . text . ParseException e ) { com . welflex . client . ProductClientImpl . LOG . error ( "Error<sp>Parsing<sp>returned<sp>cache<sp>date..no<sp>caching<sp>will<sp>occur" , e ) ; } } } else { throw new java . lang . RuntimeException ( "Error<sp>Getting<sp>Products...." ) ; } } return productList ; } | org . junit . Assert . assertFalse ( ( ( products . getProducts ( ) . size ( ) ) == 0 ) ) |
testOperator_lessEqualsThanNullNull ( ) { java . lang . Number x = null ; java . lang . Number y = null ; boolean resutl = org . eclipse . smarthome . model . script . lib . NumberExtensions . operator_lessEqualsThan ( x , y ) ; "<AssertPlaceHolder>" ; } operator_lessEqualsThan ( java . lang . Number , java . lang . Number ) { java . math . BigDecimal leftValue = org . eclipse . smarthome . model . script . lib . NumberExtensions . numberToBigDecimal ( left ) ; java . math . BigDecimal rightValue = org . eclipse . smarthome . model . script . lib . NumberExtensions . numberToBigDecimal ( right ) ; if ( leftValue == null ) { return true ; } else if ( rightValue == null ) { return false ; } else { return ( leftValue . compareTo ( rightValue ) ) <= 0 ; } } | org . junit . Assert . assertTrue ( resutl ) |
testMergeDefaultLocale ( ) { this . previousDocument . setDefaultLocale ( Locale . ENGLISH ) ; this . nextDocument . setDefaultLocale ( Locale . FRENCH ) ; this . currentDocument . setDefaultLocale ( Locale . ENGLISH ) ; merge ( ) ; "<AssertPlaceHolder>" ; } getDefaultLocale ( ) { return ( this . defaultLocale ) != null ? this . defaultLocale : java . util . Locale . ROOT ; } | org . junit . Assert . assertEquals ( Locale . FRENCH , this . currentDocument . getDefaultLocale ( ) ) |
test_parse_simple_empty_array ( ) { java . lang . Object [ ] objs = org . nutz . json . Json . fromJson ( java . lang . Object [ ] . class , "[]" ) ; "<AssertPlaceHolder>" ; } fromJson ( java . lang . Class , java . io . Reader ) { return ( ( T ) ( org . nutz . json . Json . parse ( type , reader ) ) ) ; } | org . junit . Assert . assertEquals ( 0 , objs . length ) |
getInsertableColumnsInsertableFalseWithId ( ) { @ com . zaxxer . sansorm . internal . Table ( name = "Test_Class" ) class TestClass { @ com . zaxxer . sansorm . internal . Id @ com . zaxxer . sansorm . internal . Column ( insertable = false ) java . lang . String id ; @ com . zaxxer . sansorm . internal . Column ( name = "\"Delimited<sp>Field<sp>Name\"" , insertable = false ) java . lang . String delimitedFieldName ; @ com . zaxxer . sansorm . internal . Column ( name = "Default_Case" , insertable = false ) java . lang . String defaultCase ; } com . zaxxer . sansorm . internal . Introspected introspected = com . zaxxer . sansorm . internal . Introspector . getIntrospected ( TestClass . class ) ; java . lang . String [ ] cols = introspected . getInsertableColumns ( ) ; "<AssertPlaceHolder>" ; } getInsertableColumns ( ) { return insertableColumns ; } | org . junit . Assert . assertArrayEquals ( new java . lang . String [ ] { } , cols ) |
testDeleteMessageCollection ( ) { sut . deleteMessageCollection ( testId , org . eurekastreams . server . service . opensocial . spi . MessageServiceTest . TEST_MESSAGE_COLL_ID , org . eurekastreams . server . service . opensocial . spi . MessageServiceTest . FAKETOKEN ) ; "<AssertPlaceHolder>" ; } deleteMessageCollection ( org . apache . shindig . social . opensocial . spi . UserId , java . lang . String , org . apache . shindig . auth . SecurityToken ) { return null ; } | org . junit . Assert . assertTrue ( true ) |
testProcessor ( ) { com . manning . sbia . ch14 . domain . Product p1 = new com . manning . sbia . ch14 . domain . Product ( ) ; p1 . setPrice ( new java . math . BigDecimal ( "100.0" ) ) ; com . manning . sbia . ch14 . domain . Product p2 = processor . process ( p1 ) ; "<AssertPlaceHolder>" ; } process ( com . manning . sbia . ch06 . mail . Customer ) { org . springframework . mail . SimpleMailMessage msg = new org . springframework . mail . SimpleMailMessage ( ) ; msg . setFrom ( "springbatchinaction@test.com" ) ; msg . setTo ( item . getEmail ( ) ) ; msg . setSubject ( "Welcome<sp>message<sp>!!" ) ; msg . setText ( ( "Hello<sp>" + ( item . getName ( ) ) ) ) ; return msg ; } | org . junit . Assert . assertNotNull ( p2 ) |
testRegionOpenProcedureIsNotHandledByDispatcher ( ) { org . apache . hadoop . hbase . TableName tableName = org . apache . hadoop . hbase . TableName . valueOf ( "testRegionOpenProcedureIsNotHandledByDisPatcher" ) ; org . apache . hadoop . hbase . client . RegionInfo hri = org . apache . hadoop . hbase . client . RegionInfoBuilder . newBuilder ( tableName ) . setStartKey ( org . apache . hadoop . hbase . util . Bytes . toBytes ( 1 ) ) . setEndKey ( org . apache . hadoop . hbase . util . Bytes . toBytes ( 2 ) ) . setSplit ( false ) . setRegionId ( 0 ) . build ( ) ; org . apache . hadoop . hbase . master . procedure . MasterProcedureEnv env = master . getMasterProcedureExecutor ( ) . getEnvironment ( ) ; env . getAssignmentManager ( ) . getRegionStates ( ) . getOrCreateRegionStateNode ( hri ) ; org . apache . hadoop . hbase . master . assignment . TransitRegionStateProcedure proc = org . apache . hadoop . hbase . master . assignment . TransitRegionStateProcedure . assign ( env , hri , null ) ; org . apache . hadoop . hbase . ServerName worker = master . getServerManager ( ) . getOnlineServersList ( ) . get ( 0 ) ; org . apache . hadoop . hbase . master . assignment . OpenRegionProcedure openRegionProcedure = new org . apache . hadoop . hbase . master . assignment . OpenRegionProcedure ( proc , hri , worker ) ; java . util . concurrent . Future < byte [ ] > future = submitProcedure ( openRegionProcedure ) ; java . lang . Thread . sleep ( 2000 ) ; rsDispatcher . removeNode ( worker ) ; try { future . get ( 2000 , TimeUnit . MILLISECONDS ) ; org . junit . Assert . fail ( ) ; } catch ( java . util . concurrent . TimeoutException e ) { org . apache . hadoop . hbase . master . procedure . TestServerRemoteProcedure . LOG . info ( "timeout<sp>is<sp>expected" ) ; } "<AssertPlaceHolder>" ; } isFinished ( ) { throw new java . lang . UnsupportedOperationException ( "Asynchonous<sp>operation<sp>is<sp>not<sp>supported." ) ; } | org . junit . Assert . assertFalse ( openRegionProcedure . isFinished ( ) ) |
deleteNonExistentSubEquipment ( ) { cern . c2mon . shared . client . configuration . api . equipment . SubEquipment subEquipment = buildDeleteSubEquipment ( 10L ) ; java . util . List < cern . c2mon . shared . client . configuration . api . equipment . SubEquipment > equipmentRemoveList = java . util . Arrays . asList ( subEquipment ) ; cern . c2mon . shared . client . configuration . api . Configuration config = new cern . c2mon . shared . client . configuration . api . Configuration ( 1L ) ; config . setEntities ( equipmentRemoveList ) ; org . easymock . EasyMock . expect ( subEquipmentCache . hasKey ( 10L ) ) . andReturn ( false ) ; org . easymock . EasyMock . replay ( subEquipmentCache ) ; "<AssertPlaceHolder>" ; org . easymock . EasyMock . verify ( subEquipmentCache ) ; } parse ( cern . c2mon . shared . client . configuration . api . Configuration ) { if ( ( ( configuration . getEntities ( ) ) != null ) && ( ! ( configuration . getEntities ( ) . isEmpty ( ) ) ) ) { return parseConfigurationList ( configuration . getEntities ( ) ) ; } else { throw new cern . c2mon . server . configuration . parser . exception . ConfigurationParseException ( "Empty<sp>configuration<sp>received!" ) ; } } | org . junit . Assert . assertEquals ( 0 , parser . parse ( config ) . size ( ) ) |
expectedOutputShouldBeUsed ( ) { org . mule . runtime . core . api . event . CoreEvent jsonMessage = eventBuilder ( muleContext ) . message ( org . mule . runtime . api . message . Message . builder ( ) . value ( "{\"student\":<sp>false}" ) . mediaType ( org . mule . runtime . core . internal . el . dataweave . APPLICATION_JSON ) . build ( ) ) . build ( ) ; org . mule . runtime . api . metadata . TypedValue result = expressionLanguage . evaluate ( "payload.student" , org . mule . runtime . core . internal . el . dataweave . BOOLEAN , jsonMessage , org . mule . runtime . api . el . BindingContext . builder ( ) . build ( ) ) ; "<AssertPlaceHolder>" ; } getValue ( ) { return value ; } | org . junit . Assert . assertThat ( result . getValue ( ) , org . hamcrest . Matchers . is ( false ) ) |
testEvictionListenerThatThrowsReadWriteLocked ( ) { org . terracotta . offheapstore . eviction . EvictionListenerIT . ThrowingEvictionListener listener = new org . terracotta . offheapstore . eviction . EvictionListenerIT . ThrowingEvictionListener ( ) ; org . terracotta . offheapstore . paging . PageSource source = new org . terracotta . offheapstore . paging . UpfrontAllocatingPageSource ( new org . terracotta . offheapstore . buffersource . OffHeapBufferSource ( ) , ( 16 * 4096 ) , 2048 ) ; java . util . Map < java . lang . Long , java . lang . String > map = new org . terracotta . offheapstore . concurrent . ConcurrentOffHeapClockCache ( source , org . terracotta . offheapstore . storage . LongStorageEngine . createFactory ( org . terracotta . offheapstore . storage . OffHeapBufferHalfStorageEngine . createFactory ( source , 128 , StringPortability . INSTANCE ) ) , listener ) ; for ( long i = 0 ; i < 2000L ; i ++ ) { try { map . put ( i , java . lang . Long . toString ( i ) ) ; } catch ( java . lang . NullPointerException e ) { } } java . util . Set < java . lang . Long > evictedKeys = listener . evictedKeys ( ) ; for ( java . lang . Long l : evictedKeys ) { "<AssertPlaceHolder>" ; } } containsKey ( java . lang . Object ) { return segmentFor ( key ) . containsKey ( key ) ; } | org . junit . Assert . assertFalse ( map . containsKey ( l ) ) |
evaluateRandomForestOzone ( ) { org . jpmml . rattle . Batch batch = new org . jpmml . rattle . RattleBatch ( "RandomForest" , "Ozone" ) ; "<AssertPlaceHolder>" ; } evaluate ( org . jpmml . evaluator . Batch ) { org . jpmml . evaluator . PMML pmml = org . jpmml . evaluator . IOUtil . unmarshal ( batch . getModel ( ) ) ; org . jpmml . evaluator . PMMLManager pmmlManager = new org . jpmml . evaluator . PMMLManager ( pmml ) ; org . jpmml . evaluator . ModelManager < ? > modelManager = pmmlManager . getModelManager ( null , org . jpmml . evaluator . ModelEvaluatorFactory . getInstance ( ) ) ; org . jpmml . evaluator . List < org . jpmml . evaluator . Map < org . jpmml . evaluator . FieldName , java . lang . String > > input = org . jpmml . evaluator . CsvUtil . load ( batch . getInput ( ) ) ; org . jpmml . evaluator . List < org . jpmml . evaluator . Map < org . jpmml . evaluator . FieldName , java . lang . String > > output = org . jpmml . evaluator . CsvUtil . load ( batch . getOutput ( ) ) ; org . jpmml . evaluator . Evaluator evaluator = ( ( org . jpmml . evaluator . Evaluator ) ( modelManager ) ) ; org . jpmml . evaluator . List < org . jpmml . evaluator . Map < org . jpmml . evaluator . FieldName , java . lang . Object > > table = org . jpmml . evaluator . Lists . newArrayList ( ) ; org . jpmml . evaluator . List < org . jpmml . evaluator . FieldName > activeFields = evaluator . getActiveFields ( ) ; org . jpmml . evaluator . List < org . jpmml . evaluator . FieldName > groupFields = evaluator . getGroupFields ( ) ; org . jpmml . evaluator . List < org . jpmml . evaluator . FieldName > predictedFields = evaluator . getPredictedFields ( ) ; org . jpmml . evaluator . List < org . jpmml . evaluator . FieldName > outputFields = evaluator . getOutputFields ( ) ; org . jpmml . evaluator . List < org . jpmml . evaluator . FieldName > inputFields = org . jpmml . evaluator . Lists . newArrayList ( ) ; inputFields . addAll ( activeFields ) ; inputFields . addAll ( groupFields ) ; for ( int i = 0 ; i < ( input . size ( ) ) ; i ++ ) { org . jpmml . evaluator . Map < org . jpmml . evaluator . FieldName , java . lang . String > inputRow = input . get ( i ) ; org . jpmml . evaluator . Map < org . jpmml . evaluator . FieldName , java . lang . Object > arguments = org . jpmml . evaluator . Maps . newLinkedHashMap ( ) ; for ( org . jpmml . evaluator . FieldName inputField : inputFields ) { java . lang . String inputCell = inputRow . get ( inputField ) ; java . lang . Object inputValue = evaluator . prepare ( inputField , inputCell ) ; arguments . put ( inputField , inputValue ) ; } table . add ( arguments ) ; } if ( ( groupFields . size ( ) ) == 1 ) { org . jpmml . evaluator . FieldName groupField = groupFields . get ( 0 ) ; table = org . jpmml . evaluator . EvaluatorUtil . groupRows ( groupField , table ) ; } else if ( ( groupFields . size ( ) ) > 1 ) { throw new org . jpmml . evaluator . EvaluationException ( ) ; } if ( output . isEmpty ( ) ) { for ( int i = 0 ; i < ( table . size ( ) ) ; i ++ ) { org . jpmml . evaluator . Map < org . jpmml . evaluator . FieldName , ? > arguments = table . get ( i ) ; evaluator . evaluate ( arguments ) ; } return true ; } else { if ( ( table . size ( ) ) != ( output . size ( ) ) ) { throw new org . jpmml . evaluator . EvaluationException ( ) ; } boolean success = true ; for ( int i = 0 ; i < ( output . size ( ) ) ; i ++ ) { org . jpmml . evaluator . Map < org . jpmml . evaluator . FieldName , java . lang . String > outputRow = output . get ( i ) ; org . jpmml . evaluator . Map < org . jpmml . evaluator . FieldName , ? > arguments = table . get ( i ) ; org . jpmml . evaluator . Map < org . jpmml . evaluator . FieldName , ? > result = evaluator . evaluate ( arguments ) ; for ( org . jpmml . evaluator . FieldName predictedField : predictedFields ) { java . lang . String outputCell = outputRow . get ( predictedField ) ; java . lang . Object predictedValue = org . jpmml . evaluator . EvaluatorUtil . decode ( result . get ( predictedField ) ) ; success &= org . jpmml . evaluator . BatchUtil . acceptable ( outputCell , predictedValue ) ; } for ( org . jpmml . evaluator . FieldName outputField : outputFields ) { java . lang . String outputCell = outputRow . get ( outputField ) ; java . lang . Object computedValue = result . get ( outputField ) ; success &= ( outputCell != null ) ? org . jpmml . evaluator . BatchUtil . acceptable ( outputCell , computedValue | org . junit . Assert . assertTrue ( org . jpmml . rattle . BatchUtil . evaluate ( batch ) ) |
testIsConfigChangedNetConfigsDiffer ( ) { org . eclipse . kura . net . admin . monitor . EthernetMonitorServiceImpl svc = new org . eclipse . kura . net . admin . monitor . EthernetMonitorServiceImpl ( ) ; java . util . List < org . eclipse . kura . net . NetInterfaceAddressConfig > newConfig = new java . util . ArrayList ( ) ; java . util . List < org . eclipse . kura . net . NetInterfaceAddressConfig > currentConfig = new java . util . ArrayList ( ) ; org . eclipse . kura . core . net . NetInterfaceAddressConfigImpl nic = new org . eclipse . kura . core . net . NetInterfaceAddressConfigImpl ( ) ; java . util . ArrayList < org . eclipse . kura . net . NetConfig > newNets = new java . util . ArrayList ( ) ; org . eclipse . kura . net . NetConfig nc = new org . eclipse . kura . net . firewall . FirewallAutoNatConfig ( ) ; newNets . add ( nc ) ; nc = mock ( org . eclipse . kura . net . dhcp . DhcpServerConfigIP4 . class ) ; newNets . add ( nc ) ; nc = new org . eclipse . kura . net . NetConfigIP4 ( org . eclipse . kura . net . NetInterfaceStatus . netIPv4StatusEnabledLAN , true ) ; newNets . add ( nc ) ; nic . setNetConfigs ( newNets ) ; newConfig . add ( nic ) ; nic = new org . eclipse . kura . core . net . NetInterfaceAddressConfigImpl ( ) ; java . util . ArrayList < org . eclipse . kura . net . NetConfig > currentNets = new java . util . ArrayList ( ) ; nc = new org . eclipse . kura . net . firewall . FirewallAutoNatConfig ( ) ; currentNets . add ( nc ) ; nc = mock ( org . eclipse . kura . net . dhcp . DhcpServerConfigIP4 . class ) ; currentNets . add ( nc ) ; nc = new org . eclipse . kura . net . NetConfigIP4 ( org . eclipse . kura . net . NetInterfaceStatus . netIPv4StatusEnabledWAN , true ) ; currentNets . add ( nc ) ; nic . setNetConfigs ( currentNets ) ; currentConfig . add ( nic ) ; boolean result = ( ( boolean ) ( org . eclipse . kura . core . testutil . TestUtil . invokePrivate ( svc , "isConfigChanged" , newConfig , currentConfig ) ) ) ; "<AssertPlaceHolder>" ; } invokePrivate ( java . lang . Object , java . lang . String , org . eclipse . kura . core . testutil . Class [ ] , java . lang . Object [ ] ) { java . lang . reflect . Method method = org . eclipse . kura . core . testutil . TestUtil . getMethod ( svc , methodName , paramTypes ) ; method . setAccessible ( true ) ; try { java . lang . Object result = method . invoke ( svc , params ) ; return result ; } catch ( java . lang . IllegalAccessException e ) { org . eclipse . kura . core . testutil . TestUtil . logger . warn ( e . getMessage ( ) , e ) ; } catch ( java . lang . IllegalArgumentException e ) { org . eclipse . kura . core . testutil . TestUtil . logger . warn ( e . getMessage ( ) , e ) ; } catch ( java . lang . reflect . InvocationTargetException e ) { throw e . getCause ( ) ; } return null ; } | org . junit . Assert . assertTrue ( result ) |
testDatasetWithoutAnId ( ) { dataset = new edu . harvard . iq . dataverse . Dataset ( ) ; edu . harvard . iq . dataverse . privateurl . PrivateUrl privateUrl = testEngine . submit ( new edu . harvard . iq . dataverse . engine . command . impl . GetPrivateUrlCommand ( null , dataset ) ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertNull ( privateUrl ) |
testCheckTimeoutCommitting ( ) { final java . lang . Exception [ ] checkTimeoutException = new com . sun . sgs . test . impl . service . transaction . Exception [ ] { null } ; com . sun . sgs . test . util . DummyTransactionParticipant participant = new com . sun . sgs . test . util . DummyTransactionParticipant ( ) { public void commit ( com . sun . sgs . service . Transaction txn ) { try { txn . checkTimeout ( ) ; } catch ( java . lang . RuntimeException e ) { checkTimeoutException [ 0 ] = e ; throw e ; } } } ; txn . join ( participant ) ; handle . commit ( ) ; "<AssertPlaceHolder>" ; } commit ( ) { isCommitted = true ; checkFlush ( ) ; } | org . junit . Assert . assertNull ( checkTimeoutException [ 0 ] ) |
readErrorMessage ( ) { long session = 1L ; int code = - 200 ; java . lang . String message = "Failed!" ; cocaine . message . Message msg = cocaine . message . Messages . error ( session , code , message ) ; byte [ ] bytes = pack . write ( java . util . Arrays . asList ( 5 , session , java . util . Arrays . asList ( code , message ) ) ) ; cocaine . message . Message result = pack . read ( bytes , cocaine . msgpack . MessageTemplate . getInstance ( ) ) ; "<AssertPlaceHolder>" ; } getInstance ( ) { return cocaine . msgpack . MessageTemplate . instance ; } | org . junit . Assert . assertEquals ( msg , result ) |
testAppendNullStringArray ( ) { com . liferay . petra . string . StringBundler sb = new com . liferay . petra . string . StringBundler ( ) ; sb . append ( ( ( java . lang . String [ ] ) ( null ) ) ) ; "<AssertPlaceHolder>" ; } index ( ) { for ( Map . Entry < java . lang . String , java . util . Set < java . lang . Long > > primaryKeysEntry : _primaryKeys . entrySet ( ) ) { java . lang . String className = primaryKeysEntry . getKey ( ) ; com . liferay . portal . kernel . search . Indexer < ? > indexer = indexerRegistry . getIndexer ( className ) ; if ( indexer == null ) { if ( com . liferay . exportimport . resources . importer . internal . util . FileSystemImporter . _log . isWarnEnabled ( ) ) { com . liferay . exportimport . resources . importer . internal . util . FileSystemImporter . _log . warn ( ( "No<sp>indexer<sp>for<sp>" + className ) ) ; } continue ; } if ( com . liferay . exportimport . resources . importer . internal . util . FileSystemImporter . _log . isDebugEnabled ( ) ) { com . liferay . exportimport . resources . importer . internal . util . FileSystemImporter . _log . debug ( ( "Indexing<sp>" + className ) ) ; } for ( long primaryKey : primaryKeysEntry . getValue ( ) ) { try { indexer . reindex ( className , primaryKey ) ; } catch ( com . liferay . portal . kernel . search . SearchException se ) { if ( com . liferay . exportimport . resources . importer . internal . util . FileSystemImporter . _log . isWarnEnabled ( ) ) { com . liferay . exportimport . resources . importer . internal . util . FileSystemImporter . _log . warn ( com . liferay . petra . string . StringBundler . concat ( "Unable<sp>to<sp>index<sp>entry<sp>for<sp>class<sp>name<sp>" , className , "<sp>and<sp>primary<sp>key<sp>" , primaryKey ) , se ) ; } } } } if ( _ddmStructureKeys . isEmpty ( ) ) { return ; } java . util . Set < java . lang . Long > primaryKeys = _primaryKeys . get ( com . liferay . journal . model . JournalArticle . class . getName ( ) ) ; com . liferay . portal . kernel . search . Indexer < ? > indexer = indexerRegistry . getIndexer ( com . liferay . journal . model . JournalArticle . class . getName ( ) ) ; for ( java . lang . String ddmStructureKey : _ddmStructureKeys ) { java . util . List < com . liferay . journal . model . JournalArticle > journalArticles = journalArticleLocalService . getArticlesByStructureId ( getGroupId ( ) , ddmStructureKey , QueryUtil . ALL_POS , QueryUtil . ALL_POS , null ) ; for ( com . liferay . journal . model . JournalArticle journalArticle : journalArticles ) { if ( ( primaryKeys != null ) && ( primaryKeys . contains ( journalArticle . getPrimaryKey ( ) ) ) ) { continue ; } try { indexer . reindex ( com . liferay . journal . model . JournalArticle . class . getName ( ) , journalArticle . getPrimaryKey ( ) ) ; } catch ( com . liferay . portal . kernel . search . SearchException se ) { if ( com . liferay . exportimport . resources . importer . internal . util . FileSystemImporter . _log . isWarnEnabled ( ) ) { com . liferay . petra . string . StringBundler sb = new com . liferay . petra . string . StringBundler ( 4 ) ; sb . append ( "Cannot<sp>index<sp>entry:<sp>className=" ) ; sb . append ( com . liferay . journal . model . JournalArticle . class . getName ( ) ) ; sb . append ( ",<sp>primaryKey=" ) ; sb . append ( journalArticle . getPrimaryKey ( ) ) ; com . liferay . exportimport . resources . importer . internal . util . FileSystemImporter . _log . warn ( sb . toString ( ) , se ) ; } } } } } | org . junit . Assert . assertEquals ( 0 , sb . index ( ) ) |
testReadLong ( ) { java . nio . ByteBuffer byteBuffer = java . nio . ByteBuffer . allocate ( ( ( com . liferay . petra . io . DeserializerTest . _COUNT ) * 8 ) ) ; byteBuffer . order ( ByteOrder . BIG_ENDIAN ) ; java . nio . LongBuffer longBuffer = byteBuffer . asLongBuffer ( ) ; long [ ] longs = new long [ com . liferay . petra . io . DeserializerTest . _COUNT ] ; for ( int i = 0 ; i < ( com . liferay . petra . io . DeserializerTest . _COUNT ) ; i ++ ) { longs [ i ] = _random . nextLong ( ) ; longBuffer . put ( longs [ i ] ) ; } com . liferay . petra . io . Deserializer deserializer = new com . liferay . petra . io . Deserializer ( byteBuffer ) ; for ( int i = 0 ; i < ( com . liferay . petra . io . DeserializerTest . _COUNT ) ; i ++ ) { "<AssertPlaceHolder>" ; } } readLong ( ) { _detectBufferUnderflow ( 8 ) ; long l = com . liferay . petra . io . BigEndianCodec . getLong ( _buffer , _index ) ; _index += 8 ; return l ; } | org . junit . Assert . assertEquals ( longs [ i ] , deserializer . readLong ( ) ) |
filter_items_in_list_with_apache_commons ( ) { java . util . List < com . levelup . java . collections . FilterACollection . NFLTeam > nflTeams = com . google . common . collect . Lists . newArrayList ( ) ; nflTeams . add ( new com . levelup . java . collections . FilterACollection . NFLTeam ( "Green<sp>Bay<sp>Packers" , true ) ) ; nflTeams . add ( new com . levelup . java . collections . FilterACollection . NFLTeam ( "Chicago<sp>Bears" , true ) ) ; nflTeams . add ( new com . levelup . java . collections . FilterACollection . NFLTeam ( "Detroit<sp>Lions" , false ) ) ; org . apache . commons . collections . CollectionUtils . filter ( nflTeams , new org . apache . commons . collections . Predicate ( ) { public boolean evaluate ( java . lang . Object nflTeam ) { return ( ( com . levelup . java . collections . FilterACollection . NFLTeam ) ( nflTeam ) ) . hasWonSuperBowl ; } } ) ; com . levelup . java . collections . FilterACollection . logger . info ( nflTeams ) ; "<AssertPlaceHolder>" ; } evaluate ( java . lang . Object ) { com . levelup . java . collections . ListContainsAll . Camera camera = ( ( com . levelup . java . collections . ListContainsAll . Camera ) ( object ) ) ; return ( camera . focalLength ) >= 80 ; } | org . junit . Assert . assertTrue ( ( ( nflTeams . size ( ) ) == 2 ) ) |
testQueryIterableWithTransformer ( ) { createDocs ( ) ; org . nuxeo . ecm . core . api . IterableQueryResult res ; res = session . queryAndFetch ( "SELECT<sp>*<sp>FROM<sp>Document<sp>WHERE<sp>ecm:isVersion<sp>=<sp>0" , "NXQL" ) ; "<AssertPlaceHolder>" ; res . close ( ) ; } size ( ) { return getCollectedDocumentIds ( ) . size ( ) ; } | org . junit . Assert . assertEquals ( 4 , res . size ( ) ) |
validate_shouldPassValidationIfAutoExpireDateAndDurationUnitsAreNotSet ( ) { org . openmrs . DrugOrder drugOrder = createValidDrugOrder ( ) ; drugOrder . setDurationUnits ( null ) ; drugOrder . setAutoExpireDate ( null ) ; org . springframework . validation . Errors errors = new org . springframework . validation . BindException ( drugOrder , "drugOrder" ) ; new org . openmrs . SimpleDosingInstructions ( ) . validate ( drugOrder , errors ) ; "<AssertPlaceHolder>" ; } hasErrors ( ) { return erroneous ; } | org . junit . Assert . assertFalse ( errors . hasErrors ( ) ) |
shouldGetAllUsers_Annotated ( ) { sqlSessionFactory . getConfiguration ( ) . addMapper ( org . apache . ibatis . submitted . duplicate_statements . AnnotatedMapper . class ) ; org . apache . ibatis . session . SqlSession sqlSession = sqlSessionFactory . openSession ( ) ; try { org . apache . ibatis . submitted . duplicate_statements . AnnotatedMapper mapper = sqlSession . getMapper ( org . apache . ibatis . submitted . duplicate_statements . AnnotatedMapper . class ) ; java . util . List < org . apache . ibatis . submitted . duplicate_statements . User > users = mapper . getAllUsers ( ) ; "<AssertPlaceHolder>" ; } finally { sqlSession . close ( ) ; } } size ( ) { return loaderMap . size ( ) ; } | org . junit . Assert . assertEquals ( 10 , users . size ( ) ) |
testBuildFolderPathRootServicePathNoEncoding ( ) { System . out . println ( ( ( ( getTestTraceHead ( "[NGSIHDFSSinkTest.buildFolderPath]" ) ) + "--------<sp>When<sp>no<sp>encoding<sp>and<sp>when<sp>a<sp>root<sp>service-path<sp>is<sp>notified/defaulted<sp>the<sp>HDFS<sp>folder<sp>path<sp>" ) + "--------<sp>When<sp>no<sp>encoding<sp>and<sp>when<sp>a<sp>root<sp>service-path<sp>is<sp>notified/defaulted<sp>the<sp>HDFS<sp>folder<sp>path<sp>" 1 ) ) ; java . lang . String backendImpl = null ; java . lang . String backendMaxConns = null ; java . lang . String backendMaxConnsPerRoute = null ; java . lang . String batchSize = null ; java . lang . String batchTime = null ; java . lang . String batchTTL = null ; java . lang . String csvSeparator = null ; java . lang . String dataModel = null ; java . lang . String enableEncoding = "--------<sp>When<sp>no<sp>encoding<sp>and<sp>when<sp>a<sp>root<sp>service-path<sp>is<sp>notified/defaulted<sp>the<sp>HDFS<sp>folder<sp>path<sp>" 6 ; java . lang . String enableGrouping = null ; java . lang . String enableLowercase = null ; java . lang . String fileFormat = null ; java . lang . String host = null ; java . lang . String password = "mypassword" ; java . lang . String port = null ; java . lang . String username = "--------<sp>When<sp>no<sp>encoding<sp>and<sp>when<sp>a<sp>root<sp>service-path<sp>is<sp>notified/defaulted<sp>the<sp>HDFS<sp>folder<sp>path<sp>" 2 ; java . lang . String hive = "--------<sp>When<sp>no<sp>encoding<sp>and<sp>when<sp>a<sp>root<sp>service-path<sp>is<sp>notified/defaulted<sp>the<sp>HDFS<sp>folder<sp>path<sp>" 6 ; java . lang . String krb5 = "--------<sp>When<sp>no<sp>encoding<sp>and<sp>when<sp>a<sp>root<sp>service-path<sp>is<sp>notified/defaulted<sp>the<sp>HDFS<sp>folder<sp>path<sp>" 6 ; java . lang . String token = "--------<sp>When<sp>no<sp>encoding<sp>and<sp>when<sp>a<sp>root<sp>service-path<sp>is<sp>notified/defaulted<sp>the<sp>HDFS<sp>folder<sp>path<sp>" 3 ; java . lang . String serviceAsNamespace = null ; com . telefonica . iot . cygnus . sinks . NGSIHDFSSink sink = new com . telefonica . iot . cygnus . sinks . NGSIHDFSSink ( ) ; sink . configure ( createContext ( backendImpl , backendMaxConns , backendMaxConnsPerRoute , batchSize , batchTime , batchTTL , csvSeparator , dataModel , enableEncoding , enableGrouping , enableLowercase , fileFormat , host , password , port , username , hive , krb5 , token , serviceAsNamespace ) ) ; java . lang . String service = "someService" ; java . lang . String servicePath = "--------<sp>When<sp>no<sp>encoding<sp>and<sp>when<sp>a<sp>root<sp>service-path<sp>is<sp>notified/defaulted<sp>the<sp>HDFS<sp>folder<sp>path<sp>" 5 ; java . lang . String entity = "someId=someType" ; try { java . lang . String builtTableName = sink . buildFolderPath ( service , servicePath , entity ) ; java . lang . String expecetedTableName = "someService/someId_someType" ; try { "<AssertPlaceHolder>" ; System . out . println ( ( ( ( ( ( getTestTraceHead ( "[NGSIHDFSSinkTest.buildFolderPath]" ) ) + "-<sp>OK<sp>-<sp>'" ) + builtTableName ) + "'<sp>is<sp>equals<sp>to<sp>the<sp>encoding<sp>of<sp>" ) + "--------<sp>When<sp>no<sp>encoding<sp>and<sp>when<sp>a<sp>root<sp>service-path<sp>is<sp>notified/defaulted<sp>the<sp>HDFS<sp>folder<sp>path<sp>" 4 ) ) ; } catch ( java . lang . AssertionError e ) { System . out . println ( ( ( ( ( ( getTestTraceHead ( "[NGSIHDFSSinkTest.buildFolderPath]" ) ) + "-<sp>FAIL<sp>-<sp>'" ) + builtTableName ) + "'<sp>is<sp>not<sp>equals<sp>to<sp>the<sp>encoding<sp>of<sp>" ) + "--------<sp>When<sp>no<sp>encoding<sp>and<sp>when<sp>a<sp>root<sp>service-path<sp>is<sp>notified/defaulted<sp>the<sp>HDFS<sp>folder<sp>path<sp>" 4 ) ) ; throw e ; } } catch ( java . lang . Exception e ) { System . out . println ( ( ( getTestTraceHead ( "[NGSIHDFSSinkTest.buildFolderPath]" ) ) + "--------<sp>When<sp>no<sp>encoding<sp>and<sp>when<sp>a<sp>root<sp>service-path<sp>is<sp>notified/defaulted<sp>the<sp>HDFS<sp>folder<sp>path<sp>" 0 ) ) ; throw e ; } createContext ( java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String , java . lang . String ) { org . apache . flume . Context context = new org . apache . flume . Context ( ) ; context . put ( "hdfs_port" 8 , backendImpl ) ; context . put ( "backend.max_conns" , backendMaxConns ) ; context . put ( "hdfs_port" 7 , backendMaxConnsPerRoute ) ; context . put ( "batchSize" , batchSize ) ; context . put ( "batchTime" , batchTime ) ; context . put ( "batchTTL" , batchTTL ) ; context . put ( "hdfs_port" 4 , csvSeparator ) ; context . put ( "hdfs_port" 5 , dataModel ) ; context . put ( "hdfs_port" 0 , enableEncoding ) ; context . put ( "hdfs_port" 2 , enableGrouping ) ; context . put ( "hdfs_port" 2 , enableLowercase ) ; context . put ( "hdfs_port" 1 , fileFormat ) ; context . put ( "hdfs_host" , host ) ; context . put ( "hdfs_password" , password ) ; context . put ( "hdfs_port" , port ) ; context . put ( "hdfs_username" , username ) ; context . put ( "hive" , hive ) ; context . put ( "krb5_auth" , krb5 ) ; context . put ( "hdfs_port" 3 , token ) ; context . put ( "hdfs_port" 6 , serviceAsNamespace ) ; return context ; } | org . junit . Assert . assertEquals ( expecetedTableName , builtTableName ) |
file_is_regular_file_java ( ) { java . io . File file = source . toFile ( ) ; boolean isRegularFile = file . isFile ( ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertTrue ( isRegularFile ) |
test2 ( ) { long [ ] [ ] lhs0 = new long [ ] [ ] { new long [ ] { 14945154 , 0 , 0 , 0 , 0 , 0 , 0 , 0 } , new long [ ] { 15840518 , 0 , 0 , 0 , 0 , 0 , 0 , 23072526 } , new long [ ] { 0 , 1 , 0 , 0 , 0 , 0 , 0 , 0 } , new long [ ] { 0 , 1 , 0 , 0 , 0 , 0 , 0 , 18556231 } , new long [ ] { 0 , 0 , 7274769 , 0 , 0 , 0 , 0 , 0 } , new long [ ] { 0 , 0 , 1285016 , 0 , 0 , 0 , 0 , 2411651 } , new long [ ] { 0 , 0 , 0 , 9614891 , 7274769 , 0 , 0 , 0 } , new long [ ] { 0 , 0 , 0 , 4514307 , 1285016 , 0 , 0 , 17488741 } , new long [ ] { 0 , 0 , 0 , 0 , 0 , 9614891 , 0 , 0 } , new long [ ] { 0 , 0 , 0 , 0 , 0 , 4514307 , 0 , 7852752 } , new long [ ] { 0 , 0 , 0 , 0 , 0 , 0 , 9614891 , 0 } , new long [ ] { 0 , 0 , 0 , 0 , 0 , 0 , 4514307 , 22089485 } } ; long [ ] rhs0 = new long [ ] { 1 , 0 , 22655879 , 0 , 12777324 , 0 , 1128298 , 0 , 20152010 , 0 , 4506067 , 0 } ; long modulus = 23072527 ; cc . redberry . rings . bigint . BigInteger [ ] [ ] lhs = cc . redberry . rings . linear . LinearSolverTest . convert ( lhs0 ) ; cc . redberry . rings . bigint . BigInteger [ ] rhs = cc . redberry . rings . linear . LinearSolverTest . convert ( rhs0 ) ; cc . redberry . rings . IntegersZp domain = new cc . redberry . rings . IntegersZp ( modulus ) ; cc . redberry . rings . bigint . BigInteger [ ] solution = solve ( domain , lhs , rhs ) ; long [ ] expected = new long [ ] { 16402965 , 22655879 , 11505290 , 2916536 , 13894224 , 7600529 , 2132874 , 14945154 } ; "<AssertPlaceHolder>" ; } convert ( long [ ] ) { cc . redberry . rings . bigint . BigInteger [ ] r = new cc . redberry . rings . bigint . BigInteger [ arr . length ] ; for ( int i = 0 ; i < ( arr . length ) ; i ++ ) r [ i ] = cc . redberry . rings . bigint . BigInteger . valueOf ( arr [ i ] ) ; return r ; } | org . junit . Assert . assertArrayEquals ( cc . redberry . rings . linear . LinearSolverTest . convert ( expected ) , solution ) |
titleFoundIfCommitteeMissing ( ) { java . util . List < java . util . List < java . lang . String > > pages = gov . nysenate . openleg . processor . hearing . PublicHearingTestHelper . getPagesFromFileName ( "01-03-13<sp>HurricaneSandy_NYS<sp>TaskForce<sp>Roundtable_Final.txt" ) ; java . lang . String expected = "ROUNDTABLE<sp>DISCUSSION<sp>HELD<sp>BY<sp>THE<sp>NEW<sp>YORK<sp>STATE<sp>SENATE<sp>" + "BIPARTISAN<sp>TASK<sp>FORCE<sp>FOR<sp>\"HURRICANE<sp>SANDY\"<sp>RECOVERY" ; java . lang . String actual = titleParser . parse ( pages . get ( 0 ) ) ; "<AssertPlaceHolder>" ; } get ( java . lang . Object ) { if ( ( gov . nysenate . openleg . service . shiro . shiroCacheManager . cache . get ( key ) ) == null ) return null ; else return gov . nysenate . openleg . service . shiro . shiroCacheManager . cache . get ( key ) . getObjectValue ( ) ; } | org . junit . Assert . assertThat ( actual , org . hamcrest . Matchers . is ( expected ) ) |
testGetParametersWithDefaultEntity ( ) { org . lnu . is . domain . enrolment . benefit . EnrolmentBenefit entity = new org . lnu . is . domain . enrolment . benefit . EnrolmentBenefit ( ) ; java . util . Map < java . lang . String , java . lang . Object > expected = new java . util . HashMap < java . lang . String , java . lang . Object > ( ) ; expected . put ( "status" , RowStatus . ACTIVE ) ; expected . put ( "userGroups" , groups ) ; java . util . Map < java . lang . String , java . lang . Object > actual = unit . getParameters ( entity ) ; "<AssertPlaceHolder>" ; } getParameters ( org . springframework . web . context . request . NativeWebRequest ) { java . util . Map < java . lang . String , java . lang . Object > resultMap = new java . util . HashMap < java . lang . String , java . lang . Object > ( ) ; java . util . Map < java . lang . String , java . lang . String > pathVariables = ( ( java . util . Map < java . lang . String , java . lang . String > ) ( webRequest . getAttribute ( HandlerMapping . URI_TEMPLATE_VARIABLES_ATTRIBUTE , RequestAttributes . SCOPE_REQUEST ) ) ) ; java . util . Map < java . lang . String , java . lang . Object > requestParams = getRequestParameterMap ( webRequest ) ; for ( Map . Entry < java . lang . String , java . lang . Object > entry : requestParams . entrySet ( ) ) { resultMap . put ( entry . getKey ( ) , entry . getValue ( ) ) ; } resultMap . putAll ( pathVariables ) ; return resultMap ; } | org . junit . Assert . assertEquals ( expected , actual ) |
testBaseDirectory ( ) { org . apache . uima . jcas . tcas . DocumentAnnotation da = getDocumentAnnotation ( ) ; da . setSourceUri ( uk . gov . dstl . baleen . annotators . misc . DocumentTypeByLocationTest . tmp . getAbsolutePath ( ) ) ; processJCas ( uk . gov . dstl . baleen . annotators . misc . DocumentTypeByLocationTest . BASE_DIRECTORY , uk . gov . dstl . baleen . annotators . misc . DocumentTypeByLocationTest . childDir . getAbsolutePath ( ) ) ; "<AssertPlaceHolder>" ; } getDocType ( ) { if ( ( DocumentAnnotation_Type . featOkTst ) && ( ( ( ( org . apache . uima . jcas . tcas . DocumentAnnotation_Type ) ( jcasType ) ) . casFeat_docType ) == null ) ) jcasType . jcas . throwFeatMissing ( "docType" , "uima.tcas.DocumentAnnotation" ) ; return jcasType . ll_cas . ll_getStringValue ( addr , ( ( org . apache . uima . jcas . tcas . DocumentAnnotation_Type ) ( jcasType ) ) . casFeatCode_docType ) ; } | org . junit . Assert . assertEquals ( "" , da . getDocType ( ) ) |
testCloneArrayForMultidimensionalArray ( ) { java . lang . Double [ ] [ ] expected = new java . lang . Double [ ] [ ] { new cz . zcu . kiv . jop . util . Double [ ] { 0.0 , 1.0 , 2.0 } , new cz . zcu . kiv . jop . util . Double [ ] { 3.0 } , new cz . zcu . kiv . jop . util . Double [ ] { 4.0 , 5.0 } } ; "<AssertPlaceHolder>" ; } cloneArray ( java . lang . Object ) { if ( array == null ) { return null ; } java . lang . Class < ? > objType = array . getClass ( ) ; if ( ! ( objType . isArray ( ) ) ) { throw new java . lang . IllegalArgumentException ( "Given<sp>object<sp>has<sp>to<sp>be<sp>array" ) ; } java . lang . Class < ? > compType = objType . getComponentType ( ) ; int len = java . lang . reflect . Array . getLength ( array ) ; java . lang . Object clone = java . lang . reflect . Array . newInstance ( compType , len ) ; if ( compType . isArray ( ) ) { for ( int i = 0 ; i < len ; i ++ ) { java . lang . reflect . Array . set ( clone , i , cz . zcu . kiv . jop . util . ArrayUtils . cloneArray ( java . lang . reflect . Array . get ( array , i ) ) ) ; } } else { java . lang . System . arraycopy ( array , 0 , clone , 0 , len ) ; } return clone ; } | org . junit . Assert . assertArrayEquals ( expected , ( ( java . lang . Double [ ] [ ] ) ( cz . zcu . kiv . jop . util . ArrayUtils . cloneArray ( expected ) ) ) ) |
testClone ( ) { org . apache . tajo . catalog . statistics . ColumnStats stat = new org . apache . tajo . catalog . statistics . ColumnStats ( new org . apache . tajo . catalog . Column ( "test" , org . apache . tajo . common . TajoDataTypes . Type . INT8 ) ) ; stat . setNumDistVals ( 1000 ) ; stat . setNumNulls ( 999 ) ; stat . setMinValue ( org . apache . tajo . datum . DatumFactory . createInt8 ( 5 ) ) ; stat . setMaxValue ( org . apache . tajo . datum . DatumFactory . createInt8 ( 10 ) ) ; org . apache . tajo . catalog . statistics . ColumnStats stat2 = ( ( org . apache . tajo . catalog . statistics . ColumnStats ) ( stat . clone ( ) ) ) ; "<AssertPlaceHolder>" ; } clone ( ) { org . apache . tajo . catalog . statistics . Stat stat = ( ( org . apache . tajo . catalog . statistics . Stat ) ( super . clone ( ) ) ) ; stat . type = type ; stat . val = val ; return stat ; } | org . junit . Assert . assertEquals ( stat , stat2 ) |
testSeparator ( ) { org . stringtemplate . v4 . STGroup group = new org . stringtemplate . v4 . STGroup ( ) ; group . defineTemplate ( "test" , "names" , "<names:{n<sp>|<sp>case<sp><n>};<sp>separator=\",<sp>\">" ) ; org . stringtemplate . v4 . ST st = group . getInstanceOf ( "test" ) ; st . add ( "names" , "Ter" ) ; st . add ( "names" , "Tom" ) ; java . lang . String expected = "case<sp>Ter,<sp>case<sp>Tom" ; java . lang . String result = st . render ( ) ; "<AssertPlaceHolder>" ; } render ( ) { return render ( java . util . Locale . getDefault ( ) ) ; } | org . junit . Assert . assertEquals ( expected , result ) |
testCloseCaseInstance ( ) { final org . jbpm . workbench . cm . model . CaseInstanceSummary cis = caseInstanceSummaryList . remove ( 0 ) ; presenter . closeCaseInstance ( cis ) ; verify ( caseManagementService ) . closeCaseInstance ( cis . getContainerId ( ) , cis . getCaseId ( ) , null ) ; verify ( caseManagementService ) . getCaseInstances ( any ( org . jbpm . workbench . cm . util . CaseInstanceSearchRequest . class ) ) ; final org . mockito . ArgumentCaptor < java . util . List > captor = org . mockito . ArgumentCaptor . forClass ( java . util . List . class ) ; verify ( view ) . setCaseInstanceList ( captor . capture ( ) ) ; "<AssertPlaceHolder>" ; } getValue ( ) { return dataBinder . getModel ( ) ; } | org . junit . Assert . assertEquals ( caseInstanceSummaryList . size ( ) , captor . getValue ( ) . size ( ) ) |
testSimpleCommand ( ) { ddf . platform . scheduler . FirstArgumentAnswer captureInput = new ddf . platform . scheduler . FirstArgumentAnswer ( ) ; org . apache . karaf . shell . api . console . Session session = mock ( org . apache . karaf . shell . api . console . Session . class ) ; when ( session . execute ( isA ( ddf . platform . scheduler . CharSequence . class ) ) ) . then ( captureInput ) ; ddf . platform . scheduler . CommandJob commandJob = new ddf . platform . scheduler . CommandJob ( ) { @ ddf . platform . scheduler . Override public ddf . security . Subject getSystemSubject ( ) { return ddf . platform . scheduler . CommandJobTest . createMockSubject ( ) ; } @ ddf . platform . scheduler . Override protected org . apache . karaf . shell . api . console . SessionFactory getSessionFactory ( ) { org . apache . karaf . shell . api . console . SessionFactory sessionFactory = mock ( org . apache . karaf . shell . api . console . SessionFactory . class ) ; when ( sessionFactory . create ( notNull ( java . io . InputStream . class ) , any ( ) , any ( ) ) ) . thenReturn ( session ) ; return sessionFactory ; } } ; java . lang . String command = ddf . platform . scheduler . CommandJobTest . VALID_COMMAND ; commandJob . execute ( createMockJobExecutionContext ( command ) ) ; "<AssertPlaceHolder>" ; verify ( session , times ( 1 ) ) . execute ( command ) ; verify ( session , times ( 1 ) ) . close ( ) ; } getInputArg ( ) { return inputArg ; } | org . junit . Assert . assertThat ( captureInput . getInputArg ( ) , org . hamcrest . Matchers . is ( command ) ) |
testOnPortUpdatePreSuccess ( ) { java . util . ArrayList < java . lang . String > attrList = new java . util . ArrayList < java . lang . String > ( ) ; org . o3project . odenos . core . component . network . topology . Port prev = new org . o3project . odenos . core . component . network . topology . Port ( "0" , "port1" , "node1" , "" , "" , new java . util . HashMap < java . lang . String , java . lang . String > ( ) ) ; org . o3project . odenos . core . component . network . topology . Port curr = new org . o3project . odenos . core . component . network . topology . Port ( "0" , "port2" , "node1" , "" , "" , new java . util . HashMap < java . lang . String , java . lang . String > ( ) ) ; org . o3project . odenos . core . component . ConversionTable conversionTable = org . powermock . api . mockito . PowerMockito . spy ( new org . o3project . odenos . core . component . ConversionTable ( ) ) ; org . powermock . api . mockito . PowerMockito . doReturn ( "original" ) . when ( conversionTable , "getConnectionType" , org . o3project . odenos . component . aggregator . AggregatorTest . ORIGINAL_NW_ID ) ; org . powermock . api . mockito . PowerMockito . doReturn ( conversionTable ) . when ( target , "conversionTable" ) ; "<AssertPlaceHolder>" ; } onPortUpdatePre ( java . lang . String , org . o3project . odenos . core . component . network . topology . Port , org . o3project . odenos . core . component . network . topology . Port , java . util . ArrayList ) { org . o3project . odenos . component . federator . Federator . log . debug ( "" ) ; java . lang . String fedNwId = getNetworkIdByType ( org . o3project . odenos . component . federator . Federator . FEDERATED_NETWORK ) ; if ( fedNwId == null ) { return false ; } java . lang . String connType = conversionTable ( ) . getConnectionType ( networkId ) ; if ( connType . equals ( org . o3project . odenos . component . federator . Federator . ORIGINAL_NETWORK ) ) { if ( java . lang . Boolean . valueOf ( curr . getAttribute ( Port . IS_BOUNDARY ) ) ) { java . util . ArrayList < java . lang . String > dstPort = conversionTable ( ) . getPort ( networkId , curr . getNode ( ) , curr . getId ( ) ) ; java . lang . String [ ] portId = dstPort . get ( 0 ) . split ( "::" ) ; org . o3project . odenos . core . component . NetworkInterface fedNw = networkInterfaces ( ) . get ( fedNwId ) ; org . o3project . odenos . core . component . network . topology . Port fedPort = fedNw . getPort ( portId [ 1 ] , portId [ 2 ] ) ; org . o3project . odenos . core . component . network . topology . Link link = fedNw . getInLink ( fedPort ) ; if ( link != null ) { updateBoundaryLinkAttribute ( link , curr ) ; } } } return true ; } | org . junit . Assert . assertThat ( target . onPortUpdatePre ( org . o3project . odenos . component . aggregator . AggregatorTest . ORIGINAL_NW_ID , prev , curr , attrList ) , org . hamcrest . CoreMatchers . is ( true ) ) |
testGetUserPreferredIIOImplementation ( ) { edu . illinois . library . cantaloupe . config . Configuration config = edu . illinois . library . cantaloupe . config . Configuration . getInstance ( ) ; config . setProperty ( GIFImageReader . IMAGEIO_PLUGIN_CONFIG_KEY , "cats" ) ; "<AssertPlaceHolder>" ; } getUserPreferredIIOImplementation ( ) { edu . illinois . library . cantaloupe . config . Configuration config = edu . illinois . library . cantaloupe . config . Configuration . getInstance ( ) ; return config . getString ( edu . illinois . library . cantaloupe . processor . codec . png . PNGImageWriter . IMAGEIO_PLUGIN_CONFIG_KEY ) ; } | org . junit . Assert . assertEquals ( "cats" , ( ( edu . illinois . library . cantaloupe . processor . codec . GIFImageReader ) ( instance ) ) . getUserPreferredIIOImplementation ( ) ) |
testElimina ( ) { log . debug ( "Test<sp>'elimina'<sp>paquete" ) ; mx . edu . um . mateo . general . model . Usuario usuario = obtieneUsuario ( ) ; mx . edu . um . mateo . inscripciones . model . AFEPlaza afePlaza = new mx . edu . um . mateo . inscripciones . model . AFEPlaza ( ) ; afePlaza . setClave ( "1110475" ) ; afePlaza . setDias ( "23" ) ; afePlaza . setEmail ( "samuel.9401@gmail.com" ) ; afePlaza . setEmpresa ( usuario . getEmpresa ( ) ) ; afePlaza . setFechaAlta ( new java . util . Date ( ) ) ; afePlaza . setFechaModificacion ( new java . util . Date ( ) ) ; afePlaza . setPrimerIngreso ( false ) ; afePlaza . setRequisitos ( "Estudiante" ) ; afePlaza . setTipoPlaza ( "Ayudante<sp>General" ) ; afePlaza . setTurno ( "Matutino" ) ; afePlaza . setIndustrial ( true ) ; afePlaza . setObservaciones ( "prueba" ) ; afePlaza . setUsuarioAlta ( usuario ) ; afePlaza . setUsuarioModificacion ( usuario ) ; currentSession ( ) . save ( afePlaza ) ; "<AssertPlaceHolder>" ; this . authenticate ( usuario , usuario . getPassword ( ) , new java . util . ArrayList < org . springframework . security . core . GrantedAuthority > ( usuario . getRoles ( ) ) ) ; this . mockMvc . perform ( post ( Constantes . PATH_AFEPLAZA_ELIMINA ) . param ( "id" , afePlaza . getId ( ) . toString ( ) ) ) . andExpect ( flash ( ) . attributeExists ( Constantes . CONTAINSKEY_MESSAGE ) ) . andExpect ( flash ( ) . attribute ( Constantes . CONTAINSKEY_MESSAGE , "afePlaza.elimina.message" ) ) . andExpect ( redirectedUrl ( Constantes . PATH_AFEPLAZA_LISTA ) ) ; } getId ( ) { return id ; } | org . junit . Assert . assertNotNull ( afePlaza . getId ( ) ) |
testSetReactionCenter ( ) { org . openscience . cdk . reaction . type . parameters . SetReactionCenter react = new org . openscience . cdk . reaction . type . parameters . SetReactionCenter ( ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertNotNull ( react ) |
getInstance ( ) { io . robe . hibernate . RobeHibernateBundle bundle = io . robe . hibernate . RobeHibernateBundle . getInstance ( ) ; "<AssertPlaceHolder>" ; } getInstance ( ) { io . robe . hibernate . RobeHibernateBundle bundle = io . robe . hibernate . RobeHibernateBundle . getInstance ( ) ; org . junit . Assert . assertNotNull ( bundle ) ; } | org . junit . Assert . assertNotNull ( bundle ) |
g_V_selectXfirst_aX ( ) { final org . apache . tinkerpop . gremlin . process . traversal . Traversal < org . apache . tinkerpop . gremlin . structure . Vertex , java . lang . Object > traversal = get_g_V_selectXfirst_aX ( ) ; printTraversalForm ( traversal ) ; "<AssertPlaceHolder>" ; } toList ( ) { return this . fill ( new java . util . ArrayList ( ) ) ; } | org . junit . Assert . assertEquals ( java . util . Collections . emptyList ( ) , traversal . toList ( ) ) |
testZoekPersoon ( ) { initMocks ( nl . bzk . brp . model . bevraging . levering . ZoekPersoonBericht . class , SoortBericht . LVG_BVG_ZOEK_PERSOON ) ; initBerichtVerwerker ( new java . util . ArrayList < nl . bzk . brp . model . validatie . Melding > ( ) , nl . bzk . brp . business . dto . bevraging . BevragingResultaat . class , true ) ; final nl . bzk . brp . model . bevraging . levering . ZoekPersoonAntwoordBericht testBericht = new nl . bzk . brp . model . bevraging . levering . ZoekPersoonAntwoordBericht ( ) ; when ( getAntwoordBerichtFactory ( ) . bouwAntwoordBericht ( any ( nl . bzk . brp . model . logisch . ber . Bericht . class ) , any ( nl . bzk . brp . business . dto . bevraging . BevragingResultaat . class ) ) ) . thenReturn ( testBericht ) ; final nl . bzk . brp . model . bevraging . levering . ZoekPersoonAntwoordBericht resultaat = ( ( nl . bzk . brp . web . service . LeveringBevragingService ) ( getWebService ( ) ) ) . zoekPersoon ( ( ( nl . bzk . brp . model . bevraging . levering . ZoekPersoonBericht ) ( getBericht ( ) ) ) ) ; "<AssertPlaceHolder>" ; } getBericht ( ) { return bericht ; } | org . junit . Assert . assertEquals ( testBericht , resultaat ) |
applyMatch ( ) { setup ( "name" ) ; org . apache . hadoop . fs . shell . PathData item = new org . apache . hadoop . fs . shell . PathData ( "/directory/path/name" , mockFs . getConf ( ) ) ; "<AssertPlaceHolder>" ; } apply ( org . apache . hadoop . fs . shell . PathData , int ) { java . lang . String name = getPath ( item ) . getName ( ) ; if ( ! ( caseSensitive ) ) { name = org . apache . hadoop . util . StringUtils . toLowerCase ( name ) ; } if ( globPattern . matches ( name ) ) { return Result . PASS ; } else { return Result . FAIL ; } } | org . junit . Assert . assertEquals ( Result . PASS , name . apply ( item , ( - 1 ) ) ) |
test_writeFloat ( float , java . lang . String ) { outputCompact . writeFloat ( input ) ; "<AssertPlaceHolder>" ; } toString ( ) { return getClass ( ) . getSimpleName ( ) ; } | org . junit . Assert . assertEquals ( buf . toString ( ) , expected ) |
expected_and_current_data_coparison_should_succeed ( ) { com . lordofthejars . nosqlunit . marklogic . MarkLogicOperation markLogicOperation = new com . lordofthejars . nosqlunit . marklogic . MarkLogicOperation ( com . lordofthejars . nosqlunit . marklogic . ManagedMarkLogicConfigurationBuilder . marklogic ( ) . build ( ) ) ; markLogicOperation . insert ( new java . io . ByteArrayInputStream ( com . lordofthejars . nosqlunit . marklogic . integration . WhenManagedMarkLogicOperationsAreRequired . DATA . getBytes ( ) ) ) ; boolean result = markLogicOperation . databaseIs ( new java . io . ByteArrayInputStream ( com . lordofthejars . nosqlunit . marklogic . integration . WhenManagedMarkLogicOperationsAreRequired . DATA . getBytes ( ) ) ) ; "<AssertPlaceHolder>" ; } databaseIs ( java . io . InputStream ) { return compareData ( contentStream ) ; } | org . junit . Assert . assertThat ( result , org . hamcrest . CoreMatchers . is ( true ) ) |
testMsgid1 ( ) { com . zx . sms . codec . smgp . msg . MsgId m = new com . zx . sms . codec . smgp . msg . MsgId ( "12345607091206000017" ) ; byte [ ] arr = com . zx . sms . codec . smgp . util . SMGPMsgIdUtil . msgId2Bytes ( m ) ; "<AssertPlaceHolder>" ; } bytes2MsgId ( byte [ ] ) { assert ( bytes . length ) == 10 ; java . lang . String str = java . lang . String . valueOf ( org . apache . commons . codec . binary . Hex . encodeHex ( bytes ) ) ; try { return new com . zx . sms . codec . smgp . msg . MsgId ( str ) ; } catch ( java . lang . Exception ex ) { com . zx . sms . codec . smgp . util . SMGPMsgIdUtil . logger . warn ( "Err<sp>MsgID<sp>:<sp>0x{}" , str ) ; return new com . zx . sms . codec . smgp . msg . MsgId ( ) ; } } | org . junit . Assert . assertEquals ( m , com . zx . sms . codec . smgp . util . SMGPMsgIdUtil . bytes2MsgId ( arr ) ) |
getCreatedAtReferenceMutableObject ( ) { org . eclipse . egit . github . core . PullRequest pullRequest = new org . eclipse . egit . github . core . PullRequest ( ) ; pullRequest . setCreatedAt ( new java . util . Date ( 10000 ) ) ; pullRequest . getCreatedAt ( ) . setTime ( 0 ) ; "<AssertPlaceHolder>" ; } getCreatedAt ( ) { return org . eclipse . egit . github . core . util . DateUtils . clone ( createdAt ) ; } | org . junit . Assert . assertTrue ( ( ( pullRequest . getCreatedAt ( ) . getTime ( ) ) != 0 ) ) |
testPlaceHolder ( ) { definition . setPlaceHolder ( org . kie . workbench . common . dmn . client . property . dmn . QNameFieldDefinitionTest . PLACE_HOLDER ) ; "<AssertPlaceHolder>" ; } getPlaceHolder ( ) { return placeHolder ; } | org . junit . Assert . assertEquals ( org . kie . workbench . common . dmn . client . property . dmn . QNameFieldDefinitionTest . PLACE_HOLDER , definition . getPlaceHolder ( ) ) |
testGetParametersWithEmptyFields ( ) { org . lnu . is . domain . asset . address . AssetAddress entity = new org . lnu . is . domain . asset . address . AssetAddress ( ) ; java . util . Map < java . lang . String , java . lang . Object > expected = new java . util . HashMap < java . lang . String , java . lang . Object > ( ) ; expected . put ( "status" , RowStatus . ACTIVE ) ; expected . put ( "userGroups" , groups ) ; java . util . Map < java . lang . String , java . lang . Object > actual = unit . getParameters ( entity ) ; verify ( adminUnitDao , times ( 0 ) ) . getEntityById ( anyLong ( ) ) ; verify ( assetDao , times ( 0 ) ) . getEntityById ( anyLong ( ) ) ; verify ( addressTypeDao , times ( 0 ) ) . getEntityById ( anyLong ( ) ) ; verify ( streetTypeDao , times ( 0 ) ) . getEntityById ( anyLong ( ) ) ; "<AssertPlaceHolder>" ; } getEntityById ( KEY ) { org . lnu . is . dao . dao . DefaultDao . LOG . info ( "Getting<sp>{}.entity<sp>wit<sp>id" , getEntityClass ( ) . getSimpleName ( ) , id ) ; return persistenceManager . findById ( getEntityClass ( ) , id ) ; } | org . junit . Assert . assertEquals ( expected , actual ) |
preserveTimestampDirCanBeDisabled ( ) { org . springframework . context . annotation . AnnotationConfigApplicationContext context = new org . springframework . context . annotation . AnnotationConfigApplicationContext ( ) ; org . springframework . boot . test . EnvironmentTestUtils . addEnvironment ( context , "sftp.preserveTimestamp:false" ) ; context . register ( org . springframework . cloud . stream . app . sftp . source . SftpSourcePropertiesTests . Conf . class ) ; context . refresh ( ) ; org . springframework . cloud . stream . app . sftp . source . SftpSourceProperties properties = context . getBean ( org . springframework . cloud . stream . app . sftp . source . SftpSourceProperties . class ) ; "<AssertPlaceHolder>" ; } isPreserveTimestamp ( ) { return preserveTimestamp ; } | org . junit . Assert . assertTrue ( ( ! ( properties . isPreserveTimestamp ( ) ) ) ) |
appendJsonAttribute_good_case_quoted_notlast ( ) { java . lang . StringBuilder stringBuilder = new java . lang . StringBuilder ( ) ; stringBuilder . append ( "prefix_" ) ; java . lang . String expResult = "prefix_\"keyname\":\"value\"," ; com . microsoft . azure . sdk . iot . service . Tools . appendJsonAttribute ( stringBuilder , "keyname" , "value" , true , false ) ; "<AssertPlaceHolder>" ; } toString ( ) { com . google . gson . Gson gson = new com . google . gson . GsonBuilder ( ) . disableHtmlEscaping ( ) . setPrettyPrinting ( ) . create ( ) ; return gson . toJson ( this ) ; } | org . junit . Assert . assertEquals ( expResult , stringBuilder . toString ( ) ) |
testHasPageAction ( ) { com . eclipsesource . tabris . internal . ui . PageDescriptor pageDescriptor = new com . eclipsesource . tabris . internal . ui . PageDescriptor ( "bar" , com . eclipsesource . tabris . internal . ui . TestPage . class ) ; pageDescriptor . setTopLevel ( true ) ; pageDescriptor . addAction ( new com . eclipsesource . tabris . ui . ActionConfiguration ( "foo" , com . eclipsesource . tabris . internal . ui . TestAction . class ) ) ; uiDescriptor . add ( pageDescriptor ) ; com . eclipsesource . tabris . internal . ui . RemoteUI remoteUI = mock ( com . eclipsesource . tabris . internal . ui . RemoteUI . class ) ; when ( remoteUI . getPageParent ( ) ) . thenReturn ( shell ) ; com . eclipsesource . tabris . internal . ui . Controller controller = new com . eclipsesource . tabris . internal . ui . Controller ( remoteUI , uiDescriptor ) ; controller . createGlobalActions ( ui ) ; controller . createRootPages ( ui ) ; controller . showPage ( ui , pageDescriptor , mock ( com . eclipsesource . tabris . ui . PageData . class ) ) ; boolean hasAction = controller . hasAction ( "foo" ) ; "<AssertPlaceHolder>" ; } hasAction ( java . lang . String ) { com . eclipsesource . tabris . internal . ui . rendering . ActionRenderer result = findActionInGlobalActions ( id ) ; if ( result == null ) { result = findActionInPageActions ( id ) ; } return result != null ; } | org . junit . Assert . assertTrue ( hasAction ) |
test1 ( ) { org . n3r . eql . EqlTran eqlTran = new org . n3r . eql . Eql ( ) . newTran ( ) ; eqlTran . start ( ) ; org . n3r . eql . impl . EqlBatch eqlBatch = new org . n3r . eql . impl . EqlBatch ( ) ; for ( int i = 0 ; i < 10 ; ++ i ) { java . lang . String orderNo = randLetters ( 10 ) ; java . lang . String userId = randLetters ( 10 ) ; int prizeItem = randInt ( 10 ) ; int ret = new org . n3r . eql . Eql ( ) . useBatch ( eqlBatch ) . useTran ( eqlTran ) . insert ( "insertPrizeBingoo" ) . params ( orderNo , "Olympic" , ( "" + prizeItem ) , userId ) . execute ( ) ; "<AssertPlaceHolder>" ; } eqlBatch . executeBatch ( ) ; eqlTran . commit ( ) ; org . n3r . eql . util . Closes . closeQuietly ( eqlTran ) ; } insert ( java . lang . String ) { this . sqlId = sqlId ; return this ; } | org . junit . Assert . assertThat ( ret , org . hamcrest . CoreMatchers . is ( 0 ) ) |
testReflect20 ( ) { com . ibm . wala . ipa . callgraph . AnalysisScope scope = com . ibm . wala . core . tests . callGraph . ReflectionTest . findOrCreateAnalysisScope ( ) ; com . ibm . wala . ipa . cha . IClassHierarchy cha = com . ibm . wala . core . tests . callGraph . ReflectionTest . findOrCreateCHA ( scope ) ; java . lang . Iterable < com . ibm . wala . ipa . callgraph . Entrypoint > entrypoints = com . ibm . wala . ipa . callgraph . impl . Util . makeMainEntrypoints ( scope , cha , TestConstants . REFLECT20_MAIN ) ; com . ibm . wala . ipa . callgraph . AnalysisOptions options = com . ibm . wala . core . tests . callGraph . CallGraphTestUtil . makeAnalysisOptions ( scope , entrypoints ) ; com . ibm . wala . ipa . callgraph . CallGraph cg = com . ibm . wala . core . tests . callGraph . CallGraphTestUtil . buildZeroOneCFA ( options , new com . ibm . wala . ipa . callgraph . AnalysisCacheImpl ( ) , cha , scope , false ) ; com . ibm . wala . types . TypeReference tr = com . ibm . wala . types . TypeReference . findOrCreate ( ClassLoaderReference . Application , "Lreflection/Helper" ) ; com . ibm . wala . types . MethodReference mr = com . ibm . wala . types . MethodReference . findOrCreate ( tr , "o" , "(Ljava/lang/Object;Ljava/lang/Object;)V" ) ; java . util . Set < com . ibm . wala . ipa . callgraph . CGNode > nodes = cg . getNodes ( mr ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return ( top ) == ( - 1 ) ; } | org . junit . Assert . assertFalse ( nodes . isEmpty ( ) ) |
testFolderExtraction ( ) { java . io . File tessDataFolder = null ; try { java . lang . String result = instance . doOCR ( imageFile ) ; net . sourceforge . tess4j . TestFolderExtraction . logger . info ( result ) ; } catch ( net . sourceforge . tess4j . TesseractException e ) { net . sourceforge . tess4j . TestFolderExtraction . logger . error ( e . getMessage ( ) ) ; net . sourceforge . tess4j . TestFolderExtraction . logger . error ( e . getMessage ( ) , e ) ; } catch ( java . net . URISyntaxException e ) { net . sourceforge . tess4j . TestFolderExtraction . logger . error ( e . getMessage ( ) , e ) ; } "<AssertPlaceHolder>" ; } getMessage ( ) { return outputMessage . toString ( ) ; } | org . junit . Assert . assertTrue ( ( ( tessDataFolder != null ) && ( tessDataFolder . exists ( ) ) ) ) |
testEncodeAndWriteStream ( ) { org . apache . fop . pdf . PDFFilterList filters = stream . getFilterList ( ) ; filters . addFilter ( "null" ) ; byte [ ] bytes = createSampleData ( ) ; stream . setData ( bytes ) ; java . io . ByteArrayOutputStream actual = new java . io . ByteArrayOutputStream ( ) ; org . apache . fop . pdf . PDFNumber number = new org . apache . fop . pdf . PDFNumber ( ) ; stream . encodeAndWriteStream ( actual , number ) ; "<AssertPlaceHolder>" ; } createSampleStreamData ( ) { java . io . ByteArrayOutputStream stream = new java . io . ByteArrayOutputStream ( ) ; stream . write ( "\nstream\n" . getBytes ( "US-ASCII" ) ) ; stream . write ( createSampleData ( ) ) ; stream . write ( "\nendstream" . getBytes ( "US-ASCII" ) ) ; return stream . toByteArray ( ) ; } | org . junit . Assert . assertArrayEquals ( createSampleStreamData ( ) , actual . toByteArray ( ) ) |
testWithUniqueNameWhenNameIsUnique ( ) { final org . kie . workbench . common . dmn . client . editors . types . common . DataType dataType = makeDataType ( "uuid1" ) ; doReturn ( dataType ) . when ( manager ) . get ( ) ; doReturn ( asList ( makeDataType ( "uuid2" , "tPerson" ) , makeDataType ( "uuid3" , "tCity" ) ) ) . when ( dataTypeNameValidator ) . siblings ( dataType ) ; final java . lang . String actualDataTypeName = manager . withDataType ( dataType ) . withUniqueName ( "tCompany" ) . get ( ) . getName ( ) ; final java . lang . String expectedDataTypeName = "tCompany" ; "<AssertPlaceHolder>" ; } getName ( ) { return name ; } | org . junit . Assert . assertEquals ( expectedDataTypeName , actualDataTypeName ) |
shouldAlwaysReturnSameInstanceOfBackgroundExecutor ( ) { "<AssertPlaceHolder>" ; } getBackgroundExecutor ( ) { return getExecutorInstance ( ) ; } | org . junit . Assert . assertSame ( util . getBackgroundExecutor ( ) , util . getBackgroundExecutor ( ) ) |
testOne ( ) { org . terrier . matching . MatchingQueryTerms mqt = new org . terrier . matching . MatchingQueryTerms ( ) ; mqt . add ( org . terrier . querying . parser . Query . QTPBuilder . of ( new org . terrier . matching . matchops . SingleTermOp ( "a" ) ) . build ( ) ) ; new org . terrier . querying . DependenceModelPreProcess ( ) . process ( mqt , "pBiL" ) ; System . out . println ( mqt . toString ( ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return location . length ; } | org . junit . Assert . assertEquals ( 1 , mqt . size ( ) ) |
testBuildFromEncodedMapComplex2 ( ) { java . util . Map < java . lang . String , java . lang . Object > maps = new java . util . HashMap ( ) ; maps . put ( "x" , "w" 2 ) ; maps . put ( "y" , "/path-absolute/test1" ) ; maps . put ( "w" 0 , "fred@example.com" ) ; maps . put ( "w" , "path-rootless/test2" ) ; maps . put ( "w" 1 , "extra" ) ; java . lang . String expectedPath = "path-rootless/test2/x%25yz//path-absolute/test1/fred@example.com/x%25yz" ; java . net . URI uri = javax . ws . rs . core . UriBuilder . fromPath ( "" ) . path ( "{w}/{x}/{y}/{z}/{x}" ) . buildFromEncodedMap ( maps ) ; java . lang . String rawPath = uri . getRawPath ( ) ; "<AssertPlaceHolder>" ; } buildFromEncodedMap ( java . util . Map ) { java . util . Map < java . lang . String , java . lang . String > decodedMap = new java . util . HashMap ( map . size ( ) ) ; for ( Map . Entry < java . lang . String , ? extends java . lang . Object > entry : map . entrySet ( ) ) { if ( ( entry . getValue ( ) ) == null ) { throw new java . lang . IllegalArgumentException ( "Value<sp>is<sp>null" ) ; } java . lang . String theValue = entry . getValue ( ) . toString ( ) ; if ( theValue . contains ( "/" ) ) { java . lang . StringBuilder buf = new java . lang . StringBuilder ( ) ; java . lang . String [ ] values = theValue . split ( "/" ) ; for ( int i = 0 ; i < ( values . length ) ; i ++ ) { buf . append ( org . apache . cxf . jaxrs . utils . HttpUtils . encodePartiallyEncoded ( values [ i ] , false ) ) ; if ( ( i + 1 ) < ( values . length ) ) { buf . append ( '/' ) ; } } decodedMap . put ( entry . getKey ( ) , buf . toString ( ) ) ; } else { decodedMap . put ( entry . getKey ( ) , org . apache . cxf . jaxrs . utils . HttpUtils . encodePartiallyEncoded ( theValue , false ) ) ; } } return doBuildFromMap ( decodedMap , true , false ) ; } | org . junit . Assert . assertEquals ( expectedPath , rawPath ) |
testConditionTask ( ) { com . jim . framework . web . SpringbootApplicationTests . ConditionTask conditionTask = new com . jim . framework . web . SpringbootApplicationTests . ConditionTask ( ) ; conditionTask . run ( ) ; java . lang . Thread . sleep ( 3000 ) ; "<AssertPlaceHolder>" ; } run ( ) { try { if ( ! ( this . netInitFlag . get ( ) ) ) { initNet ( ) ; } } catch ( java . lang . Exception e ) { com . jim . framework . configcenter . service . impl . ConfigCenterServiceImpl . logger . error ( "init<sp>zk<sp>error" , e ) ; } } | org . junit . Assert . assertTrue ( true ) |
testAssignedAction ( ) { org . eclipse . xtend2 . lib . StringConcatenation _builder = new org . eclipse . xtend2 . lib . StringConcatenation ( ) ; _builder . append ( "Rule:<sp>{Foo}<sp>{Action.feat=current};" ) ; _builder . newLine ( ) ; final java . lang . String actual = this . toPda ( _builder ) ; org . eclipse . xtend2 . lib . StringConcatenation _builder_1 = new org . eclipse . xtend2 . lib . StringConcatenation ( ) ; _builder_1 . append ( "Rule:" ) ; _builder_1 . newLine ( ) ; _builder_1 . append ( "\t" ) ; _builder_1 . append ( "start<sp>-><sp>{Foo}" ) ; _builder_1 . newLine ( ) ; _builder_1 . append ( "\t" ) ; _builder_1 . append ( "{Action.feat=}<sp>-><sp>stop" ) ; _builder_1 . newLine ( ) ; _builder_1 . append ( "\t" ) ; _builder_1 . append ( "{Foo}<sp>-><sp>{Action.feat=}" ) ; _builder_1 . newLine ( ) ; final java . lang . String expected = _builder_1 . toString ( ) ; "<AssertPlaceHolder>" ; } toString ( ) { return com . google . common . base . Joiner . on ( "\n" ) . join ( getList ( ) ) ; } | org . junit . Assert . assertEquals ( expected , actual ) |
testFindIndexLessThan ( ) { org . eclipse . january . dataset . Dataset dataset = org . eclipse . january . dataset . DatasetFactory . createFromObject ( new java . lang . Double [ ] { 5.2 , 2.0 , 4.8 , 13.0 , 0.8 , 96.0 , 13.0 } ) ; int expected = 4 ; int actual = org . eclipse . january . dataset . DatasetUtils . findIndexLessThan ( dataset , 1.0 ) ; "<AssertPlaceHolder>" ; } findIndexLessThan ( org . eclipse . january . dataset . Dataset , double ) { org . eclipse . january . dataset . IndexIterator iter = a . getIterator ( ) ; while ( iter . hasNext ( ) ) { if ( ( a . getElementDoubleAbs ( iter . index ) ) < n ) break ; } return iter . index ; } | org . junit . Assert . assertEquals ( expected , actual ) |
testGetView ( ) { screen . init ( ) ; "<AssertPlaceHolder>" ; } getView ( ) { return view ; } | org . junit . Assert . assertEquals ( view , screen . getView ( ) ) |
sort ( ) { final com . yandex . yoctodb . mutable . DatabaseBuilder dbBuilder = com . yandex . yoctodb . DatabaseFormat . getCurrent ( ) . newDatabaseBuilder ( ) ; for ( int i = 0 ; i < ( DOCS ) ; i ++ ) { dbBuilder . merge ( com . yandex . yoctodb . DatabaseFormat . getCurrent ( ) . newDocumentBuilder ( ) . withField ( "id" , i , com . yandex . yoctodb . FULL ) . withPayload ( ( "payload" + i ) . getBytes ( ) ) ) ; } final java . io . ByteArrayOutputStream os = new java . io . ByteArrayOutputStream ( ) ; dbBuilder . buildWritable ( ) . writeTo ( os ) ; final com . yandex . yoctodb . immutable . Database db = com . yandex . yoctodb . DatabaseFormat . getCurrent ( ) . getDatabaseReader ( ) . from ( com . yandex . yoctodb . util . buf . Buffer . from ( os . toByteArray ( ) ) ) ; final com . yandex . yoctodb . query . Query q1 = select ( ) . where ( gte ( "id" , from ( 0 ) ) ) . and ( lte ( "id" , from ( DOCS ) ) ) . orderBy ( desc ( "id" ) ) ; final java . util . List < java . lang . String > results = new java . util . ArrayList ( DOCS ) ; db . execute ( q1 , new com . yandex . yoctodb . StringProcessor ( results ) ) ; int i = ( DOCS ) - 1 ; for ( java . lang . String result : results ) { "<AssertPlaceHolder>" ; i -- ; } } execute ( com . yandex . yoctodb . query . Query , com . yandex . yoctodb . query . DocumentProcessor ) { throw new java . lang . IllegalStateException ( ) ; } | org . junit . Assert . assertEquals ( ( "payload" + i ) , result ) |
testWrongFormatOrder ( ) { boolean raises = false ; java . util . Date now = new java . util . Date ( ) ; try { org . nuxeo . ecm . platform . audit . api . query . DateRangeParser . parseDateRangeQuery ( now , "2h2m" ) ; } catch ( org . nuxeo . ecm . platform . audit . api . query . AuditQueryException e ) { raises = true ; } "<AssertPlaceHolder>" ; } parseDateRangeQuery ( java . util . Date , java . lang . String ) { try { java . util . Calendar calendar = java . util . Calendar . getInstance ( ) ; calendar . setTime ( now ) ; if ( dateRangeQuery != null ) { java . util . Map < java . lang . String , java . lang . Integer > parsed = org . nuxeo . ecm . platform . audit . api . query . DateRangeParser . parseQuery ( dateRangeQuery ) ; if ( parsed . containsKey ( DateRangeQueryConstants . HOUR ) ) { calendar . add ( Calendar . HOUR_OF_DAY , ( - ( parsed . get ( DateRangeQueryConstants . HOUR ) ) ) ) ; } if ( parsed . containsKey ( DateRangeQueryConstants . MIN ) ) { calendar . add ( Calendar . MINUTE , ( - ( parsed . get ( DateRangeQueryConstants . MIN ) ) ) ) ; } } return calendar . getTime ( ) ; } catch ( java . lang . NumberFormatException nfe ) { throw new org . nuxeo . ecm . platform . audit . api . query . AuditQueryException ( "Invalid<sp>query<sp>format..." , nfe ) ; } } | org . junit . Assert . assertTrue ( raises ) |
testEquals ( ) { org . bff . javampd . output . MPDOutput output1 = new org . bff . javampd . output . MPDOutput ( 1 ) ; org . bff . javampd . output . MPDOutput output2 = new org . bff . javampd . output . MPDOutput ( 1 ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertEquals ( output1 , output2 ) |
testError ( ) { java . lang . Exception exception = new java . lang . Exception ( ) ; org . mockito . Mockito . when ( transactionWorker . getError ( ) ) . thenReturn ( exception ) ; "<AssertPlaceHolder>" ; } getError ( ) { return error ; } | org . junit . Assert . assertEquals ( exception , transaction . getError ( ) ) |
determineSegmentCount_boostConcurrency ( ) { int _segs = org . cache2k . core . InternalCache2kBuilder . determineSegmentCount ( false , 12 , true , 1000000 , 0 ) ; "<AssertPlaceHolder>" ; } determineSegmentCount ( boolean , int , boolean , long , int ) { int _segmentCount = 1 ; if ( _availableProcessors > 1 ) { _segmentCount = 2 ; if ( _boostConcurrency ) { _segmentCount = 2 << ( 31 - ( java . lang . Integer . numberOfLeadingZeros ( _availableProcessors ) ) ) ; } } if ( _segmentCountOverride > 0 ) { _segmentCount = 1 << ( 32 - ( java . lang . Integer . numberOfLeadingZeros ( ( _segmentCountOverride - 1 ) ) ) ) ; } else { int _maxSegments = _availableProcessors * 2 ; _segmentCount = java . lang . Math . min ( _segmentCount , _maxSegments ) ; } if ( ( _entryCapacity >= 0 ) && ( _entryCapacity < 1000 ) ) { _segmentCount = 1 ; } if ( _strictEviction ) { _segmentCount = 1 ; } return _segmentCount ; } | org . junit . Assert . assertEquals ( 16 , _segs ) |
testGetSetName ( ) { java . lang . String expectedName = "expectedName" ; com . microsoft . windowsazure . services . media . models . ContentKeyAuthorizationPolicyOptionInfo contentKeyAuthorizationPolicyOptionInfo = new com . microsoft . windowsazure . services . media . models . ContentKeyAuthorizationPolicyOptionInfo ( null , new com . microsoft . windowsazure . services . media . implementation . content . ContentKeyAuthorizationPolicyOptionType ( ) . setName ( expectedName ) ) ; java . lang . String actualName = contentKeyAuthorizationPolicyOptionInfo . getName ( ) ; "<AssertPlaceHolder>" ; } getName ( ) { return name ; } | org . junit . Assert . assertEquals ( expectedName , actualName ) |
testHandelingIsNietVerwerktEnLockKanWordenGeplaatst ( ) { final boolean resultaat = administratieveHandelingLockRepository . vergrendelAlsNogNietIsVerwerkt ( nl . bzk . brp . levering . dataaccess . repository . lezenschrijven . AdministratieveHandelingVergrendelRepositoryTest . ADMINISTRATIEVE_HANDELING_ID ) ; "<AssertPlaceHolder>" ; } vergrendelAlsNogNietIsVerwerkt ( java . lang . Long ) { if ( isNietVerwerktEnVergrendelingGeplaatst ( administratieveHandelingId ) ) { return true ; } return false ; } | org . junit . Assert . assertTrue ( resultaat ) |
testFormatSqsQueueName_ReplacesNonAlphaNumeric ( ) { java . lang . String expected = "prod-foo-bar" ; java . lang . String actual = com . streamreduce . util . SqsQueueNameFormatter . formatSqsQueueName ( "foo*bar" , "prod" ) ; "<AssertPlaceHolder>" ; } formatSqsQueueName ( java . lang . String , java . lang . String ) { if ( ( org . apache . commons . lang . StringUtils . isBlank ( originalQueueName ) ) || ( org . apache . commons . lang . StringUtils . isBlank ( environmentPrefix ) ) ) { throw new java . lang . IllegalArgumentException ( "queueName<sp>and<sp>environmentPrefix<sp>must<sp>be<sp>non-blank" ) ; } java . lang . String queueNameWithPrefix = ( ( com . streamreduce . util . SqsQueueNameFormatter . addMachineNameToPrefixIfNeeded ( environmentPrefix ) ) + "-" ) + originalQueueName ; java . lang . String modifiedQueueName = queueNameWithPrefix . trim ( ) . replaceAll ( "[^a-zA-Z1-9_-]" , "-" ) ; return org . apache . commons . lang . StringUtils . substring ( modifiedQueueName , 0 , com . streamreduce . util . SqsQueueNameFormatter . MAX_LENGTH_OF_SQS_QUEUE ) ; } | org . junit . Assert . assertEquals ( expected , actual ) |
test_createBuildExportHandler_nullVersion ( ) { java . util . Map < java . lang . String , java . lang . Object > parameters = new java . util . HashMap < java . lang . String , java . lang . Object > ( ) ; parameters . put ( IBuildParametes . ITEM , PropertiesFactory . eINSTANCE . createProcessItem ( ) ) ; org . talend . core . runtime . repository . build . RepositoryObjectTypeBuildProvider provider = createTestBuildProvider ( ) ; "<AssertPlaceHolder>" ; } createBuildExportHandler ( java . util . Map ) { if ( ( parameters == null ) || ( parameters . isEmpty ( ) ) ) { return null ; } final java . lang . Object item = parameters . get ( org . talend . designer . core . build . ITEM ) ; if ( ( item == null ) || ( ! ( item instanceof org . talend . core . model . properties . ProcessItem ) ) ) { return null ; } final java . lang . Object version = parameters . get ( org . talend . designer . core . build . VERSION ) ; if ( version == null ) { return null ; } final java . lang . Object contextGroup = parameters . get ( org . talend . designer . core . build . CONTEXT_GROUP ) ; if ( contextGroup == null ) { return null ; } java . lang . Object choiceOption = parameters . get ( org . talend . designer . core . build . CHOICE_OPTION ) ; if ( choiceOption == null ) { choiceOption = java . util . Collections . emptyMap ( ) ; } if ( ! ( choiceOption instanceof java . util . Map ) ) { return null ; } org . talend . core . runtime . process . IBuildJobHandler buildHandler = new org . talend . repository . ui . wizards . exportjob . handler . BuildOSGiBundleHandler ( ( ( org . talend . core . model . properties . ProcessItem ) ( item ) ) , version . toString ( ) , contextGroup . toString ( ) , ( ( java . util . Map < org . talend . repository . ui . wizards . exportjob . scriptsmanager . JobScriptsManager . ExportChoice , java . lang . Object > ) ( choiceOption ) ) ) ; return buildHandler ; } | org . junit . Assert . assertNull ( provider . createBuildExportHandler ( parameters ) ) |
testExtraEndpoint ( ) { java . lang . String extraEndpoint = ( "http://localhost:" + ( port ) ) + "/rest/extra" ; org . kie . server . common . rest . KieServerHttpRequest httpRequest = org . kie . server . common . rest . KieServerHttpRequest . newRequest ( extraEndpoint , user , password ) . followRedirects ( true ) . timeout ( 1000 ) . contentType ( "application/json" ) . accept ( "application/json" ) ; httpRequest . get ( ) ; org . kie . server . common . rest . KieServerHttpResponse response = httpRequest . response ( ) ; int responseCode = response . code ( ) ; "<AssertPlaceHolder>" ; } get ( ) { return result ; } | org . junit . Assert . assertEquals ( 200 , responseCode ) |
testOnPistonExtendFlagSetOffIsland ( ) { when ( island . onIsland ( org . mockito . Mockito . any ( ) ) ) . thenReturn ( false ) ; org . bukkit . event . block . BlockPistonExtendEvent e = new org . bukkit . event . block . BlockPistonExtendEvent ( block , blocks , org . bukkit . block . BlockFace . EAST ) ; new world . bentobox . bentobox . listeners . flags . worldsettings . PistonPushListener ( ) . onPistonExtend ( e ) ; "<AssertPlaceHolder>" ; } isCancelled ( ) { return cancelled ; } | org . junit . Assert . assertTrue ( e . isCancelled ( ) ) |
isPoint_overEdgeEntries_resultIsFalse ( ) { drawingMotion . addPoint ( 0 , 0 ) ; drawingMotion . addPoint ( java . lang . Math . nextUp ( DrawingMotion . POINT_TOLERANCE ) , 0 ) ; drawingMotion . addPoint ( ( - ( java . lang . Math . nextUp ( DrawingMotion . POINT_TOLERANCE ) ) ) , 0 ) ; drawingMotion . addPoint ( 0 , java . lang . Math . nextUp ( DrawingMotion . POINT_TOLERANCE ) ) ; drawingMotion . addPoint ( 0 , ( - ( java . lang . Math . nextUp ( DrawingMotion . POINT_TOLERANCE ) ) ) ) ; "<AssertPlaceHolder>" ; } isPoint ( ) { if ( ( this . getPathSize ( ) ) == 0 ) { return false ; } for ( io . github . data4all . model . drawing . Point p : points ) { if ( ( io . github . data4all . model . drawing . DrawingMotion . delta ( this . getStart ( ) , p ) ) > ( io . github . data4all . model . drawing . DrawingMotion . POINT_TOLERANCE ) ) { return false ; } } return true ; } | org . junit . Assert . assertThat ( drawingMotion . isPoint ( ) , org . hamcrest . CoreMatchers . is ( false ) ) |
testConvertEmlContributorUserIdWithMissingSchemeURI ( ) { org . gbif . metadata . eml . UserId userId1 = new org . gbif . metadata . eml . UserId ( "" , "0000-0099-6824-9999" ) ; "<AssertPlaceHolder>" ; } convertEmlUserIdIntoContributorNameIdentifier ( org . gbif . metadata . eml . UserId ) { if ( ( ! ( com . google . common . base . Strings . isNullOrEmpty ( userId . getIdentifier ( ) ) ) ) && ( ! ( com . google . common . base . Strings . isNullOrEmpty ( userId . getDirectory ( ) ) ) ) ) { java . lang . String directory = com . google . common . base . Strings . nullToEmpty ( userId . getDirectory ( ) ) . toLowerCase ( ) ; if ( ( directory . contains ( org . gbif . ipt . utils . DataCiteMetadataBuilder . ORCID_NAME_IDENTIFIER_SCHEME . toLowerCase ( ) ) ) || ( directory . contains ( org . gbif . ipt . utils . DataCiteMetadataBuilder . RESEARCHERID_NAME_IDENTIFIER_SCHEME . toLowerCase ( ) ) ) ) { org . gbif . doi . metadata . datacite . DataCiteMetadata . Contributors . Contributor . NameIdentifier nid = org . gbif . ipt . utils . DataCiteMetadataBuilder . FACTORY . createDataCiteMetadataContributorsContributorNameIdentifier ( ) ; nid . setValue ( userId . getIdentifier ( ) ) ; nid . setSchemeURI ( userId . getDirectory ( ) ) ; nid . setNameIdentifierScheme ( ( directory . contains ( org . gbif . ipt . utils . DataCiteMetadataBuilder . ORCID_NAME_IDENTIFIER_SCHEME . toLowerCase ( ) ) ? org . gbif . ipt . utils . DataCiteMetadataBuilder . ORCID_NAME_IDENTIFIER_SCHEME : org . gbif . ipt . utils . DataCiteMetadataBuilder . RESEARCHERID_NAME_IDENTIFIER_SCHEME ) ) ; return nid ; } else { org . gbif . ipt . utils . DataCiteMetadataBuilder . LOG . debug ( ( ( "UserId<sp>has<sp>unrecognized<sp>directory<sp>(" + directory ) + "),<sp>only<sp>ORCID<sp>and<sp>ResearcherID<sp>are<sp>supported" ) ) ; return null ; } } return null ; } | org . junit . Assert . assertNull ( org . gbif . ipt . utils . DataCiteMetadataBuilder . convertEmlUserIdIntoContributorNameIdentifier ( userId1 ) ) |
testManyFalseFilters ( ) { com . northconcepts . eventbus . EventFilter filter = new com . northconcepts . eventbus . filter . AndFilter ( FALSE_FILTER , FALSE_FILTER , FALSE_FILTER ) ; "<AssertPlaceHolder>" ; } allow ( com . northconcepts . eventbus . Event , java . lang . Object ) { return true ; } | org . junit . Assert . assertFalse ( filter . allow ( event , listener ) ) |
testFetchByPrimaryKeysWithNoPrimaryKeys ( ) { java . util . Set < java . io . Serializable > primaryKeys = new java . util . HashSet < java . io . Serializable > ( ) ; java . util . Map < java . io . Serializable , com . liferay . polls . model . PollsChoice > pollsChoices = _persistence . fetchByPrimaryKeys ( primaryKeys ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return _portalCacheListeners . isEmpty ( ) ; } | org . junit . Assert . assertTrue ( pollsChoices . isEmpty ( ) ) |
testOperationIdResolver ( ) { final com . github . phiz71 . vertx . swagger . router . ServiceIdResolver resolver = new com . github . phiz71 . vertx . swagger . router . OperationIdServiceIdResolver ( ) ; final java . lang . String serviceId = resolver . resolve ( HttpMethod . GET , "/some/path" , new io . swagger . models . Operation ( ) . operationId ( "testOperationIdResolver" ) ) ; "<AssertPlaceHolder>" ; } resolve ( io . swagger . models . HttpMethod , java . lang . String , io . swagger . models . Operation ) { return ( httpMethod . name ( ) ) + ( pathname . replaceAll ( "-" , "_" ) . replaceAll ( "/" , "_" ) . replaceAll ( "[{}]" , "" ) ) ; } | org . junit . Assert . assertEquals ( "testOperationIdResolver" , serviceId ) |
getLoad ( ) { org . jboss . modcluster . container . Engine engine = mock ( org . jboss . modcluster . container . Engine . class ) ; org . jboss . modcluster . container . Connector connector = mock ( org . jboss . modcluster . container . Connector . class ) ; org . jboss . modcluster . load . metric . impl . DeterministicLoadState state = mock ( org . jboss . modcluster . load . metric . impl . DeterministicLoadState . class ) ; org . jboss . modcluster . load . metric . LoadMetric metric = new org . jboss . modcluster . load . metric . impl . RequestCountLoadMetric ( state ) ; when ( engine . getConnectors ( ) ) . thenReturn ( java . util . Collections . singleton ( connector ) ) ; when ( connector . getRequestCount ( ) ) . thenReturn ( 10L ) ; when ( state . delta ( 10.0 ) ) . thenReturn ( 9.0 ) ; double load = metric . getLoad ( engine ) ; "<AssertPlaceHolder>" ; } getLoad ( org . jboss . modcluster . container . Engine ) { double load = this . bean . getSystemLoadAverage ( ) ; if ( load < 0 ) { ModClusterLogger . LOGGER . notSupportedOnSystem ( this . getClass ( ) . getSimpleName ( ) ) ; this . setWeight ( 0 ) ; return 0 ; } return load / ( this . bean . getAvailableProcessors ( ) ) ; } | org . junit . Assert . assertEquals ( 9.0 , load , 0.0 ) |
testBounceString ( ) { "<AssertPlaceHolder>" ; } bounce_string ( java . lang . String ) { return value ; } | org . junit . Assert . assertEquals ( line , server . bounce_string ( line ) ) |
testUserSets ( ) { java . lang . String code = "" ; try { code = _setupTestUser ( true ) ; _checkUserIntoDb ( code ) ; } catch ( java . lang . Exception e ) { e . printStackTrace ( ) ; "<AssertPlaceHolder>" ; } return ; } _checkUserIntoDb ( java . lang . String ) { org . isf . menu . model . User foundUser ; foundUser = ( ( org . isf . menu . model . User ) ( org . isf . menu . test . Tests . jpa . find ( org . isf . menu . model . User . class , code ) ) ) ; org . isf . menu . test . Tests . testUser . check ( foundUser ) ; return ; } | org . junit . Assert . assertEquals ( true , false ) |
testMapCache ( ) { org . opentripplanner . model . FareRule result1 = subject . map ( org . opentripplanner . gtfs . mapping . FareRuleMapperTest . FARE_RULE ) ; org . opentripplanner . model . FareRule result2 = subject . map ( org . opentripplanner . gtfs . mapping . FareRuleMapperTest . FARE_RULE ) ; "<AssertPlaceHolder>" ; } map ( org . onebusaway . gtfs . model . StopTime ) { return orginal == null ? null : mappedStopTimes . computeIfAbsent ( orginal , this :: doMap ) ; } | org . junit . Assert . assertTrue ( ( result1 == result2 ) ) |
testListMessagesResultNoMessagesListed ( ) { try { final java . lang . String listResult = "00003" + "410" ; final java . lang . String listMessagesResult = listResult + ( nl . bzk . migratiebrp . voisc . mailbox . client . impl . MailboxServerProxyTest . TERMINATOR ) ; org . mockito . Mockito . when ( mbsConnection . read ( ) ) . thenReturn ( listMessagesResult . getBytes ( ) ) ; int nextMSSequenceNumber = mbsProxy . listMessages ( mbsConnection , 0 , new java . util . ArrayList ( ) , nl . bzk . migratiebrp . voisc . mailbox . client . impl . MailboxServerProxyTest . MAX_LIST_LIMIT_NR , nl . bzk . migratiebrp . voisc . mailbox . client . impl . MailboxServerProxyTest . MSSTATUS , nl . bzk . migratiebrp . voisc . mailbox . client . impl . MailboxServerProxyTest . PRIORITY ) ; "<AssertPlaceHolder>" ; } catch ( final java . lang . Exception e ) { org . junit . Assert . fail ( ( ( nl . bzk . migratiebrp . voisc . mailbox . client . impl . MailboxServerProxyTest . MELDING_GEEN_EXCEPTION_VERWACHT ) + ( e . getMessage ( ) ) ) ) ; } } listMessages ( nl . bzk . migratiebrp . voisc . mailbox . client . Connection , int , java . util . List , int , java . lang . String , nl . bzk . migratiebrp . voisc . spd . SpdConstants$Priority ) { final nl . bzk . migratiebrp . voisc . spd . Operation request = new nl . bzk . migratiebrp . voisc . spd . Operation . Builder ( ) . add ( new nl . bzk . migratiebrp . voisc . spd . ListMessages . Builder ( ) . limit ( listLimitNr ) . status ( msStatus ) . priority ( prio ) . fromMSSequenceNumber ( nextSequenceNr ) . build ( ) ) . build ( ) ; sendOperation ( connection , request ) ; int nextMSSequenceNr = 0 ; nl . bzk . migratiebrp . voisc . spd . exception . VoaException voaException = null ; try { final nl . bzk . migratiebrp . voisc . spd . Operation response = receiveOperation ( connection ) ; for ( final nl . bzk . migratiebrp . voisc . spd . OperationRecord record : response . records ( ) ) { if ( record instanceof nl . bzk . migratiebrp . voisc . spd . ListMessagesConfirmation ) { voaException = processListMessageConfirmation ( ( ( nl . bzk . migratiebrp . voisc . spd . ListMessagesConfirmation ) ( record ) ) ) ; } else if ( record instanceof nl . bzk . migratiebrp . voisc . spd . ListResult ) { nextMSSequenceNr = processListResult ( ( ( nl . bzk . migratiebrp . voisc . spd . ListResult ) ( record ) ) ) ; } else if ( record instanceof nl . bzk . migratiebrp . voisc . spd . MSList ) { procesMSList ( sequenceNumbers , ( ( nl . bzk . migratiebrp . voisc . spd . MSList ) ( record ) ) ) ; } else { voaException = new nl . bzk . migratiebrp . voisc . spd . exception . SpdProtocolException ( nl . bzk . migratiebrp . voisc . spd . exception . MessagesCodes . ERRMSG_VOSPG_SPD_LISTMESSAGES_CONF_OPCODE , new java . lang . Object [ ] { record . operationCode ( ) } ) ; } } } catch ( final java . lang . NumberFormatException ex ) { voaException = new nl . bzk . migratiebrp . voisc . spd . exception . SpdProtocolException ( nl . bzk . migratiebrp . voisc . spd . exception . MessagesCodes . ERRMSG_VOSPG_SPD_NOT_NUMERIC_FIELD , null , ex ) ; } catch ( final nl . bzk . migratiebrp . voisc . spd . exception . UnknownOpcodeException ex ) { voaException = new nl . bzk . migratiebrp . voisc . spd . exception . SpdProtocolException ( nl . bzk . migratiebrp . voisc . spd . exception . MessagesCodes . ERRMSG_VOSPG_SPD_LISTMESSAGES_CONF_OPCODE , new java . lang . Object [ ] { ex . getOpcode ( ) } , ex ) ; } catch ( java . lang . IllegalArgumentException | nl . bzk . migratiebrp . voisc . spd . exception . ParseException ex ) { voaException = new nl . bzk . migratiebrp . voisc . spd . exception . SpdProtocolException ( nl . bzk . migratiebrp . voisc . spd . exception . MessagesCodes . ERRMSG_VOSPG_SPD_RCV_LISTMESSAGES , null , ex ) ; } if ( voaException != null ) { throw voaException ; } return nextMSSequenceNr ; } | org . junit . Assert . assertEquals ( 0 , nextMSSequenceNumber ) |
in_memory_configuration_should_use_targeted_instance ( ) { org . neo4j . graphdb . GraphDatabaseService graphDatabaseService1 = mock ( org . neo4j . graphdb . GraphDatabaseService . class ) ; org . neo4j . graphdb . GraphDatabaseService graphDatabaseService2 = mock ( org . neo4j . graphdb . GraphDatabaseService . class ) ; com . lordofthejars . nosqlunit . neo4j . EmbeddedNeo4jInstances . getInstance ( ) . addGraphDatabaseService ( graphDatabaseService1 , "a" ) ; com . lordofthejars . nosqlunit . neo4j . EmbeddedNeo4jInstances . getInstance ( ) . addGraphDatabaseService ( graphDatabaseService2 , "b" ) ; com . lordofthejars . nosqlunit . neo4j . EmbeddedNeoServerConfigurationBuilder embeddedNeoServerConfiguration = com . lordofthejars . nosqlunit . neo4j . EmbeddedNeoServerConfigurationBuilder . newEmbeddedNeoServerConfiguration ( ) ; com . lordofthejars . nosqlunit . neo4j . Neo4jConfiguration embeddedConfiguration = embeddedNeoServerConfiguration . buildFromTargetPath ( "b" ) ; com . lordofthejars . nosqlunit . neo4j . EmbeddedNeo4jInstances . getInstance ( ) . removeGraphDatabaseService ( "a" ) ; com . lordofthejars . nosqlunit . neo4j . EmbeddedNeo4jInstances . getInstance ( ) . removeGraphDatabaseService ( "b" ) ; "<AssertPlaceHolder>" ; } getGraphDatabaseService ( ) { return graphDatabaseService ; } | org . junit . Assert . assertThat ( embeddedConfiguration . getGraphDatabaseService ( ) , org . hamcrest . CoreMatchers . is ( graphDatabaseService2 ) ) |
getCompressionTypeTarFromMetadataSuccess ( ) { when ( obj . getKey ( ) ) . thenReturn ( ( ( ( ( "A" + ( java . io . File . separator ) ) + "File" ) + ( java . io . File . separator ) ) + "XK321K" ) ) ; when ( metadata . getContentType ( ) ) . thenReturn ( "application/tar" ) ; final com . amazonaws . codepipeline . jenkinsplugin . CodePipelineStateModel . CompressionType compressionType = com . amazonaws . codepipeline . jenkinsplugin . ExtractionTools . getCompressionType ( obj , null ) ; "<AssertPlaceHolder>" ; } getCompressionType ( com . amazonaws . services . s3 . model . S3Object , hudson . model . TaskListener ) { final java . lang . String key = sessionObject . getKey ( ) ; com . amazonaws . codepipeline . jenkinsplugin . CodePipelineStateModel . CompressionType compressionType = com . amazonaws . codepipeline . jenkinsplugin . CodePipelineStateModel . CompressionType . None ; if ( com . amazonaws . codepipeline . jenkinsplugin . ExtractionTools . endsWithLowerCase ( key , ".zip" ) ) { compressionType = com . amazonaws . codepipeline . jenkinsplugin . CodePipelineStateModel . CompressionType . Zip ; } else if ( com . amazonaws . codepipeline . jenkinsplugin . ExtractionTools . endsWithLowerCase ( key , ".tar.gz" ) ) { compressionType = com . amazonaws . codepipeline . jenkinsplugin . CodePipelineStateModel . CompressionType . TarGz ; } else if ( com . amazonaws . codepipeline . jenkinsplugin . ExtractionTools . endsWithLowerCase ( key , ".tar" ) ) { compressionType = com . amazonaws . codepipeline . jenkinsplugin . CodePipelineStateModel . CompressionType . Tar ; } if ( compressionType == ( com . amazonaws . codepipeline . jenkinsplugin . CodePipelineStateModel . CompressionType . None ) ) { final java . lang . String contentType = sessionObject . getObjectMetadata ( ) . getContentType ( ) ; if ( "application/zip" . equalsIgnoreCase ( contentType ) ) { compressionType = com . amazonaws . codepipeline . jenkinsplugin . CodePipelineStateModel . CompressionType . Zip ; } else if ( ( "application/gzip" . equalsIgnoreCase ( contentType ) ) || ( "application/x-gzip" . equalsIgnoreCase ( contentType ) ) ) { compressionType = com . amazonaws . codepipeline . jenkinsplugin . CodePipelineStateModel . CompressionType . TarGz ; } else if ( ( "application/tar" . equalsIgnoreCase ( contentType ) ) || ( "application/x-tar" . equalsIgnoreCase ( contentType ) ) ) { compressionType = com . amazonaws . codepipeline . jenkinsplugin . CodePipelineStateModel . CompressionType . Tar ; } } com . amazonaws . codepipeline . jenkinsplugin . LoggingHelper . log ( l , "Detected<sp>compression<sp>type:<sp>%s" , compressionType . name ( ) ) ; return compressionType ; } | org . junit . Assert . assertEquals ( CompressionType . Tar , compressionType ) |
testAllocateAndSuccessorSize ( ) { final tlc2 . tool . liveness . GraphNode node = new tlc2 . tool . liveness . GraphNode ( 0 , 0 ) ; node . addTransition ( 0 , 0 , 0 , 0 , null , 0 , 100 ) ; "<AssertPlaceHolder>" ; } succSize ( ) { if ( ( this . offset ) != ( tlc2 . tool . liveness . GraphNode . NO_FREE_SLOTS ) ) { return ( this . offset ) / ( tlc2 . tool . liveness . GraphNode . NNODE_RECORD_SIZE ) ; } return ( this . nnodes . length ) / ( tlc2 . tool . liveness . GraphNode . NNODE_RECORD_SIZE ) ; } | org . junit . Assert . assertEquals ( 1 , node . succSize ( ) ) |
shouldNotHandleRequest_noBearerValue ( ) { io . gravitee . common . http . HttpHeaders headers = new io . gravitee . common . http . HttpHeaders ( ) ; io . gravitee . gateway . api . Request request = mock ( io . gravitee . gateway . api . Request . class ) ; when ( request . headers ( ) ) . thenReturn ( headers ) ; headers . add ( HttpHeaders . AUTHORIZATION , ( ( OAuth2AuthenticationHandler . BEARER_AUTHORIZATION_TYPE ) + "<sp>" ) ) ; boolean handle = authenticationHandler . canHandle ( request ) ; "<AssertPlaceHolder>" ; } canHandle ( io . gravitee . gateway . repository . plugins . Plugin ) { return ( plugin . type ( ) ) == ( PluginType . REPOSITORY ) ; } | org . junit . Assert . assertFalse ( handle ) |
shouldNotTransferFromUnknownAccount ( ) { try { accountService . transfer ( ( - 1 ) , 2 , 50 ) ; entityManager . flush ( ) ; org . junit . Assert . fail ( "Expected<sp>UnknownAccountException" ) ; } catch ( com . jayway . service . UnknownAccountException e ) { } int secondBalance = getBalance ( 2 ) ; "<AssertPlaceHolder>" ; } getBalance ( java . lang . Integer ) { return jdbcTemplate . queryForObject ( "SELECT<sp>balance<sp>FROM<sp>account_t<sp>WHERE<sp>account_number<sp>=<sp>?" , com . jayway . repository . Integer . class , accountNumber ) ; } | org . junit . Assert . assertThat ( secondBalance , org . hamcrest . CoreMatchers . is ( 200 ) ) |
calculateListWithNulls ( ) { cucumber . runtime . formatter . UsageFormatter . AverageUsageStatisticStrategy averageUsageStatisticStrategy = new cucumber . runtime . formatter . UsageFormatter . AverageUsageStatisticStrategy ( ) ; java . lang . Long result = averageUsageStatisticStrategy . calculate ( java . util . Arrays . asList ( 3L , null ) ) ; "<AssertPlaceHolder>" ; } | org . junit . Assert . assertEquals ( result , java . lang . Long . valueOf ( 0 ) ) |
testParseString ( ) { java . math . BigInteger bigInteger = new java . math . BigInteger ( "200" ) ; "<AssertPlaceHolder>" ; } parseString ( java . math . BigInteger ) { return value . toString ( ) ; } | org . junit . Assert . assertEquals ( handler . parseString ( bigInteger ) , "200" ) |
createInnerNode_elementNonNull_canRetrieveElement ( ) { java . lang . String element = "element" ; org . codefx . libfx . collection . tree . stream . SimpleTreeNode < java . lang . String > node = org . codefx . libfx . collection . tree . stream . SimpleTreeNode . innerNode ( element , 0 ) ; "<AssertPlaceHolder>" ; } getElement ( ) { return element ; } | org . junit . Assert . assertSame ( element , node . getElement ( ) ) |
testIsJavaVersionAtLeast ( ) { "<AssertPlaceHolder>" ; } isJavaVersionAtLeast ( int ) { return ( org . apache . hadoop . util . Shell . JAVA_SPEC_VER ) >= version ; } | org . junit . Assert . assertTrue ( org . apache . hadoop . util . Shell . Shell . isJavaVersionAtLeast ( 8 ) ) |
testInitWithBuilder_and_reachable_check ( ) { try ( net . sourceforge . jwbf . JettyServer server = new net . sourceforge . jwbf . JettyServer ( ) . started ( net . sourceforge . jwbf . JettyServer . echoHandler ( ) ) ) { java . lang . String url = server . getTestUrl ( ) ; testee = new net . sourceforge . jwbf . mediawiki . bots . MediaWikiBot ( net . sourceforge . jwbf . JWBF . newURL ( url ) , true ) ; "<AssertPlaceHolder>" ; } catch ( java . lang . Exception e ) { throw new java . lang . IllegalStateException ( e ) ; } } newURL ( java . lang . String ) { try { return new java . net . URL ( url ) ; } catch ( java . net . MalformedURLException e ) { throw new java . lang . IllegalArgumentException ( ( ( "\"" + url ) + "\"" ) , e ) ; } } | org . junit . Assert . assertNotNull ( testee ) |
testGetSupportedIIIF1Qualities ( ) { edu . illinois . library . cantaloupe . processor . Processor proc = newInstance ( ) ; proc . setSourceFormat ( getAnySupportedSourceFormat ( proc ) ) ; java . util . Set < edu . illinois . library . cantaloupe . resource . iiif . v1 . Quality > expectedQualities = java . util . EnumSet . of ( edu . illinois . library . cantaloupe . processor . edu . illinois . library . cantaloupe . resource . iiif . v1 . Quality , edu . illinois . library . cantaloupe . processor . edu . illinois . library . cantaloupe . resource . iiif . v1 . Quality , edu . illinois . library . cantaloupe . processor . edu . illinois . library . cantaloupe . resource . iiif . v1 . Quality , edu . illinois . library . cantaloupe . processor . edu . illinois . library . cantaloupe . resource . iiif . v1 . Quality ) ; "<AssertPlaceHolder>" ; } getSupportedIIIF1Qualities ( ) { return new java . util . HashSet ( ) ; } | org . junit . Assert . assertEquals ( expectedQualities , proc . getSupportedIIIF1Qualities ( ) ) |
testCreatePastMedicalHistorySectionConsultBuilder ( ) { org . openhealthtools . mdht . uml . cda . builder . SectionBuilder < org . openhealthtools . mdht . uml . cda . cdt . PastMedicalHistorySection > sectionBuilder = org . openhealthtools . mdht . uml . cda . cdt . builder . CDTBuilderFactory . createPastMedicalHistorySectionBuilder ( ) ; org . openhealthtools . mdht . uml . cda . cdt . PastMedicalHistorySection section = sectionBuilder . buildSection ( ) ; "<AssertPlaceHolder>" ; Diagnostician . INSTANCE . validate ( section ) ; org . openhealthtools . mdht . uml . cda . util . CDAUtil . saveSnippet ( section , System . out ) ; } buildSection ( ) { org . openhealthtools . mdht . uml . cda . Section section = CDAFactory . eINSTANCE . createSection ( ) ; construct ( section ) ; return section ; } | org . junit . Assert . assertNotNull ( section ) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.