input
stringlengths 28
18.7k
| output
stringlengths 39
1.69k
|
|---|---|
testGetIntLE ( ) { for ( int i = 0 ; i < ( ( bytes . length ) - 3 ) ; i ++ ) { int expected = byteBufferLE . getInt ( i ) ; int value = primitiveArrayUtils . getIntLE ( bytes , i ) ; "<AssertPlaceHolder>" ; } }
|
org . junit . Assert . assertEquals ( expected , value )
|
testTotalItems ( ) { int totalItems = com . liferay . talend . runtime . apio . jsonld . ApioResourceCollectionTest . _apioJsonLDResource . getTotalItems ( ) ; "<AssertPlaceHolder>" ; } getTotalItems ( ) { java . util . List < com . liferay . adaptive . media . image . configuration . AMImageConfigurationEntry > selectedConfigurationEntries = getSelectedConfigurationEntries ( ) ; return selectedConfigurationEntries . size ( ) ; }
|
org . junit . Assert . assertThat ( totalItems , org . hamcrest . CoreMatchers . equalTo ( 10 ) )
|
testToCollectionWithResultsBag ( ) { org . apache . geode . cache . query . internal . ResultsBag mockResultsBag = org . mockito . Mockito . mock ( org . apache . geode . cache . query . internal . ResultsBag . class , "testToCollectionWithResultsBag.ResultsBag" ) ; java . util . List < java . lang . String > expectedList = java . util . Arrays . asList ( "a" , "b" , "c" ) ; org . mockito . Mockito . when ( mockResultsBag . asList ( ) ) . thenReturn ( expectedList ) ; java . util . Collection < ? > actualList = repositoryQuery . toCollection ( mockResultsBag ) ; "<AssertPlaceHolder>" ; } toCollection ( java . lang . Object ) { if ( source instanceof org . apache . geode . cache . query . SelectResults ) { return ( ( org . apache . geode . cache . query . SelectResults ) ( source ) ) . asList ( ) ; } if ( source instanceof java . util . Collection ) { return ( ( java . util . Collection < ? > ) ( source ) ) ; } if ( source == null ) { return java . util . Collections . emptyList ( ) ; } return source . getClass ( ) . isArray ( ) ? org . springframework . util . CollectionUtils . arrayToList ( source ) : java . util . Collections . singletonList ( source ) ; }
|
org . junit . Assert . assertSame ( expectedList , actualList )
|
testGetIdentifiable ( ) { unwrap ( service . put ( "key" , "value" ) ) ; com . google . appengine . api . memcache . MemcacheService . IdentifiableValue identifiable = unwrap ( service . getIdentifiable ( "key" ) ) ; "<AssertPlaceHolder>" ; } unwrap ( java . util . concurrent . Future ) { try { return f . get ( ) ; } catch ( java . lang . InterruptedException e ) { java . lang . Thread . currentThread ( ) . interrupt ( ) ; throw new java . lang . IllegalStateException ( e ) ; } catch ( java . util . concurrent . ExecutionException e ) { throw new java . lang . IllegalStateException ( e ) ; } }
|
org . junit . Assert . assertEquals ( "value" , identifiable . getValue ( ) )
|
testMinWithRangeBeginningEarlierThanTable ( ) { int TIME_BASELINE = ( ( int ) ( ( new java . util . GregorianCalendar ( 2014 , 10 , 9 , 23 , 0 , 0 ) . getTime ( ) . getTime ( ) ) / 1000 ) ) ; int TIME_LIMIT = ( ( int ) ( ( new java . util . GregorianCalendar ( 2014 , 10 , 10 , 2 , 0 , 0 ) . getTime ( ) . getTime ( ) ) / 1000 ) ) ; org . apache . hadoop . hbase . coprocessor . client . TimeseriesAggregationClient aClient = new org . apache . hadoop . hbase . coprocessor . client . TimeseriesAggregationClient ( org . apache . hadoop . hbase . coprocessor . TestTimeseriesAggregateProtocol . conf , 900 , TIME_BASELINE , TIME_LIMIT , org . apache . hadoop . hbase . coprocessor . TestTimeseriesAggregateProtocol . KEY_FILTER_PATTERN ) ; org . apache . hadoop . hbase . client . Scan scan = new org . apache . hadoop . hbase . client . Scan ( ) ; scan . addFamily ( org . apache . hadoop . hbase . coprocessor . TestTimeseriesAggregateProtocol . TEST_FAMILY ) ; final org . apache . hadoop . hbase . coprocessor . ColumnInterpreter < java . lang . Long , java . lang . Long , org . apache . hadoop . hbase . protobuf . generated . HBaseProtos . EmptyMsg , org . apache . hadoop . hbase . protobuf . generated . HBaseProtos . LongMsg , org . apache . hadoop . hbase . protobuf . generated . HBaseProtos . LongMsg > ci = new org . apache . hadoop . hbase . client . coprocessor . LongColumnInterpreter ( ) ; java . util . Map < java . lang . Long , java . lang . Long > results = new java . util . concurrent . ConcurrentSkipListMap < java . lang . Long , java . lang . Long > ( ) ; results . put ( 1415574000000L , 0L ) ; results . put ( 1415574900000L , 25L ) ; results . put ( 1415575800000L , 50L ) ; results . put ( 1415576700000L , 75L ) ; results . put ( 1415577600000L , 0L ) ; results . put ( 1415578500000L , 25L ) ; results . put ( 1415579400000L , 50L ) ; results . put ( 1415580300000L , 75L ) ; results . put ( 1415581200000L , 0L ) ; java . util . concurrent . ConcurrentSkipListMap < java . lang . Long , java . lang . Long > minimum = aClient . min ( org . apache . hadoop . hbase . coprocessor . TestTimeseriesAggregateProtocol . TEST_TABLE , ci , scan ) ; "<AssertPlaceHolder>" ; aClient . close ( ) ; } min ( org . apache . hadoop . hbase . client . Table , org . apache . hadoop . hbase . coprocessor . ColumnInterpreter , org . apache . hadoop . hbase . client . Scan ) { final org . apache . hadoop . hbase . protobuf . generated . TimeseriesAggregateProtos . TimeseriesAggregateRequest requestArg = validateArgAndGetPB ( scan , ci , false , intervalSeconds , timestampSecondsMin , timestampSecondsMax , keyFilterPattern ) ; class MinCallBack implements org . apache . hadoop . hbase . client . coprocessor . Batch . Callback < org . apache . hadoop . hbase . protobuf . generated . TimeseriesAggregateProtos . TimeseriesAggregateResponse > { java . util . concurrent . ConcurrentSkipListMap < java . lang . Long , R > min = new java . util . concurrent . ConcurrentSkipListMap < java . lang . Long , R > ( ) ; java . util . concurrent . ConcurrentSkipListMap < java . lang . Long , R > getMin ( ) { return min ; } @ org . apache . hadoop . hbase . coprocessor . client . Override public synchronized void update ( byte [ ] region , byte [ ] row , org . apache . hadoop . hbase . protobuf . generated . TimeseriesAggregateProtos . TimeseriesAggregateResponse result ) { java . util . List < org . apache . hadoop . hbase . protobuf . generated . TimeseriesAggregateProtos . TimeseriesAggregateResponseMapEntry > results = ( ( org . apache . hadoop . hbase . protobuf . generated . TimeseriesAggregateProtos . TimeseriesAggregateResponse ) ( result ) ) . getEntryList ( ) ; for ( org . apache . hadoop . hbase . protobuf . generated . TimeseriesAggregateProtos . TimeseriesAggregateResponseMapEntry entry : results ) { R candidate ; if ( ( entry . getValue ( ) . getFirstPartCount ( ) ) > 0 ) { com . google . protobuf . ByteString b = entry . getValue ( ) . getFirstPart ( 0 ) ; Q q = null ; try { q = org . apache . hadoop . hbase . protobuf . ProtobufUtil . getParsedGenericInstance ( ci . getClass ( ) , 3 , b ) ; } catch ( java . io . IOException e ) { e . printStackTrace ( ) ; } candidate = ci . getCellValueFromProto ( q ) ; if ( null != q ) { if ( min . containsKey ( entry . getKey ( ) ) ) { R current = min . get ( entry . getKey ( ) ) ; min . put ( entry . getKey ( ) , ( ( current == null ) || ( ( candidate != null ) && ( ( ci . compare ( current , candidate ) ) < 0 ) ) ? current : candidate ) ) ; } else { min . put ( entry . getKey ( ) , ci . getCellValueFromProto ( q ) ) ; } } } } } } MinCallBack aMinCallBack = new MinCallBack ( ) ; table . coprocessorService ( org . apache . hadoop . hbase . protobuf . generated . TimeseriesAggregateProtos . TimeseriesAggregateService .
|
org . junit . Assert . assertEquals ( results , minimum )
|
shouldDecodeCommonFlagsLong ( ) { com . couchbase . client . deps . io . netty . buffer . ByteBuf content = com . couchbase . client . deps . io . netty . buffer . Unpooled . copiedBuffer ( "9223372036854775807" , CharsetUtil . UTF_8 ) ; com . couchbase . client . java . document . JsonLongDocument decoded = converter . decode ( "id" , content , 0 , 0 , TranscoderUtils . JSON_COMPAT_FLAGS , ResponseStatus . SUCCESS ) ; "<AssertPlaceHolder>" ; } content ( ) { return content ; }
|
org . junit . Assert . assertEquals ( Long . MAX_VALUE , ( ( long ) ( decoded . content ( ) ) ) )
|
testGetGeometryMetadata ( ) { System . out . println ( ( ( ( com . telefonica . iot . cygnus . utils . CommonUtilsForTests . getTestTraceHead ( "[Utils.getLocation]" ) ) + "--------<sp>When<sp>getting<sp>a<sp>geometry,<sp>a<sp>CartoDB<sp>point<sp>is<sp>obtained<sp>when<sp>passing<sp>an<sp>attribute<sp>" ) + "ST_SetSRID(ST_MakePoint(-3.7167,40.3833),<sp>4326)" 2 ) ) ; org . json . simple . JSONObject metadataJson = new org . json . simple . JSONObject ( ) ; metadataJson . put ( "name" , "location" ) ; metadataJson . put ( "ST_SetSRID(ST_MakePoint(-3.7167,40.3833),<sp>4326)" 1 , "ST_SetSRID(ST_MakePoint(-3.7167,40.3833),<sp>4326)" 5 ) ; metadataJson . put ( "value" , "WGS84" ) ; org . json . simple . JSONArray metadatasJson = new org . json . simple . JSONArray ( ) ; metadatasJson . add ( metadataJson ) ; java . lang . String attrMetadataStr = metadatasJson . toJSONString ( ) ; java . lang . String attrValue = "-3.7167,<sp>40.3833" ; java . lang . String attrType = "ST_SetSRID(ST_MakePoint(-3.7167,40.3833),<sp>4326)" 3 ; boolean swapCoordinates = false ; org . apache . commons . lang3 . tuple . ImmutablePair < java . lang . String , java . lang . Boolean > geometry = com . telefonica . iot . cygnus . utils . NGSIUtils . getGeometry ( attrValue , attrType , attrMetadataStr , swapCoordinates ) ; try { "<AssertPlaceHolder>" ; System . out . println ( ( ( ( ( ( ( ( ( com . telefonica . iot . cygnus . utils . CommonUtilsForTests . getTestTraceHead ( "[Utils.getLocation]" ) ) + "ST_SetSRID(ST_MakePoint(-3.7167,40.3833),<sp>4326)" 4 ) + ( geometry . getLeft ( ) ) ) + "'<sp>obtained<sp>for<sp>an<sp>attribute<sp>with<sp>metadata<sp>'" ) + attrMetadataStr ) + "'<sp>and<sp>value<sp>'" ) + attrValue ) + "ST_SetSRID(ST_MakePoint(-3.7167,40.3833),<sp>4326)" 0 ) ) ; } catch ( java . lang . AssertionError e ) { System . out . println ( ( ( ( ( ( ( ( ( com . telefonica . iot . cygnus . utils . CommonUtilsForTests . getTestTraceHead ( "[Utils.getLocation]" ) ) + "ST_SetSRID(ST_MakePoint(-3.7167,40.3833),<sp>4326)" 6 ) + ( geometry . getLeft ( ) ) ) + "'<sp>obtained<sp>for<sp>an<sp>attribute<sp>with<sp>metadata<sp>'" ) + attrMetadataStr ) + "'<sp>and<sp>value<sp>'" ) + attrValue ) + "ST_SetSRID(ST_MakePoint(-3.7167,40.3833),<sp>4326)" 0 ) ) ; throw e ; } getGeometry ( java . lang . String , java . lang . String , java . lang . String , boolean ) { if ( attrType . equals ( "geo:point" ) ) { java . lang . String [ ] split = attrValue . split ( "value" 3 ) ; if ( swapCoordinates ) { return new org . apache . commons . lang3 . tuple . ImmutablePair ( ( ( ( ( "ST_SetSRID(ST_MakePoint(" + ( split [ 1 ] . trim ( ) ) ) + "value" 3 ) + ( split [ 0 ] . trim ( ) ) ) + "),<sp>4326)" ) , true ) ; } else { return new org . apache . commons . lang3 . tuple . ImmutablePair ( ( ( ( ( "ST_SetSRID(ST_MakePoint(" + ( split [ 0 ] . trim ( ) ) ) + "value" 3 ) + ( split [ 1 ] . trim ( ) ) ) + "),<sp>4326)" ) , true ) ; } } if ( attrType . equals ( "value" 2 ) ) { return new org . apache . commons . lang3 . tuple . ImmutablePair ( ( ( "ST_GeomFromGeoJSON(" + attrValue ) + ")" ) , true ) ; } org . json . simple . parser . JSONParser parser = new org . json . simple . parser . JSONParser ( ) ; org . json . simple . JSONArray mds ; try { mds = ( ( org . json . simple . JSONArray ) ( parser . parse ( metadata ) ) ) ; } catch ( org . json . simple . parser . ParseException e ) { com . telefonica . iot . cygnus . utils . NGSIUtils . LOGGER . error ( ( "Error<sp>while<sp>parsing<sp>the<sp>metadata.<sp>Details:<sp>" + ( e . getMessage ( ) ) ) ) ; return new org . apache . commons . lang3 . tuple . ImmutablePair ( attrValue , false ) ; } for ( java . lang . Object mdObject : mds ) { org . json . simple . JSONObject md = ( ( org . json . simple . JSONObject ) ( mdObject ) ) ; java . lang . String mdName = ( ( java . lang . String ) ( md . get ( "name" ) ) ) ; java . lang . String mdType = ( ( java . lang . String ) ( md . get ( "value" 0 ) ) ) ; java . lang . String mdValue = ( ( java . lang . String ) ( md . get ( "value" ) ) ) ; if ( ( ( mdName . equals ( "location" ) ) && ( mdType . equals ( "value" 1 ) ) ) && ( mdValue . equals ( "WGS84" ) ) ) { java . lang . String [ ] split = attrValue . split ( "value" 3 ) ; if ( swapCoordinates ) { return new org . apache . commons . lang3 . tuple . ImmutablePair ( ( ( ( ( "ST_SetSRID(ST_MakePoint(" + ( split [ 1 ] . trim ( ) ) ) + "value" 3 ) + ( split [ 0 ] . trim ( ) ) ) + "),<sp>4326)" ) , true ) ; } else { return new org . apache . commons . lang3 . tuple . ImmutablePair ( ( ( ( ( "ST_SetSRID(ST_MakePoint(" + ( split [ 0 ] . trim ( ) ) ) + "value" 3 ) + ( split [ 1 ] . trim ( ) ) ) + "),<sp>4326)" ) , true ) ; } } } return new org . apache . commons . lang3 . tuple . ImmutablePair ( attrValue , false ) ; }
|
org . junit . Assert . assertEquals ( "ST_SetSRID(ST_MakePoint(-3.7167,40.3833),<sp>4326)" , geometry . getLeft ( ) )
|
simulator_runs_until_all_stories_are_finished ( ) { com . bigvisible . kanbansimulator . Simulator stimuator = new com . bigvisible . kanbansimulator . SimulatorEngine ( ) ; stimuator . addStories ( 88 ) ; stimuator . run ( null ) ; "<AssertPlaceHolder>" ; } getStoriesCompleted ( ) { return storiesCompleted ; }
|
org . junit . Assert . assertEquals ( 88 , stimuator . getStoriesCompleted ( ) )
|
convertNullResultReturnsNull ( ) { org . codehaus . jettison . json . JSONArray results = this . converterNotPaged . convert ( null ) ; "<AssertPlaceHolder>" ; } convert ( java . lang . Object ) { if ( object == null ) { return null ; } else if ( ( ( object instanceof java . lang . String ) || ( object instanceof java . lang . Number ) ) || ( object instanceof java . lang . Boolean ) ) { return object ; } else if ( object instanceof com . tinkerpop . blueprints . Element ) { try { final com . tinkerpop . blueprints . Element element = ( ( com . tinkerpop . blueprints . Element ) ( object ) ) ; final java . util . Set < java . lang . String > propertyKeys = element . getPropertyKeys ( ) ; final boolean isVertex = ! ( element instanceof com . tinkerpop . blueprints . Edge ) ; java . util . HashMap < java . lang . Object , java . lang . Object > outMap = new java . util . HashMap < java . lang . Object , java . lang . Object > ( ) ; outMap . put ( Tokens . _ID , com . tinkerpop . rexster . gremlin . converter . SerializedResultConverter . serializeElementId ( element ) ) ; if ( isVertex ) { outMap . put ( Tokens . _TYPE , Tokens . VERTEX ) ; } else { final com . tinkerpop . blueprints . Edge edge = ( ( com . tinkerpop . blueprints . Edge ) ( element ) ) ; outMap . put ( Tokens . _TYPE , Tokens . EDGE ) ; outMap . put ( Tokens . _IN_V , com . tinkerpop . rexster . gremlin . converter . SerializedResultConverter . serializeElementId ( edge . getVertex ( Direction . IN ) ) ) ; outMap . put ( Tokens . _OUT_V , com . tinkerpop . rexster . gremlin . converter . SerializedResultConverter . serializeElementId ( edge . getVertex ( Direction . OUT ) ) ) ; outMap . put ( Tokens . _LABEL , edge . getLabel ( ) ) ; } if ( ( propertyKeys . size ( ) ) > 0 ) { java . util . HashMap < java . lang . Object , java . lang . Object > propertyMap = new java . util . HashMap < java . lang . Object , java . lang . Object > ( ) ; final java . util . Iterator < java . lang . String > itty = propertyKeys . iterator ( ) ; while ( itty . hasNext ( ) ) { final java . lang . String propertyKey = itty . next ( ) ; propertyMap . put ( propertyKey , com . tinkerpop . rexster . gremlin . converter . SerializedResultConverter . convert ( element . getProperty ( propertyKey ) ) ) ; } outMap . put ( Tokens . _PROPERTIES , propertyMap ) ; } return outMap ; } catch ( java . lang . Exception e ) { return null ; } } else if ( object instanceof java . util . Map ) { final java . util . Map map = ( ( java . util . Map ) ( object ) ) ; java . util . HashMap < java . lang . Object , java . lang . Object > outMap = new java . util . HashMap < java . lang . Object , java . lang . Object > ( ) ; for ( java . lang . Object key : map . keySet ( ) ) { if ( key instanceof com . tinkerpop . blueprints . Element ) { final com . tinkerpop . blueprints . Element element = ( ( com . tinkerpop . blueprints . Element ) ( key ) ) ; final java . util . HashMap < java . lang . String , java . lang . Object > m = new java . util . HashMap < java . lang . String , java . lang . Object > ( ) ; m . put ( Tokens . _KEY , element ) ; m . put ( Tokens . _VALUE , map . get ( key ) ) ; outMap . put ( element . getId ( ) . toString ( ) , com . tinkerpop . rexster . gremlin . converter . SerializedResultConverter . convert ( m ) ) ; } else { outMap . put ( key , com . tinkerpop . rexster . gremlin . converter . SerializedResultConverter . convert ( map . get ( key ) ) ) ; } } return outMap ; } else if ( object instanceof com . tinkerpop . pipes . util . structures . Table ) { final com . tinkerpop . pipes . util . structures . Table table = ( ( com . tinkerpop . pipes . util . structures . Table ) ( object ) ) ; final java . util . Iterator < com . tinkerpop . pipes . util . structures . Row > rows = table . iterator ( ) ; java . util . ArrayList < java . lang . Object > outArray = new java . util . ArrayList < java . lang . Object > ( ) ; while ( rows . hasNext ( ) ) { outArray . add ( com . tinkerpop . rexster . gremlin . converter . SerializedResultConverter . convert ( rows . next ( ) ) ) ; } return outArray ; } else if ( object instanceof com . tinkerpop . pipes . util . structures . Row ) { final com . tinkerpop . pipes . util . structures . Row row = ( ( com . tinkerpop . pipes . util . structures . Row ) ( object
|
org . junit . Assert . assertNull ( results )
|
testGetCurrentListeners ( ) { final com . ctrip . platform . dal . dao . client . DalHints hints = new com . ctrip . platform . dal . dao . client . DalHints ( ) ; final com . ctrip . platform . dal . dao . client . DalTransactionListener testListener = new com . ctrip . platform . dal . dao . client . DalTransactionListener ( ) { @ com . ctrip . platform . dal . dao . client . Override public void beforeCommit ( ) { } @ com . ctrip . platform . dal . dao . client . Override public void beforeRollback ( ) { } @ com . ctrip . platform . dal . dao . client . Override public void afterCommit ( ) { } @ com . ctrip . platform . dal . dao . client . Override public void afterRollback ( ) { } } ; try { final com . ctrip . platform . dal . dao . client . DalTransactionManager test = new com . ctrip . platform . dal . dao . client . DalTransactionManager ( com . ctrip . platform . dal . dao . client . DalTransactionManagerTest . getDalConnectionManager ( ) ) ; com . ctrip . platform . dal . dao . client . DalTransactionManager . getCurrentListeners ( ) ; org . junit . Assert . fail ( ) ; } catch ( java . lang . Exception e ) { } try { final com . ctrip . platform . dal . dao . client . DalTransactionManager test = new com . ctrip . platform . dal . dao . client . DalTransactionManager ( com . ctrip . platform . dal . dao . client . DalTransactionManagerTest . getDalConnectionManager ( ) ) ; com . ctrip . platform . dal . dao . client . ConnectionAction < ? > action = new com . ctrip . platform . dal . dao . client . ConnectionAction < java . lang . Object > ( ) { public java . lang . Object execute ( ) throws com . ctrip . platform . dal . dao . client . Exception { com . ctrip . platform . dal . dao . client . DalTransactionManager . register ( testListener ) ; com . ctrip . platform . dal . dao . client . DalTransactionManager . register ( testListener ) ; "<AssertPlaceHolder>" ; return null ; } } ; action . operation = DalEventEnum . EXECUTE ; test . doInTransaction ( action , hints ) ; try { com . ctrip . platform . dal . dao . client . DalTransactionManager . getCurrentListeners ( ) ; org . junit . Assert . fail ( ) ; } catch ( java . lang . Exception e ) { } } catch ( java . lang . Exception e ) { e . printStackTrace ( ) ; org . junit . Assert . fail ( ) ; } } getCurrentListeners ( ) { com . ctrip . platform . dal . dao . client . DalTransactionManager . reqiresTransaction ( ) ; return com . ctrip . platform . dal . dao . client . DalTransactionManager . transactionHolder . get ( ) . getListeners ( ) ; }
|
org . junit . Assert . assertEquals ( 2 , com . ctrip . platform . dal . dao . client . DalTransactionManager . getCurrentListeners ( ) . size ( ) )
|
values_returnsSingletonCollectionWithInstance ( ) { com . amazonaws . services . dynamodbv2 . util . LockClientUtils [ ] values = new com . amazonaws . services . dynamodbv2 . util . LockClientUtils [ ] { LockClientUtils . INSTANCE } ; "<AssertPlaceHolder>" ; }
|
org . junit . Assert . assertArrayEquals ( values , com . amazonaws . services . dynamodbv2 . util . LockClientUtils . values ( ) )
|
testLoadGoTaxon ( ) { owltools . io . ParserWrapper pw = new owltools . io . ParserWrapper ( ) ; pw . addIRIMapper ( new owltools . io . CatalogXmlIRIMapper ( owltools . solrj . GoLoaderIntegrationRunner . catalogXml ) ) ; owltools . graph . OWLGraphWrapper g = pw . parseToOWLGraph ( owltools . solrj . GoLoaderIntegrationRunner . goFile ) ; owltools . solrj . GafSolrDocumentLoaderIntegrationRunner . printMemoryStats ( ) ; owltools . solrj . GafSolrDocumentLoaderIntegrationRunner . gc ( ) ; owltools . solrj . GafSolrDocumentLoaderIntegrationRunner . printMemoryStats ( ) ; owltools . yaml . golrconfig . ConfigManager configManager = new owltools . yaml . golrconfig . ConfigManager ( ) ; configManager . add ( "src/test/resources/test-ont-config.yaml" ) ; org . apache . commons . lang3 . time . StopWatch watch = new org . apache . commons . lang3 . time . StopWatch ( ) ; watch . start ( ) ; owltools . flex . FlexCollection c = new owltools . flex . FlexCollection ( configManager , g ) ; owltools . solrj . GafSolrDocumentLoaderIntegrationRunner . printMemoryStats ( ) ; owltools . solrj . GafSolrDocumentLoaderIntegrationRunner . gc ( ) ; owltools . solrj . GafSolrDocumentLoaderIntegrationRunner . printMemoryStats ( ) ; owltools . solrj . FlexSolrDocumentLoader loader = new owltools . solrj . FlexSolrDocumentLoader ( ( ( org . apache . solr . client . solrj . SolrServer ) ( null ) ) , c ) { @ owltools . solrj . Override protected void addToServer ( java . util . Collection < org . apache . solr . common . SolrInputDocument > docs ) throws java . io . IOException , org . apache . solr . client . solrj . SolrServerException { owltools . solrj . GoLoaderIntegrationRunner . solrCounter += docs . size ( ) ; } } ; loader . load ( ) ; watch . stop ( ) ; owltools . solrj . GafSolrDocumentLoaderIntegrationRunner . printMemoryStats ( ) ; System . out . println ( ( ( ( "Loaded<sp>" + ( owltools . solrj . GoLoaderIntegrationRunner . solrCounter ) ) + "<sp>Solr<sp>docs<sp>in<sp>" ) + watch ) ) ; "<AssertPlaceHolder>" ; } printMemoryStats ( ) { int mb = 1024 * 1024 ; java . lang . Runtime runtime = java . lang . Runtime . getRuntime ( ) ; System . out . println ( ( ( ( ( "Heap<sp>Used<sp>Memory:" + ( ( ( runtime . totalMemory ( ) ) - ( runtime . freeMemory ( ) ) ) / mb ) ) + "<sp>of<sp>max<sp>" ) + ( ( runtime . maxMemory ( ) ) / mb ) ) + "<sp>[MB]" ) ) ; }
|
org . junit . Assert . assertTrue ( ( ( owltools . solrj . GoLoaderIntegrationRunner . solrCounter ) > 0 ) )
|
unmarshall_json_to_list_with_jackson ( ) { com . fasterxml . jackson . databind . ObjectMapper objectMapper = new com . fasterxml . jackson . databind . ObjectMapper ( ) ; java . util . List < com . levelup . java . json . JsontoJavaObject . Plane > planes = objectMapper . readValue ( ARRAY_AS_JSON , objectMapper . getTypeFactory ( ) . constructCollectionType ( java . util . List . class , com . levelup . java . json . JsontoJavaObject . Plane . class ) ) ; com . levelup . java . json . JsontoJavaObject . logger . info ( planes ) ; "<AssertPlaceHolder>" ; }
|
org . junit . Assert . assertEquals ( 5 , planes . size ( ) )
|
simple ( ) { put ( 100 ) ; put ( 300 ) ; put ( 200 ) ; put ( 400 ) ; com . asakusafw . utils . io . Source < org . apache . hadoop . io . IntWritable > results = sort ( ) ; check ( results , 100 ) ; check ( results , 200 ) ; check ( results , 300 ) ; check ( results , 400 ) ; "<AssertPlaceHolder>" ; } next ( ) { try { java . lang . Object object = input . readObject ( ) ; next = type . cast ( object ) ; canGet = true ; return true ; } catch ( java . lang . ClassNotFoundException e ) { throw new java . io . IOException ( e ) ; } catch ( java . io . EOFException e ) { next = null ; canGet = true ; return false ; } catch ( java . io . OptionalDataException e ) { if ( e . eof ) { next = null ; canGet = true ; return false ; } throw e ; } }
|
org . junit . Assert . assertThat ( results . next ( ) , is ( false ) )
|
testFetchByPrimaryKeyExisting ( ) { com . liferay . journal . model . JournalArticleResource newJournalArticleResource = addJournalArticleResource ( ) ; com . liferay . journal . model . JournalArticleResource existingJournalArticleResource = _persistence . fetchByPrimaryKey ( newJournalArticleResource . getPrimaryKey ( ) ) ; "<AssertPlaceHolder>" ; } getPrimaryKey ( ) { return _amImageEntryId ; }
|
org . junit . Assert . assertEquals ( existingJournalArticleResource , newJournalArticleResource )
|
getSetId ( ) { java . lang . String expected = "expectedId" ; com . microsoft . windowsazure . services . media . models . AccessPolicyInfo policy = new com . microsoft . windowsazure . services . media . models . AccessPolicyInfo ( null , new com . microsoft . windowsazure . services . media . implementation . content . AccessPolicyType ( ) . setId ( expected ) ) ; java . lang . String actual = policy . getId ( ) ; "<AssertPlaceHolder>" ; } getId ( ) { return id ; }
|
org . junit . Assert . assertEquals ( expected , actual )
|
testGetCurrentQueryCount ( ) { org . apache . blur . user . UserContext . setUser ( new org . apache . blur . user . User ( "testUser" , null ) ) ; org . apache . blur . thrift . generated . Blur . Iface client = org . apache . blur . console . util . Config . getClient ( ) ; org . apache . blur . thrift . generated . BlurQuery query = new org . apache . blur . thrift . generated . BlurQuery ( new org . apache . blur . thrift . generated . Query ( "fam0.col0:*" , true , org . apache . blur . thrift . generated . ScoreType . SUPER , null , null ) , null , null , false , 0 , 10 , 1 , 2000 , java . util . UUID . randomUUID ( ) . toString ( ) , "testUser" , false , java . lang . System . currentTimeMillis ( ) , null , null ) ; int currentCount = org . apache . blur . console . util . QueryUtil . getCurrentQueryCount ( ) ; client . query ( "queryUnitTable" , query ) ; "<AssertPlaceHolder>" ; org . apache . blur . user . UserContext . reset ( ) ; } getCurrentQueryCount ( ) { org . apache . blur . console . util . CachingBlurClient client = org . apache . blur . console . util . Config . getCachingBlurClient ( ) ; int count = 0 ; java . util . List < java . lang . String > tableList = client . enabledTables ( ) ; for ( java . lang . String table : tableList ) { java . util . List < java . lang . String > queries = client . queryStatusIdList ( table ) ; count += queries . size ( ) ; } return count ; }
|
org . junit . Assert . assertEquals ( ( currentCount + 1 ) , org . apache . blur . console . util . QueryUtil . getCurrentQueryCount ( ) )
|
testOrderByOnValues ( ) { com . splicemachine . derby . impl . sql . execute . operations . List < java . lang . Object [ ] > expected = com . splicemachine . derby . impl . sql . execute . operations . Arrays . asList ( o ( 0 , 0 , 1 ) , o ( 0 , 1 , 0 ) , o ( 1 , 0 , 0 ) , o ( 1 , 0 , 1 ) ) ; java . sql . ResultSet rs = methodWatcher . executeQuery ( "values<sp>(1,0,1),(1,0,0),(0,0,1),(0,1,0)<sp>order<sp>by<sp>1,2,3" ) ; com . splicemachine . derby . impl . sql . execute . operations . List result = com . splicemachine . homeless . TestUtils . resultSetToArrays ( rs ) ; "<AssertPlaceHolder>" ; } toArray ( ) { return java . util . Arrays . copyOf ( buffer , size ) ; }
|
org . junit . Assert . assertArrayEquals ( expected . toArray ( ) , result . toArray ( ) )
|
onlyDays ( ) { java . util . Calendar cal = java . util . Calendar . getInstance ( ) ; cal . add ( Calendar . DATE , ( - 3 ) ) ; int expected = cal . get ( Calendar . DATE ) ; java . util . Date from = org . candlepin . resource . util . ResourceDateParser . getFromDate ( null , null , "3" ) ; cal . setTime ( from ) ; "<AssertPlaceHolder>" ; } get ( java . lang . String ) { return curator . findFile ( id ) ; }
|
org . junit . Assert . assertEquals ( expected , cal . get ( Calendar . DATE ) )
|
entrySet_toArray ( ) { com . gs . collections . api . map . MutableMap < java . lang . Integer , java . lang . String > map = this . newMapWithKeyValue ( 1 , "One" ) ; java . lang . Object [ ] entries = map . entrySet ( ) . toArray ( ) ; "<AssertPlaceHolder>" ; } of ( T1 , T2 ) { return new com . gs . collections . impl . tuple . ImmutableEntry < T1 , T2 > ( key , value ) ; }
|
org . junit . Assert . assertArrayEquals ( new java . util . Map . Entry [ ] { com . gs . collections . impl . tuple . ImmutableEntry . of ( 1 , "One" ) } , entries )
|
testLocalUserFallThrough ( ) { org . sakaiproject . user . api . User user = userDirectoryService . addUser ( null , "local" , null , null , null , "localPW" , null , null ) ; org . sakaiproject . user . api . User authUser = userDirectoryService . authenticate ( "local" , "localPW" ) ; "<AssertPlaceHolder>" ; } getId ( ) { return id ; }
|
org . junit . Assert . assertTrue ( authUser . getId ( ) . equals ( user . getId ( ) ) )
|
testHashSetMode ( ) { final int numMemSegments = needNumMemSegments ( org . apache . flink . table . runtime . aggregate . BytesHashMapTest . NUM_ENTRIES , rowLength ( new org . apache . flink . table . type . RowType ( valueTypes ) ) , rowLength ( new org . apache . flink . table . type . RowType ( keyTypes ) ) , org . apache . flink . table . runtime . aggregate . BytesHashMapTest . PAGE_SIZE ) ; int memorySize = numMemSegments * ( org . apache . flink . table . runtime . aggregate . BytesHashMapTest . PAGE_SIZE ) ; org . apache . flink . runtime . memory . MemoryManager memoryManager = new org . apache . flink . runtime . memory . MemoryManager ( ( numMemSegments * ( org . apache . flink . table . runtime . aggregate . BytesHashMapTest . PAGE_SIZE ) ) , 32 ) ; org . apache . flink . table . runtime . aggregate . BytesHashMap table = new org . apache . flink . table . runtime . aggregate . BytesHashMap ( this , memoryManager , memorySize , keyTypes , new org . apache . flink . table . type . InternalType [ ] { } ) ; "<AssertPlaceHolder>" ; final java . util . Random rnd = new java . util . Random ( org . apache . flink . table . runtime . aggregate . BytesHashMapTest . RANDOM_SEED ) ; org . apache . flink . table . dataformat . BinaryRow [ ] keys = getRandomizedInput ( org . apache . flink . table . runtime . aggregate . BytesHashMapTest . NUM_ENTRIES , rnd , true ) ; verifyKeyInsert ( keys , table ) ; verifyKeyPresent ( keys , table ) ; table . free ( ) ; } isHashSetMode ( ) { return hashSetMode ; }
|
org . junit . Assert . assertTrue ( table . isHashSetMode ( ) )
|
testAddQuotesIfNotExist_Case_1 ( ) { java . lang . String input = null ; java . lang . String output = org . talend . core . utils . TalendQuoteUtils . addQuotesIfNotExist ( input , org . talend . core . utils . TalendQuoteUtilsTest . QUOTES ) ; "<AssertPlaceHolder>" ; } addQuotesIfNotExist ( java . lang . String , java . lang . String ) { if ( text == null ) { return null ; } if ( ! ( text . startsWith ( quote ) ) ) { text = quote + text ; } if ( ( ! ( text . endsWith ( quote ) ) ) || ( ( quote != null ) && ( quote . equals ( text ) ) ) ) { text = text + quote ; } return text ; }
|
org . junit . Assert . assertTrue ( ( output == null ) )
|
terminateSubscription_Broker_NotAuthorized ( ) { java . lang . reflect . Method method = beanClass . getMethod ( "terminateSubscription" , org . oscm . internal . vo . VOSubscription . class , java . lang . String . class ) ; boolean isBrokerRoleAllowed = isRoleAllowed ( method , UserRoleType . BROKER_MANAGER ) ; "<AssertPlaceHolder>" ; } isRoleAllowed ( java . lang . reflect . Method , org . oscm . internal . types . enumtypes . UserRoleType ) { javax . annotation . security . RolesAllowed rolesAllowed = method . getAnnotation ( javax . annotation . security . RolesAllowed . class ) ; if ( rolesAllowed == null ) { return true ; } for ( java . lang . String role : rolesAllowed . value ( ) ) { if ( role . equals ( roleType . name ( ) ) ) { return true ; } } return false ; }
|
org . junit . Assert . assertFalse ( isBrokerRoleAllowed )
|
getLastUpdatedTimeDateGets ( ) { final java . util . Date actualLastUpdatedTimeDate = new java . util . Date ( ) ; com . microsoft . azure . sdk . iot . service . FileUploadNotification testFileUploadNotification = new com . microsoft . azure . sdk . iot . service . FileUploadNotification ( mockDeviceId , mockBlobUri , mockBlobName , actualLastUpdatedTimeDate , mockBlobSizeInBytes , mockEnqueuedTimeUtcDate ) ; java . util . Date testLastUpdatedTimeDate = testFileUploadNotification . getLastUpdatedTimeDate ( ) ; "<AssertPlaceHolder>" ; } getLastUpdatedTimeDate ( ) { return lastUpdatedTimeDate ; }
|
org . junit . Assert . assertEquals ( testLastUpdatedTimeDate , actualLastUpdatedTimeDate )
|
testOperator_equalsNullNumber ( ) { java . lang . Number x = null ; java . lang . Number y = 123 ; boolean resutl = org . eclipse . smarthome . model . script . lib . NumberExtensions . operator_equals ( x , y ) ; "<AssertPlaceHolder>" ; } operator_equals ( java . lang . Number , java . lang . Number ) { if ( org . eclipse . smarthome . model . script . lib . NumberExtensions . oneIsQuantity ( left , right ) ) { return false ; } java . math . BigDecimal leftValue = org . eclipse . smarthome . model . script . lib . NumberExtensions . numberToBigDecimal ( left ) ; java . math . BigDecimal rightValue = org . eclipse . smarthome . model . script . lib . NumberExtensions . numberToBigDecimal ( right ) ; if ( leftValue == null ) { return rightValue == null ; } else if ( rightValue == null ) { return false ; } else { return ( leftValue . compareTo ( rightValue ) ) == 0 ; } }
|
org . junit . Assert . assertFalse ( resutl )
|
testGetSelectionEnabled_Initial ( ) { "<AssertPlaceHolder>" ; } getSelectionEnabled ( ) { checkWidget ( ) ; return selectionEnabled ; }
|
org . junit . Assert . assertTrue ( grid . getSelectionEnabled ( ) )
|
testAccessToken_UsesAgentIdentities ( ) { final com . spotify . sshagentproxy . AgentProxy proxy = mock ( com . spotify . sshagentproxy . AgentProxy . class ) ; final com . spotify . sshagentproxy . Identity identity = mockIdentity ( ) ; final com . spotify . helios . client . AuthenticatingHttpConnector authConnector = createAuthenticatingConnectorWithAccessToken ( com . google . common . base . Optional . of ( proxy ) , com . google . common . collect . ImmutableList . of ( identity ) ) ; final java . lang . String path = "/foo/bar" ; final javax . net . ssl . HttpsURLConnection connection = mock ( javax . net . ssl . HttpsURLConnection . class ) ; when ( connector . connect ( argThat ( matchesAnyEndpoint ( path ) ) , eq ( method ) , eq ( entity ) , argThat ( hasKeys ( java . util . Collections . singletonList ( "Authorization" ) ) ) ) ) . thenReturn ( connection ) ; when ( connection . getResponseCode ( ) ) . thenReturn ( 200 ) ; final java . net . URI uri = new java . net . URI ( ( "https://helios" + path ) ) ; final java . net . HttpURLConnection returnedConnection = authConnector . connect ( uri , method , entity , headers ) ; "<AssertPlaceHolder>" ; verify ( connector ) . setExtraHttpsHandler ( isA ( com . spotify . sshagenttls . HttpsHandler . class ) ) ; } connect ( java . net . URI , java . lang . String , byte [ ] , java . util . Map ) { final com . spotify . helios . client . Endpoint endpoint = endpointIterator . next ( ) ; final java . lang . String endpointHost = endpoint . getUri ( ) . getHost ( ) ; try { final java . net . HttpURLConnection connection = connect0 ( uri , method , entity , headers , endpointHost ) ; if ( ( connection . getResponseCode ( ) ) == ( HTTP_BAD_GATEWAY ) ) { throw new com . spotify . helios . common . HeliosException ( java . lang . String . format ( "Request<sp>to<sp>%s<sp>returned<sp>%s,<sp>master<sp>is<sp>down" , uri , connection . getResponseCode ( ) ) ) ; } return connection ; } catch ( java . net . ConnectException | java . net . SocketTimeoutException | java . net . UnknownHostException e ) { com . spotify . helios . client . DefaultHttpConnector . log . debug ( e . toString ( ) ) ; throw new com . spotify . helios . common . HeliosException ( ( "Unable<sp>to<sp>connect<sp>to<sp>master:<sp>" + uri ) , e ) ; } catch ( java . io . IOException e ) { throw new com . spotify . helios . common . HeliosException ( ( "Unexpected<sp>error<sp>connecting<sp>to<sp>" + uri ) , e ) ; } }
|
org . junit . Assert . assertSame ( returnedConnection , connection )
|
testPread ( ) { org . apache . hadoop . fs . Path zonePath = new org . apache . hadoop . fs . Path ( "/TestEncryptionZone" ) ; fsWrapper . mkdir ( zonePath , org . apache . hadoop . fs . permission . FsPermission . getDirDefault ( ) , false ) ; dfsAdmin . createEncryptionZone ( zonePath , TEST_KEY , org . apache . hadoop . hdfs . TestEncryptionZones . NO_TRASH ) ; final org . apache . hadoop . fs . Path encryptedFilePath = new org . apache . hadoop . fs . Path ( "/TestEncryptionZone/foo" ) ; org . apache . hadoop . hdfs . web . WebHdfsFileSystem webfs = org . apache . hadoop . hdfs . web . WebHdfsTestUtil . getWebHdfsFileSystem ( conf , WebHdfsConstants . WEBHDFS_SCHEME ) ; org . apache . hadoop . hdfs . DFSTestUtil . createFile ( webfs , encryptedFilePath , 1024 , ( ( short ) ( 1 ) ) , 65261 ) ; byte [ ] data = org . apache . hadoop . hdfs . DFSTestUtil . readFileAsBytes ( fs , encryptedFilePath ) ; org . apache . hadoop . fs . FSDataInputStream in = webfs . open ( encryptedFilePath ) ; for ( int i = 0 ; i < 1024 ; i ++ ) { in . seek ( i ) ; "<AssertPlaceHolder>" ; } } read ( ) { try { return sb . toString ( ) ; } finally { hasData = false ; sb . setLength ( 0 ) ; } }
|
org . junit . Assert . assertEquals ( ( ( data [ i ] ) & 255 ) , in . read ( ) )
|
testReadWriteComponentType ( ) { java . io . InputStream is = getClass ( ) . getResourceAsStream ( "CalculatorImpl.componentType" ) ; org . apache . tuscany . sca . assembly . ComponentType componentType = ( ( org . apache . tuscany . sca . assembly . ComponentType ) ( org . apache . tuscany . sca . interfacedef . java . xml . WriteTestCase . staxProcessor . read ( org . apache . tuscany . sca . interfacedef . java . xml . WriteTestCase . inputFactory . createXMLStreamReader ( is ) , org . apache . tuscany . sca . interfacedef . java . xml . WriteTestCase . context ) ) ) ; "<AssertPlaceHolder>" ; java . io . ByteArrayOutputStream bos = new java . io . ByteArrayOutputStream ( ) ; org . apache . tuscany . sca . interfacedef . java . xml . WriteTestCase . staxProcessor . write ( componentType , org . apache . tuscany . sca . interfacedef . java . xml . WriteTestCase . outputFactory . createXMLStreamWriter ( bos ) , org . apache . tuscany . sca . interfacedef . java . xml . WriteTestCase . context ) ; } createXMLStreamReader ( java . net . URL ) { try { javax . xml . transform . stream . StreamSource scdlSource = new javax . xml . transform . stream . StreamSource ( org . apache . tuscany . sca . common . xml . stax . StAXHelper . openStream ( url ) , url . toString ( ) ) ; return inputFactory . createXMLStreamReader ( scdlSource ) ; } catch ( java . io . IOException e ) { throw new javax . xml . stream . XMLStreamException ( e ) ; } }
|
org . junit . Assert . assertNotNull ( componentType )
|
serializable ( ) { model . setMySerializable ( new org . slim3 . datastore . model . MySerializable ( "aaa" ) ) ; com . google . appengine . api . datastore . Entity entity = meta . modelToEntity ( model ) ; com . google . appengine . api . datastore . Key key = ds . put ( entity ) ; com . google . appengine . api . datastore . Entity entity2 = ds . get ( key ) ; org . slim3 . datastore . model . Hoge model2 = meta . entityToModel ( entity2 ) ; "<AssertPlaceHolder>" ; } getMySerializable ( ) { return mySerializable ; }
|
org . junit . Assert . assertThat ( model2 . getMySerializable ( ) , org . hamcrest . CoreMatchers . is ( new org . slim3 . datastore . model . MySerializable ( "aaa" ) ) )
|
testTimestampMinusTimestampPositiveValues ( ) { java . lang . String sqlText = "values<sp>timestamp('2011-12-26',<sp>'17:13:30')<sp>-<sp>timestamp('2011-06-05',<sp>'05:06:00')" ; java . sql . ResultSet rs = com . splicemachine . derby . impl . sql . execute . operations . SimpleDateArithmeticIT . spliceClassWatcher . executeQuery ( sqlText ) ; java . lang . String expected = "1<sp>|\n" + ( "-----\n" + "204<sp>|" ) ; "<AssertPlaceHolder>" ; rs . close ( ) ; } toStringUnsorted ( com . splicemachine . homeless . ResultSet ) { return com . splicemachine . homeless . TestUtils . FormattedResult . ResultFactory . convert ( "" , rs , false ) . toString ( ) . trim ( ) ; }
|
org . junit . Assert . assertEquals ( ( ( "\n" + sqlText ) + "\n" ) , expected , TestUtils . FormattedResult . ResultFactory . toStringUnsorted ( rs ) )
|
testNewInstance ( ) { org . pentaho . di . trans . steps . getxmldata . GetXMLDataStepAnalyzer analyzer = new org . pentaho . di . trans . steps . getxmldata . GetXMLDataStepAnalyzer ( ) ; "<AssertPlaceHolder>" ; } newInstance ( ) { return new org . pentaho . di . trans . steps . jsoninput . analyzer . JsonInputAnalyzer ( ) ; }
|
org . junit . Assert . assertTrue ( analyzer . newInstance ( ) . getClass ( ) . equals ( org . pentaho . di . trans . steps . getxmldata . GetXMLDataStepAnalyzer . class ) )
|
readJSONReturnsNullWhenPassedEmptyJSONObject ( ) { "<AssertPlaceHolder>" ; } readJSON ( org . json . JSONObject ) { return null ; }
|
org . junit . Assert . assertNull ( this . component . readJSON ( new org . json . JSONObject ( ) ) )
|
testIsModal_Is_False_For_Non_Modal_Window ( ) { when ( window . isModal ( ) ) . thenReturn ( false ) ; mmarquee . automation . uiautomation . IUIAutomation mocked_automation = org . mockito . Mockito . mock ( mmarquee . automation . uiautomation . IUIAutomation . class ) ; mmarquee . automation . UIAutomation instance = new mmarquee . automation . UIAutomation ( mocked_automation ) ; mmarquee . automation . controls . AutomationWindow windw = new mmarquee . automation . controls . AutomationWindow ( new mmarquee . automation . controls . ElementBuilder ( element ) . addPattern ( container , window ) . automation ( instance ) ) ; "<AssertPlaceHolder>" ; } isModal ( ) { com . sun . jna . ptr . IntByReference ibr = new com . sun . jna . ptr . IntByReference ( ) ; final int res = this . getPattern ( ) . getCurrentIsModal ( ibr ) ; if ( res != 0 ) { throw new mmarquee . automation . AutomationException ( res ) ; } return ( ibr . getValue ( ) ) == 1 ; }
|
org . junit . Assert . assertFalse ( windw . isModal ( ) )
|
testUncompressBytesThatDontNeedUncompressed ( ) { byte [ ] bytes = new byte [ ] { 5 , 16 , 16 } ; byte [ ] result = com . cedarsoftware . util . IOUtilities . uncompressBytes ( bytes ) ; "<AssertPlaceHolder>" ; } uncompressBytes ( byte [ ] ) { return com . cedarsoftware . util . IOUtilities . uncompressBytes ( bytes , 0 , bytes . length ) ; }
|
org . junit . Assert . assertSame ( bytes , result )
|
testEquals ( ) { com . github . jqudt . Unit unit1 = new com . github . jqudt . Unit ( ) ; java . net . URI resource1 = new java . net . URI ( "http://qudt.org/vocab/unit#Kelvin" ) ; unit1 . setResource ( resource1 ) ; com . github . jqudt . Unit unit2 = new com . github . jqudt . Unit ( ) ; java . net . URI resource2 = new java . net . URI ( "http://qudt.org/vocab/unit#Kelvin" ) ; unit2 . setResource ( resource2 ) ; "<AssertPlaceHolder>" ; } setResource ( java . net . URI ) { this . resource = resource ; }
|
org . junit . Assert . assertEquals ( unit1 , unit2 )
|
testMixed ( ) { eu . infomas . annotation . FileIterator iter = new eu . infomas . annotation . FileIterator ( new java . io . File ( "./src/test/java/eu/infomas/util/TestSupport.java" ) , new java . io . File ( "./src/test/java/eu/infomas/annotation/" ) ) ; "<AssertPlaceHolder>" ; } countFiles ( eu . infomas . annotation . FileIterator ) { int counter = 0 ; while ( ( iter . next ( ) ) != null ) { ++ counter ; } return counter ; }
|
org . junit . Assert . assertEquals ( 8 , countFiles ( iter ) )
|
testFindYoungestAgeOfWhoBoughtACarForMoreThan50000 ( ) { int ageIterative = Integer . MAX_VALUE ; for ( ch . lambdaj . demo . Sale sale : db . getSales ( ) ) { if ( ( sale . getCost ( ) ) > 50000.0 ) { int buyerAge = sale . getBuyer ( ) . getAge ( ) ; if ( buyerAge < ageIterative ) ageIterative = buyerAge ; } } int age = min ( forEach ( select ( db . getSales ( ) , having ( on ( ch . lambdaj . demo . Sale . class ) . getCost ( ) , greaterThan ( 50000.0 ) ) ) ) . getBuyer ( ) , on ( ch . lambdaj . demo . Person . class ) . getAge ( ) ) ; "<AssertPlaceHolder>" ; } getAge ( ) { return ( getCurrentYear ( ) ) - ( getBirthYear ( ) ) ; }
|
org . junit . Assert . assertEquals ( age , ageIterative )
|
testProxy ( ) { org . jinstagram . auth . InstagramAuthService fixture = new org . jinstagram . auth . InstagramAuthService ( ) ; fixture . apiSecret ( "" ) ; fixture . callback ( "" ) ; fixture . scope ( "" ) ; fixture . proxy ( new java . net . Proxy ( java . net . Proxy . Type . DIRECT , new java . net . InetSocketAddress ( 1 ) ) ) ; fixture . apiKey ( "" ) ; fixture . display ( "" ) ; java . net . Proxy requestProxy = new java . net . Proxy ( java . net . Proxy . Type . DIRECT , new java . net . InetSocketAddress ( 1 ) ) ; org . jinstagram . auth . InstagramAuthService result = fixture . proxy ( requestProxy ) ; "<AssertPlaceHolder>" ; } proxy ( java . net . Proxy ) { this . requestProxy = requestProxy ; return this ; }
|
org . junit . Assert . assertNotNull ( result )
|
TestEmptyGet ( ) { java . util . List r = com . groupon . mapreduce . mongo . JobUtil . get ( m , "z" ) ; "<AssertPlaceHolder>" ; } get ( java . util . Map , java . lang . String ) { java . lang . String [ ] fields = path . split ( "\\." ) ; java . util . List found = new java . util . ArrayList ( ) ; com . groupon . mapreduce . mongo . JobUtil . get ( map , fields , 0 , found ) ; return found ; }
|
org . junit . Assert . assertEquals ( r . size ( ) , 0 )
|
testMultiPolygons ( ) { org . springframework . mock . web . MockHttpServletResponse resp = getAsServletResponse ( "wfs?request=GetFeature&version=1.1.0&typeName=MPolygons&outputFormat=dxf" ) ; java . lang . String sResponse = testBasicResult ( resp , "MPolygons" ) ; int pos = getGeometrySearchStart ( sResponse ) ; "<AssertPlaceHolder>" ; checkSequence ( sResponse , new java . lang . String [ ] { "LWPOLYLINE" , "LWPOLYLINE" } , pos ) ; } getGeometrySearchStart ( java . lang . String ) { return response . indexOf ( "BLOCKS" ) ; }
|
org . junit . Assert . assertTrue ( ( pos != ( - 1 ) ) )
|
testFetchByPrimaryKeyMissing ( ) { long pk = com . liferay . portal . kernel . test . util . RandomTestUtil . nextLong ( ) ; com . liferay . portal . tools . service . builder . test . model . LocalizedEntryLocalization missingLocalizedEntryLocalization = _persistence . fetchByPrimaryKey ( pk ) ; "<AssertPlaceHolder>" ; } fetchByPrimaryKey ( long ) { return com . liferay . adaptive . media . image . service . persistence . AMImageEntryUtil . getPersistence ( ) . fetchByPrimaryKey ( amImageEntryId ) ; }
|
org . junit . Assert . assertNull ( missingLocalizedEntryLocalization )
|
testGetAllDrugInfo_shouldReturnAllDrugInfoIncludingRetired ( ) { java . util . List < org . raxa . module . raxacore . DrugInfo > allDrugInfo = s . getAllDrugInfo ( true ) ; "<AssertPlaceHolder>" ; } getAllDrugInfo ( boolean ) { org . hibernate . Criteria criteria = sessionFactory . getCurrentSession ( ) . createCriteria ( org . raxa . module . raxacore . DrugInfo . class ) ; if ( includeRetired == false ) { criteria . add ( org . hibernate . criterion . Restrictions . eq ( "retired" , false ) ) ; } return criteria . list ( ) ; }
|
org . junit . Assert . assertEquals ( allDrugInfo . size ( ) , 3 )
|
test2 ( ) { introclassJava . median_68eb0bb0_000 mainClass = new introclassJava . median_68eb0bb0_000 ( ) ; java . lang . String expected = "Please<sp>enter<sp>3<sp>numbers<sp>separated<sp>by<sp>spaces<sp>><sp>6<sp>is<sp>the<sp>median" ; mainClass . scanner = new java . util . Scanner ( "2<sp>8<sp>6" ) ; mainClass . exec ( ) ; java . lang . String out = mainClass . output . replace ( "\n" , "<sp>" ) . trim ( ) ; "<AssertPlaceHolder>" ; } replace ( int , java . lang . Object ) { org . mozilla . javascript . xmlimpl . XMLList xlChildToReplace = child ( index ) ; if ( ( xlChildToReplace . length ( ) ) > 0 ) { org . mozilla . javascript . xmlimpl . XML childToReplace = xlChildToReplace . item ( 0 ) ; insertChildAfter ( childToReplace , xml ) ; removeChild ( index ) ; } return this ; }
|
org . junit . Assert . assertEquals ( expected . replace ( "<sp>" , "" ) , out . replace ( "<sp>" , "" ) )
|
testNodeCommunicationTypeLengths ( ) { final int MAX_LENGTH_IN_DB = 10 ; for ( org . jumpmind . symmetric . model . NodeCommunication . CommunicationType communicationType : NodeCommunication . CommunicationType . values ( ) ) { java . lang . String msg = ( ( ( ( communicationType . name ( ) ) + "<sp>is<sp>too<sp>long<sp>for<sp>DB.<sp>" ) + ( communicationType . name ( ) . length ( ) ) ) + "<sp><=<sp>" ) + MAX_LENGTH_IN_DB ; "<AssertPlaceHolder>" ; } }
|
org . junit . Assert . assertTrue ( msg , ( ( communicationType . name ( ) . length ( ) ) <= MAX_LENGTH_IN_DB ) )
|
test_plus_Period_zero ( ) { java . time . LocalDate t = TEST_2007_07_15 . plus ( MockSimplePeriod . ZERO_DAYS ) ; "<AssertPlaceHolder>" ; } plus ( int ) { return compileOperation ( new org . apache . xpath . operations . Plus ( ) , opPos ) ; }
|
org . junit . Assert . assertSame ( t , TEST_2007_07_15 )
|
testDivisionByZeroB ( ) { java . util . Map < java . lang . String , java . lang . Double > values = new java . util . HashMap < java . lang . String , java . lang . Double > ( ) ; values . put ( "ifInOctets" , 200000.0 ) ; values . put ( "ifSpeed" , 0.0 ) ; values . put ( "ifHighSpeed" , 0.0 ) ; double value = wrapper . evaluate ( values ) ; "<AssertPlaceHolder>" ; } evaluate ( javax . servlet . http . HttpServletRequest ) { if ( request . isUserInRole ( Authentication . ROLE_ADMIN ) ) { return super . evaluate ( request ) ; } else { return DisplayStatus . NO_DISPLAY ; } }
|
org . junit . Assert . assertTrue ( ( value == 0.0 ) )
|
addAndRemoveOffers ( ) { fr . gouv . vitam . storage . engine . server . spi . DriverManager . addOfferToDriver ( fr . gouv . vitam . storage . engine . server . spi . DriverManagerTest . MON_MODULE_DRIVER_MON_DRIVER , fr . gouv . vitam . storage . engine . server . spi . DriverManagerTest . OFFER_3 ) ; fr . gouv . vitam . storage . driver . Driver driver = fr . gouv . vitam . storage . engine . server . spi . DriverManager . getDriverFor ( fr . gouv . vitam . storage . engine . server . spi . DriverManagerTest . OFFER_3 ) ; "<AssertPlaceHolder>" ; fr . gouv . vitam . storage . engine . server . spi . DriverManager . removeOffer ( fr . gouv . vitam . storage . engine . server . spi . DriverManagerTest . OFFER_3 ) ; try { driver = fr . gouv . vitam . storage . engine . server . spi . DriverManager . getDriverFor ( fr . gouv . vitam . storage . engine . server . spi . DriverManagerTest . OFFER_3 ) ; org . junit . Assert . fail ( "Excepted<sp>Storage<sp>Driver<sp>Not<sp>Found<sp>Exception" ) ; } catch ( final fr . gouv . vitam . storage . engine . common . exception . StorageDriverNotFoundException exc ) { } } getDriverFor ( java . lang . String ) { for ( java . lang . String driverName : fr . gouv . vitam . storage . engine . server . spi . DriverManager . drivers . keySet ( ) ) { final fr . gouv . vitam . storage . driver . Driver driver = fr . gouv . vitam . storage . engine . server . spi . DriverManager . drivers . get ( driverName ) ; if ( driver . hasOffer ( offerId ) ) { return driver ; } } fr . gouv . vitam . storage . engine . server . spi . DriverManager . LOGGER . error ( ( "No<sp>suitable<sp>driver<sp>for<sp>offer<sp>ID<sp>:<sp>" + offerId ) ) ; throw new fr . gouv . vitam . storage . engine . common . exception . StorageDriverNotFoundException ( ( "No<sp>suitable<sp>driver<sp>for<sp>offer<sp>ID<sp>:<sp>" + offerId ) ) ; }
|
org . junit . Assert . assertNotNull ( driver )
|
iremove17 ( ) { org . roaringbitmap . buffer . MappeableContainer rc = new org . roaringbitmap . buffer . MappeableRunContainer ( ) ; rc . iadd ( 37543 , 65536 ) ; rc . iremove ( 9795 , 65536 ) ; "<AssertPlaceHolder>" ; } getCardinality ( ) { return ( ( int ) ( getLongCardinality ( ) ) ) ; }
|
org . junit . Assert . assertEquals ( rc . getCardinality ( ) , 0 )
|
test_parse_instantNoZone_Instant ( java . time . format . DateTimeFormatter , java . lang . String , java . time . Instant ) { java . time . temporal . TemporalAccessor actual = formatter . parse ( text ) ; "<AssertPlaceHolder>" ; } from ( java . time . temporal . TemporalAccessor ) { if ( temporal instanceof java . time . Instant ) { return ( ( java . time . Instant ) ( temporal ) ) ; } java . util . Objects . requireNonNull ( temporal , "temporal" ) ; try { long instantSecs = temporal . getLong ( java . time . INSTANT_SECONDS ) ; int nanoOfSecond = temporal . get ( java . time . NANO_OF_SECOND ) ; return java . time . Instant . ofEpochSecond ( instantSecs , nanoOfSecond ) ; } catch ( java . time . DateTimeException ex ) { throw new java . time . DateTimeException ( ( ( ( "Unable<sp>to<sp>obtain<sp>Instant<sp>from<sp>TemporalAccessor:<sp>" + temporal ) + "<sp>of<sp>type<sp>" ) + ( temporal . getClass ( ) . getName ( ) ) ) , ex ) ; } }
|
org . junit . Assert . assertEquals ( java . time . Instant . from ( actual ) , expected )
|
comments ( ) { "<AssertPlaceHolder>" ; } comments ( ) { org . junit . Assert . assertThat ( facebook . comments ( ) , org . hamcrest . CoreMatchers . instanceOf ( facebook4j . CommentMethods . class ) ) ; }
|
org . junit . Assert . assertThat ( facebook . comments ( ) , org . hamcrest . CoreMatchers . instanceOf ( facebook4j . CommentMethods . class ) )
|
test_isSubgraphIsomorphism_true ( ) { com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Vertex a = new com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Vertex ( 1 , "a" ) ; com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Vertex b = new com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Vertex ( 2 , "b" ) ; com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Vertex c = new com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Vertex ( 3 , "c" ) ; com . pengyifan . commons . collections . graph . IndexGraph < com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Vertex , com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Edge > graph = new com . pengyifan . commons . collections . graph . IndexGraph ( com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Edge . class ) ; graph . addVertex ( a ) ; graph . addVertex ( b ) ; graph . addVertex ( c ) ; graph . addEdge ( b , a , new com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Edge ( 4 , "NSUBJ" ) ) ; graph . addEdge ( b , c , new com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Edge ( 5 , "DOBJ" ) ) ; java . util . function . Predicate < com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Vertex > predA = ( v ) -> v . word ( ) . equals ( "a" ) ; java . util . function . Predicate < com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Vertex > predB = ( v ) -> v . word ( ) . equals ( "b" ) ; org . jgrapht . DirectedGraph < java . util . function . Predicate < com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Vertex > , java . util . function . Predicate < com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Edge > > subgraph = new org . jgrapht . graph . DefaultDirectedGraph ( new com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Factory ( ) ) ; subgraph . addVertex ( predA ) ; subgraph . addVertex ( predB ) ; subgraph . addEdge ( predB , predA , ( e ) -> e . word ( ) . equals ( "NSUBJ" ) ) ; com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicate < com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Vertex , com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicateTest . Edge > esm = new com . pengyifan . commons . collections . graph . ExactSubgraphMatchingPredicate ( subgraph , graph ) ; "<AssertPlaceHolder>" ; } isSubgraphIsomorphism ( ) { if ( ! ( isSubgraphSmaller ( ) ) ) { System . err . println ( ( ( ( ( "The<sp>size<sp>of<sp>the<sp>subgraph:<sp>" + ( subgraph . vertexSet ( ) . size ( ) ) ) + "<sp>is<sp>bigger<sp>the<sp>size<sp>of<sp>the<sp>graph<sp>" ) + ( graph . vertexSet ( ) . size ( ) ) ) + ".<sp>Please<sp>check." ) ) ; return false ; } boolean isSubgraphIsomorphism = false ; for ( int i = 0 ; i < ( graphStartNodes . size ( ) ) ; i ++ ) { java . util . Map < V , V > subgraphToGraph = com . google . common . collect . Maps . newHashMap ( ) ; java . util . Map < V , V > graphToSubgraph = com . google . common . collect . Maps . newHashMap ( ) ; java . util . List < V > toMatch = java . util . Arrays . asList ( subgraphStartNode , graphStartNodes . get ( i ) ) ; if ( matchNodeForSingleMatch ( toMatch , subgraphToGraph , graphToSubgraph , subgraph , graph ) ) { isSubgraphIsomorphism = true ; break ; } } return isSubgraphIsomorphism ; }
|
org . junit . Assert . assertTrue ( esm . isSubgraphIsomorphism ( ) )
|
callCreateSchemas ( ) { java . util . List < org . apache . olingo . commons . api . edm . EdmSchema > schemas = edm . getSchemas ( ) ; "<AssertPlaceHolder>" ; } getSchemas ( ) { java . util . List < org . apache . olingo . commons . api . edm . provider . CsdlSchema > schemas = new java . util . ArrayList < org . apache . olingo . commons . api . edm . provider . CsdlSchema > ( ) ; org . apache . olingo . commons . api . edm . provider . CsdlSchema schema = new org . apache . olingo . commons . api . edm . provider . CsdlSchema ( ) ; schema . setNamespace ( org . apache . olingo . server . sample . edmprovider . CarsEdmProvider . NAMESPACE ) ; java . util . List < org . apache . olingo . commons . api . edm . provider . CsdlEntityType > entityTypes = new java . util . ArrayList < org . apache . olingo . commons . api . edm . provider . CsdlEntityType > ( ) ; entityTypes . add ( getEntityType ( org . apache . olingo . server . sample . edmprovider . CarsEdmProvider . ET_CAR ) ) ; entityTypes . add ( getEntityType ( org . apache . olingo . server . sample . edmprovider . CarsEdmProvider . ET_MANUFACTURER ) ) ; schema . setEntityTypes ( entityTypes ) ; java . util . List < org . apache . olingo . commons . api . edm . provider . CsdlComplexType > complexTypes = new java . util . ArrayList < org . apache . olingo . commons . api . edm . provider . CsdlComplexType > ( ) ; complexTypes . add ( getComplexType ( org . apache . olingo . server . sample . edmprovider . CarsEdmProvider . CT_ADDRESS ) ) ; schema . setComplexTypes ( complexTypes ) ; schema . setEntityContainer ( getEntityContainer ( ) ) ; schemas . add ( schema ) ; return schemas ; }
|
org . junit . Assert . assertNotNull ( schemas )
|
testConvertWithOrderBy ( ) { org . lnu . is . domain . benefit . BenefitType context = new org . lnu . is . domain . benefit . BenefitType ( ) ; org . lnu . is . pagination . OrderBy orderBy1 = new org . lnu . is . pagination . OrderBy ( "name" , org . lnu . is . pagination . OrderByType . ASC ) ; org . lnu . is . pagination . OrderBy orderBy2 = new org . lnu . is . pagination . OrderBy ( "priority" , org . lnu . is . pagination . OrderByType . DESC ) ; java . util . List < org . lnu . is . pagination . OrderBy > orders = java . util . Arrays . asList ( orderBy1 , orderBy2 ) ; java . lang . String expected = "SELECT<sp>e<sp>FROM<sp>BenefitType<sp>e<sp>WHERE<sp>e.status=:status<sp>AND<sp>e.crtUserGroup<sp>IN<sp>(:userGroups)<sp>ORDER<sp>BY<sp>e.name<sp>ASC,<sp>e.priority<sp>DESC" ; org . lnu . is . pagination . MultiplePagedSearch < org . lnu . is . domain . benefit . BenefitType > pagedSearch = new org . lnu . is . pagination . MultiplePagedSearch ( ) ; pagedSearch . setEntity ( context ) ; pagedSearch . setOrders ( orders ) ; java . lang . String actualQuery = unit . build ( pagedSearch ) ; "<AssertPlaceHolder>" ; } setOrders ( java . util . List ) { this . orders = orders ; }
|
org . junit . Assert . assertEquals ( expected , actualQuery )
|
createsPrimitiveValue ( java . lang . Class , javax . ws . rs . core . MultivaluedMap , java . lang . String , java . lang . Object ) { org . everrest . core . impl . method . PrimitiveTypeProducer producer = new org . everrest . core . impl . method . PrimitiveTypeProducer ( primitiveTypeWrapper ) ; java . lang . Object result = producer . createValue ( "value" , values , defaultValue ) ; "<AssertPlaceHolder>" ; } createValue ( java . lang . String , javax . ws . rs . core . MultivaluedMap , java . lang . String ) { java . lang . String value = values . getFirst ( param ) ; if ( value != null ) { return createValue ( value ) ; } else if ( defaultValue != null ) { return createValue ( defaultValue ) ; } return this . defaultDefaultValue ; }
|
org . junit . Assert . assertEquals ( expected , result )
|
shouldHaveCommentForAllRootPathsInConfigurationData ( ) { java . util . Set < java . lang . String > rootPaths = java . util . Arrays . stream ( fr . xephi . authme . message . MessageKey . values ( ) ) . map ( ( key ) -> key . getKey ( ) . split ( "\\." ) [ 0 ] ) . collect ( java . util . stream . Collectors . toSet ( ) ) ; java . util . Map < java . lang . String , java . util . List < java . lang . String > > comments = fr . xephi . authme . message . updater . MessageUpdater . createConfigurationData ( ) . getAllComments ( ) ; "<AssertPlaceHolder>" ; } createConfigurationData ( ) { java . util . Map < java . lang . String , java . lang . String > comments = com . google . common . collect . ImmutableMap . < java . lang . String , java . lang . String > builder ( ) . put ( "misc" 9 , "recovery" 4 ) . put ( "recovery" 3 , "misc" 0 ) . put ( "recovery" 9 , "Login" ) . put ( "misc" 5 , "recovery" 0 ) . put ( "misc" 8 , "AntiBot" ) . put ( "Password<sp>recovery<sp>by<sp>email" 0 , "misc" 7 ) . put ( "misc" , "misc" 3 ) . put ( "misc" 1 , "misc" 2 ) . put ( "recovery" 6 , "Error<sp>messages<sp>when<sp>joining" ) . put ( "recovery" 8 , "recovery" 2 ) . put ( "recovery" , "Password<sp>recovery<sp>by<sp>email" ) . put ( "recovery" 5 , "Captcha" ) . put ( "misc" 6 , "recovery" 1 ) . put ( "time" , "Time<sp>units" ) . put ( "misc" 4 , "recovery" 7 ) . build ( ) ; java . util . Set < java . lang . String > addedKeys = new java . util . HashSet ( ) ; fr . xephi . authme . message . updater . MessageUpdater . MessageKeyPropertyListBuilder builder = new fr . xephi . authme . message . updater . MessageUpdater . MessageKeyPropertyListBuilder ( ) ; for ( java . lang . String path : comments . keySet ( ) ) { fr . xephi . authme . message . MessageKey key = java . util . Arrays . stream ( fr . xephi . authme . message . MessageKey . values ( ) ) . filter ( ( p ) -> p . getKey ( ) . startsWith ( ( path + "." ) ) ) . findFirst ( ) . orElseThrow ( ( ) -> new java . lang . IllegalStateException ( path ) ) ; builder . addMessageKey ( key ) ; addedKeys . add ( key . getKey ( ) ) ; } java . util . Arrays . stream ( fr . xephi . authme . message . MessageKey . values ( ) ) . filter ( ( key ) -> ! ( addedKeys . contains ( key . getKey ( ) ) ) ) . forEach ( builder :: addMessageKey ) ; java . util . Map < java . lang . String , java . util . List < java . lang . String > > commentsMap = comments . entrySet ( ) . stream ( ) . collect ( java . util . stream . Collectors . toMap ( ( e ) -> e . getKey ( ) , ( e ) -> singletonList ( e . getValue ( ) ) ) ) ; return new fr . xephi . authme . message . updater . MessageKeyConfigurationData ( builder , commentsMap ) ; }
|
org . junit . Assert . assertThat ( comments . keySet ( ) , org . hamcrest . Matchers . equalTo ( rootPaths ) )
|
testPingTCPPool ( ) { net . tomp2p . p2p . Peer sender = null ; net . tomp2p . p2p . Peer recv1 = null ; net . tomp2p . connection . ChannelCreator cc = null ; try { sender = new net . tomp2p . p2p . PeerBuilder ( new net . tomp2p . peers . Number160 ( "0x9876" ) ) . p2pId ( 55 ) . ports ( 2424 ) . start ( ) ; recv1 = new net . tomp2p . p2p . PeerBuilder ( new net . tomp2p . peers . Number160 ( "0x1234" ) ) . p2pId ( 55 ) . ports ( 8088 ) . start ( ) ; java . util . List < net . tomp2p . futures . FutureResponse > list = new java . util . ArrayList < net . tomp2p . futures . FutureResponse > ( 50 ) ; net . tomp2p . futures . FutureChannelCreator fcc = recv1 . connectionBean ( ) . reservation ( ) . create ( 0 , 50 ) ; fcc . awaitUninterruptibly ( ) ; cc = fcc . channelCreator ( ) ; for ( int i = 0 ; i < 50 ; i ++ ) { net . tomp2p . futures . FutureResponse fr = sender . pingRPC ( ) . pingTCP ( recv1 . peerAddress ( ) , cc , new net . tomp2p . connection . DefaultConnectionConfiguration ( ) ) ; list . add ( fr ) ; } for ( net . tomp2p . futures . FutureResponse fr2 : list ) { fr2 . awaitUninterruptibly ( ) ; "<AssertPlaceHolder>" ; } } finally { if ( cc != null ) { cc . shutdown ( ) . await ( ) ; } if ( sender != null ) { sender . shutdown ( ) . await ( ) ; } if ( recv1 != null ) { recv1 . shutdown ( ) . await ( ) ; } } } isSuccess ( ) { if ( ! ( super . isSuccess ( ) ) ) { return false ; } return checkResults ( result ( ) , rawResult . size ( ) , dataSize ) ; }
|
org . junit . Assert . assertTrue ( fr2 . isSuccess ( ) )
|
testNonTokenQop02 ( ) { java . lang . String header = "Digest<sp>qop=auth{" ; java . io . StringReader input = new java . io . StringReader ( header ) ; java . util . Map < java . lang . String , java . lang . String > result = org . apache . tomcat . util . http . parser . Authorization . parseAuthorizationDigest ( input ) ; "<AssertPlaceHolder>" ; } parseAuthorizationDigest ( java . io . StringReader ) { java . util . Map < java . lang . String , java . lang . String > result = new java . util . HashMap ( ) ; if ( ( org . apache . tomcat . util . http . parser . HttpParser . skipConstant ( input , "Digest" ) ) != ( SkipResult . FOUND ) ) { return null ; } java . lang . String field = org . apache . tomcat . util . http . parser . HttpParser . readToken ( input ) ; if ( field == null ) { return null ; } while ( ! ( field . equals ( "" ) ) ) { if ( ( org . apache . tomcat . util . http . parser . HttpParser . skipConstant ( input , "=" ) ) != ( SkipResult . FOUND ) ) { return null ; } java . lang . String value ; java . lang . Integer type = org . apache . tomcat . util . http . parser . Authorization . fieldTypes . get ( field . toLowerCase ( Locale . ENGLISH ) ) ; if ( type == null ) { type = org . apache . tomcat . util . http . parser . Authorization . FIELD_TYPE_TOKEN_OR_QUOTED_STRING ; } switch ( type . intValue ( ) ) { case 0 : value = org . apache . tomcat . util . http . parser . HttpParser . readToken ( input ) ; break ; case 1 : value = org . apache . tomcat . util . http . parser . HttpParser . readQuotedString ( input , false ) ; break ; case 2 : value = org . apache . tomcat . util . http . parser . HttpParser . readTokenOrQuotedString ( input , false ) ; break ; case 3 : value = org . apache . tomcat . util . http . parser . HttpParser . readLhex ( input ) ; break ; case 4 : value = org . apache . tomcat . util . http . parser . HttpParser . readQuotedToken ( input ) ; break ; default : throw new java . lang . IllegalArgumentException ( "TODO<sp>i18n:<sp>Unsupported<sp>type" ) ; } if ( value == null ) { return null ; } result . put ( field , value ) ; if ( ( org . apache . tomcat . util . http . parser . HttpParser . skipConstant ( input , "," ) ) == ( SkipResult . NOT_FOUND ) ) { return null ; } field = org . apache . tomcat . util . http . parser . HttpParser . readToken ( input ) ; if ( field == null ) { return null ; } } return result ; }
|
org . junit . Assert . assertNull ( result )
|
testFindByPrimaryKeyExisting ( ) { com . liferay . mobile . device . rules . model . MDRRule newMDRRule = addMDRRule ( ) ; com . liferay . mobile . device . rules . model . MDRRule existingMDRRule = _persistence . findByPrimaryKey ( newMDRRule . getPrimaryKey ( ) ) ; "<AssertPlaceHolder>" ; } getPrimaryKey ( ) { return _amImageEntryId ; }
|
org . junit . Assert . assertEquals ( existingMDRRule , newMDRRule )
|
testDST ( ) { java . lang . String s1 = "2005-04-03<sp>02:01:00<sp>America/Los_Angeles" ; java . lang . String s2 = "2005-04-03<sp>03:01:00<sp>America/Los_Angeles" ; "<AssertPlaceHolder>" ; } parse ( java . lang . String ) { return org . apache . hadoop . hive . common . type . TimestampTZUtil . parse ( s , null ) ; }
|
org . junit . Assert . assertEquals ( org . apache . hadoop . hive . common . type . TimestampTZUtil . parse ( s1 ) , org . apache . hadoop . hive . common . type . TimestampTZUtil . parse ( s2 ) )
|
testReverse ( ) { org . stringtemplate . v4 . org . stringtemplate . v4 . ST e = new org . stringtemplate . v4 . org . stringtemplate . v4 . ST ( "<reverse(names);<sp>separator=\",<sp>\">" ) ; e . add ( "names" , "Ter" ) ; e . add ( "names" , "Tom" ) ; e . add ( "names" , "Sriram" ) ; java . lang . String expecting = "Sriram,<sp>Tom,<sp>Ter" ; "<AssertPlaceHolder>" ; } render ( ) { return render ( java . util . Locale . getDefault ( ) ) ; }
|
org . junit . Assert . assertEquals ( expecting , e . render ( ) )
|
shouldIgnoreUnknownNonPolicySettings ( ) { org . neo4j . kernel . configuration . ssl . SslPolicyConfigValidator validator = new org . neo4j . kernel . configuration . ssl . SslPolicyConfigValidator ( ) ; java . util . Map < java . lang . String , java . lang . String > originalParams = org . neo4j . kernel . configuration . ssl . SslPolicyConfigValidatorTest . params ( "dbms.ssl.unknown" , "xyz" , "dbms.ssl.something" , "xyz" , "dbms.unrelated.totally" , "xyz" ) ; java . util . Map < java . lang . String , java . lang . String > validatedParams = validator . validate ( originalParams , warnings ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { try ( org . neo4j . cursor . RawCursor < org . neo4j . index . internal . gbptree . Hit < org . neo4j . kernel . impl . index . labelscan . LabelScanKey , org . neo4j . kernel . impl . index . labelscan . LabelScanValue > , java . io . IOException > cursor = index . seek ( new org . neo4j . kernel . impl . index . labelscan . LabelScanKey ( 0 , 0 ) , new org . neo4j . kernel . impl . index . labelscan . LabelScanKey ( Integer . MAX_VALUE , Long . MAX_VALUE ) ) ) { return ! ( cursor . next ( ) ) ; } }
|
org . junit . Assert . assertTrue ( validatedParams . isEmpty ( ) )
|
testCallLowHighWithResultSet ( ) { org . apache . ibatis . session . SqlSession sqlSession = org . apache . ibatis . submitted . sptests . SPTest . sqlSessionFactory . openSession ( ) ; try { org . apache . ibatis . submitted . sptests . SPMapper spMapper = sqlSession . getMapper ( org . apache . ibatis . submitted . sptests . SPMapper . class ) ; java . util . List < org . apache . ibatis . submitted . sptests . Name > names = spMapper . getNamesAnnotatedLowHighWithXMLResultMap ( 1 , 1 ) ; "<AssertPlaceHolder>" ; } finally { sqlSession . close ( ) ; } } size ( ) { return loaderMap . size ( ) ; }
|
org . junit . Assert . assertEquals ( 1 , names . size ( ) )
|
testFetchByPrimaryKeyExisting ( ) { com . liferay . portal . workflow . kaleo . model . KaleoInstanceToken newKaleoInstanceToken = addKaleoInstanceToken ( ) ; com . liferay . portal . workflow . kaleo . model . KaleoInstanceToken existingKaleoInstanceToken = _persistence . fetchByPrimaryKey ( newKaleoInstanceToken . getPrimaryKey ( ) ) ; "<AssertPlaceHolder>" ; } getPrimaryKey ( ) { return _amImageEntryId ; }
|
org . junit . Assert . assertEquals ( existingKaleoInstanceToken , newKaleoInstanceToken )
|
testComplexField ( ) { final java . lang . String expectedJson = org . apache . commons . io . IOUtils . toString ( this . getClass ( ) . getResource ( "map2-2" 5 ) , org . apache . commons . io . Charsets . UTF_8 ) ; java . io . StringWriter writer = new java . io . StringWriter ( ) ; com . streamsets . datacollector . json . JsonRecordWriterImpl jsonRecordWriter = new com . streamsets . datacollector . json . JsonRecordWriterImpl ( writer , com . streamsets . pipeline . api . ext . json . Mode . ARRAY_OBJECTS ) ; com . streamsets . pipeline . api . Record record = new com . streamsets . datacollector . record . RecordImpl ( "list3-1" 8 , "list2-1" 6 , null , null ) ; final com . streamsets . testing . fieldbuilder . MapFieldBuilder builder = com . streamsets . testing . fieldbuilder . MapFieldBuilder . builder ( ) ; builder . startMap ( "list2-1" 5 ) . startMap ( "map2-2" 1 ) . add ( "list2-1" 9 , "list2-1" 2 ) . add ( "list3-1" 5 , "b" ) . end ( ) . startMap ( "map2-2" 2 ) . add ( "list2-1" 7 , "list3-1" 0 ) . add ( "list3-1" 9 , "list2-1" 4 ) . end ( ) . end ( ) . startList ( "lists" ) . startList ( "list3-1" 4 ) . add ( "list1-1" ) . add ( "list3-1" 1 ) . add ( "list3-1" 2 ) . end ( ) . startList ( "list3-1" 6 ) . add ( "list2-1" ) . add ( "list2-2" ) . add ( "list2-3" ) . end ( ) . end ( ) . startMap ( "map2-2" 6 ) . startMap ( "list2-1" 0 ) . startMap ( "map2-1-1" ) . add ( "list2-1" 8 , "list2-1" 2 ) . add ( "map2-2" 3 , "b" ) . end ( ) . startMap ( "map2-2" 0 ) . add ( "map2-2" 8 , "list3-1" 0 ) . add ( "list3-1" 3 , "list2-1" 4 ) . end ( ) . end ( ) . startMap ( "map2-2" ) . startMap ( "map2-2" 4 ) . add ( "list2-1" 3 , 1 ) . add ( "map2-2" 7 , new java . util . Date ( 1500000000 ) ) . end ( ) . startMap ( "list3-1" 7 ) . add ( "map2-2" 9 , new byte [ ] { 0 , 1 , 2 } ) . add ( "map2-2-2-d-s" , ( ( short ) ( 222 ) ) ) . end ( ) . end ( ) . end ( ) . startList ( "list2-1" 1 ) . add ( "list3-1" ) . add ( 3.2 ) . add ( 33L ) . end ( ) ; record . set ( builder . build ( ) ) ; jsonRecordWriter . write ( record ) ; jsonRecordWriter . close ( ) ; "<AssertPlaceHolder>" ; } toString ( ) { return com . streamsets . pipeline . api . impl . Utils . format ( "DelayedRecordWriter[path='{}'<sp>expiresInSecs='{}'" , writer . getPath ( ) , getDelay ( TimeUnit . SECONDS ) ) ; }
|
org . junit . Assert . assertEquals ( expectedJson , writer . toString ( ) )
|
testAppendStringWithZeroCountToNullBuffer ( ) { buffer . appendString ( org . smpp . util . ByteBufferTest . ABC , 0 ) ; "<AssertPlaceHolder>" ; } getBuffer ( ) { return buffer ; }
|
org . junit . Assert . assertNull ( buffer . getBuffer ( ) )
|
rebaseWithFailedMerge ( ) { org . apache . jackrabbit . oak . spi . state . NodeBuilder rootBuilder = store . getRoot ( ) . builder ( ) ; rootBuilder . child ( "foo" ) ; org . apache . jackrabbit . oak . spi . state . NodeBuilder b = store . getRoot ( ) . builder ( ) ; b . child ( "bar" ) ; store . merge ( b , EmptyHook . INSTANCE , CommitInfo . EMPTY ) ; try { store . merge ( rootBuilder , new org . apache . jackrabbit . oak . spi . commit . CommitHook ( ) { @ javax . annotation . Nonnull @ org . apache . jackrabbit . oak . kernel . Override public org . apache . jackrabbit . oak . spi . state . NodeState processCommit ( org . apache . jackrabbit . oak . spi . state . NodeState before , org . apache . jackrabbit . oak . spi . state . NodeState after , org . apache . jackrabbit . oak . spi . commit . CommitInfo info ) throws org . apache . jackrabbit . oak . api . CommitFailedException { throw new org . apache . jackrabbit . oak . api . CommitFailedException ( "" , 0 , "commit<sp>rejected" ) ; } } , CommitInfo . EMPTY ) ; org . junit . Assert . fail ( "must<sp>throw<sp>CommitFailedException" ) ; } catch ( org . apache . jackrabbit . oak . api . CommitFailedException e ) { } org . apache . jackrabbit . oak . spi . state . NodeState root = store . merge ( rootBuilder , EmptyHook . INSTANCE , CommitInfo . EMPTY ) ; "<AssertPlaceHolder>" ; } hasChildNode ( java . lang . String ) { init ( ) ; if ( childNames . contains ( name ) ) { return true ; } else if ( ( getChildNodeCount ( org . apache . jackrabbit . oak . kernel . KernelNodeState . MAX_CHILD_NAMES ) ) <= ( org . apache . jackrabbit . oak . kernel . KernelNodeState . MAX_CHILD_NAMES ) ) { return false ; } else { return ( isValidName ( name ) ) && ( getChildNode ( name ) . exists ( ) ) ; } }
|
org . junit . Assert . assertTrue ( root . hasChildNode ( "bar" ) )
|
doesTargetStorageSupportDiskOfferingTestDiskOfferingTagsEmptyAndStorageTagsNotEmpty ( ) { com . cloud . storage . DiskOfferingVO diskOfferingVoMock = org . mockito . Mockito . mock ( com . cloud . storage . DiskOfferingVO . class ) ; org . mockito . Mockito . doReturn ( "" ) . when ( diskOfferingVoMock ) . getTags ( ) ; com . cloud . storage . StoragePool storagePoolMock = org . mockito . Mockito . mock ( com . cloud . storage . StoragePool . class ) ; org . mockito . Mockito . doReturn ( "A,B,C,D,X,Y" ) . when ( volumeApiServiceImpl ) . getStoragePoolTags ( storagePoolMock ) ; boolean result = volumeApiServiceImpl . doesTargetStorageSupportDiskOffering ( storagePoolMock , diskOfferingVoMock ) ; "<AssertPlaceHolder>" ; } doesTargetStorageSupportDiskOffering ( com . cloud . storage . StoragePool , java . lang . String ) { if ( org . apache . commons . lang . StringUtils . isBlank ( diskOfferingTags ) ) { return true ; } java . lang . String storagePoolTags = getStoragePoolTags ( destPool ) ; if ( org . apache . commons . lang . StringUtils . isBlank ( storagePoolTags ) ) { return false ; } java . lang . String [ ] storageTagsAsStringArray = org . apache . commons . lang . StringUtils . split ( storagePoolTags , "," ) ; java . lang . String [ ] newDiskOfferingTagsAsStringArray = org . apache . commons . lang . StringUtils . split ( diskOfferingTags , "," ) ; return org . apache . commons . collections . CollectionUtils . isSubCollection ( java . util . Arrays . asList ( newDiskOfferingTagsAsStringArray ) , java . util . Arrays . asList ( storageTagsAsStringArray ) ) ; }
|
org . junit . Assert . assertTrue ( result )
|
testGetVirtualSizeWhenVirtualSizeThrowsException ( ) { final long virtualSize = 2000 ; final long actualSize = 1000 ; final java . io . File mockFile = org . mockito . Mockito . mock ( java . io . File . class ) ; org . mockito . Mockito . when ( mockFile . length ( ) ) . thenReturn ( actualSize ) ; org . mockito . Mockito . doThrow ( new java . io . IOException ( "virtual<sp>size<sp>calculation<sp>failed" ) ) . when ( this . processor ) . getTemplateVirtualSize ( ( ( java . io . File ) ( org . mockito . Mockito . any ( ) ) ) ) ; "<AssertPlaceHolder>" ; org . mockito . Mockito . verify ( mockFile , org . mockito . Mockito . times ( 1 ) ) . length ( ) ; } getVirtualSize ( java . io . File ) { try { final long size = getTemplateVirtualSize ( file ) ; return size ; } catch ( final java . lang . Exception e ) { com . cloud . common . storageprocessor . QCOW2Processor . s_logger . info ( ( ( "[ignored]" + "failed<sp>to<sp>get<sp>template<sp>virtual<sp>size<sp>for<sp>QCOW2:<sp>" ) + ( e . getLocalizedMessage ( ) ) ) ) ; } return file . length ( ) ; }
|
org . junit . Assert . assertEquals ( actualSize , this . processor . getVirtualSize ( mockFile ) )
|
test_equals_otherClass ( ) { org . threeten . extra . PeriodDuration test = org . threeten . extra . PeriodDuration . of ( org . threeten . extra . TestPeriodDuration . P1Y2M3D , org . threeten . extra . TestPeriodDuration . DUR_5 ) ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ( this ) == otherAmount ) { return true ; } if ( otherAmount instanceof org . threeten . extra . Seconds ) { org . threeten . extra . Seconds other = ( ( org . threeten . extra . Seconds ) ( otherAmount ) ) ; return ( this . seconds ) == ( other . seconds ) ; } return false ; }
|
org . junit . Assert . assertEquals ( false , test . equals ( "" ) )
|
shouldReturnFalseIfServiceProvidedDateIsBlank ( ) { org . ei . drishti . util . SafeMap safeMap = new org . ei . drishti . util . SafeMap ( ) ; safeMap . put ( "entityId" , "child<sp>id<sp>1" ) ; safeMap . put ( "serviceProvidedDate" , "" ) ; safeMap . put ( "dateOfBirth" , "2011-01-01" ) ; boolean didRuleApply = rule . apply ( safeMap ) ; "<AssertPlaceHolder>" ; } apply ( org . ei . drishti . util . SafeMap ) { return new org . ei . drishti . service . reporting . ChildImmunization ( ) . isImmunizedWith ( ChildImmunizationFields . BCG_VALUE , reportFields ) ; }
|
org . junit . Assert . assertFalse ( didRuleApply )
|
getStringValueFromScopeWithoutDefaultValueTest ( ) { mockStorageRead ( org . uberfire . preferences . backend . PreferenceStoreImplTest . VALUE ) ; final java . lang . String value = preferenceStore . get ( allUsersEntireApplication , org . uberfire . preferences . backend . PreferenceStoreImplTest . KEY ) ; "<AssertPlaceHolder>" ; verify ( storage ) . read ( allUsersEntireApplication , org . uberfire . preferences . backend . PreferenceStoreImplTest . KEY ) ; } get ( java . lang . String , java . lang . String [ ] ) { return org . uberfire . java . nio . file . Paths . get ( first , more ) ; }
|
org . junit . Assert . assertEquals ( org . uberfire . preferences . backend . PreferenceStoreImplTest . VALUE , value )
|
testuint2bytes ( ) { int control ; byte [ ] observed ; byte [ ] expected ; control = 2145845231 ; expected = new byte [ ] { ( ( byte ) ( 127 ) ) , ( ( byte ) ( 230 ) ) , ( ( byte ) ( 255 ) ) , ( ( byte ) ( 239 ) ) } ; observed = org . openstack . atlas . util . converters . BitConverters . uint2bytes ( control ) ; "<AssertPlaceHolder>" ; } bytesEqual ( byte [ ] , byte [ ] ) { int i ; if ( ( a . length ) != ( b . length ) ) { return false ; } for ( i = 0 ; i < ( a . length ) ; i ++ ) { if ( ( a [ i ] ) != ( b [ i ] ) ) { return false ; } } return true ; }
|
org . junit . Assert . assertTrue ( bytesEqual ( expected , observed ) )
|
testInternalStorageGetsReused ( ) { org . neo4j . smack . pipeline . http . HttpHeaderContainer headers = new org . neo4j . smack . pipeline . http . HttpHeaderContainer ( ) ; headers . addHeader ( HttpHeaderNames . CONTENT_LENGTH , new org . neo4j . smack . gcfree . MutableString ( "0" ) ) ; headers . addHeader ( HttpHeaderNames . CONTENT_LENGTH , new org . neo4j . smack . gcfree . MutableString ( "0" ) ) ; headers . addHeader ( HttpHeaderNames . CONTENT_LENGTH , new org . neo4j . smack . gcfree . MutableString ( "0" ) ) ; headers . addHeader ( HttpHeaderNames . CONTENT_LENGTH , new org . neo4j . smack . gcfree . MutableString ( "0" ) ) ; int preClearCapacity = headers . getHeaders ( HttpHeaderNames . CONTENT_LENGTH ) . currentCapacity ( ) ; headers . clear ( ) ; headers . addHeader ( HttpHeaderNames . CONTENT_LENGTH , new org . neo4j . smack . gcfree . MutableString ( "0" ) ) ; headers . addHeader ( HttpHeaderNames . CONTENT_LENGTH , new org . neo4j . smack . gcfree . MutableString ( "0" ) ) ; headers . addHeader ( HttpHeaderNames . CONTENT_LENGTH , new org . neo4j . smack . gcfree . MutableString ( "0" ) ) ; headers . addHeader ( HttpHeaderNames . CONTENT_LENGTH , new org . neo4j . smack . gcfree . MutableString ( "0" ) ) ; "<AssertPlaceHolder>" ; } getHeaders ( org . neo4j . smack . pipeline . http . HttpHeaderName ) { return headers . get ( name ) ; }
|
org . junit . Assert . assertThat ( headers . getHeaders ( HttpHeaderNames . CONTENT_LENGTH ) . currentCapacity ( ) , org . hamcrest . Matchers . is ( preClearCapacity ) )
|
testDoGetDirAsZipWithPathId ( ) { java . lang . String path = "path" ; doReturn ( path ) . when ( fileResource . fileService ) . idToPath ( org . pentaho . platform . web . http . api . resources . FileResourceTest . PATH_ID ) ; doReturn ( true ) . when ( fileResource ) . isPathValid ( path ) ; doReturn ( true ) . when ( fileResource . policy ) . isAllowed ( PublishAction . NAME ) ; org . pentaho . platform . api . repository2 . unified . RepositoryFile mockFile = mock ( org . pentaho . platform . api . repository2 . unified . RepositoryFile . class ) ; doReturn ( mockFile ) . when ( fileResource . repository ) . getFile ( path ) ; javax . ws . rs . core . Response mockResponse = mock ( javax . ws . rs . core . Response . class ) ; doReturn ( mockResponse ) . when ( fileResource ) . doGetDirAsZip ( mockFile ) ; javax . ws . rs . core . Response testResponse = fileResource . doGetDirAsZip ( org . pentaho . platform . web . http . api . resources . FileResourceTest . PATH_ID ) ; "<AssertPlaceHolder>" ; verify ( fileResource . repository , times ( 1 ) ) . getFile ( path ) ; verify ( fileResource . policy , times ( 1 ) ) . isAllowed ( PublishAction . NAME ) ; verify ( fileResource , times ( 1 ) ) . isPathValid ( path ) ; verify ( fileResource . fileService , times ( 1 ) ) . idToPath ( org . pentaho . platform . web . http . api . resources . FileResourceTest . PATH_ID ) ; } doGetDirAsZip ( org . pentaho . platform . api . repository2 . unified . RepositoryFile ) { java . lang . String path = repositoryFile . getPath ( ) ; final java . io . InputStream is ; try { org . pentaho . platform . plugin . services . importexport . Exporter exporter = getExporter ( ) ; exporter . setRepoPath ( path ) ; exporter . setRepoWs ( org . pentaho . platform . web . http . api . resources . FileResource . repoWs ) ; java . io . File zipFile = exporter . doExportAsZip ( repositoryFile ) ; is = getFileInputStream ( zipFile ) ; } catch ( java . lang . Exception e ) { return buildServerErrorResponse ( e . toString ( ) ) ; } javax . ws . rs . core . StreamingOutput streamingOutput = getStreamingOutput ( is ) ; return buildOkResponse ( streamingOutput , org . pentaho . platform . web . http . api . resources . FileResource . APPLICATION_ZIP ) ; }
|
org . junit . Assert . assertEquals ( mockResponse , testResponse )
|
testTreeHeightOddLeaves ( ) { org . provebit . merkle . FileMerkle mTree = new org . provebit . merkle . FileMerkle ( HashType . SHA256 ) ; mTree . addTracking ( org . provebit . merkle . FileMerkleTest . incompleteDirPath , false ) ; mTree . makeTree ( ) ; "<AssertPlaceHolder>" ; } getHeight ( ) { return height ; }
|
org . junit . Assert . assertTrue ( ( ( mTree . getHeight ( ) ) == 4 ) )
|
testDiskNull ( ) { java . util . Map < java . lang . String , java . lang . Object > parameters = createBasicParams ( ) ; parameters . remove ( "disk" ) ; org . opennms . netmgt . poller . PollStatus status = monitor . poll ( createMonitor ( ) , parameters ) ; "<AssertPlaceHolder>" ; } isAvailable ( ) { return available ; }
|
org . junit . Assert . assertFalse ( status . isAvailable ( ) )
|
runTest ( ) { boolean result = checkNoError ( "Social_Profiles_Check_Colleague" ) ; "<AssertPlaceHolder>" ; } getNoErrorMsg ( ) { return noErrorMsg ; }
|
org . junit . Assert . assertTrue ( getNoErrorMsg ( ) , result )
|
testHashBaseForItemAnswersSkipsMissingResourcesForSimpleExtendedMatchingItems ( ) { final org . sakaiproject . tool . assessment . data . dao . assessment . ItemData item = newExtendedMatchingItem ( ) ; expectServerUrlLookup ( ) ; failResourceLookup ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 9 ] ) ; expectResourceLookupUnchecked ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 10 ] ) ; failResourceLookup ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 12 ] ) ; failResourceLookup ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 13 ] ) ; expectResourceLookupUnchecked ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 14 ] ) ; expectResourceLookupUnchecked ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 15 ] ) ; failResourceLookup ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 16 ] ) ; expectResourceLookupUnchecked ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 17 ] ) ; final java . lang . StringBuilder expectedHashBase = new java . lang . StringBuilder ( ) . append ( labeled ( "EmiLabel" , resourceDocTemplate1 ( fullUrlForContentResource ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 9 ] ) ) ) ) . append ( labeled ( "EmiLabel" , resourceDocTemplate1 ( expectedContentResourceHash1 ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 10 ] ) ) ) ) . append ( labeled ( "EmiCorrectOptionLabels" , "Answer<sp>Label<sp>3Answer<sp>Label<sp>5" ) ) . append ( labeled ( "EmiSequence" , ( "" + ( Long . MAX_VALUE ) ) ) ) . append ( labeled ( "EmiText" , resourceDocTemplate1 ( fullUrlForContentResource ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 12 ] ) ) ) ) . append ( expectedContentResourceHash1 ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 14 ] ) ) . append ( labeled ( "EmiCorrectOptionLabels" , "Answer<sp>Label<sp>6Answer<sp>Label<sp>8" ) ) . append ( labeled ( "EmiSequence" , ( "" + ( Long . MAX_VALUE ) ) ) ) . append ( labeled ( "EmiText" , resourceDocTemplate1 ( expectedContentResourceHash1 ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 15 ] ) ) ) ) . append ( expectedContentResourceHash1 ( org . sakaiproject . tool . assessment . facade . ItemHashUtilTest . CONTENT_RESOURCES [ 17 ] ) ) ; java . lang . StringBuilder actualHashBase = new java . lang . StringBuilder ( ) ; itemHashUtil . hashBaseForItemAnswers ( item , actualHashBase ) ; "<AssertPlaceHolder>" ; } toString ( ) { if ( name ( ) . equals ( "sessionId" ) ) { return "session-id" ; } else { return name ( ) ; } }
|
org . junit . Assert . assertThat ( actualHashBase . toString ( ) , org . hamcrest . CoreMatchers . equalTo ( expectedHashBase . toString ( ) ) )
|
testGetVserverType ( ) { parameters . put ( PropertyHandler . VSERVER_TYPE , new org . oscm . app . v2_0 . data . Setting ( PropertyHandler . VSERVER_TYPE , "type" ) ) ; propertyHandler = new org . oscm . app . iaas . PropertyHandler ( settings ) ; java . lang . String type = propertyHandler . getVserverType ( ) ; "<AssertPlaceHolder>" ; } getVserverType ( ) { return props . getValidatedProperty ( org . oscm . app . iaas . PropertyHandler . VSERVER_TYPE ) ; }
|
org . junit . Assert . assertEquals ( "type" , type )
|
testSnappyCompressionSimple ( ) { if ( checkNativeSnappy ( ) ) { return ; } java . io . File snappyFile = new java . io . File ( testMeta . getDir ( ) , "snappyTestFile.snappy" ) ; java . io . BufferedOutputStream os = new java . io . BufferedOutputStream ( new java . io . FileOutputStream ( snappyFile ) ) ; org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . conf . Configuration ( ) ; org . apache . hadoop . io . compress . CompressionCodec codec = ( ( org . apache . hadoop . io . compress . CompressionCodec ) ( org . apache . hadoop . util . ReflectionUtils . newInstance ( org . apache . hadoop . io . compress . SnappyCodec . class , conf ) ) ) ; org . apache . apex . malhar . lib . io . fs . FilterStreamCodec . SnappyFilterStream filterStream = new org . apache . apex . malhar . lib . io . fs . FilterStreamCodec . SnappyFilterStream ( codec . createOutputStream ( os ) ) ; int ONE_MB = 1024 * 1024 ; java . lang . String testStr = "TestSnap-16bytes" ; for ( int i = 0 ; i < ONE_MB ; i ++ ) { filterStream . write ( testStr . getBytes ( ) ) ; } filterStream . flush ( ) ; filterStream . close ( ) ; org . apache . hadoop . io . compress . CompressionInputStream is = codec . createInputStream ( new java . io . FileInputStream ( snappyFile ) ) ; byte [ ] recovered = new byte [ testStr . length ( ) ] ; int bytesRead = is . read ( recovered ) ; is . close ( ) ; "<AssertPlaceHolder>" ; } close ( ) { isAlive = false ; holdingBuffer . clear ( ) ; }
|
org . junit . Assert . assertEquals ( testStr , new java . lang . String ( recovered ) )
|
testModifyTicketRemoveUserNullTicketId ( ) { org . irods . jargon . ticket . packinstr . TicketAdminInp pi = org . irods . jargon . ticket . packinstr . TicketAdminInp . instanceForModifyRemoveAccess ( null , TicketModifyAddOrRemoveTypeEnum . TICKET_MODIFY_USER , userName ) ; "<AssertPlaceHolder>" ; } instanceForModifyRemoveAccess ( java . lang . String , org . irods . jargon . ticket . packinstr . TicketModifyAddOrRemoveTypeEnum , java . lang . String ) { if ( ( ticketId == null ) || ( ticketId . isEmpty ( ) ) ) { throw new java . lang . IllegalArgumentException ( "null<sp>or<sp>empty<sp>ticket<sp>id" ) ; } if ( addTypeEnum == null ) { throw new java . lang . IllegalArgumentException ( "null<sp>modify<sp>remove<sp>permission<sp>type<sp>not<sp>set" ) ; } if ( ( modObject == null ) || ( modObject . isEmpty ( ) ) ) { throw new java . lang . IllegalArgumentException ( "null<sp>or<sp>empty<sp>modify<sp>remove<sp>-<sp>user,<sp>group,<sp>or<sp>host" ) ; } return new org . irods . jargon . ticket . packinstr . TicketAdminInp ( TICKET_ADMIN_INP_API_NBR , "mod" , ticketId , "remove" , addTypeEnum . getTextValue ( ) , modObject , BLANK ) ; }
|
org . junit . Assert . assertNotNull ( pi )
|
testIntMethodCalledFromScriptWithResult ( ) { com . eclipsesource . v8 . V8CallbackTest . ICallback callback = mock ( com . eclipsesource . v8 . V8CallbackTest . ICallback . class ) ; doReturn ( 7 ) . when ( callback ) . intMethodNoParameters ( ) ; v8 . registerJavaMethod ( callback , "intMethodNoParameters" , "foo" , new java . lang . Class < ? > [ 0 ] ) ; int result = v8 . executeIntegerScript ( "foo();" ) ; "<AssertPlaceHolder>" ; } executeIntegerScript ( java . lang . String ) { return executeIntegerScript ( script , null , 0 ) ; }
|
org . junit . Assert . assertEquals ( 7 , result )
|
testGenericMapList ( ) { org . apache . servicecomb . it . schema . Generic < java . util . Map < java . lang . String , java . util . List < java . lang . String > > > mapListGeneric = new org . apache . servicecomb . it . schema . Generic ( ) ; java . util . Map < java . lang . String , java . util . List < java . lang . String > > map = new java . util . HashMap ( ) ; java . util . List < java . lang . String > list = new java . util . ArrayList ( ) ; list . add ( "hello" ) ; map . put ( "test" , list ) ; mapListGeneric . value = map ; @ org . apache . servicecomb . it . testcase . SuppressWarnings ( "unchecked" ) org . apache . servicecomb . it . schema . Generic < java . util . Map < java . lang . String , java . util . List < java . lang . String > > > result = org . apache . servicecomb . it . testcase . TestGenericEdge . client . postForObject ( "/genericMapList" , mapListGeneric , org . apache . servicecomb . it . schema . Generic . class ) ; java . lang . String test = result . value . get ( "test" ) . get ( 0 ) ; "<AssertPlaceHolder>" ; } get ( T [ ] ) { return instance [ idx ] ; }
|
org . junit . Assert . assertEquals ( "hello" , test )
|
normalizeTime ( ) { double value = com . att . aro . core . util . Util . normalizeTime ( ( - 0.0 ) , 1.2 ) ; "<AssertPlaceHolder>" ; } normalizeTime ( double , double ) { double tmpTime ; tmpTime = ( time > ( com . att . aro . core . util . Util . TIME_CORRECTION ) ) ? time - pcapTime : time ; if ( tmpTime < 0 ) { tmpTime = 0.0 ; } return tmpTime ; }
|
org . junit . Assert . assertNotNull ( ( value > 0 ) )
|
testReadCSV ( ) { java . util . List < io . github . biezhi . excel . plus . model . Book > books = io . github . biezhi . excel . plus . Reader . create ( io . github . biezhi . excel . plus . model . Book . class ) . from ( new java . io . File ( ( ( classPath ( ) ) + "/book.csv" ) ) ) . start ( 0 ) . asList ( ) ; log . info ( "{}" , books ) ; "<AssertPlaceHolder>" ; } asList ( ) { java . util . stream . Stream < T > stream = this . asStream ( ) ; return stream . collect ( io . github . biezhi . excel . plus . Reader . toList ( ) ) ; }
|
org . junit . Assert . assertEquals ( 5 , books . size ( ) )
|
testGetVersionNo_fileNotExistent ( ) { org . seasar . extension . jdbc . gen . internal . version . DdlInfoFileImpl ddlInfoFile = new org . seasar . extension . jdbc . gen . internal . version . DdlInfoFileImpl ( new java . io . File ( "notExistent" ) ) ; "<AssertPlaceHolder>" ; } getCurrentVersionNo ( ) { return getCurrentVersionNoInternal ( ) ; }
|
org . junit . Assert . assertEquals ( 0 , ddlInfoFile . getCurrentVersionNo ( ) )
|
testRootTokenInStarLoop ( ) { org . antlr . tool . Grammar g = new org . antlr . tool . Grammar ( ( "grammar<sp>Expr;\n" + ( "options<sp>{<sp>output=AST;<sp>backtrack=true;<sp>}\n" + "a<sp>:<sp>(\'*\'^)*<sp>;\n" ) ) ) ; java . lang . String expecting = "(rule<sp>synpred1_Expr<sp>ARG<sp>RET<sp>scope<sp>(BLOCK<sp>(ALT<sp>'*'<sp><end-of-alt>)<sp><end-of-block>)<sp><end-of-rule>)" ; java . lang . String found = g . getRule ( "synpred1_Expr" ) . tree . toStringTree ( ) ; "<AssertPlaceHolder>" ; } toStringTree ( ) { if ( ( ( children ) == null ) || ( children . isEmpty ( ) ) ) { return this . toString ( ) ; } java . lang . StringBuilder buf = new java . lang . StringBuilder ( ) ; if ( ! ( isNil ( ) ) ) { buf . append ( "(" ) ; buf . append ( this . toString ( ) ) ; buf . append ( '<sp>' ) ; } for ( int i = 0 ; ( ( children ) != null ) && ( i < ( children . size ( ) ) ) ; i ++ ) { org . antlr . runtime . tree . Tree t = ( ( org . antlr . runtime . tree . Tree ) ( children . get ( i ) ) ) ; if ( i > 0 ) { buf . append ( '<sp>' ) ; } buf . append ( t . toStringTree ( ) ) ; } if ( ! ( isNil ( ) ) ) { buf . append ( ")" ) ; } return buf . toString ( ) ; }
|
org . junit . Assert . assertEquals ( expecting , found )
|
testSplitURL_String ( ) { logger . info ( "splitURL" ) ; java . lang . String URLString = "http://user:password@www.example.com:8080/to/path/document?arg1=val1&arg2=val2#part" ; java . util . Map < com . datumbox . framework . core . common . text . parsers . URLParser . URLParts , java . lang . String > expResult = new java . util . HashMap ( ) ; expResult . put ( URLParser . URLParts . PROTOCOL , "www.example.com" 0 ) ; expResult . put ( URLParser . URLParts . PATH , "/to/path/document" ) ; expResult . put ( URLParser . URLParts . HOST , "www.example.com" ) ; expResult . put ( URLParser . URLParts . PORT , "8080" ) ; expResult . put ( URLParser . URLParts . USERINFO , "user:password" ) ; expResult . put ( URLParser . URLParts . FILENAME , "/to/path/document?arg1=val1&arg2=val2" ) ; expResult . put ( URLParser . URLParts . QUERY , "arg1=val1&arg2=val2" ) ; expResult . put ( URLParser . URLParts . AUTHORITY , "user:password@www.example.com:8080" ) ; expResult . put ( URLParser . URLParts . REF , "part" ) ; java . util . Map < com . datumbox . framework . core . common . text . parsers . URLParser . URLParts , java . lang . String > result = com . datumbox . framework . core . common . text . parsers . URLParser . splitURL ( URLString ) ; "<AssertPlaceHolder>" ; } splitURL ( java . lang . String ) { try { return com . datumbox . framework . core . common . text . parsers . URLParser . splitURL ( new java . net . URL ( URLString ) ) ; } catch ( java . net . MalformedURLException ex ) { throw new java . lang . RuntimeException ( ex ) ; } }
|
org . junit . Assert . assertEquals ( expResult , result )
|
openUiMessages ( ) { java . io . InputStream stream = org . oscm . ui . resources . DefaultMessages . openUiMessages ( Locale . ENGLISH ) ; java . lang . String result = org . oscm . string . Strings . toString ( org . oscm . stream . Streams . readFrom ( stream ) ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return false ; }
|
org . junit . Assert . assertFalse ( result . isEmpty ( ) )
|
searchUncategorizedProcessDeploymentInfosSupervisedBy ( ) { final org . bonitasoft . engine . persistence . QueryOptions options = new org . bonitasoft . engine . persistence . QueryOptions ( 0 , 10 ) ; final long userId = 9 ; when ( persistenceService . searchEntity ( org . bonitasoft . engine . core . process . definition . model . SProcessDefinitionDeployInfo . class , "UncategorizedAndWithSupervisor" , options , java . util . Collections . singletonMap ( "userId" , ( ( java . lang . Object ) ( userId ) ) ) ) ) . thenReturn ( new java . util . ArrayList < org . bonitasoft . engine . core . process . definition . model . SProcessDefinitionDeployInfo > ( ) ) ; final java . util . List < org . bonitasoft . engine . core . process . definition . model . SProcessDefinitionDeployInfo > result = processDefinitionServiceImpl . searchUncategorizedProcessDeploymentInfosSupervisedBy ( userId , options ) ; "<AssertPlaceHolder>" ; } searchUncategorizedProcessDeploymentInfosSupervisedBy ( long , org . bonitasoft . engine . search . SearchOptions ) { final org . bonitasoft . engine . service . TenantServiceAccessor tenantAccessor = getTenantAccessor ( ) ; final org . bonitasoft . engine . search . descriptor . SearchEntitiesDescriptor searchEntitiesDescriptor = tenantAccessor . getSearchEntitiesDescriptor ( ) ; final org . bonitasoft . engine . core . process . definition . ProcessDefinitionService processDefinitionService = tenantAccessor . getProcessDefinitionService ( ) ; final org . bonitasoft . engine . search . descriptor . SearchProcessDefinitionsDescriptor searchDescriptor = searchEntitiesDescriptor . getSearchProcessDefinitionsDescriptor ( ) ; final org . bonitasoft . engine . search . process . SearchUncategorizedProcessDeploymentInfosSupervisedBy transactionSearch = new org . bonitasoft . engine . search . process . SearchUncategorizedProcessDeploymentInfosSupervisedBy ( processDefinitionService , searchDescriptor , searchOptions , userId ) ; try { transactionSearch . execute ( ) ; } catch ( final org . bonitasoft . engine . commons . exceptions . SBonitaException e ) { throw new org . bonitasoft . engine . exception . SearchException ( "Problem<sp>encountered<sp>while<sp>searching<sp>for<sp>Uncategorized<sp>Process<sp>Definitions<sp>for<sp>a<sp>supervisor" , e ) ; } return transactionSearch . getResult ( ) ; }
|
org . junit . Assert . assertNotNull ( result )
|
encryptInTryWithResources_decryptInTryWithRessources_yieldsOriginalPlaintext ( ) { java . io . ByteArrayOutputStream result = new java . io . ByteArrayOutputStream ( ) ; try ( java . io . BufferedOutputStream bufferedOutputStream = new java . io . BufferedOutputStream ( result , 16384 ) ; final java . io . OutputStream outputStream = name . neuhalfen . projects . crypto . bouncycastle . openpgp . BouncyGPG . encryptToStream ( ) . withConfig ( name . neuhalfen . projects . crypto . bouncycastle . openpgp . testtooling . Configs . keyringConfigFromFilesForSender ( ) ) . withAlgorithms ( algorithmSuite ) . toRecipient ( "recipient@example.com" ) . andSignWith ( "sender@example.com" ) . binaryOutput ( ) . andWriteTo ( bufferedOutputStream ) ; final java . io . InputStream is = new java . io . ByteArrayInputStream ( ExampleMessages . IMPORTANT_QUOTE_TEXT . getBytes ( ) ) ) { org . bouncycastle . util . io . Streams . pipeAll ( is , outputStream ) ; } final byte [ ] ciphertext = result . toByteArray ( ) ; final java . io . ByteArrayOutputStream plainBA = new java . io . ByteArrayOutputStream ( ) ; try ( final java . io . InputStream plainIS = name . neuhalfen . projects . crypto . bouncycastle . openpgp . BouncyGPG . decryptAndVerifyStream ( ) . withConfig ( name . neuhalfen . projects . crypto . bouncycastle . openpgp . testtooling . Configs . keyringConfigFromFilesForRecipient ( ) ) . andRequireSignatureFromAllKeys ( "sender@example.com" ) . fromEncryptedInputStream ( new java . io . ByteArrayInputStream ( ciphertext ) ) ) { org . bouncycastle . util . io . Streams . pipeAll ( plainIS , plainBA ) ; } "<AssertPlaceHolder>" ; } keyringConfigFromFilesForRecipient ( ) { return name . neuhalfen . projects . crypto . bouncycastle . openpgp . testtooling . Configs . keyringConfigFromFilesForRecipient ( name . neuhalfen . projects . crypto . bouncycastle . openpgp . keys . callbacks . KeyringConfigCallbacks . withPasswordsFromMap ( ExampleMessages . ALL_KEYRINGS_PASSWORDS ) ) ; }
|
org . junit . Assert . assertArrayEquals ( ExampleMessages . IMPORTANT_QUOTE_TEXT . getBytes ( ) , plainBA . toByteArray ( ) )
|
test2502355_zoomOutBoth ( ) { org . jfree . data . xy . DefaultXYDataset dataset = new org . jfree . data . xy . DefaultXYDataset ( ) ; org . jfree . chart . JFreeChart chart = org . jfree . chart . ChartFactory . createXYLineChart ( "TestChart" , "X" , "Y" , dataset ) ; org . jfree . chart . ChartPanel panel = new org . jfree . chart . ChartPanel ( chart ) ; chart . addChangeListener ( this ) ; this . chartChangeEvents . clear ( ) ; panel . zoomOutBoth ( 1.0 , 2.0 ) ; "<AssertPlaceHolder>" ; } size ( ) { return this . tickUnits . size ( ) ; }
|
org . junit . Assert . assertEquals ( 1 , this . chartChangeEvents . size ( ) )
|
ifThereisNotPathReturnEmptyArray ( ) { org . zkoss . ganttz . util . MutableTreeModelTest . Prueba root = new org . zkoss . ganttz . util . MutableTreeModelTest . Prueba ( ) ; org . zkoss . ganttz . util . MutableTreeModel < org . zkoss . ganttz . util . MutableTreeModelTest . Prueba > model = org . zkoss . ganttz . util . MutableTreeModel . create ( org . zkoss . ganttz . util . MutableTreeModelTest . Prueba . class , root ) ; org . zkoss . ganttz . util . MutableTreeModelTest . Prueba child = new org . zkoss . ganttz . util . MutableTreeModelTest . Prueba ( ) ; model . add ( root , child ) ; "<AssertPlaceHolder>" ; } equalTo ( T extends java . lang . Comparable ) { return org . zkoss . ganttz . data . constraint . ConstraintOnComparableValues . instantiate ( org . zkoss . ganttz . data . constraint . ConstraintOnComparableValues . ComparisonType . EQUAL_TO , value ) ; }
|
org . junit . Assert . assertThat ( model . getPath ( child , root ) , org . hamcrest . CoreMatchers . equalTo ( new int [ 0 ] ) )
|
testRefineToRuntime ( ) { java . lang . String expected = "project/runtime" ; org . sonatype . aether . graph . DependencyNode node = parser . parseLiteral ( "gid:aid:ext:ver:runtime" ) ; node . setRequestContext ( "project" ) ; org . sonatype . aether . graph . DependencyNode refinedNode = refiner . transformGraph ( node , context ) ; "<AssertPlaceHolder>" ; } getRequestContext ( ) { return context ; }
|
org . junit . Assert . assertEquals ( expected , refinedNode . getRequestContext ( ) )
|
shouldSerialiseAndDeserialiseOperationChain ( ) { final uk . gov . gchq . gaffer . operation . OperationChain opChain = service . execute ( ) ; byte [ ] bytes = uk . gov . gchq . gaffer . jsonserialisation . JSONSerialiser . serialise ( opChain ) ; final uk . gov . gchq . gaffer . operation . OperationChain deserialisedOp = uk . gov . gchq . gaffer . jsonserialisation . JSONSerialiser . deserialise ( bytes , opChain . getClass ( ) ) ; "<AssertPlaceHolder>" ; } deserialise ( java . lang . String , java . lang . Class ) { try { return uk . gov . gchq . gaffer . jsonserialisation . JSONSerialiser . getInstance ( ) . mapper . readValue ( json , clazz ) ; } catch ( final java . io . IOException e ) { throw new uk . gov . gchq . gaffer . exception . SerialisationException ( e . getMessage ( ) , e ) ; } }
|
org . junit . Assert . assertNotNull ( deserialisedOp )
|
databaseMappingDefaultsToPrimaryWhenNothingMatches ( ) { com . hotels . bdp . waggledance . mapping . model . DatabaseMapping databaseMapping = service . databaseMapping ( "some_unknown_prefix_db" ) ; "<AssertPlaceHolder>" ; } getDatabasePrefix ( ) { return databasePrefix ; }
|
org . junit . Assert . assertThat ( databaseMapping . getDatabasePrefix ( ) , org . hamcrest . CoreMatchers . is ( "" ) )
|
testRoundRobinMap ( ) { org . stringtemplate . v4 . STGroup group = new org . stringtemplate . v4 . STGroup ( ) ; group . defineTemplate ( "a" , "x" , "[<x>]" ) ; group . defineTemplate ( "b" , "x" , "(<x>)" ) ; group . defineTemplate ( "test" , "name" , "hi<sp><name:a(),b()>!" ) ; org . stringtemplate . v4 . ST st = group . getInstanceOf ( "test" ) ; st . add ( "name" , "Ter" ) ; st . add ( "name" , "Tom" ) ; st . add ( "name" , "a" 0 ) ; java . lang . String expected = "a" 1 ; java . lang . String result = st . render ( ) ; "<AssertPlaceHolder>" ; } render ( ) { return render ( java . util . Locale . getDefault ( ) ) ; }
|
org . junit . Assert . assertEquals ( expected , result )
|
canWrite ( ) { org . sqlite . SQLiteConfig config = new org . sqlite . SQLiteConfig ( ) ; config . enforceForeignKeys ( true ) ; java . sql . Connection conn = java . sql . DriverManager . getConnection ( "jdbc:sqlite:" , config . toProperties ( ) ) ; java . sql . Statement stat = conn . createStatement ( ) ; try { "<AssertPlaceHolder>" ; } finally { stat . close ( ) ; conn . close ( ) ; } } isReadOnly ( ) { return conn . isReadOnly ( ) ; }
|
org . junit . Assert . assertFalse ( conn . isReadOnly ( ) )
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.