input
stringlengths
28
18.7k
output
stringlengths
39
1.69k
testAddFormulaEmpty ( ) { final org . prop4j . explain . solvers . SatProblem instance = getInstance ( ) ; instance . addFormula ( new org . prop4j . And ( ) ) ; final java . util . List < org . prop4j . Node > expected = java . util . Collections . emptyList ( ) ; final java . util . List < org . prop4j . Node > actual = instance . getClauses ( ) ; "<AssertPlaceHolder>" ; } getClauses ( ) { return clauses ; }
org . junit . Assert . assertEquals ( expected , actual )
testArraySelectWithALL ( ) { long ts = nextTimestamp ( ) ; java . lang . String tenantId = getOrganizationId ( ) ; org . apache . phoenix . end2end . ArrayIT . createTableWithArray ( getUrl ( ) , getDefaultSplits ( tenantId ) , null , ( ts - 2 ) ) ; java . sql . Connection conn = null ; try { org . apache . phoenix . end2end . ArrayIT . createSimpleTableWithArray ( getUrl ( ) , getDefaultSplits ( tenantId ) , null , ( ts - 2 ) ) ; org . apache . phoenix . end2end . ArrayIT . initSimpleArrayTable ( tenantId , null , ts , false ) ; java . util . Properties props = org . apache . phoenix . util . PropertiesUtil . deepCopy ( org . apache . phoenix . end2end . TEST_PROPERTIES ) ; props . setProperty ( PhoenixRuntime . CURRENT_SCN_ATTRIB , java . lang . Long . toString ( ( ts + 2 ) ) ) ; conn = java . sql . DriverManager . getConnection ( getUrl ( ) , props ) ; java . lang . String query = ( "SELECT<sp>a_double_array[1]<sp>FROM<sp>" + ( org . apache . phoenix . end2end . ArrayIT . SIMPLE_TABLE_WITH_ARRAY ) ) + "<sp>WHERE<sp>64.87d<sp>=<sp>ALL(a_double_array)" ; java . sql . PreparedStatement statement = conn . prepareStatement ( query ) ; java . sql . ResultSet rs = statement . executeQuery ( ) ; "<AssertPlaceHolder>" ; } finally { if ( conn != null ) { conn . close ( ) ; } } } next ( ) { if ( ! ( hasNext ( ) ) ) { throw new java . util . NoSuchElementException ( ) ; } return children . get ( ( ( position ) ++ ) ) ; }
org . junit . Assert . assertFalse ( rs . next ( ) )
testMissingOldVersion ( ) { japicmp . maven . Parameter parameter = new japicmp . maven . Parameter ( ) ; parameter . setIgnoreMissingOldVersion ( "true" ) ; japicmp . maven . VersionChange vc = new japicmp . maven . VersionChange ( java . util . Collections . singletonList ( new japicmp . cmp . JApiCmpArchive ( new java . io . File ( "lib-1.2.3.jar" ) , "1.2.3" ) ) , java . util . Collections . singletonList ( new japicmp . cmp . JApiCmpArchive ( new java . io . File ( "lib-1.2.3.jar" ) , "1.2.3" ) ) , parameter ) ; "<AssertPlaceHolder>" ; } computeChangeType ( ) { if ( this . oldVersions . isEmpty ( ) ) { if ( ! ( ignoreMissingOldVersion ) ) { throw new japicmp . exception . JApiCmpException ( JApiCmpException . Reason . IllegalArgument , ( "Could<sp>not<sp>extract<sp>semantic<sp>version<sp>for<sp>at<sp>least<sp>one<sp>old<sp>version.<sp>Please<sp>" + "follow<sp>the<sp>rules<sp>for<sp>semantic<sp>versioning." ) ) ; } else { return japicmp . util . Optional . absent ( ) ; } } if ( this . newVersions . isEmpty ( ) ) { if ( ! ( ignoreMissingNewVersion ) ) { throw new japicmp . exception . JApiCmpException ( JApiCmpException . Reason . IllegalArgument , ( "Could<sp>not<sp>extract<sp>semantic<sp>version<sp>for<sp>at<sp>least<sp>one<sp>new<sp>version.<sp>Please<sp>" + "follow<sp>the<sp>rules<sp>for<sp>semantic<sp>versioning." ) ) ; } else { return japicmp . util . Optional . absent ( ) ; } } if ( ( allVersionsTheSame ( oldVersions ) ) && ( allVersionsTheSame ( newVersions ) ) ) { japicmp . versioning . SemanticVersion oldVersion = oldVersions . get ( 0 ) ; japicmp . versioning . SemanticVersion newVersion = newVersions . get ( 0 ) ; return oldVersion . computeChangeType ( newVersion ) ; } else { if ( ( oldVersions . size ( ) ) != ( newVersions . size ( ) ) ) { throw new japicmp . exception . JApiCmpException ( JApiCmpException . Reason . IllegalArgument , "Cannot<sp>compare<sp>versions<sp>because<sp>the<sp>number<sp>of<sp>old<sp>versions<sp>is<sp>different<sp>than<sp>the<sp>number<sp>of<sp>new<sp>versions." ) ; } else { java . util . List < japicmp . versioning . SemanticVersion . ChangeType > changeTypes = new java . util . ArrayList ( ) ; for ( int i = 0 ; i < ( oldVersions . size ( ) ) ; i ++ ) { japicmp . versioning . SemanticVersion oldVersion = oldVersions . get ( i ) ; japicmp . versioning . SemanticVersion newVersion = newVersions . get ( i ) ; japicmp . util . Optional < japicmp . versioning . SemanticVersion . ChangeType > changeTypeOptional = oldVersion . computeChangeType ( newVersion ) ; if ( changeTypeOptional . isPresent ( ) ) { changeTypes . add ( changeTypeOptional . get ( ) ) ; } } japicmp . versioning . SemanticVersion . ChangeType maxRank = SemanticVersion . ChangeType . UNCHANGED ; for ( japicmp . versioning . SemanticVersion . ChangeType changeType : changeTypes ) { if ( ( changeType . getRank ( ) ) > ( maxRank . getRank ( ) ) ) { maxRank = changeType ; } } return japicmp . util . Optional . fromNullable ( maxRank ) ; } } }
org . junit . Assert . assertThat ( vc . computeChangeType ( ) . get ( ) , org . hamcrest . core . Is . is ( SemanticVersion . ChangeType . UNCHANGED ) )
testCreationAndReopen ( ) { com . puresoltechnologies . ductiledb . storage . spi . StorageConfiguration configuration = com . puresoltechnologies . ductiledb . logstore . LogStructuredStoreTestUtils . createStorageConfiguration ( ) ; com . puresoltechnologies . ductiledb . storage . spi . Storage storage = com . puresoltechnologies . ductiledb . storage . api . StorageFactory . getStorageInstance ( configuration ) ; java . io . File directory = new java . io . File ( "LogStoreCreateAndReopenIT.testCreationAndReopen" ) ; if ( storage . exists ( directory ) ) { storage . removeDirectory ( directory , true ) ; } try ( com . puresoltechnologies . ductiledb . logstore . LogStructuredStore store = com . puresoltechnologies . ductiledb . logstore . LogStructuredStore . create ( storage , directory , new com . puresoltechnologies . ductiledb . logstore . LogStoreConfiguration ( ) ) ) { store . open ( ) ; store . put ( com . puresoltechnologies . ductiledb . logstore . Key . of ( "Key" ) , com . puresoltechnologies . ductiledb . commons . Bytes . fromString ( "Value" ) ) ; } try ( com . puresoltechnologies . ductiledb . logstore . LogStructuredStore store = com . puresoltechnologies . ductiledb . logstore . LogStructuredStore . open ( storage , directory ) ) { store . open ( ) ; "<AssertPlaceHolder>" ; } } toString ( byte [ ] ) { return com . puresoltechnologies . ductiledb . commons . Bytes . converter . toString ( bytes , com . puresoltechnologies . ductiledb . commons . Bytes . defaultCharset ) ; }
org . junit . Assert . assertEquals ( "Value" , com . puresoltechnologies . ductiledb . commons . Bytes . toString ( store . get ( com . puresoltechnologies . ductiledb . logstore . Key . of ( "Key" ) ) ) )
getOpNum ( ) { "<AssertPlaceHolder>" ; } getOpNum ( ) { return opNum ; }
org . junit . Assert . assertEquals ( 10 , request . getOpNum ( ) )
testRelativePathForJarLaunch ( ) { openwhisk . java . local . Launcher launcher = new openwhisk . java . local . Launcher ( ) ; java . nio . file . Path jarPath = java . nio . file . Paths . get ( "./src/test/resources/serverlessJava.jar" ) ; launcher . setBinaryPath ( jarPath ) ; launcher . setEntryClassName ( "aproject.App" ) ; com . google . gson . JsonObject result = launcher . launch ( ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertNotNull ( result )
test_SingleDir ( ) { de . vandermeer . svg2vector . applications . is . Svg2Vector_IS app = new de . vandermeer . svg2vector . applications . is . Svg2Vector_IS ( ) ; java . lang . String [ ] args = org . apache . commons . lang3 . ArrayUtils . addAll ( de . vandermeer . svg2vector . applications . is . Test_Svg2Vector_IS_EPS . STD_OPTIONS , "-f" , "src/test/resources/svg-files/chomsky-hierarchy.svgz" , "-d" , de . vandermeer . svg2vector . applications . is . Test_Svg2Vector_IS_EPS . OUT_DIR_PREFIX ) ; "<AssertPlaceHolder>" ; } executeApplication ( java . lang . String [ ] ) { int ret = super . executeApplication ( args ) ; if ( ret != 0 ) { return ret ; } de . vandermeer . svg2vector . applications . base . SvgTargets target = this . getProps ( ) . getTarget ( ) ; java . lang . String fn = this . optionInkscapeExec . getValue ( ) ; if ( ( ret = this . checkIsExecutable ( fn ) ) < 0 ) { return ret ; } this . printDetailMessage ( ( "Inkscape<sp>exec:<sp>" + fn ) ) ; this . setWarnings ( target ) ; de . vandermeer . svg2vector . applications . is . IsCmd isCmd = new de . vandermeer . svg2vector . applications . is . IsCmd ( fn , target , this . getProps ( ) ) ; isCmd . appendTargetSettings ( target , this . optionExpDpi , this . optionExpPdfver , this . optionExpPslevel ) ; de . vandermeer . svg2vector . applications . is . IsCmd isTmpCmd = new de . vandermeer . svg2vector . applications . is . IsCmd ( fn , de . vandermeer . svg2vector . applications . base . SvgTargets . svg , this . getProps ( ) ) ; if ( this . optionSvgFirst . inCli ( ) ) { this . printProgressMessage ( "converting<sp>to<sp>temporary<sp>SVG<sp>first" ) ; this . printDetailMessage ( ( "Inkscape<sp>cmd<sp>tmp:<sp>" + isTmpCmd ) ) ; } else { this . printProgressMessage ( "converting<sp>directly<sp>to<sp>target" ) ; this . printDetailMessage ( ( "Inkscape<sp>cmd:<sp>" + isCmd ) ) ; } ret = this . createTempArtifacts ( isTmpCmd ) ; if ( ret < 0 ) { return ret ; } ret = this . convertInput ( isCmd , target ) ; if ( ret < 0 ) { return ret ; } this . removeTempArtifacts ( ) ; this . printProgressMessage ( "finished<sp>successfully" ) ; return 0 ; }
org . junit . Assert . assertEquals ( 0 , app . executeApplication ( args ) )
test_readCurrentState ( ) { test_addstate ( ) ; org . apache . eagle . alert . coordination . model . ScheduleState state = org . apache . eagle . alert . metadata . impl . JdbcImplTest . dao . getScheduleState ( ) ; "<AssertPlaceHolder>" ; org . apache . eagle . alert . metadata . impl . JdbcImplTest . LOG . debug ( state . getVersion ( ) ) ; org . apache . eagle . alert . metadata . impl . JdbcImplTest . LOG . debug ( state . getGenerateTime ( ) ) ; } getScheduleState ( ) { org . bson . BsonDocument sort = new org . bson . BsonDocument ( ) ; sort . append ( "generateTime" , new org . bson . BsonInt32 ( ( - 1 ) ) ) ; org . apache . eagle . alert . coordination . model . ScheduleState state = scheduleStates . find ( ) . sort ( sort ) . map ( new com . mongodb . Function < org . bson . Document , org . apache . eagle . alert . coordination . model . ScheduleState > ( ) { @ org . apache . eagle . alert . metadata . impl . Override public org . apache . eagle . alert . coordination . model . ScheduleState apply ( org . bson . Document t ) { java . lang . String json = t . toJson ( ) ; try { return org . apache . eagle . alert . metadata . impl . MongoMetadataDaoImpl . mapper . readValue ( json , org . apache . eagle . alert . coordination . model . ScheduleState . class ) ; } catch ( java . io . IOException e ) { org . apache . eagle . alert . metadata . impl . MongoMetadataDaoImpl . LOG . error ( "deserialize<sp>config<sp>item<sp>failed!" , e ) ; } return null ; } } ) . first ( ) ; if ( state != null ) { java . lang . String version = state . getVersion ( ) ; state = addDetailForScheduleState ( state , version ) ; } return state ; }
org . junit . Assert . assertNotNull ( state )
testRmLeftSide ( ) { org . antlr . v4 . runtime . misc . IntervalSet s = org . antlr . v4 . runtime . misc . IntervalSet . of ( 1 , 10 ) ; s . add ( ( - 3 ) , ( - 3 ) ) ; s . remove ( 1 ) ; java . lang . String expecting = "{-3,<sp>2..10}" ; java . lang . String result = s . toString ( ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( getTarget ( ) ) + ":" ) + ( getTestName ( ) ) ; }
org . junit . Assert . assertEquals ( expecting , result )
testSortedEntries4 ( ) { tl . lin . data . map . HMapIV < java . lang . String > m = new tl . lin . data . map . HMapIV < java . lang . String > ( ) ; tl . lin . data . map . MapIV < java . lang . String > [ ] e = m . getEntriesSortedByValue ( ) ; "<AssertPlaceHolder>" ; } getEntriesSortedByValue ( ) { if ( ( this . size ( ) ) == 0 ) return null ; it . unimi . dsi . fastutil . ints . Int2FloatMap [ ] entries = new it . unimi . dsi . fastutil . ints . Int2FloatMap . Entry [ this . size ( ) ] ; entries = this . int2FloatEntrySet ( ) . toArray ( entries ) ; java . util . Arrays . sort ( entries , new java . util . Comparator < it . unimi . dsi . fastutil . ints . Int2FloatMap . Entry > ( ) { public int compare ( it . unimi . dsi . fastutil . ints . Int2FloatMap . Entry e1 , it . unimi . dsi . fastutil . ints . Int2FloatMap . Entry e2 ) { if ( ( e1 . getFloatValue ( ) ) > ( e2 . getFloatValue ( ) ) ) { return - 1 ; } else if ( ( e1 . getFloatValue ( ) ) < ( e2 . getFloatValue ( ) ) ) { return 1 ; } if ( ( e1 . getIntKey ( ) ) == ( e2 . getIntKey ( ) ) ) return 0 ; return ( e1 . getIntKey ( ) ) > ( e2 . getIntKey ( ) ) ? 1 : - 1 ; } } ) ; return entries ; }
org . junit . Assert . assertTrue ( ( e == null ) )
jarUriWithFileStringifiesToOriginalForm ( ) { org . eclipse . rdf4j . common . net . ParsedIRI uri = new org . eclipse . rdf4j . common . net . ParsedIRI ( "jar:file:///some-file.jar!/another-file" ) ; "<AssertPlaceHolder>" ; } toString ( ) { return getName ( ) ; }
org . junit . Assert . assertEquals ( "jar:file:///some-file.jar!/another-file" , uri . toString ( ) )
testAddHiveStepToEmrCluster ( ) { org . finra . herd . model . api . xml . EmrHiveStepAddRequest emrHiveStepAddRequest = new org . finra . herd . model . api . xml . EmrHiveStepAddRequest ( NAMESPACE , EMR_CLUSTER_DEFINITION_NAME , EMR_CLUSTER_NAME , EMR_STEP_NAME , EMR_STEP_SCRIPT_LOCATION , java . util . Arrays . asList ( org . finra . herd . rest . ATTRIBUTE_NAME_1_MIXED_CASE ) , CONTINUE_ON_ERROR , EMR_CLUSTER_ID , AWS_ACCOUNT_ID ) ; org . finra . herd . model . api . xml . EmrHiveStep emrHiveStep = new org . finra . herd . model . api . xml . EmrHiveStep ( EMR_STEP_ID , NAMESPACE , EMR_CLUSTER_DEFINITION_NAME , EMR_CLUSTER_NAME , EMR_STEP_NAME , EMR_STEP_SCRIPT_LOCATION , java . util . Arrays . asList ( org . finra . herd . rest . ATTRIBUTE_NAME_1_MIXED_CASE ) , CONTINUE_ON_ERROR , EMR_CLUSTER_ID ) ; when ( emrService . addStepToCluster ( emrHiveStepAddRequest ) ) . thenReturn ( emrHiveStep ) ; org . finra . herd . model . api . xml . EmrHiveStep result = emrRestController . addHiveStepToEmrCluster ( emrHiveStepAddRequest ) ; verify ( emrService ) . addStepToCluster ( emrHiveStepAddRequest ) ; verifyNoMoreInteractionsHelper ( ) ; "<AssertPlaceHolder>" ; } verifyNoMoreInteractionsHelper ( ) { verifyNoMoreInteractions ( awsHelper , javaPropertiesHelper , retryPolicyFactory , s3Operations ) ; }
org . junit . Assert . assertEquals ( emrHiveStep , result )
testForDateArray ( ) { java . sql . Date [ ] dateArr = new java . sql . Date [ 2 ] ; dateArr [ 0 ] = new java . sql . Date ( java . lang . System . currentTimeMillis ( ) ) ; dateArr [ 1 ] = new java . sql . Date ( ( ( java . lang . System . currentTimeMillis ( ) ) + ( java . lang . System . currentTimeMillis ( ) ) ) ) ; com . salesforce . phoenix . schema . PhoenixArray arr = com . salesforce . phoenix . schema . PArrayDataType . instantiatePhoenixArray ( PDataType . DATE , dateArr ) ; PDataType . DATE_ARRAY . toObject ( arr , PDataType . DATE_ARRAY ) ; byte [ ] bytes = PDataType . DATE_ARRAY . toBytes ( arr ) ; com . salesforce . phoenix . schema . PhoenixArray resultArr = ( ( com . salesforce . phoenix . schema . PhoenixArray ) ( PDataType . DATE_ARRAY . toObject ( bytes , 0 , bytes . length ) ) ) ; "<AssertPlaceHolder>" ; } toObject ( java . lang . Object , com . salesforce . phoenix . schema . PDataType , com . salesforce . phoenix . schema . ColumnModifier ) { return toObject ( object , actualType ) ; }
org . junit . Assert . assertEquals ( arr , resultArr )
testSpliterator ( ) { final com . iluwatar . fluentinterface . fluentiterable . List < java . lang . Integer > integers = com . iluwatar . fluentinterface . fluentiterable . Arrays . asList ( 1 , 2 , 3 ) ; final com . iluwatar . fluentinterface . fluentiterable . Spliterator < java . lang . Integer > split = createFluentIterable ( integers ) . spliterator ( ) ; "<AssertPlaceHolder>" ; } spliterator ( ) { return iterable . spliterator ( ) ; }
org . junit . Assert . assertNotNull ( split )
shouldNotHaveNextIfAllCachedItemsDeleted ( ) { final byte [ ] [ ] bytes = new byte [ ] [ ] { new byte [ ] { 0 } , new byte [ ] { 1 } , new byte [ ] { 2 } } ; for ( final byte [ ] aByte : bytes ) { final org . apache . kafka . common . utils . Bytes aBytes = org . apache . kafka . common . utils . Bytes . wrap ( aByte ) ; store . put ( aBytes , aByte ) ; cache . put ( namespace , aBytes , new org . apache . kafka . streams . state . internals . LRUCacheEntry ( null ) ) ; } "<AssertPlaceHolder>" ; } createIterator ( ) { final org . apache . kafka . streams . state . internals . ThreadCache . MemoryLRUCacheBytesIterator cacheIterator = cache . all ( namespace ) ; final org . apache . kafka . streams . state . KeyValueIterator < org . apache . kafka . common . utils . Bytes , byte [ ] > storeIterator = new org . apache . kafka . streams . state . internals . DelegatingPeekingKeyValueIterator ( "store" , store . all ( ) ) ; return new org . apache . kafka . streams . state . internals . MergedSortedCacheKeyValueBytesStoreIterator ( cacheIterator , storeIterator ) ; }
org . junit . Assert . assertFalse ( createIterator ( ) . hasNext ( ) )
testMvDestExistsOverwrite ( ) { java . io . File src = mock ( java . io . File . class ) ; java . io . File dest = mock ( java . io . File . class ) ; java . io . File parent = mock ( java . io . File . class ) ; expect ( src . exists ( ) ) . andReturn ( true ) . anyTimes ( ) ; expect ( src . isDirectory ( ) ) . andReturn ( false ) . anyTimes ( ) ; expect ( src . canRead ( ) ) . andReturn ( true ) ; expect ( src . renameTo ( dest ) ) . andReturn ( true ) ; replay ( src ) ; expect ( parent . exists ( ) ) . andReturn ( true ) ; expect ( parent . isDirectory ( ) ) . andReturn ( true ) ; expect ( parent . canRead ( ) ) . andReturn ( true ) ; expect ( parent . canExecute ( ) ) . andReturn ( true ) ; expect ( parent . canWrite ( ) ) . andReturn ( true ) ; replay ( parent ) ; expect ( dest . exists ( ) ) . andReturn ( true ) . once ( ) ; expect ( dest . exists ( ) ) . andReturn ( true ) . once ( ) ; expect ( dest . exists ( ) ) . andReturn ( false ) . once ( ) ; expect ( dest . isDirectory ( ) ) . andReturn ( false ) . anyTimes ( ) ; expect ( dest . canRead ( ) ) . andReturn ( true ) ; expect ( dest . canWrite ( ) ) . andReturn ( true ) ; expect ( dest . delete ( ) ) . andReturn ( true ) ; expect ( dest . getParentFile ( ) ) . andReturn ( parent ) ; expect ( dest . getPath ( ) ) . andReturn ( "" ) . anyTimes ( ) ; replay ( dest ) ; java . lang . String ret = datawave . common . io . Files . mv ( src , dest , true ) ; "<AssertPlaceHolder>" ; } mv ( java . lang . String , java . lang . String , boolean ) { if ( ( src == null ) || ( dest == null ) ) throw new java . lang . IllegalArgumentException ( datawave . common . io . Files . NULL_PARAMS ) ; return datawave . common . io . Files . mv ( new java . io . File ( src ) , new java . io . File ( dest ) , overwrite ) ; }
org . junit . Assert . assertNull ( ret )
firstErrorPropagates ( ) { final org . apache . poi . ss . formula . functions . ValueEval [ ] args = new org . apache . poi . ss . formula . functions . ValueEval [ ] { ErrorEval . DIV_ZERO , ErrorEval . NUM_ERROR } ; final org . apache . poi . ss . formula . functions . ValueEval result = getInstance ( ) . evaluate ( args , 0 , 0 ) ; "<AssertPlaceHolder>" ; } evaluate ( java . lang . String , org . apache . poi . ss . util . CellReference , org . apache . poi . ss . util . CellRangeAddressBase ) { return evaluate ( formula , target , region , FormulaType . CELL ) ; }
org . junit . Assert . assertEquals ( ErrorEval . DIV_ZERO , result )
fromCurrentJournalEntry ( ) { org . apache . jackrabbit . oak . plugins . document . DocumentNodeStore ns = builderProvider . newBuilder ( ) . clock ( clock ) . setAsyncDelay ( 0 ) . getNodeStore ( ) ; org . apache . jackrabbit . oak . plugins . document . DocumentNodeState s1 = ns . getRoot ( ) ; org . apache . jackrabbit . oak . spi . state . NodeBuilder builder = ns . getRoot ( ) . builder ( ) ; builder . child ( "foo" ) ; org . apache . jackrabbit . oak . plugins . document . JournalDiffLoaderTest . merge ( ns , builder ) ; org . apache . jackrabbit . oak . plugins . document . DocumentNodeState s2 = ns . getRoot ( ) ; "<AssertPlaceHolder>" ; } changeChildNodes ( org . apache . jackrabbit . oak . plugins . document . DocumentNodeStore , org . apache . jackrabbit . oak . plugins . document . AbstractDocumentNodeState , org . apache . jackrabbit . oak . plugins . document . AbstractDocumentNodeState ) { java . lang . String diff = new org . apache . jackrabbit . oak . plugins . document . JournalDiffLoader ( before , after , store ) . call ( ) ; final java . util . Set < java . lang . String > changes = newHashSet ( ) ; org . apache . jackrabbit . oak . plugins . document . DiffCache . parseJsopDiff ( diff , new org . apache . jackrabbit . oak . plugins . document . DiffCache . Diff ( ) { @ org . apache . jackrabbit . oak . plugins . document . Override public boolean childNodeAdded ( java . lang . String name ) { org . junit . Assert . fail ( ) ; return true ; } @ org . apache . jackrabbit . oak . plugins . document . Override public boolean childNodeChanged ( java . lang . String name ) { changes . add ( name ) ; return true ; } @ org . apache . jackrabbit . oak . plugins . document . Override public boolean childNodeDeleted ( java . lang . String name ) { org . junit . Assert . fail ( ) ; return true ; } } ) ; return changes ; }
org . junit . Assert . assertEquals ( newHashSet ( "foo" ) , org . apache . jackrabbit . oak . plugins . document . JournalDiffLoaderTest . changeChildNodes ( ns , s1 , s2 ) )
testEntityContainer ( ) { org . apache . olingo . odata2 . api . edm . EdmAssociationSet associationSet = org . apache . olingo . odata2 . core . edm . provider . EdmAssociationSetImplProvTest . edmAssociationSet ; "<AssertPlaceHolder>" ; } getEntityContainer ( ) { final org . apache . http . HttpResponse response = callUri ( "$metadata" ) ; final org . apache . olingo . odata2 . api . edm . EdmEntityContainer entityContainer = org . apache . olingo . odata2 . api . ep . EntityProvider . readMetadata ( response . getEntity ( ) . getContent ( ) , false ) . getDefaultEntityContainer ( ) ; getBody ( response ) ; return entityContainer ; }
org . junit . Assert . assertNotNull ( associationSet . getEntityContainer ( ) )
testGetList ( ) { org . apache . hadoop . yarn . appcatalog . controller . AppListController ac = org . mockito . Mockito . mock ( org . apache . hadoop . yarn . appcatalog . controller . AppListController . class ) ; java . util . List < org . apache . hadoop . yarn . appcatalog . model . AppEntry > actual = new java . util . ArrayList < org . apache . hadoop . yarn . appcatalog . model . AppEntry > ( ) ; when ( ac . getList ( ) ) . thenReturn ( actual ) ; final java . util . List < org . apache . hadoop . yarn . appcatalog . model . AppEntry > result = ac . getList ( ) ; "<AssertPlaceHolder>" ; } getList ( ) { org . apache . hadoop . yarn . appcatalog . application . AppCatalogSolrClient sc = new org . apache . hadoop . yarn . appcatalog . application . AppCatalogSolrClient ( ) ; return sc . listAppEntries ( ) ; }
org . junit . Assert . assertEquals ( result , actual )
testStreamPairedMismatchPrefix ( ) { java . io . ByteArrayOutputStream outputStream = new java . io . ByteArrayOutputStream ( ) ; new org . biojava . bio . program . fastq . SangerFastqWriter ( ) . write ( outputStream , mismatchPrefix ) ; java . io . Reader mismatchPrefixReader = new java . io . StringReader ( outputStream . toString ( ) ) ; org . nmdp . ngs . reads . paired . PairedEndFastqReader . streamPaired ( firstReader , mismatchPrefixReader , new org . nmdp . ngs . reads . paired . PairedEndAdapter ( ) { @ org . nmdp . ngs . reads . paired . Override public void paired ( final org . biojava . bio . program . fastq . Fastq left , final org . biojava . bio . program . fastq . Fastq right ) { org . junit . Assert . fail ( ( ( ( "paired<sp>" + left ) + "<sp>" ) + right ) ) ; } @ org . nmdp . ngs . reads . paired . Override public void unpaired ( final org . biojava . bio . program . fastq . Fastq unpaired ) { "<AssertPlaceHolder>" ; } } ) ; } getDescription ( ) { return description . replace ( ">" , "" ) ; }
org . junit . Assert . assertTrue ( ( ( left . getDescription ( ) . equals ( unpaired . getDescription ( ) ) ) || ( mismatchPrefix . getDescription ( ) . equals ( unpaired . getDescription ( ) ) ) ) )
should_get_an_enabled_project ( ) { fr . norad . visuwall . providers . hudson . domain . HudsonJob job = new fr . norad . visuwall . providers . hudson . domain . HudsonJob ( ) ; job . setDisabled ( false ) ; when ( hudson . findJob ( anyString ( ) ) ) . thenReturn ( job ) ; fr . norad . visuwall . api . domain . SoftwareProjectId softwareProjectId = new fr . norad . visuwall . api . domain . SoftwareProjectId ( "projectId" ) ; boolean isDisabled = hudsonConnection . isProjectDisabled ( softwareProjectId ) ; "<AssertPlaceHolder>" ; } isProjectDisabled ( fr . norad . visuwall . api . domain . SoftwareProjectId ) { checkConnected ( ) ; checkSoftwareProjectId ( softwareProjectId ) ; try { java . lang . String jobName = softwareProjectId . getProjectId ( ) ; fr . norad . visuwall . providers . hudson . domain . HudsonJob job = hudson . findJob ( jobName ) ; return job . isDisabled ( ) ; } catch ( fr . norad . visuwall . providers . hudson . exception . HudsonJobNotFoundException e ) { throw new fr . norad . visuwall . api . exception . ProjectNotFoundException ( ( "Can't<sp>find<sp>job<sp>with<sp>software<sp>project<sp>id:<sp>" + softwareProjectId ) , e ) ; } }
org . junit . Assert . assertFalse ( isDisabled )
pipe ( ) { java . util . List < java . lang . Object > expected = asList ( ( ( java . lang . Object ) ( "Hello<sp>World!" ) ) ) ; doReturn ( expected . get ( 0 ) ) . when ( tupleMock ) . getValueByField ( "in" ) ; eu . icolumbo . breeze . SpringBolt subject = new eu . icolumbo . breeze . SpringBolt ( eu . icolumbo . breeze . TestBean . class , "echo(in)" , "out" ) ; subject . setOutputStreamId ( "deep" ) ; run ( subject ) ; verify ( outputFieldsDeclarerMock ) . declareStream ( eq ( "deep" ) , outputFieldsCaptor . capture ( ) ) ; "<AssertPlaceHolder>" ; org . mockito . InOrder order = inOrder ( outputCollectorMock ) ; order . verify ( outputCollectorMock ) . emit ( "deep" , tupleMock , expected ) ; order . verify ( outputCollectorMock ) . ack ( tupleMock ) ; order . verifyNoMoreInteractions ( ) ; } run ( eu . icolumbo . breeze . SpringBolt ) { subject . setApplicationContext ( applicationContextMock ) ; subject . prepare ( stormConf , topologyContextMock , outputCollectorMock ) ; subject . declareOutputFields ( outputFieldsDeclarerMock ) ; subject . execute ( tupleMock ) ; }
org . junit . Assert . assertEquals ( asList ( "out" ) , outputFieldsCaptor . getValue ( ) . toList ( ) )
shouldSelectListOfPostsLikeTwoParameters ( ) { org . apache . ibatis . session . SqlSession session = org . apache . ibatis . binding . BindingTest . sqlSessionFactory . openSession ( ) ; try { org . apache . ibatis . binding . BoundBlogMapper mapper = session . getMapper ( org . apache . ibatis . binding . BoundBlogMapper . class ) ; java . util . List < org . apache . ibatis . domain . blog . Post > posts = mapper . selectPostsLikeSubjectAndBody ( new org . apache . ibatis . session . RowBounds ( 1 , 1 ) , "%a%" , "%a%" ) ; "<AssertPlaceHolder>" ; } finally { session . close ( ) ; } } size ( ) { return loaderMap . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , posts . size ( ) )
testAndFunction ( ) { items . add ( new org . eclipse . smarthome . core . library . types . ArithmeticGroupFunctionTest . TestItem ( "TestItem1" , OpenClosedType . OPEN ) ) ; items . add ( new org . eclipse . smarthome . core . library . types . ArithmeticGroupFunctionTest . TestItem ( "TestItem2" , OpenClosedType . OPEN ) ) ; items . add ( new org . eclipse . smarthome . core . library . types . ArithmeticGroupFunctionTest . TestItem ( "TestItem3" , OpenClosedType . OPEN ) ) ; function = new org . eclipse . smarthome . core . library . types . ArithmeticGroupFunction . And ( OpenClosedType . OPEN , OpenClosedType . CLOSED ) ; org . eclipse . smarthome . core . types . State state = function . calculate ( items ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( OpenClosedType . OPEN , state )
getOpenApi ( ) { getRouter ( ) . GET ( "/tutu" ) . specs ( new org . spincast . plugins . openapi . bottomup . @ org . spincast . plugins . openapi . bottomup . Specs ( @ io . swagger . v3 . oas . annotations . Operation ( operationId = "myOperationId" , summary = "my<sp>summary" , description = "my<sp>description" , responses = @ io . swagger . v3 . oas . annotations . responses . ApiResponse ( responseCode = "200" , description = "voila!" ) ) ) SpecsObject ( ) { } ) . handle ( getTestController ( ) :: sayHello ) ; io . swagger . v3 . oas . models . OpenAPI openAPI = getSpincastOpenApiManager ( ) . getOpenApi ( ) ; "<AssertPlaceHolder>" ; } getOpenApi ( ) { getRouter ( ) . GET ( "/tutu" ) . specs ( new org . spincast . plugins . openapi . bottomup . @ org . spincast . plugins . openapi . bottomup . Specs ( @ io . swagger . v3 . oas . annotations . Operation ( operationId = "myOperationId" , summary = "my<sp>summary" , description = "my<sp>description" , responses = @ io . swagger . v3 . oas . annotations . responses . ApiResponse ( responseCode = "200" , description = "voila!" ) ) ) SpecsObject ( ) { } ) . handle ( getTestController ( ) :: sayHello ) ; io . swagger . v3 . oas . models . OpenAPI openAPI = getSpincastOpenApiManager ( ) . getOpenApi ( ) ; org . junit . Assert . assertNotNull ( openAPI ) ; }
org . junit . Assert . assertNotNull ( openAPI )
testDynamicQueryByPrimaryKeyMissing ( ) { com . liferay . portal . kernel . dao . orm . DynamicQuery dynamicQuery = com . liferay . portal . kernel . dao . orm . DynamicQueryFactoryUtil . forClass ( com . liferay . adaptive . media . image . model . AMImageEntry . class , _dynamicQueryClassLoader ) ; dynamicQuery . add ( com . liferay . portal . kernel . dao . orm . RestrictionsFactoryUtil . eq ( "amImageEntryId" , com . liferay . portal . kernel . test . util . RandomTestUtil . nextLong ( ) ) ) ; java . util . List < com . liferay . adaptive . media . image . model . AMImageEntry > result = _persistence . findWithDynamicQuery ( dynamicQuery ) ; "<AssertPlaceHolder>" ; } size ( ) { if ( ( _workflowTaskAssignees ) != null ) { return _workflowTaskAssignees . size ( ) ; } return _kaleoTaskAssignmentInstanceLocalService . getKaleoTaskAssignmentInstancesCount ( _kaleoTaskInstanceToken . getKaleoTaskInstanceTokenId ( ) ) ; }
org . junit . Assert . assertEquals ( 0 , result . size ( ) )
testWriteAccess ( ) { org . nuxeo . ecm . platform . audit . api . AuditLogger writer = org . nuxeo . runtime . api . Framework . getService ( org . nuxeo . ecm . platform . audit . api . AuditLogger . class ) ; "<AssertPlaceHolder>" ; } getService ( java . lang . Class ) { org . nuxeo . runtime . api . ServiceProvider provider = org . nuxeo . runtime . api . DefaultServiceProvider . getProvider ( ) ; if ( provider != null ) { return provider . getService ( serviceClass ) ; } org . nuxeo . runtime . api . Framework . checkRuntimeInitialized ( ) ; return org . nuxeo . runtime . api . Framework . runtime . getService ( serviceClass ) ; }
org . junit . Assert . assertNotNull ( writer )
subscribeToService_SubscriptionManager_Authorized ( ) { java . lang . reflect . Method method = beanClass . getMethod ( "subscribeToService" , org . oscm . internal . vo . VOSubscription . class , org . oscm . internal . vo . VOService . class , java . util . List . class , org . oscm . internal . vo . VOPaymentInfo . class , org . oscm . internal . vo . VOBillingContact . class , java . util . List . class ) ; boolean isSubscriptionManagerRoleAllowed = isRoleAllowed ( method , UserRoleType . SUBSCRIPTION_MANAGER ) ; "<AssertPlaceHolder>" ; } isRoleAllowed ( java . lang . reflect . Method , org . oscm . internal . types . enumtypes . UserRoleType ) { javax . annotation . security . RolesAllowed rolesAllowed = method . getAnnotation ( javax . annotation . security . RolesAllowed . class ) ; if ( rolesAllowed == null ) { return true ; } for ( java . lang . String role : rolesAllowed . value ( ) ) { if ( role . equals ( roleType . name ( ) ) ) { return true ; } } return false ; }
org . junit . Assert . assertTrue ( isSubscriptionManagerRoleAllowed )
testSelectAfter ( ) { try { cambridge . parser . TemplateTokenizer tokenizer = new cambridge . parser . TemplateTokenizer ( cambridge . ParserTest . class . getResourceAsStream ( "full.html" ) ) ; cambridge . parser . TemplateParser parser = new cambridge . parser . TemplateParser ( tokenizer , Expressions . cambridgeExpressionLanguage ) ; cambridge . model . TemplateDocument t = parser . parse ( ) ; cambridge . model . FragmentList fragments = t . select ( "after<sp>/html/body/div[3]" ) ; java . io . StringWriter builder = new java . io . StringWriter ( ) ; for ( cambridge . model . Fragment f : fragments ) { f . eval ( cambridge . ParserTest . bindings , builder ) ; } after = after . replaceAll ( "\\n" , ( ( java . lang . String ) ( java . lang . System . getProperties ( ) . get ( "line.separator" ) ) ) ) ; "<AssertPlaceHolder>" ; } catch ( java . io . IOException e ) { e . printStackTrace ( ) ; } catch ( cambridge . TemplateParsingException e ) { e . printStackTrace ( ) ; } catch ( cambridge . BehaviorInstantiationException e ) { e . printStackTrace ( ) ; } catch ( cambridge . TemplateEvaluationException e ) { e . printStackTrace ( ) ; } } toString ( ) { return ( ( ( ( ( ( getMessage ( ) ) + "<sp>(" ) + "Line:<sp>" ) + ( line ) ) + ",<sp>Col:<sp>" ) + ( col ) ) + ")" ; }
org . junit . Assert . assertEquals ( after , builder . toString ( ) )
testDeleteFiles ( ) { final com . civilizer . web . view . FileListBean fileListBean = new com . civilizer . web . view . FileListBean ( ) ; fileListBean . setFileEntities ( com . civilizer . test . web . WebFileBoxTest . fileEntityDao . findAll ( ) ) ; com . civilizer . web . view . FilePathTree filePathTree = new com . civilizer . web . view . FilePathTree ( ) ; fileListBean . setFilePathTree ( filePathTree ) ; final boolean forFolder = com . civilizer . test . helper . TestUtil . getRandom ( ) . nextBoolean ( ) ; final int srcNodeId = com . civilizer . test . web . WebFileBoxTest . getRandomFilePathId ( filePathTree . getFilePathBeans ( ) , forFolder , true ) ; final com . civilizer . web . view . FilePathBean filePathBean = fileListBean . getFilePathBean ( srcNodeId ) ; final java . lang . String filePath = filePathBean . getFullPath ( ) ; try { com . civilizer . utils . FsUtil . forceDelete ( filePathBean . toFile ( com . civilizer . test . web . WebFileBoxTest . filesHomePath ) ) ; } catch ( com . civilizer . test . web . IOException e ) { e . printStackTrace ( ) ; org . junit . Assert . fail ( java . lang . String . format ( "deleting<sp>'%s'<sp>failed!" , filePath ) ) ; } com . civilizer . test . web . List < com . civilizer . domain . FileEntity > entities = com . civilizer . test . web . Collections . emptyList ( ) ; if ( filePathBean . isFolder ( ) ) { entities = com . civilizer . test . web . WebFileBoxTest . fileEntityDao . findByNamePattern ( ( filePath + "/%" ) ) ; } else { com . civilizer . domain . FileEntity entity = com . civilizer . test . web . WebFileBoxTest . fileEntityDao . findByName ( filePath ) ; if ( entity != null ) { entities = new com . civilizer . test . web . ArrayList ( ) ; entities . add ( entity ) ; } } for ( com . civilizer . domain . FileEntity f : entities ) { final java . lang . String pathOnFileSystem = ( com . civilizer . test . web . WebFileBoxTest . filesHomePath ) + ( f . getFileName ( ) ) ; "<AssertPlaceHolder>" ; com . civilizer . test . web . WebFileBoxTest . fileEntityDao . delete ( f ) ; } com . civilizer . test . web . WebFileBoxTest . renewTestData ( ) ; } getFileName ( ) { return fileName ; }
org . junit . Assert . assertEquals ( false , new com . civilizer . test . web . File ( pathOnFileSystem ) . exists ( ) )
nextReturnsShardWithParent ( ) { com . amazonaws . services . dynamodbv2 . model . Shard first = new com . amazonaws . services . dynamodbv2 . model . Shard ( ) . withShardId ( "first_shard" ) . withParentShardId ( "other_shard_id" ) ; com . amazonaws . services . dynamodbv2 . model . Shard second = new com . amazonaws . services . dynamodbv2 . model . Shard ( ) . withParentShardId ( "first_shard" ) . withShardId ( "second_shard" ) ; org . apache . camel . component . aws . ddbstream . ShardList shards = new org . apache . camel . component . aws . ddbstream . ShardList ( ) ; shards . add ( first ) ; shards . add ( second ) ; "<AssertPlaceHolder>" ; } nextAfter ( com . amazonaws . services . dynamodbv2 . model . Shard ) { for ( com . amazonaws . services . dynamodbv2 . model . Shard shard : shards . values ( ) ) { if ( previous . getShardId ( ) . equals ( shard . getParentShardId ( ) ) ) { return shard ; } } throw new java . lang . IllegalStateException ( ( ( ( "Unable<sp>to<sp>find<sp>the<sp>next<sp>shard<sp>for<sp>" + previous ) + "<sp>in<sp>" ) + ( shards ) ) ) ; }
org . junit . Assert . assertThat ( shards . nextAfter ( first ) , org . hamcrest . CoreMatchers . is ( second ) )
getTotalUnreadPrivateMessages ( ) { new net . jforum . util . JDBCLoader ( session ( ) ) . run ( "/userdao/getTotalUnreadPrivateMessages.sql" ) ; net . jforum . entities . User user = new net . jforum . entities . User ( ) ; user . setId ( 1 ) ; "<AssertPlaceHolder>" ; } newDao ( ) { return new net . jforum . repository . AvatarRepository ( session ( ) ) ; }
org . junit . Assert . assertEquals ( 1 , this . newDao ( ) . getTotalUnreadPrivateMessages ( user ) )
shouldConfigureBindingMode ( ) { final org . apache . camel . component . rest . RestEndpoint restEndpoint = new org . apache . camel . component . rest . RestEndpoint ( "rest:GET:/path" , restComponent ) ; restEndpoint . setComponentName ( "mock-rest" ) ; restEndpoint . setParameters ( new java . util . HashMap ( ) ) ; restEndpoint . setHost ( "http://localhost" ) ; restEndpoint . setBindingMode ( "json" ) ; final org . apache . camel . component . rest . RestProducer producer = ( ( org . apache . camel . component . rest . RestProducer ) ( restEndpoint . createProducer ( ) ) ) ; "<AssertPlaceHolder>" ; } getBindingMode ( ) { return bindingMode ; }
org . junit . Assert . assertEquals ( producer . getBindingMode ( ) . name ( ) , "json" )
testQueryStringMultiParam ( ) { final java . lang . String queryString = "?foo=bar&bar=foo" ; org . switchyard . component . camel . common . QueryString qs = new org . switchyard . component . camel . common . QueryString ( ) . add ( "foo" , "bar" ) . add ( "bar" , "foo" ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( ( _mimeType ) + ";" ) + ( org . switchyard . component . http . ContentType . CHARSET ) ) + "=" ) + ( _charset ) ; }
org . junit . Assert . assertEquals ( queryString , qs . toString ( ) )
testBuildPublishMessagesURL ( ) { java . lang . String expected = "/admin/messages/rebroadcast?queueName=some-queue&startChangeNumber=345&type=ACTIVITY&limit=888" ; java . lang . String url = org . sagebionetworks . client . SynapseAdminClientImpl . buildPublishMessagesURL ( "some-queue" , new java . lang . Long ( 345 ) , ObjectType . ACTIVITY , new java . lang . Long ( 888 ) ) ; "<AssertPlaceHolder>" ; } buildPublishMessagesURL ( java . lang . String , java . lang . Long , org . sagebionetworks . repo . model . ObjectType , java . lang . Long ) { if ( queueName == null ) throw new java . lang . IllegalArgumentException ( "queueName<sp>cannot<sp>be<sp>null" ) ; if ( startChangeNumber == null ) throw new java . lang . IllegalArgumentException ( "startChangeNumber<sp>cannot<sp>be<sp>null" ) ; java . lang . StringBuilder builder = new java . lang . StringBuilder ( ) ; builder . append ( org . sagebionetworks . client . SynapseAdminClientImpl . ADMIN_PUBLISH_MESSAGES ) ; builder . append ( "?" ) ; builder . append ( "queueName=" ) . append ( queueName ) ; builder . append ( "&startChangeNumber=" ) . append ( startChangeNumber ) ; if ( type != null ) { builder . append ( "&type=" ) . append ( type . name ( ) ) ; } if ( limit != null ) { builder . append ( "&limit=" ) . append ( limit ) ; } return builder . toString ( ) ; }
org . junit . Assert . assertEquals ( expected , url )
testCreateSocket ( ) { java . net . ServerSocket socket = null ; try { socket = new java . net . ServerSocket ( 0 ) ; socket . setReuseAddress ( true ) ; int port = socket . getLocalPort ( ) ; when ( mpd . getAddress ( ) ) . thenReturn ( java . net . InetAddress . getLocalHost ( ) ) ; when ( mpd . getPort ( ) ) . thenReturn ( port ) ; when ( mpd . getTimeout ( ) ) . thenReturn ( 5000 ) ; java . net . ServerSocket finalSocket = socket ; new java . lang . Thread ( ( ) -> { java . net . Socket clientSocket = null ; try { clientSocket = finalSocket . accept ( ) ; java . io . PrintWriter pw = new java . io . PrintWriter ( clientSocket . getOutputStream ( ) , true ) ; pw . write ( "OK<sp>MPD<sp>Version\r\n" ) ; pw . flush ( ) ; } catch ( e ) { org . bff . javampd . command . e . printStackTrace ( ) ; } } ) . start ( ) ; commandExecutor . setMpd ( mpd ) ; "<AssertPlaceHolder>" ; } finally { if ( socket != null ) { socket . close ( ) ; } } } createSocket ( ) { return new org . bff . javampd . command . MPDSocket ( mpd . getAddress ( ) , mpd . getPort ( ) , mpd . getTimeout ( ) ) ; }
org . junit . Assert . assertNotNull ( commandExecutor . createSocket ( ) )
testSparseInterpolation_random2 ( ) { int nIterations = its ( 500 , 1000 ) ; int badEvaluations = 0 ; org . apache . commons . math3 . random . RandomGenerator rnd = getRandom ( ) ; cc . redberry . rings . poly . multivar . MultivariateGCDTest . GCDSampleData < cc . redberry . rings . poly . multivar . Monomial < cc . redberry . rings . bigint . BigInteger > , cc . redberry . rings . poly . multivar . MultivariatePolynomial < cc . redberry . rings . bigint . BigInteger > > sampleData = new cc . redberry . rings . poly . multivar . MultivariateGCDTest . GCDSampleDataGeneric ( cc . redberry . rings . Rings . Rings . Z , 3 , 5 , 5 , 15 , 5 , 15 , rnd ) ; for ( int n = 0 ; n < nIterations ; n ++ ) { cc . redberry . rings . poly . multivar . MultivariateGCDTest . GCDSample < cc . redberry . rings . poly . multivar . Monomial < cc . redberry . rings . bigint . BigInteger > , cc . redberry . rings . poly . multivar . MultivariatePolynomial < cc . redberry . rings . bigint . BigInteger > > gcdTriplet = sampleData . nextSample ( false , false ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 skeleton = null ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 gcd = null ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 actual = null ; cc . redberry . rings . poly . multivar . IntegersZp domain = null ; cc . redberry . rings . poly . multivar . IntegersZp domain1 = null ; long seed = - 1 ; try { cc . redberry . rings . poly . multivar . MultivariatePolynomial < cc . redberry . rings . bigint . BigInteger > contentGCD = cc . redberry . rings . poly . multivar . MultivariateGCD . MultivariateGCD . contentGCD ( gcdTriplet . a , gcdTriplet . b , 0 , cc . redberry . rings . poly . multivar . MultivariateGCD . MultivariateGCD :: PolynomialGCD ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomial < cc . redberry . rings . bigint . BigInteger > a = cc . redberry . rings . poly . multivar . MultivariateDivision . divideExact ( gcdTriplet . a , contentGCD ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomial < cc . redberry . rings . bigint . BigInteger > b = cc . redberry . rings . poly . multivar . MultivariateDivision . divideExact ( gcdTriplet . b , contentGCD ) ; domain = new cc . redberry . rings . poly . multivar . IntegersZp ( getModulusRandom ( 20 ) ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 la = cc . redberry . rings . poly . multivar . MultivariatePolynomial . asOverZp64 ( a . setRing ( domain ) ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 lb = cc . redberry . rings . poly . multivar . MultivariatePolynomial . asOverZp64 ( b . setRing ( domain ) ) ; skeleton = ZippelGCD ( la , lb ) ; if ( ( ( la . isConstant ( ) ) || ( lb . isConstant ( ) ) ) || ( ( skeleton . degree ( 0 ) ) == 0 ) ) { -- n ; continue ; } domain1 = new cc . redberry . rings . poly . multivar . IntegersZp ( getModulusRandom ( 20 ) ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 la1 = cc . redberry . rings . poly . multivar . MultivariatePolynomial . asOverZp64 ( a . setRing ( domain1 ) ) ; cc . redberry . rings . poly . multivar . MultivariatePolynomialZp64 lb1 = cc . redberry . rings . poly . multivar . MultivariatePolynomial . asOverZp64 ( b . setRing ( domain1 ) ) ; gcd = ZippelGCD ( la1 , lb1 ) ; if ( ! ( gcd . sameSkeletonQ ( skeleton ) ) ) { -- n ; continue ; } rnd . setSeed ( ( seed = rnd . nextLong ( ) ) ) ; actual = interpolateGCD ( la1 , lb1 , skeleton . setRing ( la1 . ring ) , rnd ) ; if ( actual == null ) { ++ badEvaluations ; actual = interpolateGCD ( la1 , lb1 , skeleton . setRing ( la1 . ring ) , rnd ) ; } "<AssertPlaceHolder>" ; } catch ( java . lang . Throwable thr ) { System . out . println ( seed ) ; System . out . println ( gcdTriplet . domain ) ; System . out . println ( gcdTriplet . a ) ; System . out . println ( gcdTriplet . b ) ; System . out . println ( gcdTriplet . gcd ) ; System . out . println ( domain ) ; System . out . println ( domain1 ) ; System . out . println ( skeleton ) ; System . out . println ( actual ) ; throw thr ; } } System . out . println ( ( "Bad<sp>evaluations:<sp>" + badEvaluations ) ) ; } monic ( ) { if ( isMonic ( ) ) return this ; if ( isZero ( ) ) return this ; if ( ( degree ) == 0 ) { data [ 0 ] = 1 ; return this ; } return multiply ( ring . reciprocal ( lc ( ) ) ) ; }
org . junit . Assert . assertEquals ( gcd . monic ( ) , actual . monic ( ) )
padTo_A$int$Object_minus ( ) { com . m3 . scalaflavor4j . Nil < java . lang . String > nil = com . m3 . scalaflavor4j . Nil . apply ( ) ; int len = - 1 ; com . m3 . scalaflavor4j . Seq < java . lang . String > actual = nil . padTo ( len , "a" ) ; "<AssertPlaceHolder>" ; } size ( ) { return 0 ; }
org . junit . Assert . assertThat ( actual . size ( ) , org . hamcrest . CoreMatchers . is ( org . hamcrest . CoreMatchers . equalTo ( 0 ) ) )
testLubmLikeTest ( ) { org . apache . accumulo . core . client . BatchWriter bw = null ; bw = accCon . createBatchWriter ( tablename , ( ( 500L * 1024L ) * 1024L ) , Long . MAX_VALUE , 30 ) ; for ( int i = 0 ; i < 100 ; i ++ ) { org . apache . accumulo . core . data . Mutation m1 = new org . apache . accumulo . core . data . Mutation ( new org . apache . hadoop . io . Text ( ( "ProfessorA" + i ) ) ) ; org . apache . accumulo . core . data . Mutation m2 = new org . apache . accumulo . core . data . Mutation ( new org . apache . hadoop . io . Text ( ( "ProfessorB" + i ) ) ) ; m1 . put ( new org . apache . hadoop . io . Text ( "fii" 1 ) , new org . apache . hadoop . io . Text ( ( ( ( ( ( ( null + "" ) + "fii" 0 ) + "" ) + "http://www.University" ) + i ) + ".edu" ) ) , new org . apache . accumulo . core . data . Value ( new byte [ 0 ] ) ) ; m2 . put ( new org . apache . hadoop . io . Text ( "fii" 1 ) , new org . apache . hadoop . io . Text ( ( ( ( ( ( ( null + "" ) + "fii" 0 ) + "" ) + "http://www.University" ) + i ) + ".edu" ) ) , new org . apache . accumulo . core . data . Value ( new byte [ 0 ] ) ) ; m1 . put ( new org . apache . hadoop . io . Text ( "http://swat.cse.lehigh.edu/onto/univ-bench.owl#teacherOf" ) , new org . apache . hadoop . io . Text ( ( ( ( ( ( null + "" ) + "fii" 0 ) + "" ) + "http://Course" ) + i ) ) , new org . apache . accumulo . core . data . Value ( new byte [ 0 ] ) ) ; m2 . put ( new org . apache . hadoop . io . Text ( "http://swat.cse.lehigh.edu/onto/univ-bench.owl#teacherOf" ) , new org . apache . hadoop . io . Text ( ( ( ( ( ( null + "" ) + "fii" 0 ) + "" ) + "http://Course" ) + i ) ) , new org . apache . accumulo . core . data . Value ( new byte [ 0 ] ) ) ; bw . addMutation ( m1 ) ; bw . addMutation ( m2 ) ; } org . apache . rya . accumulo . documentIndex . TextColumn tc1 = new org . apache . rya . accumulo . documentIndex . TextColumn ( new org . apache . hadoop . io . Text ( "fii" 1 ) , new org . apache . hadoop . io . Text ( ( ( ( "fii" 0 + ( "" + "http://www.University" ) ) + 30 ) + ".edu" ) ) ) ; org . apache . rya . accumulo . documentIndex . TextColumn tc2 = new org . apache . rya . accumulo . documentIndex . TextColumn ( new org . apache . hadoop . io . Text ( "http://swat.cse.lehigh.edu/onto/univ-bench.owl#teacherOf" ) , new org . apache . hadoop . io . Text ( ( ( "fii" 0 + ( "" + "http://Course" ) ) + 30 ) ) ) ; org . apache . rya . accumulo . documentIndex . TextColumn [ ] tc = new org . apache . rya . accumulo . documentIndex . TextColumn [ 2 ] ; tc [ 0 ] = tc1 ; tc [ 1 ] = tc2 ; org . apache . accumulo . core . client . IteratorSetting is = new org . apache . accumulo . core . client . IteratorSetting ( 30 , "fii" , org . apache . rya . accumulo . documentIndex . DocumentIndexIntersectingIterator . class ) ; org . apache . rya . accumulo . documentIndex . DocumentIndexIntersectingIterator . setColumnFamilies ( is , tc ) ; org . apache . accumulo . core . client . Scanner scan = accCon . createScanner ( tablename , new org . apache . accumulo . core . security . Authorizations ( "auths" ) ) ; scan . addScanIterator ( is ) ; int results = 0 ; System . out . println ( "************************Test<sp>15****************************" ) ; for ( Map . Entry < org . apache . accumulo . core . data . Key , org . apache . accumulo . core . data . Value > e : scan ) { System . out . println ( e ) ; results ++ ; } "<AssertPlaceHolder>" ; } println ( java . lang . CharSequence ) { System . out . println ( cs ) ; }
org . junit . Assert . assertEquals ( 2 , results )
shouldConfigureChannel ( ) { java . lang . String feedName = "feed1" ; java . lang . String channelName = "channel1" ; java . lang . String configPath = "plugin_config/valid_config_1.json" ; getConfiguredChannel ( configPath , feedName , channelName ) ; com . mongodb . BasicDBObject config = ( ( com . mongodb . BasicDBObject ) ( channelSvc . getChannelConfiguration ( feedName , channelName ) ) ) ; config . removeField ( "_id" ) ; com . mongodb . hvdf . util . JSONParam configReturn = new com . mongodb . hvdf . util . JSONParam ( config ) ; com . mongodb . hvdf . util . JSONParam originalParam = new com . mongodb . hvdf . util . JSONParam ( com . yammer . dropwizard . testing . JsonHelpers . jsonFixture ( configPath ) ) ; "<AssertPlaceHolder>" ; } getChannelConfiguration ( java . lang . String , java . lang . String ) { com . mongodb . DBCollection configColl = getChannelConfigCollection ( feedName ) ; return configColl . findOne ( new com . mongodb . BasicDBObject ( "_id" , channelName ) ) ; }
org . junit . Assert . assertEquals ( originalParam , configReturn )
testCanSetupDefaultValueWhenItIsNotEnabled ( ) { when ( plugin . getFactField ( ) ) . thenReturn ( "factField" ) ; when ( plugin . getFactType ( ) ) . thenReturn ( "factType" ) ; when ( plugin . doesOperatorNeedValue ( ) ) . thenReturn ( true ) ; when ( model . getTableFormat ( ) ) . thenReturn ( GuidedDecisionTable52 . TableFormat . EXTENDED_ENTRY ) ; "<AssertPlaceHolder>" ; } canSetupDefaultValue ( ) { if ( ! ( isDefaultValueEnabled ( ) ) ) { return false ; } if ( ( tableFormat ( ) ) != ( GuidedDecisionTable52 . TableFormat . EXTENDED_ENTRY ) ) { return false ; } if ( doesNotHaveFactTypeOrFactField ( ) ) { return false ; } if ( ! ( org . drools . workbench . screens . guided . dtable . client . wizard . column . pages . ValueOptionsPage . plugin ( ) . doesOperatorNeedValue ( ) ) ) { return false ; } return true ; }
org . junit . Assert . assertFalse ( page . canSetupDefaultValue ( ) )
shouldFindPlayerWhenGet ( ) { com . codenjoy . dojo . services . Player newPlayer = createPlayer ( com . codenjoy . dojo . services . PlayerServiceImplTest . VASYA ) ; com . codenjoy . dojo . services . Player player = playerService . get ( com . codenjoy . dojo . services . PlayerServiceImplTest . VASYA ) ; "<AssertPlaceHolder>" ; } get ( com . codenjoy . dojo . sokoban . client . Board ) { if ( board . isGameOver ( ) ) return "" ; java . util . List < com . codenjoy . dojo . services . Direction > result = getDirections ( board ) ; if ( result . isEmpty ( ) ) return "" ; return ( result . get ( 0 ) . toString ( ) ) + ( getBombIfNeeded ( board ) ) ; }
org . junit . Assert . assertSame ( newPlayer , player )
acceptCommentsSuccess ( ) { java . lang . String script = "#<sp>Accept<sp>script\n" + ( ( ( ( ( ( ( ( ( "\taccept<sp>\'tcp://localhost:8080\'<sp>#commentagain\n" + "\t#comment<sp>#1\n" ) + "accepted\n" ) + "#comment<sp>#2\n" ) + "#comment<sp>#5\n" 0 ) + "#comment<sp>#3\n" ) + "close\n" ) + "#comment<sp>#4\n" ) + "closed\n" ) + "#comment<sp>#5\n" ) ; org . kaazing . k3po . lang . internal . parser . ScriptParser parser = new org . kaazing . k3po . lang . internal . parser . ScriptParserImpl ( ) ; org . kaazing . k3po . lang . internal . ast . AstScriptNode scriptAST = parser . parse ( new java . io . ByteArrayInputStream ( script . getBytes ( org . kaazing . k3po . driver . internal . behavior . UTF_8 ) ) ) ; org . kaazing . k3po . lang . internal . RegionInfo scriptInfo = scriptAST . getRegionInfo ( ) ; org . kaazing . k3po . driver . internal . behavior . ScriptProgress progress = new org . kaazing . k3po . driver . internal . behavior . ScriptProgress ( scriptInfo , script ) ; java . lang . String observedScript = progress . getObservedScript ( ) ; "<AssertPlaceHolder>" ; } getObservedScript ( ) { return ( progress ) != null ? progress . getObservedScript ( ) : null ; }
org . junit . Assert . assertEquals ( script , observedScript )
testParseArgsWithSubparsersAlias ( ) { net . sourceforge . argparse4j . inf . Subparsers subparsers = ap . addSubparsers ( ) ; net . sourceforge . argparse4j . inf . Subparser checkout = subparsers . addParser ( "checkout" ) . aliases ( "co" ) ; checkout . setDefault ( "func" , "checkout" ) ; net . sourceforge . argparse4j . inf . Namespace res = ap . parseArgs ( "co" . split ( "<sp>" ) ) ; "<AssertPlaceHolder>" ; } get ( java . lang . String ) { return ( ( T ) ( attrs_ . get ( dest ) ) ) ; }
org . junit . Assert . assertEquals ( "checkout" , res . get ( "func" ) )
testUpdateContent ( ) { final com . orientechnologies . orient . core . db . document . ODatabaseDocumentTx db = new com . orientechnologies . orient . core . db . document . ODatabaseDocumentTx ( "memory:OCommandExecutorSQLUpdateTestContent" ) ; db . create ( ) ; try { db . command ( new com . orientechnologies . orient . core . sql . OCommandSQL ( "insert<sp>into<sp>V<sp>(name)<sp>values<sp>('bar')" ) ) . execute ( ) ; db . command ( new com . orientechnologies . orient . core . sql . OCommandSQL ( "UPDATE<sp>V<sp>content<sp>{\"value\":\"foo\"}" ) ) . execute ( ) ; java . lang . Iterable result = db . query ( new com . orientechnologies . orient . core . sql . query . OSQLSynchQuery < java . lang . Object > ( "select<sp>from<sp>V" ) ) ; com . orientechnologies . orient . core . record . impl . ODocument doc = ( ( com . orientechnologies . orient . core . record . impl . ODocument ) ( result . iterator ( ) . next ( ) ) ) ; "<AssertPlaceHolder>" ; } finally { db . close ( ) ; } } field ( java . lang . Object ) { if ( ! ( fields . contains ( iField ) ) ) fields . add ( iField ) ; return this ; }
org . junit . Assert . assertEquals ( doc . field ( "value" ) , "foo" )
testCharPlus ( ) { org . antlr . tool . Grammar g = new org . antlr . tool . Grammar ( ( "grammar<sp>P;\n" + "a<sp>:<sp>'a'+;" ) ) ; java . lang . String expecting = "(rule<sp>a<sp>ARG<sp>RET<sp>scope<sp>(BLOCK<sp>(ALT<sp>(+<sp>(BLOCK<sp>(ALT<sp>'a'<sp><end-of-alt>)<sp><end-of-block>))<sp><end-of-alt>)<sp><end-of-block>)<sp><end-of-rule>)" ; java . lang . String found = g . getRule ( "a" ) . tree . toStringTree ( ) ; "<AssertPlaceHolder>" ; } toStringTree ( ) { if ( ( ( children ) == null ) || ( children . isEmpty ( ) ) ) { return this . toString ( ) ; } java . lang . StringBuilder buf = new java . lang . StringBuilder ( ) ; if ( ! ( isNil ( ) ) ) { buf . append ( "(" ) ; buf . append ( this . toString ( ) ) ; buf . append ( '<sp>' ) ; } for ( int i = 0 ; ( ( children ) != null ) && ( i < ( children . size ( ) ) ) ; i ++ ) { org . antlr . runtime . tree . Tree t = ( ( org . antlr . runtime . tree . Tree ) ( children . get ( i ) ) ) ; if ( i > 0 ) { buf . append ( '<sp>' ) ; } buf . append ( t . toStringTree ( ) ) ; } if ( ! ( isNil ( ) ) ) { buf . append ( ")" ) ; } return buf . toString ( ) ; }
org . junit . Assert . assertEquals ( expecting , found )
isInternalImportExternalImportType ( ) { final org . kie . soup . project . datamodel . imports . Imports imports = new org . kie . soup . project . datamodel . imports . Imports ( ) ; final org . kie . soup . project . datamodel . imports . Import importType = new org . kie . soup . project . datamodel . imports . Import ( "External1" ) ; imports . addImport ( importType ) ; presenter . setContent ( dmo , imports , false ) ; "<AssertPlaceHolder>" ; } isInternalImport ( org . kie . soup . project . datamodel . imports . Import ) { return getInternalFactTypes ( ) . contains ( importType ) ; }
org . junit . Assert . assertFalse ( presenter . isInternalImport ( importType ) )
testRead ( ) { com . github . sardine . Sardine sardine = com . github . sardine . SardineFactory . begin ( ) ; final java . lang . String url = "http://sardine.googlecode.com/svn/trunk/README.html" ; final java . io . InputStream in = sardine . get ( url ) ; "<AssertPlaceHolder>" ; in . close ( ) ; } get ( java . lang . String ) { return this . get ( url , java . util . Collections . < java . lang . String , java . lang . String > emptyMap ( ) ) ; }
org . junit . Assert . assertNotNull ( in )
return_empty_list_guava ( ) { java . util . List < java . lang . String > emptyList = com . google . common . collect . ImmutableList . of ( ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertTrue ( emptyList . isEmpty ( ) )
testName ( ) { "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( "" , name )
testRegisterDuplicateObserver ( ) { final org . restcomm . media . core . control . mgcp . message . MgcpMessageObserver observer = mock ( org . restcomm . media . core . control . mgcp . message . MgcpMessageObserver . class ) ; final org . restcomm . media . core . control . mgcp . transaction . MgcpTransactionManagerProvider provider = mock ( org . restcomm . media . core . control . mgcp . transaction . MgcpTransactionManagerProvider . class ) ; final org . restcomm . media . core . control . mgcp . transaction . GlobalMgcpTransactionManager transactionManager = new org . restcomm . media . core . control . mgcp . transaction . GlobalMgcpTransactionManager ( provider ) ; final java . util . Collection < org . restcomm . media . core . control . mgcp . message . MgcpMessageObserver > observers = ( ( java . util . Collection < org . restcomm . media . core . control . mgcp . message . MgcpMessageObserver > ) ( org . mockito . internal . util . reflection . Whitebox . getInternalState ( transactionManager , "observers" ) ) ) ; transactionManager . observe ( observer ) ; transactionManager . observe ( observer ) ; "<AssertPlaceHolder>" ; } size ( ) { return this . parameters . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , observers . size ( ) )
testDoSend_2 ( ) { org . jinstagram . http . Request fixture = new org . jinstagram . http . Request ( Verbs . DELETE , "" ) ; fixture . setConnectionKeepAlive ( true ) ; fixture . setCharset ( "UTF-8" ) ; fixture . addPayload ( "Dummy<sp>payload" ) ; fixture . setConnection ( mockHttpConnection ) ; fixture . setProxy ( org . jinstagram . http . RequestTest . proxy ) ; org . jinstagram . http . Response result = fixture . doSend ( ) ; "<AssertPlaceHolder>" ; } doSend ( ) { connection . setRequestMethod ( this . verb . name ( ) ) ; addHeaders ( connection ) ; if ( ( verb . equals ( Verbs . PUT ) ) || ( verb . equals ( Verbs . POST ) ) ) { addBody ( connection , getByteBodyContents ( ) ) ; } return new org . jinstagram . http . Response ( connection ) ; }
org . junit . Assert . assertNotNull ( result )
testRegister_requiredPropertyNotPresent ( ) { com . hortonworks . streamline . streams . cluster . catalog . Cluster cluster = getTestCluster ( 1L ) ; com . hortonworks . streamline . streams . cluster . register . impl . HDFSServiceRegistrar registrar = initializeServiceRegistrar ( ) ; try ( java . io . InputStream coreSiteIs = getClass ( ) . getClassLoader ( ) . getResourceAsStream ( com . hortonworks . streamline . streams . cluster . register . impl . HDFSServiceRegistrarTest . CORE_SITE_XML_BADCASE_FILE_PATH ) ; java . io . InputStream hdfsSiteIs = getClass ( ) . getClassLoader ( ) . getResourceAsStream ( com . hortonworks . streamline . streams . cluster . register . impl . HDFSServiceRegistrarTest . HDFS_SITE_XML_BADCASE_FILE_PATH ) ) { com . hortonworks . streamline . streams . cluster . register . ManualServiceRegistrar . ConfigFileInfo coreSiteXml = new com . hortonworks . streamline . streams . cluster . register . ManualServiceRegistrar . ConfigFileInfo ( com . hortonworks . streamline . streams . cluster . register . impl . HDFSServiceRegistrarTest . CORE_SITE_XML , coreSiteIs ) ; com . hortonworks . streamline . streams . cluster . register . ManualServiceRegistrar . ConfigFileInfo hdfsSiteXml = new com . hortonworks . streamline . streams . cluster . register . ManualServiceRegistrar . ConfigFileInfo ( com . hortonworks . streamline . streams . cluster . register . impl . HDFSServiceRegistrarTest . HDFS_SITE_XML , hdfsSiteIs ) ; registrar . register ( cluster , new com . hortonworks . streamline . common . Config ( ) , com . google . common . collect . Lists . newArrayList ( coreSiteXml , hdfsSiteXml ) ) ; org . junit . Assert . fail ( "Should<sp>throw<sp>IllegalArgumentException" ) ; } catch ( java . lang . IllegalArgumentException e ) { com . hortonworks . streamline . streams . cluster . catalog . Service hdfsService = environmentService . getServiceByName ( cluster . getId ( ) , Constants . HDFS . SERVICE_NAME ) ; "<AssertPlaceHolder>" ; } } getId ( ) { return id ; }
org . junit . Assert . assertNull ( hdfsService )
clearFilters ( ) { org . apache . log4j . Appender appender = getAppender ( ) ; appender . addFilter ( mock ( org . apache . log4j . spi . Filter . class ) ) ; appender . clearFilters ( ) ; "<AssertPlaceHolder>" ; } getFilter ( ) { return filter ; }
org . junit . Assert . assertNull ( appender . getFilter ( ) )
testParsePath ( ) { java . lang . Object [ ] tv = new java . lang . Object [ ] { "M<sp>/<sp>44H<sp>/<sp>0H<sp>/<sp>0H<sp>/<sp>1<sp>/<sp>1" , com . google . common . collect . ImmutableList . of ( new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 44 , true ) , new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 0 , true ) , new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 0 , true ) , new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 1 , false ) , new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 1 , false ) ) , "M/7H/3/3/1H/" , com . google . common . collect . ImmutableList . of ( new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 7 , true ) , new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 3 , false ) , new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 3 , false ) , new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 1 , true ) ) , "1<sp>H<sp>/<sp>2<sp>H<sp>/<sp>3<sp>H<sp>/" , com . google . common . collect . ImmutableList . of ( new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 1 , true ) , new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 2 , true ) , new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 3 , true ) ) , "1<sp>/<sp>2<sp>/<sp>3<sp>/" , com . google . common . collect . ImmutableList . of ( new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 1 , false ) , new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 2 , false ) , new com . matthewmitchell . peercoinj . crypto . ChildNumber ( 3 , false ) ) } ; for ( int i = 0 ; i < ( tv . length ) ; i += 2 ) { java . lang . String strPath = ( ( java . lang . String ) ( tv [ i ] ) ) ; java . util . List < com . matthewmitchell . peercoinj . crypto . ChildNumber > expectedPath = ( ( java . util . List < com . matthewmitchell . peercoinj . crypto . ChildNumber > ) ( tv [ ( i + 1 ) ] ) ) ; java . util . List < com . matthewmitchell . peercoinj . crypto . ChildNumber > path = com . matthewmitchell . peercoinj . crypto . HDUtils . parsePath ( strPath ) ; "<AssertPlaceHolder>" ; } } parsePath ( java . lang . String ) { java . lang . String [ ] parsedNodes = path . replace ( "M" , "" ) . split ( "/" ) ; java . util . List < com . matthewmitchell . peercoinj . crypto . ChildNumber > nodes = new java . util . ArrayList < com . matthewmitchell . peercoinj . crypto . ChildNumber > ( ) ; for ( java . lang . String n : parsedNodes ) { n = n . replaceAll ( "<sp>" , "" ) ; if ( ( n . length ( ) ) == 0 ) continue ; boolean isHard = n . endsWith ( "H" ) ; if ( isHard ) n = n . substring ( 0 , ( ( n . length ( ) ) - 1 ) ) ; int nodeNumber = java . lang . Integer . parseInt ( n ) ; nodes . add ( new com . matthewmitchell . peercoinj . crypto . ChildNumber ( nodeNumber , isHard ) ) ; } return nodes ; }
org . junit . Assert . assertEquals ( path , expectedPath )
testValideerGezagHouderGeenCurateleTeJong ( ) { final nl . bzk . brp . bijhouding . bericht . model . RegistratieAdresActieElement actie = createRegistratieAdresValideerActie ( false , false , nl . bzk . algemeenbrp . util . common . DatumUtil . vandaag ( ) , nl . bzk . brp . bijhouding . bericht . model . RegistratieAdresActieElementTest . GEZAGHOUDER ) ; final java . util . List < nl . bzk . brp . bijhouding . bericht . model . MeldingElement > meldingen = actie . valideerSpecifiekeInhoud ( ) ; "<AssertPlaceHolder>" ; } size ( ) { return elementen . size ( ) ; }
org . junit . Assert . assertEquals ( 0 , meldingen . size ( ) )
testResolveMessageLocaleNotPassed ( ) { org . terasoluna . gfw . common . message . ResultMessage message = mock ( org . terasoluna . gfw . common . message . ResultMessage . class ) ; org . springframework . context . MessageSource messageSource = mock ( org . springframework . context . MessageSource . class ) ; when ( message . getCode ( ) ) . thenReturn ( "MSG001" ) ; when ( message . getArgs ( ) ) . thenReturn ( null ) ; when ( messageSource . getMessage ( "MSG001" , null , java . util . Locale . getDefault ( ) ) ) . thenReturn ( "MESSAGE_TEXT" ) ; java . lang . String msg = org . terasoluna . gfw . common . message . ResultMessageUtils . resolveMessage ( message , messageSource ) ; "<AssertPlaceHolder>" ; } resolveMessage ( org . terasoluna . gfw . common . message . ResultMessage , org . springframework . context . MessageSource ) { return org . terasoluna . gfw . common . message . ResultMessageUtils . resolveMessage ( message , messageSource , java . util . Locale . getDefault ( ) ) ; }
org . junit . Assert . assertThat ( msg , org . hamcrest . CoreMatchers . is ( "MESSAGE_TEXT" ) )
validateByteSerialization ( ) { byte [ ] payload = org . apache . directmemory . serialization . StandardSerializerTest . serializer . serialize ( ( ( byte ) ( 127 ) ) ) ; byte res = org . apache . directmemory . serialization . StandardSerializerTest . serializer . deserialize ( payload , org . apache . directmemory . serialization . Byte . class ) ; "<AssertPlaceHolder>" ; } deserialize ( byte [ ] , java . lang . Class ) { java . io . ByteArrayInputStream bis = new java . io . ByteArrayInputStream ( source ) ; java . io . ObjectInputStream ois = new java . io . ObjectInputStream ( bis ) { @ org . apache . directmemory . serialization . Override protected org . apache . directmemory . serialization . Class < ? > resolveClass ( java . io . ObjectStreamClass objectStreamClass ) throws java . io . IOException , org . apache . directmemory . serialization . ClassNotFoundException { java . lang . ClassLoader classLoader = clazz . getClassLoader ( ) ; return classLoader != null ? classLoader . loadClass ( objectStreamClass . getName ( ) ) : java . lang . Class . forName ( objectStreamClass . getName ( ) ) ; } } ; T obj = clazz . cast ( ois . readObject ( ) ) ; ois . close ( ) ; return obj ; }
org . junit . Assert . assertEquals ( ( ( byte ) ( 127 ) ) , res )
invalidBooleanDefaultShouldCreateViolations ( ) { java . util . Set < javax . validation . ConstraintViolation < org . alien4cloud . tosca . model . definitions . PropertyDefinition > > violations = validator . validate ( createDefinitions ( ToscaTypes . BOOLEAN , "not<sp>a<sp>boolean" ) ) ; "<AssertPlaceHolder>" ; } createDefinitions ( java . lang . String , org . alien4cloud . tosca . model . definitions . PropertyConstraint ) { org . alien4cloud . tosca . model . definitions . PropertyDefinition propertyDefinition = new org . alien4cloud . tosca . model . definitions . PropertyDefinition ( ) ; propertyDefinition . setType ( propertyType ) ; propertyDefinition . setConstraints ( com . google . common . collect . Lists . newArrayList ( constraint ) ) ; return propertyDefinition ; }
org . junit . Assert . assertEquals ( 1 , violations . size ( ) )
testPanel ( ) { final com . google . gwt . user . client . ui . Panel expectedPanel = mock ( com . google . gwt . user . client . ui . AbsolutePanel . class ) ; doReturn ( expectedPanel ) . when ( gridLienzoScrollHandler ) . getScrollPanel ( ) ; final com . google . gwt . user . client . ui . Panel actualPanel = gridLienzoScrollBars . scrollPanel ( ) ; "<AssertPlaceHolder>" ; } scrollPanel ( ) { return gridLienzoScrollHandler . getScrollPanel ( ) ; }
org . junit . Assert . assertEquals ( expectedPanel , actualPanel )
getCollectionFieldsIsSuccessful ( ) { com . ibm . watson . discovery . v1 . model . ListCollectionFieldsOptions getOptions = new com . ibm . watson . discovery . v1 . model . ListCollectionFieldsOptions . Builder ( com . ibm . watson . discovery . v1 . DiscoveryServiceIT . environmentId , com . ibm . watson . discovery . v1 . DiscoveryServiceIT . collectionId ) . build ( ) ; com . ibm . watson . discovery . v1 . model . ListCollectionFieldsResponse getResponse = discovery . listCollectionFields ( getOptions ) . execute ( ) . getResult ( ) ; "<AssertPlaceHolder>" ; } getFields ( ) { return fields ; }
org . junit . Assert . assertFalse ( getResponse . getFields ( ) . isEmpty ( ) )
shouldApplyRelationshipGroupCommandToTheStore ( ) { final org . neo4j . kernel . impl . api . BatchTransactionApplier applier = newApplier ( false ) ; final org . neo4j . kernel . impl . store . record . RelationshipGroupRecord before = new org . neo4j . kernel . impl . store . record . RelationshipGroupRecord ( 42 , 1 ) ; final org . neo4j . kernel . impl . store . record . RelationshipGroupRecord after = new org . neo4j . kernel . impl . store . record . RelationshipGroupRecord ( 42 , 1 , 2 , 3 , 4 , 5 , 6 , true ) ; final org . neo4j . kernel . impl . transaction . command . Command command = new org . neo4j . kernel . impl . transaction . command . Command . RelationshipGroupCommand ( before , after ) ; final boolean result = apply ( applier , command :: handle , transactionToApply ) ; "<AssertPlaceHolder>" ; verify ( relationshipGroupStore , times ( 1 ) ) . updateRecord ( after ) ; } apply ( org . neo4j . kernel . api . proc . Context , java . lang . Object [ ] , org . neo4j . kernel . api . ResourceTracker ) { try { java . lang . Thread . sleep ( 50 ) ; } catch ( java . lang . InterruptedException e ) { throw new org . neo4j . internal . kernel . api . exceptions . ProcedureException ( Status . General . UnknownError , e , "Interrupted" ) ; } return org . neo4j . collection . RawIterator . empty ( ) ; }
org . junit . Assert . assertFalse ( result )
testSorting ( ) { java . util . Collections . sort ( files , updateComparator ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( expected , files )
checksumAdlerMissingFileTest ( ) { java . lang . String nonExistingFile = "nonExistingFile" ; long checksum = org . pentaho . di . core . row . ValueDataUtil . ChecksumAdler32 ( new org . pentaho . di . core . row . value . ValueMetaString ( ) , nonExistingFile ) ; "<AssertPlaceHolder>" ; } ChecksumAdler32 ( org . pentaho . di . core . row . ValueMetaInterface , java . lang . Object ) { long checksum = 0 ; try { checksum = org . pentaho . di . core . row . ValueDataUtil . checksumAdler32 ( metaA , dataA , false ) ; } catch ( org . pentaho . di . core . exception . KettleFileNotFoundException e ) { } return checksum ; }
org . junit . Assert . assertEquals ( 0 , checksum )
test_minus_Iterable_zero ( ) { java . lang . Iterable < org . joda . money . Money > iterable = java . util . Arrays . asList ( org . joda . money . TestMoney . GBP_0_00 ) ; org . joda . money . Money test = org . joda . money . TestMoney . GBP_2_34 . minus ( iterable ) ; "<AssertPlaceHolder>" ; } minus ( double ) { return minus ( amountToSubtract , RoundingMode . UNNECESSARY ) ; }
org . junit . Assert . assertSame ( org . joda . money . TestMoney . GBP_2_34 , test )
testGetProcessDefinitionById ( ) { java . lang . String processDefinitionId = "processDefinitionId" ; org . activiti . engine . repository . ProcessDefinitionQuery processDefinitionQuery = mock ( org . activiti . engine . repository . ProcessDefinitionQuery . class ) ; when ( activitiRepositoryService . createProcessDefinitionQuery ( ) ) . thenReturn ( processDefinitionQuery ) ; when ( processDefinitionQuery . processDefinitionId ( processDefinitionId ) ) . thenReturn ( processDefinitionQuery ) ; org . activiti . engine . repository . ProcessDefinition expectedProcessDefinition = mock ( org . activiti . engine . repository . ProcessDefinition . class ) ; when ( processDefinitionQuery . singleResult ( ) ) . thenReturn ( expectedProcessDefinition ) ; org . activiti . engine . repository . ProcessDefinition actualProcessDefinition = activitiService . getProcessDefinitionById ( processDefinitionId ) ; "<AssertPlaceHolder>" ; org . mockito . InOrder inOrder = inOrder ( processDefinitionQuery ) ; inOrder . verify ( processDefinitionQuery ) . processDefinitionId ( processDefinitionId ) ; inOrder . verify ( processDefinitionQuery ) . singleResult ( ) ; inOrder . verifyNoMoreInteractions ( ) ; } getProcessDefinitionById ( java . lang . String ) { return activitiRepositoryService . createProcessDefinitionQuery ( ) . processDefinitionId ( processDefinitionId ) . singleResult ( ) ; }
org . junit . Assert . assertSame ( expectedProcessDefinition , actualProcessDefinition )
testUserSummary ( ) { java . util . Calendar ca = java . util . Calendar . getInstance ( ) ; ca . add ( Calendar . DAY_OF_MONTH , ( - 7 ) ) ; java . util . List < ? > userSummaryList = dataApi . datacube ( DatacubeType . GETUSERSUMMARY , ca . getTime ( ) , 3 ) ; "<AssertPlaceHolder>" ; System . err . println ( userSummaryList ) ; } isEmpty ( ) { return this . headers . isEmpty ( ) ; }
org . junit . Assert . assertFalse ( userSummaryList . isEmpty ( ) )
testEmptyBucket ( ) { com . streamsets . pipeline . stage . executor . s3 . config . AmazonS3ExecutorConfig config = getConfig ( ) ; config . s3Config . bucketTemplate = "" ; com . streamsets . pipeline . stage . executor . s3 . AmazonS3Executor executor = new com . streamsets . pipeline . stage . executor . s3 . AmazonS3Executor ( config ) ; com . streamsets . pipeline . sdk . TargetRunner runner = new com . streamsets . pipeline . sdk . TargetRunner . Builder ( com . streamsets . pipeline . stage . executor . s3 . AmazonS3DExecutor . class , executor ) . setOnRecordError ( OnRecordError . TO_ERROR ) . build ( ) ; runner . runInit ( ) ; try { runner . runWrite ( com . google . common . collect . ImmutableList . of ( getTestRecord ( ) ) ) ; "<AssertPlaceHolder>" ; } finally { runner . runDestroy ( ) ; } } getErrorRecords ( ) { return errorRecords ; }
org . junit . Assert . assertEquals ( 1 , runner . getErrorRecords ( ) . size ( ) )
shouldSelectAListOfButtons ( ) { java . util . List < com . redhat . darcy . ui . api . elements . Button > shouldFind = java . util . Arrays . asList ( mock ( com . redhat . darcy . ui . api . elements . Button . class ) ) ; when ( selection . elementsOfType ( com . redhat . darcy . ui . api . elements . Button . class , com . redhat . darcy . ui . By . id ( "test" ) ) ) . thenReturn ( shouldFind ) ; "<AssertPlaceHolder>" ; } buttons ( com . redhat . darcy . ui . api . Locator ) { return com . redhat . darcy . ui . Elements . elements ( com . redhat . darcy . ui . api . elements . Button . class , locator ) ; }
org . junit . Assert . assertSame ( shouldFind , selection . buttons ( com . redhat . darcy . ui . By . id ( "test" ) ) )
testCompound ( ) { com . comphenix . protocol . utility . StreamSerializer serializer = new com . comphenix . protocol . utility . StreamSerializer ( ) ; com . comphenix . protocol . wrappers . nbt . NbtCompound initial = com . comphenix . protocol . wrappers . nbt . NbtFactory . ofCompound ( "tag" ) ; initial . put ( "name" , "Ole" ) ; initial . put ( "age" , 20 ) ; com . comphenix . protocol . utility . ByteArrayOutputStream buffer = new com . comphenix . protocol . utility . ByteArrayOutputStream ( ) ; serializer . serializeCompound ( new com . comphenix . protocol . utility . DataOutputStream ( buffer ) , initial ) ; com . comphenix . protocol . utility . DataInputStream input = new com . comphenix . protocol . utility . DataInputStream ( new com . comphenix . protocol . utility . ByteArrayInputStream ( buffer . toByteArray ( ) ) ) ; com . comphenix . protocol . wrappers . nbt . NbtCompound deserialized = serializer . deserializeCompound ( input ) ; "<AssertPlaceHolder>" ; } deserializeCompound ( java . io . DataInputStream ) { if ( input == null ) throw new java . lang . IllegalArgumentException ( "Input<sp>stream<sp>cannot<sp>be<sp>NULL." ) ; java . lang . Object nmsCompound = null ; if ( com . comphenix . protocol . utility . MinecraftReflection . isUsingNetty ( ) ) { if ( ( com . comphenix . protocol . utility . StreamSerializer . READ_NBT_METHOD ) == null ) { com . comphenix . protocol . utility . StreamSerializer . READ_NBT_METHOD = com . comphenix . protocol . reflect . accessors . Accessors . getMethodAccessor ( com . comphenix . protocol . reflect . FuzzyReflection . fromClass ( com . comphenix . protocol . utility . MinecraftReflection . getPacketDataSerializerClass ( ) , true ) . getMethodByParameters ( "readNbtCompound" , com . comphenix . protocol . utility . MinecraftReflection . getNBTCompoundClass ( ) , new java . lang . Class < ? > [ 0 ] ) ) ; } io . netty . buffer . ByteBuf buf = com . comphenix . protocol . injector . netty . NettyByteBufAdapter . packetReader ( input ) ; nmsCompound = com . comphenix . protocol . utility . StreamSerializer . READ_NBT_METHOD . invoke ( buf ) ; } else { if ( ( com . comphenix . protocol . utility . StreamSerializer . READ_NBT_METHOD ) == null ) { com . comphenix . protocol . utility . StreamSerializer . READ_NBT_METHOD = com . comphenix . protocol . reflect . accessors . Accessors . getMethodAccessor ( com . comphenix . protocol . reflect . FuzzyReflection . fromClass ( com . comphenix . protocol . utility . MinecraftReflection . getPacketClass ( ) ) . getMethod ( com . comphenix . protocol . reflect . fuzzy . FuzzyMethodContract . newBuilder ( ) . parameterCount ( 1 ) . parameterDerivedOf ( java . io . DataInput . class ) . returnDerivedOf ( com . comphenix . protocol . utility . MinecraftReflection . getNBTBaseClass ( ) ) . build ( ) ) ) ; } try { nmsCompound = com . comphenix . protocol . utility . StreamSerializer . READ_NBT_METHOD . invoke ( null , input ) ; } catch ( java . lang . Exception e ) { throw new java . io . IOException ( "Cannot<sp>read<sp>item<sp>stack." , e ) ; } } if ( nmsCompound != null ) return com . comphenix . protocol . wrappers . nbt . NbtFactory . fromNMSCompound ( nmsCompound ) ; else return null ; }
org . junit . Assert . assertEquals ( initial , deserialized )
xhasField_shouldReturnFalseIfGivenFieldIsNotDeclaredInParameterizedClassOrItsSubClass ( ) { org . openmrs . util . Reflect reflect = new org . openmrs . util . Reflect ( org . openmrs . OpenmrsObject . class ) ; java . util . List < java . lang . reflect . Field > allFields = org . openmrs . util . Reflect . getAllFields ( org . openmrs . util . OpenmrsObjectImp . class ) ; java . lang . reflect . Field normalClassField = findFieldByName ( allFields , "normalClassField" ) ; "<AssertPlaceHolder>" ; } hasField ( java . lang . reflect . Field ) { return isSuperClass ( field . getDeclaringClass ( ) ) ; }
org . junit . Assert . assertFalse ( reflect . hasField ( normalClassField ) )
mergeTextDecorationNormalizeFirstTest ( ) { java . lang . String firstValue = "<sp>underline<sp>" ; java . lang . String secondValue = "strikethrough<sp>bold" ; java . lang . String expected = "underline<sp>strikethrough<sp>bold" ; java . lang . String actual = com . itextpdf . styledxmlparser . css . resolve . CssPropertyMerger . mergeTextDecoration ( firstValue , secondValue ) ; "<AssertPlaceHolder>" ; } mergeTextDecoration ( java . lang . String , java . lang . String ) { if ( firstValue == null ) { return secondValue ; } else if ( secondValue == null ) { return firstValue ; } java . util . Set < java . lang . String > merged = com . itextpdf . styledxmlparser . css . resolve . CssPropertyMerger . normalizeTextDecoration ( firstValue ) ; merged . addAll ( com . itextpdf . styledxmlparser . css . resolve . CssPropertyMerger . normalizeTextDecoration ( secondValue ) ) ; java . lang . StringBuilder sb = new java . lang . StringBuilder ( ) ; for ( java . lang . String mergedProp : merged ) { if ( ( sb . length ( ) ) != 0 ) { sb . append ( "<sp>" ) ; } sb . append ( mergedProp ) ; } return ( sb . length ( ) ) != 0 ? sb . toString ( ) : com . itextpdf . styledxmlparser . css . CommonCssConstants . NONE ; }
org . junit . Assert . assertEquals ( expected , actual )
testKeyForTarget ( ) { org . apache . cayenne . DataObject t = mock ( org . apache . cayenne . DataObject . class ) ; when ( t . readProperty ( "abc" ) ) . thenReturn ( 44 ) ; when ( t . readNestedProperty ( "abc" ) ) . thenReturn ( 44 ) ; "<AssertPlaceHolder>" ; } keyForTarget ( org . apache . cayenne . DataObject ) { return getOrCreatePathExpression ( ) . evaluate ( target ) ; }
org . junit . Assert . assertEquals ( 44 , mapper . keyForTarget ( t ) )
testWriteRackId ( ) { java . io . File kafkaFolder = tmpFolder . newFolder ( "opt" , "kafka" ) ; java . lang . String rackFolder = ( kafkaFolder . getAbsolutePath ( ) ) + "/rack" ; new java . io . File ( rackFolder ) . mkdir ( ) ; java . util . Map < java . lang . String , java . lang . String > envVars = new java . util . HashMap ( io . strimzi . kafka . init . InitWriterTest . envVars ) ; envVars . put ( InitWriterConfig . INIT_FOLDER , rackFolder ) ; io . strimzi . kafka . init . InitWriterConfig config = io . strimzi . kafka . init . InitWriterConfig . fromMap ( envVars ) ; io . fabric8 . kubernetes . client . KubernetesClient client = mockKubernetesClient ( config . getNodeName ( ) , io . strimzi . kafka . init . InitWriterTest . labels , Collections . EMPTY_LIST ) ; io . strimzi . kafka . init . InitWriter writer = new io . strimzi . kafka . init . InitWriter ( client , config ) ; "<AssertPlaceHolder>" ; } writeRack ( ) { java . util . Map < java . lang . String , java . lang . String > nodeLabels = client . nodes ( ) . withName ( config . getNodeName ( ) ) . get ( ) . getMetadata ( ) . getLabels ( ) ; io . strimzi . kafka . init . InitWriter . log . info ( "NodeLabels<sp>=<sp>{}" , nodeLabels ) ; java . lang . String rackId = nodeLabels . get ( config . getRackTopologyKey ( ) ) ; io . strimzi . kafka . init . InitWriter . log . info ( "Rack:<sp>{}<sp>=<sp>{}" , config . getRackTopologyKey ( ) , rackId ) ; if ( rackId == null ) { io . strimzi . kafka . init . InitWriter . log . error ( "Node<sp>{}<sp>doesn't<sp>have<sp>the<sp>label<sp>{}<sp>for<sp>getting<sp>the<sp>rackid" , config . getNodeName ( ) , config . getRackTopologyKey ( ) ) ; return false ; } return write ( io . strimzi . kafka . init . InitWriter . FILE_RACK_ID , rackId ) ; }
org . junit . Assert . assertTrue ( writer . writeRack ( ) )
selectsTaskThatHasBeenUsedTheLeast ( ) { targetSelector . chooseTask ( new int [ ] { org . apache . storm . grouping . partialKeyGrouping . BalancedTargetSelectorTest . TASK_LIST [ 0 ] } ) ; targetSelector . chooseTask ( new int [ ] { org . apache . storm . grouping . partialKeyGrouping . BalancedTargetSelectorTest . TASK_LIST [ 1 ] } ) ; targetSelector . chooseTask ( new int [ ] { org . apache . storm . grouping . partialKeyGrouping . BalancedTargetSelectorTest . TASK_LIST [ 2 ] } ) ; int selectedTask = targetSelector . chooseTask ( org . apache . storm . grouping . partialKeyGrouping . BalancedTargetSelectorTest . TASK_LIST ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertThat ( selectedTask , org . hamcrest . Matchers . equalTo ( org . apache . storm . grouping . partialKeyGrouping . BalancedTargetSelectorTest . TASK_LIST [ 3 ] ) )
testHandlesResponse_1 ( ) { final org . apache . shindig . gadgets . oauth2 . OAuth2Accessor accessor = org . apache . shindig . gadgets . oauth2 . MockUtils . getOAuth2Accessor_Error ( ) ; final org . apache . shindig . gadgets . http . HttpResponse response = new org . apache . shindig . gadgets . http . HttpResponse ( ) ; final boolean result = org . apache . shindig . gadgets . oauth2 . handler . TokenAuthorizationResponseHandlerTest . tarh . handlesResponse ( accessor , response ) ; "<AssertPlaceHolder>" ; } handlesResponse ( org . apache . shindig . gadgets . oauth2 . OAuth2Accessor , org . apache . shindig . gadgets . http . HttpResponse ) { if ( ( ( accessor == null ) || ( ! ( accessor . isValid ( ) ) ) ) || ( accessor . isErrorResponse ( ) ) ) { return false ; } return response != null ; }
org . junit . Assert . assertFalse ( result )
trackable_input_post_close_04 ( ) { java . io . InputStream input = this . generateData ( 0 ) ; @ org . apache . jena . hadoop . rdf . io . input . util . SuppressWarnings ( "resource" ) org . apache . jena . hadoop . rdf . io . input . util . TrackableInputStream trackable = this . getInstance ( input ) ; trackable . close ( ) ; "<AssertPlaceHolder>" ; } read ( jena . Model , java . lang . String , java . lang . String ) { try { jena . URL url = new jena . URL ( in ) ; model . read ( in , lang ) ; } catch ( jena . java e ) { model . read ( new jena . FileInputStream ( in ) , "" , lang ) ; } }
org . junit . Assert . assertEquals ( 0 , trackable . read ( new byte [ 16 ] , 0 , 0 ) )
testBsnAlsZoekcriteria ( ) { final nl . bzk . brp . business . regels . context . BerichtRegelContext regelContext = new nl . bzk . brp . business . regels . context . BerichtRegelContext ( null , nl . bzk . brp . model . algemeen . stamgegeven . kern . SoortAdministratieveHandeling . GEEF_DETAILS_PERSOON , maakBericht ( nl . bzk . brp . business . regels . bevraging . BRBV0001Test . BSN_AANWEZIG , nl . bzk . brp . business . regels . bevraging . BRBV0001Test . ANR_NIET_AANWEZIG , nl . bzk . brp . business . regels . bevraging . BRBV0001Test . OBJECTSLEUTEL_NIET_AANWEZIG ) ) ; final java . util . List < nl . bzk . brp . model . basis . BerichtIdentificeerbaar > overtreders = brbv0001 . valideer ( regelContext ) ; "<AssertPlaceHolder>" ; } size ( ) { return elementen . size ( ) ; }
org . junit . Assert . assertEquals ( 0 , overtreders . size ( ) )
shouldTheAlgorithmReturnANumberOfSolutionsWhenSolvingASimpleProblem ( ) { org . uma . jmetal . problem . Problem < org . uma . jmetal . solution . DoubleSolution > problem ; org . uma . jmetal . algorithm . InteractiveAlgorithm < org . uma . jmetal . solution . DoubleSolution , java . util . List < org . uma . jmetal . solution . DoubleSolution > > algorithmRun ; org . uma . jmetal . operator . CrossoverOperator < org . uma . jmetal . solution . DoubleSolution > crossover ; org . uma . jmetal . operator . MutationOperator < org . uma . jmetal . solution . DoubleSolution > mutation ; org . uma . jmetal . operator . SelectionOperator < java . util . List < org . uma . jmetal . solution . DoubleSolution > , org . uma . jmetal . solution . DoubleSolution > selection ; int numberIterations = 1 ; int numberObjectives = 3 ; int numberVariables = 7 ; java . lang . String weightsName = "" ; int populationSize = 100 ; problem = new org . uma . jmetal . problem . multiobjective . dtlz . DTLZ1 ( numberVariables , numberObjectives ) ; double crossoverProbability = 0.9 ; double crossoverDistributionIndex = 20.0 ; crossover = new org . uma . jmetal . operator . impl . crossover . SBXCrossover ( crossoverProbability , crossoverDistributionIndex ) ; double mutationProbability = 1.0 / ( problem . getNumberOfVariables ( ) ) ; double mutationDistributionIndex = 20.0 ; mutation = new org . uma . jmetal . operator . impl . mutation . PolynomialMutation ( mutationProbability , mutationDistributionIndex ) ; selection = new org . uma . jmetal . operator . impl . selection . BinaryTournamentSelection < org . uma . jmetal . solution . DoubleSolution > ( new org . uma . jmetal . util . comparator . RankingAndCrowdingDistanceComparator < org . uma . jmetal . solution . DoubleSolution > ( ) ) ; org . uma . jmetal . util . point . impl . IdealPoint idealPoint = new org . uma . jmetal . util . point . impl . IdealPoint ( problem . getNumberOfObjectives ( ) ) ; idealPoint . update ( problem . createSolution ( ) . getObjectives ( ) ) ; org . uma . jmetal . util . point . impl . NadirPoint nadirPoint = new org . uma . jmetal . util . point . impl . NadirPoint ( problem . getNumberOfObjectives ( ) ) ; nadirPoint . update ( problem . createSolution ( ) . getObjectives ( ) ) ; double considerationProbability = 0.1 ; java . util . List < java . lang . Double > rankingCoeficient = new java . util . ArrayList ( ) ; for ( int i = 0 ; i < ( problem . getNumberOfObjectives ( ) ) ; i ++ ) { rankingCoeficient . add ( ( 1.0 / ( problem . getNumberOfObjectives ( ) ) ) ) ; } for ( int cont = 0 ; cont < numberIterations ; cont ++ ) { java . util . List < java . lang . Double > referencePoint = new java . util . ArrayList ( ) ; double epsilon = 0.01 ; java . util . List < java . lang . Double > asp = new java . util . ArrayList ( ) ; for ( int i = 0 ; i < ( problem . getNumberOfObjectives ( ) ) ; i ++ ) { asp . add ( 0.0 ) ; referencePoint . add ( 0.0 ) ; } algorithmRun = new org . uma . jmetal . algorithm . multiobjective . wasfga . WASFGA ( problem , populationSize , 200 , crossover , mutation , selection , new org . uma . jmetal . util . evaluator . impl . SequentialSolutionListEvaluator ( ) , epsilon , referencePoint , weightsName ) ; algorithm = new org . uma . jmetal . util . artificialdecisionmaker . impl . ArtificiallDecisionMakerBuilder ( problem , algorithmRun ) . setConsiderationProbability ( considerationProbability ) . setMaxEvaluations ( 11 ) . setTolerance ( 0.001 ) . setAsp ( asp ) . build ( ) ; new org . uma . jmetal . util . AlgorithmRunner . Executor ( algorithm ) . execute ( ) ; java . util . List < java . lang . Double > referencePoints = ( ( org . uma . jmetal . util . artificialdecisionmaker . impl . ArtificialDecisionMakerDecisionTree < org . uma . jmetal . solution . DoubleSolution > ) ( algorithm ) ) . getReferencePoints ( ) ; "<AssertPlaceHolder>" ; } } size ( ) { return map . size ( ) ; }
org . junit . Assert . assertTrue ( ( ( referencePoints . size ( ) ) >= ( numberObjectives * numberIterations ) ) )
setLowPriorityWaitFail ( ) { org . threadly . concurrent . AbstractPrioritySchedulerTest . AbstractPrioritySchedulerFactory factory = getAbstractPrioritySchedulerFactory ( ) ; long lowPriorityWait = 1000 ; org . threadly . concurrent . AbstractPriorityScheduler scheduler = factory . makeAbstractPriorityScheduler ( 1 , TaskPriority . High , lowPriorityWait ) ; try { try { scheduler . setMaxWaitForLowPriority ( ( - 1 ) ) ; org . junit . Assert . fail ( "Exception<sp>should<sp>have<sp>thrown" ) ; } catch ( java . lang . IllegalArgumentException e ) { } "<AssertPlaceHolder>" ; } finally { factory . shutdown ( ) ; } } getMaxWaitForLowPriority ( ) { return getQueueManager ( ) . getMaxWaitForLowPriority ( ) ; }
org . junit . Assert . assertEquals ( lowPriorityWait , scheduler . getMaxWaitForLowPriority ( ) )
testFetchByPrimaryKeysWithMultiplePrimaryKeysWhereNoPrimaryKeysExist ( ) { long pk1 = com . liferay . portal . kernel . test . util . RandomTestUtil . nextLong ( ) ; long pk2 = com . liferay . portal . kernel . test . util . RandomTestUtil . nextLong ( ) ; java . util . Set < java . io . Serializable > primaryKeys = new java . util . HashSet < java . io . Serializable > ( ) ; primaryKeys . add ( pk1 ) ; primaryKeys . add ( pk2 ) ; java . util . Map < java . io . Serializable , com . liferay . journal . model . JournalFolder > journalFolders = _persistence . fetchByPrimaryKeys ( primaryKeys ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return _portalCacheListeners . isEmpty ( ) ; }
org . junit . Assert . assertTrue ( journalFolders . isEmpty ( ) )
testAcceptInValidParameter ( ) { state . setInput ( "/x<sp>" ) ; "<AssertPlaceHolder>" ; } accept ( org . openspotlight . graph . query . console . ConsoleState ) { org . openspotlight . common . util . Assertions . checkNotNull ( "state" , state ) ; if ( ( ( state . getActiveCommand ( ) ) == null ) && ( state . getInput ( ) . trim ( ) . equals ( "display<sp>properties" ) ) ) { return true ; } return false ; }
org . junit . Assert . assertThat ( command . accept ( state ) , org . hamcrest . core . Is . is ( false ) )
decodeHttpRequestInvalidHeadername ( ) { org . kaazing . mina . filter . codec . ProtocolCodecSessionEx session = new org . kaazing . mina . filter . codec . ProtocolCodecSessionEx ( ) ; org . apache . mina . filter . codec . ProtocolDecoder decoder = new org . kaazing . gateway . transport . http . bridge . filter . HttpRequestDecoder ( ) ; java . nio . ByteBuffer in = java . nio . ByteBuffer . wrap ( ( "GET<sp>/echo<sp>HTTP/1.1\r\n" + "Upgrade:<sp>WebSocket\r\nConnection:<sp>Upgrade\r\n:abc\r\n" ) . getBytes ( ) ) ; org . kaazing . mina . core . buffer . IoBufferAllocatorEx < ? > allocator = session . getBufferAllocator ( ) ; org . kaazing . mina . core . buffer . IoBufferEx buf = allocator . wrap ( in ) ; decoder . decode ( session , ( ( org . apache . mina . core . buffer . IoBuffer ) ( buf ) ) , session . getDecoderOutput ( ) ) ; decoder . finishDecode ( session , session . getDecoderOutput ( ) ) ; org . kaazing . gateway . transport . http . bridge . HttpRequestMessage httpRequest = ( ( org . kaazing . gateway . transport . http . bridge . HttpRequestMessage ) ( session . getDecoderOutputQueue ( ) . poll ( ) ) ) ; "<AssertPlaceHolder>" ; } poll ( ) { if ( isEmpty ( ) ) { return null ; } java . lang . Object ret = items [ first ] ; items [ first ] = null ; decreaseSize ( ) ; if ( ( first ) == ( last ) ) { first = last = 0 ; } shrinkIfNeeded ( ) ; return ( ( E ) ( ret ) ) ; }
org . junit . Assert . assertEquals ( httpRequest , null )
verify_wrongTemplateVersionInErrorMessage ( ) { org . oscm . domobjects . Organization o = new org . oscm . domobjects . Organization ( ) ; o . setKey ( 1 ) ; org . oscm . domobjects . Organization t = new org . oscm . domobjects . Organization ( ) ; t . setKey ( 0 ) ; try { org . oscm . validation . VersionAndKeyValidator . verify ( o , t , 1 ) ; } catch ( org . oscm . internal . types . exception . ConcurrentModificationException e ) { java . lang . String versionInMessage = e . getMessage ( ) . substring ( ( ( e . getMessage ( ) . length ( ) ) - 1 ) , e . getMessage ( ) . length ( ) ) ; "<AssertPlaceHolder>" ; } } getMessage ( ) { return ( ( "EXCEPTIONID<sp>" + ( id ) ) + ":<sp>" ) + ( super . getMessage ( ) ) ; }
org . junit . Assert . assertEquals ( "1" , versionInMessage )
shouldCountNumberOfActiveLocks ( ) { clientA . acquireShared ( LockTracer . NONE , org . neo4j . kernel . impl . locking . ResourceTypes . LABEL , 0 ) ; clientA . acquireShared ( LockTracer . NONE , org . neo4j . kernel . impl . locking . ResourceTypes . RELATIONSHIP , 17 ) ; clientA . acquireShared ( LockTracer . NONE , org . neo4j . kernel . impl . locking . ResourceTypes . NODE , 12 ) ; long count = clientA . activeLockCount ( ) ; "<AssertPlaceHolder>" ; } activeLockCount ( ) { return activeLockCount ; }
org . junit . Assert . assertEquals ( 3 , count )
get ( ) { com . dexcoder . test . model . People people = new com . dexcoder . test . model . People ( ) ; people . setPeopleName ( "liyd-get" ) ; people . setPeopleAge ( 12 ) ; people . setGmtCreate ( new java . util . Date ( ) ) ; java . lang . String id = jdbcDao . insert ( people ) ; com . dexcoder . test . model . People people1 = jdbcDao . get ( com . dexcoder . test . model . People . class , id ) ; "<AssertPlaceHolder>" ; } getPeopleName ( ) { return peopleName ; }
org . junit . Assert . assertEquals ( people . getPeopleName ( ) , people1 . getPeopleName ( ) )
getSupportedVariableOrderMap ( ) { final java . lang . String sparql = "SELECT<sp>?f<sp>?m<sp>?d<sp>{<sp>" + ( ( "?f<sp><urn:talksTo><sp>?m<sp>.<sp>" + "?m<sp><uri:associatesWith><sp>?d<sp>.<sp>" ) + "}" ) ; final org . eclipse . rdf4j . query . parser . ParsedQuery query = new org . eclipse . rdf4j . query . parser . sparql . SPARQLParser ( ) . parseQuery ( sparql , null ) ; final org . eclipse . rdf4j . query . algebra . Projection pcjExpression = ( ( org . eclipse . rdf4j . query . algebra . Projection ) ( query . getTupleExpr ( ) ) ) ; final org . apache . rya . indexing . external . tupleSet . SimpleExternalTupleSet testSet = new org . apache . rya . indexing . external . tupleSet . SimpleExternalTupleSet ( pcjExpression ) ; final java . util . Map < java . lang . String , java . util . Set < java . lang . String > > expected = new java . util . HashMap ( ) ; java . lang . String varOrder = "f" ; java . util . Set < java . lang . String > vars = new java . util . HashSet ( ) ; vars . add ( "f" ) ; expected . put ( varOrder , vars ) ; varOrder = "f;m" ; vars = new java . util . HashSet ( ) ; vars . add ( "f" ) ; vars . add ( "m" ) ; expected . put ( varOrder , vars ) ; varOrder = "f;m;d" ; vars = new java . util . HashSet ( ) ; vars . add ( "f" ) ; vars . add ( "m" ) ; vars . add ( "d" ) ; expected . put ( varOrder , vars ) ; "<AssertPlaceHolder>" ; } getSupportedVariableOrderMap ( ) { final java . lang . String sparql = "SELECT<sp>?f<sp>?m<sp>?d<sp>{<sp>" + ( ( "?f<sp><urn:talksTo><sp>?m<sp>.<sp>" + "?m<sp><uri:associatesWith><sp>?d<sp>.<sp>" ) + "}" ) ; final org . eclipse . rdf4j . query . parser . ParsedQuery query = new org . eclipse . rdf4j . query . parser . sparql . SPARQLParser ( ) . parseQuery ( sparql , null ) ; final org . eclipse . rdf4j . query . algebra . Projection pcjExpression = ( ( org . eclipse . rdf4j . query . algebra . Projection ) ( query . getTupleExpr ( ) ) ) ; final org . apache . rya . indexing . external . tupleSet . SimpleExternalTupleSet testSet = new org . apache . rya . indexing . external . tupleSet . SimpleExternalTupleSet ( pcjExpression ) ; final java . util . Map < java . lang . String , java . util . Set < java . lang . String > > expected = new java . util . HashMap ( ) ; java . lang . String varOrder = "f" ; java . util . Set < java . lang . String > vars = new java . util . HashSet ( ) ; vars . add ( "f" ) ; expected . put ( varOrder , vars ) ; varOrder = "f;m" ; vars = new java . util . HashSet ( ) ; vars . add ( "f" ) ; vars . add ( "m" ) ; expected . put ( varOrder , vars ) ; varOrder = "f;m;d" ; vars = new java . util . HashSet ( ) ; vars . add ( "f" ) ; vars . add ( "m" ) ; vars . add ( "d" ) ; expected . put ( varOrder , vars ) ; org . junit . Assert . assertEquals ( expected , testSet . getSupportedVariableOrderMap ( ) ) ; }
org . junit . Assert . assertEquals ( expected , testSet . getSupportedVariableOrderMap ( ) )
rasterizePolygons ( ) { org . geotools . data . simple . SimpleFeatureCollection features = createPolys ( ) ; org . geotools . geometry . jts . ReferencedEnvelope bounds = features . getBounds ( ) ; java . awt . Dimension gridDim = new java . awt . Dimension ( ( ( int ) ( bounds . getWidth ( ) ) ) , ( ( int ) ( bounds . getHeight ( ) ) ) ) ; java . lang . String covName = "Test" ; org . opengis . util . ProgressListener monitor = null ; org . geotools . coverage . grid . GridCoverage2D cov = org . geotools . process . vector . VectorToRasterProcess . process ( features , "value" , gridDim , bounds , covName , monitor ) ; java . util . Map < java . lang . Integer , org . locationtech . jts . geom . Envelope > rects = new java . util . HashMap < java . lang . Integer , org . locationtech . jts . geom . Envelope > ( ) ; org . geotools . data . simple . SimpleFeatureIterator iter = features . features ( ) ; while ( iter . hasNext ( ) ) { org . opengis . feature . simple . SimpleFeature sf = iter . next ( ) ; rects . put ( ( ( java . lang . Integer ) ( sf . getAttribute ( "value" ) ) ) , ( ( org . locationtech . jts . geom . Geometry ) ( sf . getDefaultGeometry ( ) ) ) . getEnvelopeInternal ( ) ) ; } try { org . opengis . referencing . operation . MathTransform2D mt = cov . getGridGeometry ( ) . getGridToCRS2D ( PixelOrientation . CENTER ) ; java . awt . Point gridP = new java . awt . Point ( ) ; java . awt . geom . Point2D . Double geoP = new java . awt . geom . Point2D . Double ( ) ; int [ ] covValue = new int [ 1 ] ; for ( gridP . y = 0 ; ( gridP . y ) < ( gridDim . height ) ; ( gridP . y ) ++ ) { for ( gridP . x = 0 ; ( gridP . x ) < ( gridDim . width ) ; ( gridP . x ) ++ ) { mt . transform ( gridP , geoP ) ; } cov . evaluate ( geoP , covValue ) ; if ( ( covValue [ 0 ] ) != 0 ) { "<AssertPlaceHolder>" ; } } } catch ( org . opengis . referencing . operation . TransformException tex ) { throw new java . lang . RuntimeException ( tex ) ; } } get ( int ) { switch ( value ) { case net . opengis . wcs10 . InterpolationMethodType . NEAREST_NEIGHBOR : return net . opengis . wcs10 . InterpolationMethodType . NEAREST_NEIGHBOR_LITERAL ; case net . opengis . wcs10 . InterpolationMethodType . BILINEAR : return net . opengis . wcs10 . InterpolationMethodType . BILINEAR_LITERAL ; case net . opengis . wcs10 . InterpolationMethodType . BICUBIC : return net . opengis . wcs10 . InterpolationMethodType . BICUBIC_LITERAL ; case net . opengis . wcs10 . InterpolationMethodType . LOST_AREA : return net . opengis . wcs10 . InterpolationMethodType . LOST_AREA_LITERAL ; case net . opengis . wcs10 . InterpolationMethodType . BARYCENTRIC : return net . opengis . wcs10 . InterpolationMethodType . BARYCENTRIC_LITERAL ; case net . opengis . wcs10 . InterpolationMethodType . NONE : return net . opengis . wcs10 . InterpolationMethodType . NONE_LITERAL ; } return null ; }
org . junit . Assert . assertTrue ( rects . get ( covValue [ 0 ] ) . contains ( geoP . x , geoP . y ) )
testRemove ( ) { chain . addLast ( "A" , new org . apache . mina . core . IoFilterChainTest . EventOrderTestFilter ( 'A' ) ) ; chain . addLast ( "B" , new org . apache . mina . core . IoFilterChainTest . EventOrderTestFilter ( 'A' ) ) ; chain . addLast ( "C" , new org . apache . mina . core . IoFilterChainTest . EventOrderTestFilter ( 'A' ) ) ; chain . addLast ( "D" , new org . apache . mina . core . IoFilterChainTest . EventOrderTestFilter ( 'A' ) ) ; chain . addLast ( "E" , new org . apache . mina . core . IoFilterChainTest . EventOrderTestFilter ( 'A' ) ) ; chain . remove ( "A" ) ; chain . remove ( "E" ) ; chain . remove ( "C" ) ; chain . remove ( "B" ) ; chain . remove ( "D" ) ; "<AssertPlaceHolder>" ; } getAll ( ) { java . util . List < org . apache . mina . handler . chain . IoHandlerChain . Entry > list = new java . util . ArrayList ( ) ; org . apache . mina . handler . chain . IoHandlerChain . Entry e = head . nextEntry ; while ( e != ( tail ) ) { list . add ( e ) ; e = e . nextEntry ; } return list ; }
org . junit . Assert . assertEquals ( 0 , chain . getAll ( ) . size ( ) )
testDate ( ) { java . lang . String result = new org . sejda . model . pdf . TextStampPattern ( ) . build ( "FooBar<sp>[DATE]" ) ; java . lang . String expected = "FooBar<sp>" + ( org . sejda . model . pdf . TextStampPattern . dateNow ( ) ) ; "<AssertPlaceHolder>" ; } dateNow ( ) { return new java . text . SimpleDateFormat ( "dd/MM/yyyy" ) . format ( new java . util . Date ( ) ) ; }
org . junit . Assert . assertEquals ( expected , result )
testList ( ) { java . util . List < com . riversoft . weixin . qy . agent . bean . Agent > list = com . riversoft . weixin . qy . agent . Agents . defaultAgents ( ) . list ( ) ; "<AssertPlaceHolder>" ; } list ( ) { java . lang . String url = com . riversoft . weixin . mp . base . WxEndpoint . get ( "url.care.account.list" ) ; java . lang . String response = wxClient . get ( url ) ; com . riversoft . weixin . mp . care . Accounts . logger . debug ( "list<sp>care<sp>accounts<sp>:{}" , response ) ; com . riversoft . weixin . mp . care . Accounts . CareListResponse careListResponse = com . riversoft . weixin . common . util . JsonMapper . defaultMapper ( ) . fromJson ( response , com . riversoft . weixin . mp . care . Accounts . CareListResponse . class ) ; return careListResponse . getList ( ) ; }
org . junit . Assert . assertNotNull ( list )
testValidStream ( ) { java . io . InputStream stream = mock ( java . io . InputStream . class ) ; org . sejda . model . input . PdfStreamSource instance = org . sejda . model . input . PdfStreamSource . newInstanceWithPassword ( stream , "dsadsada" , "dsdasdsa" ) ; "<AssertPlaceHolder>" ; } newInstanceWithPassword ( java . io . InputStream , java . lang . String , java . lang . String ) { if ( stream == null ) { throw new java . lang . IllegalArgumentException ( "A<sp>not<sp>null<sp>stream<sp>instance<sp>and<sp>a<sp>not<sp>blank<sp>name<sp>are<sp>expected." ) ; } return new org . sejda . model . input . PdfStreamSource ( stream , name , password ) ; }
org . junit . Assert . assertNotNull ( instance )
testSchema ( ) { java . util . Properties props = new java . util . Properties ( ) ; java . lang . String ddl = org . teiid . translator . simpledb . visitors . TestSimpleDBMetadataProcessor . getDDL ( props ) ; java . lang . String expectedDDL = "CREATE<sp>FOREIGN<sp>TABLE<sp>G1<sp>(\n" + ( ( ( ( ( ( ( ( ( ( ( "\tItemName<sp>string<sp>NOT<sp>NULL<sp>OPTIONS<sp>(NAMEINSOURCE<sp>\'itemName()\'),\n" + "\te1<sp>string,\n" ) + "\te2<sp>string,\n" ) + "\tCONSTRAINT<sp>PK0<sp>PRIMARY<sp>KEY(ItemName)\n" ) + ")<sp>OPTIONS<sp>(UPDATABLE<sp>TRUE);\n" ) + "\n" ) + "CREATE<sp>FOREIGN<sp>TABLE<sp>G2<sp>(\n" ) + "\tItemName<sp>string<sp>NOT<sp>NULL<sp>OPTIONS<sp>(NAMEINSOURCE<sp>\'itemName()\'),\n" ) + "\te1<sp>string,\n" ) + "\te2<sp>string[],\n" ) + "\tCONSTRAINT<sp>PK0<sp>PRIMARY<sp>KEY(ItemName)\n" ) + ")<sp>OPTIONS<sp>(UPDATABLE<sp>TRUE);" ) ; "<AssertPlaceHolder>" ; } getDDL ( java . util . Properties ) { org . teiid . translator . simpledb . SimpleDBExecutionFactory translator = new org . teiid . translator . simpledb . SimpleDBExecutionFactory ( ) ; translator . start ( ) ; org . teiid . metadata . MetadataFactory mf = new org . teiid . metadata . MetadataFactory ( "vdb" , 1 , "people" , org . teiid . query . metadata . SystemMetadata . getInstance ( ) . getRuntimeTypeMap ( ) , props , null ) ; org . teiid . translator . simpledb . api . SimpleDBConnection connection = org . mockito . Mockito . mock ( org . teiid . translator . simpledb . api . SimpleDBConnection . class ) ; org . mockito . Mockito . stub ( connection . getDomains ( ) ) . toReturn ( java . util . Arrays . asList ( "G1" , "G2" ) ) ; java . util . HashSet < org . teiid . translator . simpledb . api . SimpleDBConnection . SimpleDBAttribute > cols = new java . util . HashSet < org . teiid . translator . simpledb . api . SimpleDBConnection . SimpleDBAttribute > ( ) ; cols . add ( new org . teiid . translator . simpledb . api . SimpleDBConnection . SimpleDBAttribute ( "e1" , false ) ) ; cols . add ( new org . teiid . translator . simpledb . api . SimpleDBConnection . SimpleDBAttribute ( "e2" , false ) ) ; org . mockito . Mockito . stub ( connection . getAttributeNames ( "G1" ) ) . toReturn ( cols ) ; java . util . HashSet < org . teiid . translator . simpledb . api . SimpleDBConnection . SimpleDBAttribute > cols2 = new java . util . HashSet < org . teiid . translator . simpledb . api . SimpleDBConnection . SimpleDBAttribute > ( ) ; cols2 . add ( new org . teiid . translator . simpledb . api . SimpleDBConnection . SimpleDBAttribute ( "e1" , false ) ) ; cols2 . add ( new org . teiid . translator . simpledb . api . SimpleDBConnection . SimpleDBAttribute ( "e2" , true ) ) ; org . mockito . Mockito . stub ( connection . getAttributeNames ( "G2" ) ) . toReturn ( cols2 ) ; translator . getMetadata ( mf , connection ) ; org . teiid . query . metadata . TransformationMetadata metadata = org . teiid . query . unittest . RealMetadataFactory . createTransformationMetadata ( mf . asMetadataStore ( ) , "vdb" , new org . teiid . query . function . FunctionTree ( "foo" , new org . teiid . query . function . UDFSource ( translator . getPushDownFunctions ( ) ) ) ) ; org . teiid . query . validator . ValidatorReport report = new org . teiid . query . metadata . MetadataValidator ( ) . validate ( metadata . getVdbMetaData ( ) , metadata . getMetadataStore ( ) ) ; if ( report . hasItems ( ) ) { throw new java . lang . RuntimeException ( report . getFailureMessage ( ) ) ; } java . lang . String ddl = org . teiid . query . metadata . DDLStringVisitor . getDDLString ( mf . getSchema ( ) , null , null ) ; return ddl ; }
org . junit . Assert . assertEquals ( expectedDDL , ddl )
testGetFormattedMACAddress_EmptyString ( ) { com . emc . storageos . driver . dellsc . scapi . objects . ScControllerPortIscsiConfiguration config = new com . emc . storageos . driver . dellsc . scapi . objects . ScControllerPortIscsiConfiguration ( ) ; config . macAddress = "" ; "<AssertPlaceHolder>" ; } getFormattedMACAddress ( ) { java . lang . String defaultReturn = "00:00:00:00:00:00" ; if ( ( macAddress ) == null ) { return defaultReturn ; } java . lang . String mac = macAddress . replace ( "-" , "" ) ; if ( ( mac . length ( ) ) != 12 ) { return defaultReturn ; } java . util . List < java . lang . String > parts = new java . util . ArrayList ( ) ; for ( int i = 0 ; i < 6 ; i ++ ) { int offset = 2 * i ; parts . add ( mac . substring ( offset , ( offset + 2 ) ) ) ; } return java . lang . String . join ( ":" , parts ) ; }
org . junit . Assert . assertTrue ( "00:00:00:00:00:00" . equals ( config . getFormattedMACAddress ( ) ) )
shouldRejectNullVipsObject ( ) { org . openstack . atlas . api . validation . results . ValidatorResult result = vipsValidator . validate ( null , org . openstack . atlas . api . mgmt . validation . validators . POST ) ; "<AssertPlaceHolder>" ; } resultMessage ( org . openstack . atlas . api . validation . results . ValidatorResult , java . lang . Enum ) { java . lang . StringBuilder sb = new java . lang . StringBuilder ( ) ; if ( ! ( result . passedValidation ( ) ) ) { java . util . List < org . openstack . atlas . api . validation . results . ExpectationResult > ers = result . getValidationResults ( ) ; sb . append ( java . lang . String . format ( "ON<sp>%s<sp>result.withMessage([" , ctx . toString ( ) ) ) ; for ( org . openstack . atlas . api . validation . results . ExpectationResult er : ers ) { sb . append ( java . lang . String . format ( "%s" , er . getMessage ( ) ) ) ; sb . append ( "])" ) ; } } else { sb . append ( java . lang . String . format ( "On<sp>%s<sp>All<sp>Expectations<sp>PASSED\n" , ctx . toString ( ) ) ) ; } return sb . toString ( ) ; }
org . junit . Assert . assertFalse ( resultMessage ( result , org . openstack . atlas . api . mgmt . validation . validators . POST ) , result . passedValidation ( ) )
testDoCollect ( ) { final java . util . Collection < eu . toolchain . concurrent . Stage < java . lang . Object > > futures = com . google . common . collect . ImmutableList . of ( f1 , f2 ) ; doReturn ( completable ) . when ( underTest ) . completable ( ) ; doNothing ( ) . when ( underTest ) . bindSignals ( completable , futures ) ; "<AssertPlaceHolder>" ; verify ( underTest ) . completable ( ) ; verify ( underTest ) . bindSignals ( completable , futures ) ; verify ( f1 ) . handle ( any ( eu . toolchain . concurrent . CollectHelper . class ) ) ; verify ( f2 ) . handle ( any ( eu . toolchain . concurrent . CollectHelper . class ) ) ; } doCollect ( java . util . Collection , java . util . function . Function ) { final eu . toolchain . concurrent . Completable < T > target = completable ( ) ; final eu . toolchain . concurrent . CollectHelper < ? super C , ? extends T > done = new eu . toolchain . concurrent . CollectHelper ( stages . size ( ) , collector , stages , target ) ; for ( final eu . toolchain . concurrent . Stage < ? extends C > q : stages ) { q . handle ( done ) ; } bindSignals ( target , stages ) ; return target ; }
org . junit . Assert . assertEquals ( completable , underTest . doCollect ( futures , collector ) )
header ( ) { response . setHeader ( "aaa" , "111" ) ; "<AssertPlaceHolder>" ; } getHeader ( java . lang . String ) { java . util . List < java . lang . String > values = getHeaderList ( name ) ; if ( values != null ) { return values . get ( 0 ) ; } return null ; }
org . junit . Assert . assertThat ( response . getHeader ( "aaa" ) , org . hamcrest . CoreMatchers . is ( "111" ) )
create_new_list ( ) { java . util . List < java . lang . String > myList = com . google . common . collect . Lists . newArrayList ( ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertNotNull ( myList )
testAtTimestampTypes ( ) { final java . lang . String fam = "f16bc" ; final java . lang . String qual = "q1pm2" ; final long ts = 324324L ; final java . lang . String val = "v8672194923750" ; org . apache . accumulo . core . data . Mutation expected = new org . apache . accumulo . core . data . Mutation ( "row5" ) ; expected . put ( fam , qual , ts , val ) ; org . apache . accumulo . core . data . Mutation actual = new org . apache . accumulo . core . data . Mutation ( "row5" ) ; actual . at ( ) . family ( fam ) . qualifier ( qual ) . timestamp ( ts ) . put ( val ) ; "<AssertPlaceHolder>" ; } put ( org . apache . hadoop . io . Text , org . apache . hadoop . io . Text , org . apache . accumulo . core . security . ColumnVisibility , org . apache . accumulo . core . data . Value ) { put ( columnFamily , columnQualifier , columnVisibility . getExpression ( ) , false , 0L , false , value . get ( ) ) ; }
org . junit . Assert . assertEquals ( expected , actual )