input
stringlengths
28
18.7k
output
stringlengths
39
1.69k
testStripLoadRulesWindows ( ) { java . lang . String rawCs = org . apache . commons . io . IOUtils . toString ( getClass ( ) . getResourceAsStream ( "ct-catcs-2-CRLF.log" ) ) ; hudson . plugins . clearcase . ConfigSpec cs = new hudson . plugins . clearcase . ConfigSpec ( rawCs , false ) ; cs = cs . stripLoadRules ( ) ; "<AssertPlaceHolder>" ; } getLoadRules ( ) { java . util . Set < java . lang . String > rules = new java . util . HashSet < java . lang . String > ( ) ; for ( java . lang . String row : org . apache . commons . lang . StringUtils . split ( raw , hudson . plugins . clearcase . util . PathUtil . newLineForOS ( isUnix ) ) ) { java . lang . String trimmedRow = row . trim ( ) ; if ( trimmedRow . startsWith ( "load" ) ) { java . lang . String rule = row . trim ( ) . substring ( "load" . length ( ) ) . trim ( ) ; rules . add ( rule ) ; } } return rules ; }
org . junit . Assert . assertTrue ( cs . getLoadRules ( ) . isEmpty ( ) )
testWriteIncludingIntermediateCommunities ( ) { final java . lang . String cypher = "CALL<sp>algo.louvain('',<sp>'',<sp>{concurrency:1,<sp>includeIntermediateCommunities:<sp>true})" ; final com . carrotsearch . hppc . IntIntScatterMap testMap = new com . carrotsearch . hppc . IntIntScatterMap ( ) ; org . neo4j . graphalgo . algo . LouvainClusteringIntegrationTest . DB . execute ( cypher ) . close ( ) ; java . lang . String readQuery = "MATCH<sp>(n)<sp>RETURN<sp>n.communities<sp>AS<sp>communities" ; org . neo4j . graphalgo . algo . LouvainClusteringIntegrationTest . DB . execute ( readQuery ) . accept ( ( row ) -> { final long community = ( ( int [ ] ) ( row . get ( "communities" ) ) ) [ 0 ] ; testMap . addTo ( ( ( int ) ( community ) ) , 1 ) ; return true ; } ) ; "<AssertPlaceHolder>" ; } size ( ) { return size ; }
org . junit . Assert . assertEquals ( 3 , testMap . size ( ) )
a3_startTSD ( ) { try { com . exacttarget . fuelsdk . ETTriggeredTest . ette . setKey ( tsName ) ; com . exacttarget . fuelsdk . ETTriggeredTest . ette . setStatus ( ETTriggeredEmail . Status . ACTIVE ) ; com . exacttarget . fuelsdk . ETResponse < com . exacttarget . fuelsdk . ETTriggeredEmail > response = com . exacttarget . fuelsdk . ETTriggeredTest . client . update ( com . exacttarget . fuelsdk . ETTriggeredTest . ette ) ; System . out . println ( ( "resp=" + ( response . getResponseMessage ( ) ) ) ) ; "<AssertPlaceHolder>" ; com . exacttarget . fuelsdk . ETResult < com . exacttarget . fuelsdk . ETTriggeredEmail > result = response . getResult ( ) ; System . out . println ( ( "res=" + ( result . toString ( ) ) ) ) ; } catch ( com . exacttarget . fuelsdk . ETSdkException ex ) { java . util . logging . Logger . getLogger ( com . exacttarget . fuelsdk . ETTriggeredTest . class . getName ( ) ) . log ( Level . SEVERE , null , ex ) ; ex . printStackTrace ( ) ; } } getObject ( ) { return getResult ( ) . getObject ( ) ; }
org . junit . Assert . assertEquals ( response . getObject ( ) . getStatus ( ) , ETTriggeredEmail . Status . ACTIVE )
testPersoonMetIndicatieDerdeHeeftGezagBijEenOuderGeenOuderschapEnZonderOuderlijkGezag ( ) { final java . util . List < nl . bzk . brp . model . basis . BerichtEntiteit > berichtEntiteiten = brby2017 . voerRegelUit ( maakPersoon ( SoortIndicatie . INDICATIE_DERDE_HEEFT_GEZAG , 1 , null , nl . bzk . brp . bijhouding . business . regels . impl . gegevenset . ouder . ouderlijkgezag . BRBY2017Test . GEEN_OUDERSCHAP ) , maakPersoonBericht ( ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return elementen . size ( ) ; }
org . junit . Assert . assertEquals ( 0 , berichtEntiteiten . size ( ) )
testSetsRecAsParameter ( ) { com . eclipsesource . tabris . tracking . internal . piwik . request . PiwikRequest request = new com . eclipsesource . tabris . tracking . internal . piwik . request . PiwikRequest ( new com . eclipsesource . tabris . tracking . internal . piwik . model . PiwikConfiguration ( "apiVersion" , 2 ) , new com . eclipsesource . tabris . tracking . internal . piwik . model . action . Action ( "actionUrl" ) , new com . eclipsesource . tabris . tracking . internal . piwik . request . PiwikRequestTest . TestRequestable ( "foo" , "bar" ) ) ; "<AssertPlaceHolder>" ; } assemble ( ) { java . util . Map < java . lang . String , java . lang . Object > parameters = createRequestableParameters ( ) ; parameters . putAll ( configuration . getParameter ( ) ) ; parameters . putAll ( action . getParameter ( ) ) ; parameters . put ( com . eclipsesource . tabris . tracking . internal . piwik . request . RequestKeyProvider . getRequestKey ( com . eclipsesource . tabris . tracking . internal . piwik . request . RequestKeys . RANDOM ) , java . util . UUID . randomUUID ( ) . toString ( ) ) ; parameters . put ( com . eclipsesource . tabris . tracking . internal . piwik . request . RequestKeyProvider . getRequestKey ( com . eclipsesource . tabris . tracking . internal . piwik . request . RequestKeys . REC ) , java . lang . Integer . valueOf ( 1 ) ) ; return parameters ; }
org . junit . Assert . assertEquals ( java . lang . Integer . valueOf ( 1 ) , request . assemble ( ) . get ( com . eclipsesource . tabris . tracking . internal . piwik . request . RequestKeyProvider . getRequestKey ( com . eclipsesource . tabris . tracking . internal . piwik . request . RequestKeys . REC ) ) )
testBuildWihDisabledStatusConstaintWithOrderBy ( ) { unit . setActive ( false ) ; org . lnu . is . domain . contact . type . ContactType context = new org . lnu . is . domain . contact . type . ContactType ( ) ; org . lnu . is . pagination . OrderBy orderBy1 = new org . lnu . is . pagination . OrderBy ( "abbrName" , org . lnu . is . pagination . OrderByType . ASC ) ; org . lnu . is . pagination . OrderBy orderBy2 = new org . lnu . is . pagination . OrderBy ( "name" , org . lnu . is . pagination . OrderByType . DESC ) ; java . util . List < org . lnu . is . pagination . OrderBy > orders = java . util . Arrays . asList ( orderBy1 , orderBy2 ) ; java . lang . String expectedQuery = "SELECT<sp>e<sp>FROM<sp>ContactType<sp>e<sp>WHERE<sp>e.crtUserGroup<sp>IN<sp>(:userGroups)<sp>ORDER<sp>BY<sp>e.abbrName<sp>ASC,<sp>e.name<sp>DESC" ; org . lnu . is . pagination . MultiplePagedSearch < org . lnu . is . domain . contact . type . ContactType > pagedSearch = new org . lnu . is . pagination . MultiplePagedSearch ( ) ; pagedSearch . setEntity ( context ) ; pagedSearch . setOrders ( orders ) ; java . lang . String actualQuery = unit . build ( pagedSearch ) ; "<AssertPlaceHolder>" ; } setOrders ( java . util . List ) { this . orders = orders ; }
org . junit . Assert . assertEquals ( expectedQuery , actualQuery )
testCreateSettingsPanel ( ) { System . out . println ( "createSettingsPanel" ) ; kg . apc . jmeter . graphs . AbstractMonitoringVisualizer instance = new kg . apc . jmeter . graphs . AbstractMonitoringVisualizerTest . AbstractMonitoringVisualizerImpl ( ) ; kg . apc . jmeter . vizualizers . JSettingsPanel result = instance . createSettingsPanel ( ) ; "<AssertPlaceHolder>" ; } createSettingsPanel ( ) { return new kg . apc . jmeter . vizualizers . JSettingsPanel ( this , ( ( ( ( ( ( ( JSettingsPanel . TIMELINE_OPTION ) | ( JSettingsPanel . GRADIENT_OPTION ) ) | ( JSettingsPanel . FINAL_ZEROING_OPTION ) ) | ( JSettingsPanel . LIMIT_POINT_OPTION ) ) | ( JSettingsPanel . MAXY_OPTION ) ) | ( JSettingsPanel . RELATIVE_TIME_OPTION ) ) | ( JSettingsPanel . MARKERS_OPTION ) ) ) ; }
org . junit . Assert . assertNotNull ( result )
testGetAllElementsWithViewRestrictedByGroupAndAPostAggregationFilter ( ) { final uk . gov . gchq . gaffer . graph . Graph graph = uk . gov . gchq . gaffer . mapstore . impl . GetAllElementsHandlerTest . getGraph ( ) ; final uk . gov . gchq . gaffer . operation . impl . add . AddElements addElements = new uk . gov . gchq . gaffer . operation . impl . add . AddElements . Builder ( ) . input ( uk . gov . gchq . gaffer . mapstore . impl . GetAllElementsHandlerTest . getElements ( ) ) . build ( ) ; graph . execute ( addElements , new uk . gov . gchq . gaffer . user . User ( ) ) ; final uk . gov . gchq . gaffer . operation . impl . get . GetAllElements getAllElements = new uk . gov . gchq . gaffer . operation . impl . get . GetAllElements . Builder ( ) . view ( new uk . gov . gchq . gaffer . data . elementdefinition . view . View . Builder ( ) . edge ( uk . gov . gchq . gaffer . mapstore . impl . GetAllElementsHandlerTest . BASIC_EDGE1 , new uk . gov . gchq . gaffer . data . elementdefinition . view . ViewElementDefinition . Builder ( ) . postAggregationFilter ( new uk . gov . gchq . gaffer . data . element . function . ElementFilter . Builder ( ) . select ( uk . gov . gchq . gaffer . mapstore . impl . GetAllElementsHandlerTest . COUNT ) . execute ( new uk . gov . gchq . koryphe . impl . predicate . IsMoreThan ( 5 ) ) . build ( ) ) . build ( ) ) . build ( ) ) . build ( ) ; final uk . gov . gchq . gaffer . commonutil . iterable . CloseableIterable < ? extends uk . gov . gchq . gaffer . data . element . Element > results = graph . execute ( getAllElements , new uk . gov . gchq . gaffer . user . User ( ) ) ; final java . util . Set < uk . gov . gchq . gaffer . data . element . Element > resultsSet = new java . util . HashSet ( ) ; uk . gov . gchq . gaffer . commonutil . stream . Streams . toStream ( results ) . forEach ( resultsSet :: add ) ; final java . util . Set < uk . gov . gchq . gaffer . data . element . Element > expectedResults = new java . util . HashSet ( ) ; uk . gov . gchq . gaffer . mapstore . impl . GetAllElementsHandlerTest . getElements ( ) . stream ( ) . filter ( ( e ) -> ( e . getGroup ( ) . equals ( uk . gov . gchq . gaffer . mapstore . impl . GetAllElementsHandlerTest . BASIC_EDGE1 ) ) && ( ( ( int ) ( e . getProperty ( uk . gov . gchq . gaffer . mapstore . impl . GetAllElementsHandlerTest . COUNT ) ) ) > 5 ) ) . forEach ( expectedResults :: add ) ; "<AssertPlaceHolder>" ; } forEach ( java . util . function . BiConsumer ) { delegateMap . forEach ( action ) ; }
org . junit . Assert . assertEquals ( expectedResults , resultsSet )
testMzData ( ) { java . io . File fileName = new java . io . File ( ( ( io . github . msdk . io . filetypedetection . FileTypeDetectionAlgorithmTest . TEST_DATA_PATH ) + "test.mzData" ) ) ; io . github . msdk . datamodel . FileType fileType = io . github . msdk . io . filetypedetection . FileTypeDetectionAlgorithm . detectDataFileType ( fileName ) ; "<AssertPlaceHolder>" ; } detectDataFileType ( java . io . File ) { com . google . common . base . Preconditions . checkNotNull ( fileName ) ; if ( fileName . isDirectory ( ) ) { for ( java . io . File f : fileName . listFiles ( ) ) { if ( ( f . isFile ( ) ) && ( f . getName ( ) . matches ( "_FUNC[0-9]{3}.DAT" ) ) ) return io . github . msdk . datamodel . FileType . WATERS_RAW ; } return null ; } java . io . InputStreamReader reader = new java . io . InputStreamReader ( new java . io . FileInputStream ( fileName ) , "ISO-8859-1" ) ; char [ ] buffer = new char [ 1024 ] ; reader . read ( buffer ) ; reader . close ( ) ; java . lang . String fileHeader = new java . lang . String ( buffer ) ; if ( fileHeader . startsWith ( io . github . msdk . io . filetypedetection . FileTypeDetectionAlgorithm . THERMO_HEADER ) ) { return io . github . msdk . datamodel . FileType . THERMO_RAW ; } if ( fileHeader . startsWith ( io . github . msdk . io . filetypedetection . FileTypeDetectionAlgorithm . CDF_HEADER ) ) { return io . github . msdk . datamodel . FileType . NETCDF ; } if ( fileHeader . contains ( io . github . msdk . io . filetypedetection . FileTypeDetectionAlgorithm . MZML_HEADER ) ) return io . github . msdk . datamodel . FileType . MZML ; if ( fileHeader . contains ( io . github . msdk . io . filetypedetection . FileTypeDetectionAlgorithm . MZDATA_HEADER ) ) return io . github . msdk . datamodel . FileType . MZDATA ; if ( fileHeader . contains ( io . github . msdk . io . filetypedetection . FileTypeDetectionAlgorithm . MZXML_HEADER ) ) return io . github . msdk . datamodel . FileType . MZXML ; if ( fileHeader . contains ( io . github . msdk . io . filetypedetection . FileTypeDetectionAlgorithm . MZTAB_HEADER ) ) return io . github . msdk . datamodel . FileType . MZTAB ; if ( ( fileHeader . startsWith ( io . github . msdk . io . filetypedetection . FileTypeDetectionAlgorithm . MZDB_HEADER ) ) && ( fileHeader . contains ( io . github . msdk . io . filetypedetection . FileTypeDetectionAlgorithm . MZDB_HEADER2 ) ) ) return io . github . msdk . datamodel . FileType . MZDB ; return null ; }
org . junit . Assert . assertEquals ( FileType . MZDATA , fileType )
testRemove ( ) { com . liferay . reading . time . model . ReadingTimeEntry newReadingTimeEntry = addReadingTimeEntry ( ) ; _persistence . remove ( newReadingTimeEntry ) ; com . liferay . reading . time . model . ReadingTimeEntry existingReadingTimeEntry = _persistence . fetchByPrimaryKey ( newReadingTimeEntry . getPrimaryKey ( ) ) ; "<AssertPlaceHolder>" ; } getPrimaryKey ( ) { return _amImageEntryId ; }
org . junit . Assert . assertNull ( existingReadingTimeEntry )
testCollectorActionServiceGeneration ( ) { java . lang . Boolean actionServiceRegistered = false ; java . util . List < it . geosolutions . geobatch . annotations . GenericActionService > actionList = it . geosolutions . geobatch . annotations . ActionServicePostProcessor . getActionList ( ) ; for ( it . geosolutions . geobatch . annotations . GenericActionService action : actionList ) { if ( it . geosolutions . geobatch . actions . commons . CollectorConfiguration . class . getSimpleName ( ) . equals ( action . getId ( ) ) ) { actionServiceRegistered = true ; break ; } } "<AssertPlaceHolder>" ; } getId ( ) { return id ; }
org . junit . Assert . assertTrue ( actionServiceRegistered )
testAddPartitionsForViewNullPartSd ( ) { java . lang . String tableName = "test_add_partition_view" ; createView ( tableName ) ; org . apache . hadoop . hive . metastore . api . Partition partition = buildPartition ( org . apache . hadoop . hive . metastore . client . TestAddPartitions . DB_NAME , tableName , org . apache . hadoop . hive . metastore . client . TestAddPartitions . DEFAULT_YEAR_VALUE ) ; partition . setSd ( null ) ; java . util . List < org . apache . hadoop . hive . metastore . api . Partition > partitions = com . google . common . collect . Lists . newArrayList ( partition ) ; client . add_partitions ( partitions ) ; org . apache . hadoop . hive . metastore . api . Partition part = client . getPartition ( org . apache . hadoop . hive . metastore . client . TestAddPartitions . DB_NAME , tableName , "year=2017" ) ; "<AssertPlaceHolder>" ; } getSd ( ) { return tTable . getSd ( ) ; }
org . junit . Assert . assertNull ( part . getSd ( ) )
findsClassesFromJarManifestClassPathJar ( ) { createJarTo ( new java . io . File ( getTestFolder ( ) ) ) ; org . jboss . shrinkwrap . api . spec . JavaArchive archive = org . jboss . shrinkwrap . api . ShrinkWrap . create ( org . jboss . shrinkwrap . api . spec . JavaArchive . class , "parent.jar" ) . addAsManifestResource ( com . poolik . classfinder . DEFAULT_MANIFEST_NAME ) ; java . io . File target = new java . io . File ( getTestFolder ( ) , "parent.jar" ) ; archive . as ( org . jboss . shrinkwrap . api . exporter . ZipExporter . class ) . exportTo ( target , true ) ; java . util . Collection < com . poolik . classfinder . info . ClassInfo > classes = new com . poolik . classfinder . ClassFinder ( ) . add ( target ) . findClasses ( ) ; "<AssertPlaceHolder>" ; } findClasses ( ) { return findClasses ( null ) ; }
org . junit . Assert . assertThat ( classes . size ( ) , org . hamcrest . core . Is . is ( 1 ) )
deleteNode ( ) { java . lang . String path = masterSlaveNodeApi . saveNode ( new com . zuoxiaolong . niubi . job . api . data . MasterSlaveNodeData . Data ( "192.168.1.101" ) ) ; com . zuoxiaolong . niubi . job . api . data . MasterSlaveNodeData nodeData = masterSlaveNodeApi . getNode ( path ) ; "<AssertPlaceHolder>" ; masterSlaveNodeApi . deleteNode ( path ) ; try { masterSlaveNodeApi . getNode ( path ) ; } catch ( com . zuoxiaolong . niubi . job . core . exception . NiubiException e ) { throw e . getCause ( ) ; } } getNode ( java . lang . String ) { return new com . zuoxiaolong . niubi . job . api . data . MasterSlaveNodeData ( getData ( path ) ) ; }
org . junit . Assert . assertNotNull ( nodeData )
resendAsynchronouslySingleMessageWhenMailContentStoreLoadingFails ( ) { org . xwiki . mail . MailContentStore contentStore = this . mocker . getInstance ( org . xwiki . mail . MailContentStore . class , "filesystem" ) ; when ( contentStore . load ( any ( ) , eq ( "batchId" ) , eq ( "messageId" ) ) ) . thenThrow ( new org . xwiki . mail . MailStoreException ( "error" ) ) ; try { this . mocker . getComponentUnderTest ( ) . resendAsynchronously ( "batchId" , "messageId" ) ; org . junit . Assert . fail ( "Should<sp>have<sp>thrown<sp>an<sp>exception<sp>here" ) ; } catch ( org . xwiki . mail . MailStoreException expected ) { "<AssertPlaceHolder>" ; } } getMessage ( ) { return this . message ; }
org . junit . Assert . assertEquals ( "error" , expected . getMessage ( ) )
write_Cmodes ( ) { if ( ! ( isJp2KakDriverAvailable ) ) return ; it . geosolutions . imageio . plugins . jp2kakadu . JP2KWriteTest . LOGGER . info ( "Testing<sp>JP2<sp>Write<sp>operation<sp>with<sp>CModes<sp>option<sp>setting" ) ; final java . io . File inputFile = it . geosolutions . resources . TestData . file ( this , it . geosolutions . imageio . plugins . jp2kakadu . JP2KWriteTest . testFileName ) ; "<AssertPlaceHolder>" ; final javax . media . jai . ParameterBlockJAI pbjImageRead = new javax . media . jai . ParameterBlockJAI ( "BYPASS|RESTART|CAUSAL" 6 ) ; pbjImageRead . setParameter ( "Input" , inputFile ) ; if ( it . geosolutions . imageio . plugins . jp2kakadu . JP2KWriteTest . ENABLE_SUBSAMPLING ) { javax . imageio . ImageReadParam readParam = new javax . imageio . ImageReadParam ( ) ; readParam . setSourceSubsampling ( 4 , 4 , 0 , 0 ) ; pbjImageRead . setParameter ( "readParam" , readParam ) ; } pbjImageRead . setParameter ( "Reader" , new it . geosolutions . imageio . plugins . jp2kakadu . JP2GDALKakaduImageReaderSpi ( ) . createReaderInstance ( ) ) ; javax . media . jai . RenderedOp image = javax . media . jai . JAI . create ( "BYPASS|RESTART|CAUSAL" 6 , pbjImageRead ) ; final java . lang . String [ ] createVersions = new java . lang . String [ ] { "BYPASS|RESTART|CAUSAL" 4 , "BYPASS|RESTART|CAUSAL" , "RESTART|ERTERM" , "BYPASS|RESTART|CAUSAL" 3 } ; final java . lang . String [ ] filenameVersions = new java . lang . String [ ] { "BYPASS|RESTART|CAUSAL" 4 , "BYPASSRESTARTCAUSAL" , "RESTARTERTERM" , "BYPASS|RESTART|CAUSAL" 3 } ; final int numberOfVersions = createVersions . length ; for ( int i = 0 ; i < numberOfVersions ; i ++ ) { final java . lang . String filenameVersion = filenameVersions [ i ] ; final java . lang . StringBuffer fileName = new java . lang . StringBuffer ( "BYPASS|RESTART|CAUSAL" 1 ) . append ( filenameVersion ) . append ( ".jp2" ) ; final java . io . File outputFile = it . geosolutions . resources . TestData . temp ( this , fileName . toString ( ) , it . geosolutions . imageio . plugins . jp2kakadu . JP2KWriteTest . deleteTempFilesOnExit ) ; final javax . media . jai . ParameterBlockJAI pbjImageWrite = new javax . media . jai . ParameterBlockJAI ( "BYPASS|RESTART|CAUSAL" 0 ) ; pbjImageWrite . setParameter ( "BYPASS|RESTART|CAUSAL" 2 , new it . geosolutions . imageio . stream . output . FileImageOutputStreamExtImpl ( outputFile ) ) ; javax . imageio . ImageWriter writer = new it . geosolutions . imageio . plugins . jp2kakadu . JP2GDALKakaduImageWriterSpi ( ) . createWriterInstance ( ) ; pbjImageWrite . setParameter ( "Writer" , writer ) ; pbjImageWrite . addSource ( image ) ; javax . imageio . ImageWriteParam param = writer . getDefaultWriteParam ( ) ; ( ( it . geosolutions . imageio . plugins . jp2kakadu . JP2GDALKakaduImageWriteParam ) ( param ) ) . setCmodes ( createVersions [ i ] ) ; pbjImageWrite . setParameter ( "BYPASS|RESTART|CAUSAL" 5 , param ) ; final javax . media . jai . RenderedOp op = javax . media . jai . JAI . create ( "BYPASS|RESTART|CAUSAL" 0 , pbjImageWrite ) ; } } file ( java . lang . Object , java . lang . String ) { final java . net . URL url = it . geosolutions . resources . TestData . url ( caller , path ) ; final java . io . File file = it . geosolutions . imageio . utilities . Utilities . urlToFile ( url ) ; if ( ! ( file . exists ( ) ) ) { throw new java . io . FileNotFoundException ( ( "Could<sp>not<sp>locate<sp>test-data:<sp>" + path ) ) ; } return file ; }
org . junit . Assert . assertTrue ( inputFile . exists ( ) )
testGetDoubleWithMaxValue ( ) { org . apache . activemq . command . ActiveMQMapMessage msg = new org . apache . activemq . command . ActiveMQMapMessage ( ) ; msg . setDouble ( this . name , Double . MAX_VALUE ) ; msg = ( ( org . apache . activemq . command . ActiveMQMapMessage ) ( msg . copy ( ) ) ) ; "<AssertPlaceHolder>" ; } getDouble ( java . lang . String ) { initializeReading ( ) ; java . lang . Object value = map . get ( name ) ; if ( value == null ) { return 0 ; } else if ( value instanceof java . lang . Double ) { return ( ( java . lang . Double ) ( value ) ) . doubleValue ( ) ; } else if ( value instanceof java . lang . Float ) { return ( ( java . lang . Float ) ( value ) ) . floatValue ( ) ; } else if ( value instanceof org . fusesource . hawtbuf . UTF8Buffer ) { return java . lang . Double . valueOf ( value . toString ( ) ) . doubleValue ( ) ; } else if ( value instanceof java . lang . String ) { return java . lang . Double . valueOf ( value . toString ( ) ) . doubleValue ( ) ; } else { throw new javax . jms . MessageFormatException ( ( "Cannot<sp>read<sp>a<sp>double<sp>from<sp>" + ( value . getClass ( ) . getName ( ) ) ) ) ; } }
org . junit . Assert . assertEquals ( Double . MAX_VALUE , msg . getDouble ( this . name ) , 1.0 )
dataConversionTest ( ) { org . kaaproject . kaa . common . dto . credentials . EndpointRegistrationDto endpointRegistrationDto = new org . kaaproject . kaa . common . dto . credentials . EndpointRegistrationDto ( "1" , "2" , "3" , 42 , "test" ) ; org . kaaproject . kaa . server . common . nosql . cassandra . dao . model . CassandraEndpointRegistration cassandraEndpointRegistration = new org . kaaproject . kaa . server . common . nosql . cassandra . dao . model . CassandraEndpointRegistration ( endpointRegistrationDto ) ; "<AssertPlaceHolder>" ; } toDto ( ) { java . util . List < org . kaaproject . kaa . common . dto . TopicDto > topicDtos = org . kaaproject . kaa . server . common . dao . model . sql . ModelUtils . getTopicDtos ( topicIds ) ; return new org . kaaproject . kaa . common . dto . TopicListEntryDto ( simpleHash , hash , topicDtos ) ; }
org . junit . Assert . assertEquals ( endpointRegistrationDto , cassandraEndpointRegistration . toDto ( ) )
equals_shouldReturnFalseIfGivenObjIsNull ( ) { org . openmrs . BaseOpenmrsObject o = new org . openmrs . BaseOpenmrsObjectTest . BaseOpenmrsObjectMock ( ) ; org . openmrs . BaseOpenmrsObject obj = null ; "<AssertPlaceHolder>" ; } equals ( java . lang . Object ) { if ( ! ( obj instanceof org . openmrs . ConceptReferenceTermMap ) ) { return false ; } org . openmrs . ConceptReferenceTermMap rhs = ( ( org . openmrs . ConceptReferenceTermMap ) ( obj ) ) ; if ( ( ( this . conceptReferenceTermMapId ) != null ) && ( ( rhs . conceptReferenceTermMapId ) != null ) ) { return this . conceptReferenceTermMapId . equals ( rhs . conceptReferenceTermMapId ) ; } return ( this ) == obj ; }
org . junit . Assert . assertFalse ( o . equals ( obj ) )
testToLong ( ) { byte [ ] bytes = org . apache . hadoop . hbase . util . Bytes . toBytes ( 123L ) ; "<AssertPlaceHolder>" ; } toLong ( byte [ ] ) { return 2L ; }
org . junit . Assert . assertEquals ( org . apache . hadoop . hbase . util . Bytes . toLong ( bytes ) , util . toLong ( bytes ) )
testAllGood ( ) { dataset . getLatestVersion ( ) . setVersionState ( DatasetVersion . VersionState . DRAFT ) ; edu . harvard . iq . dataverse . Dataset updatedDataset = null ; try { testEngine . submit ( new edu . harvard . iq . dataverse . engine . command . impl . AddLockCommand ( dataverseRequest , dataset , new edu . harvard . iq . dataverse . DatasetLock ( DatasetLock . Reason . InReview , dataverseRequest . getAuthenticatedUser ( ) ) ) ) ; updatedDataset = testEngine . submit ( new edu . harvard . iq . dataverse . engine . command . impl . ReturnDatasetToAuthorCommand ( dataverseRequest , dataset , "Update<sp>Your<sp>Files,<sp>Dummy" ) ) ; } catch ( edu . harvard . iq . dataverse . engine . command . exception . CommandException ex ) { System . out . println ( ( "Error<sp>updating<sp>dataset:<sp>" + ( ex . getMessage ( ) ) ) ) ; } "<AssertPlaceHolder>" ; } getMessage ( ) { return message ; }
org . junit . Assert . assertNotNull ( updatedDataset )
testParsingLogLocation ( ) { java . lang . String [ ] params = new java . lang . String [ ] { "--reporter" , org . datadog . jmxfetch . TestParsingJCommander . REPORTER_CONSOLE , "--check" , org . datadog . jmxfetch . TestParsingJCommander . SINGLE_CHECK , "--conf_directory" , org . datadog . jmxfetch . TestParsingJCommander . CONF_DIR , "--log_location" , org . datadog . jmxfetch . TestParsingJCommander . LOG_LOCATION , AppConfig . ACTION_COLLECT } ; org . datadog . jmxfetch . AppConfig appConfig = org . datadog . jmxfetch . TestParsingJCommander . testCommand ( params ) ; "<AssertPlaceHolder>" ; } getLogLocation ( ) { return logLocation ; }
org . junit . Assert . assertEquals ( org . datadog . jmxfetch . TestParsingJCommander . LOG_LOCATION , appConfig . getLogLocation ( ) )
createLinkedList_noArgs ( ) { java . util . LinkedList < java . lang . Integer > list = createLinkedList ( ) ; "<AssertPlaceHolder>" ; } createLinkedList ( ) { return new java . util . LinkedList < T > ( ) ; }
org . junit . Assert . assertTrue ( ( list instanceof java . util . LinkedList < ? > ) )
available_onPartiallyCachedStream ( ) { final java . lang . String testString = "helloWorld" ; final byte [ ] testData = testString . getBytes ( ) ; final java . io . InputStream is = new org . exist . util . io . FastByteArrayInputStream ( testData ) ; final org . exist . util . io . CachingFilterInputStream cfis = new org . exist . util . io . CachingFilterInputStream ( getNewCache ( is ) ) ; cfis . mark ( Integer . MAX_VALUE ) ; cfis . read ( ) ; cfis . read ( ) ; cfis . reset ( ) ; "<AssertPlaceHolder>" ; } available ( ) { return bis . available ( ) ; }
org . junit . Assert . assertEquals ( testData . length , cfis . available ( ) )
testRestrictedEndpointWhenRequestMethodIsNotGET ( ) { final io . undertow . server . HttpServerExchange exchange = makeHttpServerExchange ( "/containers/containerId" , "POST" ) ; final java . lang . String containerId = "containerId" ; final boolean isRestricted = restrictionPolicy . restrictedEndpoint ( exchange , containerId ) ; "<AssertPlaceHolder>" ; } restrictedEndpoint ( io . undertow . server . HttpServerExchange , java . lang . String ) { final java . lang . String relativePath = exchange . getRelativePath ( ) ; final boolean isReadOnlyRequest = isGet ( exchange . getRequestMethod ( ) ) ; final boolean isContainerResource = relativePath . endsWith ( ( "/containers/" + containerId ) ) ; final boolean isScannerResource = relativePath . endsWith ( "/scanner" ) ; final boolean isReleaseIdResource = relativePath . endsWith ( "/release-id" ) ; if ( ( ! isReadOnlyRequest ) && ( ( isContainerResource || isScannerResource ) || isReleaseIdResource ) ) { return true ; } return false ; }
org . junit . Assert . assertTrue ( isRestricted )
case1 ( ) { org . openstack . atlas . usagerefactor . List < org . openstack . atlas . service . domain . usage . entities . LoadBalancerMergedHostUsage > mergedRecords = usagePollerHelper . processExistingEvents ( lbHostMap ) ; "<AssertPlaceHolder>" ; org . openstack . atlas . usagerefactor . junit . AssertLoadBalancerMergedHostUsage . hasValues ( 1234 , 124 , 0L , 0L , 0L , 0L , 0 , 0 , 1 , 0 , UsageEvent . SSL_ONLY_ON , "2013-04-10<sp>20:03:01" , mergedRecords . get ( 0 ) ) ; org . openstack . atlas . usagerefactor . junit . AssertLoadBalancerMergedHostUsage . hasValues ( 1234 , 123 , 0L , 0L , 0L , 0L , 0 , 0 , 1 , 0 , UsageEvent . SSL_MIXED_ON , "2013-04-10<sp>20:03:00" , mergedRecords . get ( 1 ) ) ; } size ( ) { return size ; }
org . junit . Assert . assertEquals ( 2 , mergedRecords . size ( ) )
mapPut ( ) { java . util . Map < ru . lessons . lesson_10 . User , ru . lessons . lesson_10 . User > users = new java . util . HashMap < ru . lessons . lesson_10 . User , ru . lessons . lesson_10 . User > ( ) ; users . put ( new ru . lessons . lesson_10 . User ( "1" , "first" ) , new ru . lessons . lesson_10 . User ( "1" , "first" ) ) ; users . put ( new ru . lessons . lesson_10 . User ( "2" , "second" ) , new ru . lessons . lesson_10 . User ( "2" , "second" ) ) ; users . put ( new ru . lessons . lesson_10 . User ( "1" , "first" ) , new ru . lessons . lesson_10 . User ( "3" , "first" ) ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( 2 , users . size ( ) )
testSegmentWithLeftJoinsAndFilters ( ) { final java . lang . String query1 = "SELECT<sp>?e<sp>?c" 0 + ( ( ( ( ( ( ( ( ( "SELECT<sp>?e<sp>?c<sp>?l" + "{" ) + "<sp>Filter(?e<sp>=<sp><uri:s1>)<sp>" ) + "<sp>Filter(?c<sp>=<sp><uri:s2>)<sp>" ) + "<sp>?e<sp><uri:p1><sp><uri:o1>.<sp>" ) + "<sp>OPTIONAL<sp>{?e<sp><uri:p2><sp>?l}.<sp>" ) + "<sp>?c<sp><uri:p3><sp><uri:o3><sp>.<sp>" ) + "SELECT<sp>?e<sp>?c" 1 ) + "<sp>OPTIONAL<sp>{?e<sp><uri:p2><sp>?c<sp>}<sp>.<sp>" ) + "}" ) ; final java . lang . String query2 = "SELECT<sp>?e<sp>?c" 0 + ( ( ( ( ( ( "SELECT<sp>?e<sp>?c<sp>?l" + "{" ) + "<sp>Filter(?c<sp>=<sp><uri:s2>)<sp>" ) + "<sp>?e<sp><uri:p1><sp><uri:o1>.<sp>" ) + "<sp>OPTIONAL<sp>{?e<sp><uri:p2><sp>?l}.<sp>" ) + "<sp>?c<sp><uri:p3><sp><uri:o3><sp>.<sp>" ) + "}" ) ; final java . lang . String query3 = "SELECT<sp>?e<sp>?c" 0 + ( ( ( ( ( "SELECT<sp>?e<sp>?c" + "{" ) + "<sp>Filter(?e<sp>=<sp><uri:s1>)<sp>" ) + "SELECT<sp>?e<sp>?c" 1 ) + "<sp>OPTIONAL<sp>{?e<sp><uri:p2><sp>?c<sp>}<sp>.<sp>" ) + "}" ) ; final org . eclipse . rdf4j . query . parser . sparql . SPARQLParser parser = new org . eclipse . rdf4j . query . parser . sparql . SPARQLParser ( ) ; final org . eclipse . rdf4j . query . parser . ParsedQuery pq1 = parser . parseQuery ( query1 , null ) ; final org . eclipse . rdf4j . query . parser . ParsedQuery pq2 = parser . parseQuery ( query2 , null ) ; final org . eclipse . rdf4j . query . parser . ParsedQuery pq3 = parser . parseQuery ( query3 , null ) ; final org . eclipse . rdf4j . query . algebra . TupleExpr te1 = pq1 . getTupleExpr ( ) ; final org . eclipse . rdf4j . query . algebra . TupleExpr te2 = pq2 . getTupleExpr ( ) ; final org . eclipse . rdf4j . query . algebra . TupleExpr te3 = pq3 . getTupleExpr ( ) ; final org . eclipse . rdf4j . query . algebra . TupleExpr unOpt = te1 . clone ( ) ; final org . apache . rya . indexing . external . tupleSet . SimpleExternalTupleSet pcj1 = new org . apache . rya . indexing . external . tupleSet . SimpleExternalTupleSet ( ( ( org . eclipse . rdf4j . query . algebra . Projection ) ( te2 ) ) ) ; final org . apache . rya . indexing . external . tupleSet . SimpleExternalTupleSet pcj2 = new org . apache . rya . indexing . external . tupleSet . SimpleExternalTupleSet ( ( ( org . eclipse . rdf4j . query . algebra . Projection ) ( te3 ) ) ) ; final java . util . List < org . apache . rya . indexing . external . tupleSet . ExternalTupleSet > externalList = new java . util . ArrayList ( ) ; externalList . add ( pcj1 ) ; externalList . add ( pcj2 ) ; provider . setIndices ( externalList ) ; final org . apache . rya . indexing . pcj . matching . PCJOptimizer optimizer = new org . apache . rya . indexing . pcj . matching . PCJOptimizer ( externalList , false , provider ) ; optimizer . optimize ( te1 , null , null ) ; "<AssertPlaceHolder>" ; } validatePcj ( org . eclipse . rdf4j . query . algebra . TupleExpr , org . eclipse . rdf4j . query . algebra . TupleExpr , java . util . List , java . util . Set ) { final org . apache . rya . indexing . IndexPlanValidator . IndexedExecutionPlanGenerator iep = new org . apache . rya . indexing . IndexPlanValidator . IndexedExecutionPlanGenerator ( unOptTup , pcjs ) ; final java . util . List < org . apache . rya . indexing . external . tupleSet . ExternalTupleSet > indexList = iep . getNormalizedIndices ( ) ; final java . util . Set < org . eclipse . rdf4j . query . algebra . QueryModelNode > indexSet = new java . util . HashSet ( ) ; for ( final org . apache . rya . indexing . external . tupleSet . ExternalTupleSet etup : indexList ) { indexSet . add ( etup ) ; } final java . util . Set < org . eclipse . rdf4j . query . algebra . QueryModelNode > tupNodes = com . google . common . collect . Sets . newHashSet ( getNodes ( optTupleExp ) ) ; final java . util . Set < org . eclipse . rdf4j . query . algebra . QueryModelNode > diff = com . google . common . collect . Sets . difference ( tupNodes , indexSet ) ; return diff . equals ( expUnmatchedNodes ) ; }
org . junit . Assert . assertEquals ( true , validatePcj ( te1 , unOpt , externalList , new java . util . HashSet < org . eclipse . rdf4j . query . algebra . QueryModelNode > ( ) ) )
testServiceIsSame ( ) { cleancodematters . requestfactory . testutils . RequestFactoryHelperTest . TestService service1 = cleancodematters . requestfactory . testutils . RequestFactoryHelper . getService ( cleancodematters . requestfactory . testutils . RequestFactoryHelperTest . TestService . class ) ; cleancodematters . requestfactory . testutils . RequestFactoryHelperTest . TestService service2 = cleancodematters . requestfactory . testutils . RequestFactoryHelper . getService ( cleancodematters . requestfactory . testutils . RequestFactoryHelperTest . TestService . class ) ; "<AssertPlaceHolder>" ; } getService ( java . lang . Class ) { T result = ( ( T ) ( cleancodematters . requestfactory . testutils . RequestFactoryHelper . serviceLocator . getInstance ( serviceClass ) ) ) ; reset ( result ) ; return result ; }
org . junit . Assert . assertSame ( service1 , service2 )
testGetCurrentHarvestsList ( ) { javax . ws . rs . client . WebTarget target = target ( ( ( ( "/" + ( org . theeuropeanlibrary . repox . rest . pathOptions . DatasetOptionListContainer . DATASETS ) ) + "/" ) + ( org . theeuropeanlibrary . repox . rest . pathOptions . HarvestOptionListContainer . HARVESTS ) ) ) ; when ( taskManager . getRunningTasks ( ) ) . thenReturn ( new java . util . ArrayList < pt . utl . ist . task . Task > ( ) ) ; javax . ws . rs . core . Response response = target . request ( MediaType . APPLICATION_JSON ) . get ( ) ; "<AssertPlaceHolder>" ; } getStatus ( ) { return ( ( java . lang . String ) ( get ( "status" ) ) ) ; }
org . junit . Assert . assertEquals ( 200 , response . getStatus ( ) )
testSaveItem_XmlFileDuplicate ( ) { sutWithoutUserAndDuplicateHistory . saveItem ( test1Config ) ; "<AssertPlaceHolder>" ; } getHistoryLength ( ) { return test1History . list ( ) . length ; }
org . junit . Assert . assertEquals ( 6 , getHistoryLength ( ) )
testReloadRefresh ( ) { com . cloudera . flume . master . ConfigurationManager cfgman2 = flumeMaster . getSpecMan ( ) ; java . util . Map < java . lang . String , com . cloudera . flume . conf . FlumeConfigData > cfgs2 = cfgman2 . getTranslatedConfigs ( ) ; "<AssertPlaceHolder>" ; } size ( ) { return hist . size ( ) ; }
org . junit . Assert . assertEquals ( 5 , cfgs2 . size ( ) )
testEidFallthrough ( ) { org . mockito . Mockito . when ( uds . getUserByAid ( org . mockito . Mockito . anyString ( ) ) ) . thenThrow ( org . sakaiproject . user . api . UserNotDefinedException . class ) ; org . mockito . Mockito . when ( uds . getUser ( "1234" ) ) . thenReturn ( user ) ; org . mockito . Mockito . when ( dhs . getConfigurationSetting ( "user.explicit.id.only" , false ) ) . thenReturn ( false ) ; "<AssertPlaceHolder>" ; } getUserByIdEid ( java . lang . String ) { org . sakaiproject . user . api . User user = null ; if ( id != null ) { boolean doCheckForId = false ; boolean doCheckForAid = true ; java . lang . String userId = id ; if ( ( ( userId . length ( ) ) > ( org . sakaiproject . entitybroker . providers . UserEntityProvider . ID_PREFIX . length ( ) ) ) && ( userId . startsWith ( org . sakaiproject . entitybroker . providers . UserEntityProvider . ID_PREFIX ) ) ) { userId = id . substring ( org . sakaiproject . entitybroker . providers . UserEntityProvider . ID_PREFIX . length ( ) ) ; doCheckForAid = false ; doCheckForId = true ; } if ( doCheckForAid ) { try { user = userDirectoryService . getUserByAid ( id ) ; } catch ( org . sakaiproject . user . api . UserNotDefinedException e ) { user = null ; if ( ! ( userIdExplicitOnly ( ) ) ) { doCheckForId = true ; } } } if ( doCheckForId ) { try { user = userDirectoryService . getUser ( userId ) ; } catch ( org . sakaiproject . user . api . UserNotDefinedException e ) { user = null ; } } } return user ; }
org . junit . Assert . assertEquals ( user , provider . getUserByIdEid ( "1234" ) )
testShutdownFailWithoutLog ( ) { java . rmi . server . UnicastRemoteObject . unexportObject ( com . liferay . portal . kernel . resiliency . mpi . MPIHelperUtilTest . _getMPIImpl ( ) , true ) ; final java . io . IOException ioException = new java . io . IOException ( ) ; com . liferay . portal . kernel . test . ReflectionTestUtil . setFieldValue ( com . liferay . portal . kernel . resiliency . mpi . MPIHelperUtil . class , "_intraband" , new com . liferay . portal . kernel . nio . intraband . test . MockIntraband ( ) { @ com . liferay . portal . kernel . resiliency . mpi . Override public void close ( ) throws java . io . IOException { throw ioException ; } } ) ; try ( com . liferay . portal . kernel . test . CaptureHandler captureHandler = com . liferay . portal . kernel . test . JDKLoggerTestUtil . configureJDKLogger ( com . liferay . portal . kernel . resiliency . mpi . MPIHelperUtil . class . getName ( ) , Level . OFF ) ) { com . liferay . portal . kernel . resiliency . mpi . MPIHelperUtil . shutdown ( ) ; java . util . List < java . util . logging . LogRecord > logRecords = captureHandler . getLogRecords ( ) ; "<AssertPlaceHolder>" ; } } toString ( ) { com . liferay . petra . string . StringBundler sb = new com . liferay . petra . string . StringBundler ( 23 ) ; sb . append ( ",<sp>width=" 1 ) ; sb . append ( uuid ) ; sb . append ( ",<sp>width=" 0 ) ; sb . append ( amImageEntryId ) ; sb . append ( ",<sp>groupId=" ) ; sb . append ( groupId ) ; sb . append ( ",<sp>companyId=" ) ; sb . append ( companyId ) ; sb . append ( ",<sp>createDate=" ) ; sb . append ( createDate ) ; sb . append ( ",<sp>configurationUuid=" ) ; sb . append ( configurationUuid ) ; sb . append ( ",<sp>fileVersionId=" ) ; sb . append ( fileVersionId ) ; sb . append ( ",<sp>mimeType=" ) ; sb . append ( mimeType ) ; sb . append ( ",<sp>height=" ) ; sb . append ( height ) ; sb . append ( ",<sp>width=" ) ; sb . append ( width ) ; sb . append ( ",<sp>size=" ) ; sb . append ( size ) ; sb . append ( "}" ) ; return sb . toString ( ) ; }
org . junit . Assert . assertTrue ( logRecords . toString ( ) , logRecords . isEmpty ( ) )
testMapToKey ( ) { one . util . streamex . Map < java . lang . String , java . lang . Integer > expected = new one . util . streamex . HashMap ( ) ; expected . put ( "a:1" , 1 ) ; expected . put ( "bb:22" , 22 ) ; expected . put ( "ccc:33" , 33 ) ; one . util . streamex . Map < java . lang . String , java . lang . Integer > result = one . util . streamex . EntryStream . of ( one . util . streamex . EntryStreamTest . createMap ( ) ) . mapToKey ( ( str , num ) -> ( str + ":" ) + num ) . toMap ( ) ; "<AssertPlaceHolder>" ; } toMap ( ) { one . util . streamex . Map < K , V > map = ( one . util . streamex . EntryStream . isParallel ( ) ) ? new java . util . concurrent . ConcurrentHashMap ( ) : new one . util . streamex . HashMap ( ) ; forEach ( one . util . streamex . EntryStream . toMapConsumer ( map ) ) ; return map ; }
org . junit . Assert . assertEquals ( expected , result )
givenAllowBlankPathDirective_whenIsAllowed_thenReturnsTrue ( ) { parse ( "blank_allow_robots.txt" ) ; boolean allowed = service . isAllowed ( agent , java . net . URI . create ( "http://example.com/index.html" ) ) ; "<AssertPlaceHolder>" ; } isAllowed ( java . lang . String , java . net . URI ) { checkNotNull ( crawlerAgentString , "crawlerAgentString<sp>is<sp>null" ) ; checkNotNull ( resourceUri , "resourceUri<sp>is<sp>null" ) ; com . brandwatch . robots . RobotsServiceImpl . log . debug ( "Resolving<sp>robots<sp>URL<sp>for:<sp>{}" , resourceUri ) ; final java . net . URI robotsUri = utilities . getRobotsURIForResource ( resourceUri ) ; com . brandwatch . robots . RobotsServiceImpl . log . debug ( "Resolved<sp>robots<sp>URI<sp>to:<sp>{}" , robotsUri ) ; final com . brandwatch . robots . domain . Robots robots ; try { robots = loader . load ( robotsUri ) ; } catch ( java . lang . Exception e ) { com . brandwatch . robots . RobotsServiceImpl . log . debug ( "Download<sp>failure<sp>{}" , e . getMessage ( ) ) ; return allow ( resourceUri ) ; } if ( robots . getGroups ( ) . isEmpty ( ) ) { com . brandwatch . robots . RobotsServiceImpl . log . debug ( "No<sp>agent<sp>groups<sp>found" , resourceUri ) ; return allow ( resourceUri ) ; } final com . google . common . base . Optional < com . brandwatch . robots . domain . Group > group = matcherUtils . getMostSpecificMatchingGroup ( robots . getGroups ( ) , crawlerAgentString ) ; if ( group . isPresent ( ) ) { if ( com . brandwatch . robots . RobotsServiceImpl . log . isDebugEnabled ( ) ) { com . brandwatch . robots . RobotsServiceImpl . log . debug ( "Matched<sp>user-agent<sp>group:<sp>{}" , matcherUtils . getMostSpecificMatch ( group . get ( ) , crawlerAgentString ) . get ( ) . getValue ( ) ) ; } } else { com . brandwatch . robots . RobotsServiceImpl . log . debug ( "No<sp>user-agent<sp>group<sp>matched" ) ; return allow ( resourceUri ) ; } final com . google . common . base . Optional < com . brandwatch . robots . domain . PathDirective > bestMatch = matcherUtils . getMostSpecificMatch ( group . get ( ) . getDirectives ( com . brandwatch . robots . domain . PathDirective . class ) , utilities . getResourceLocalComponents ( resourceUri ) ) ; if ( ! ( bestMatch . isPresent ( ) ) ) { com . brandwatch . robots . RobotsServiceImpl . log . debug ( "No<sp>matching<sp>path<sp>directive" ) ; return allow ( resourceUri ) ; } else { final com . brandwatch . robots . domain . PathDirective directive = bestMatch . get ( ) ; com . brandwatch . robots . RobotsServiceImpl . log . debug ( "Matched<sp>path<sp>directive<sp>{}:{}" , directive . getField ( ) , directive . getValue ( ) ) ; return directive . isAllowed ( ) ? allow ( resourceUri ) : disallow ( resourceUri ) ; } }
org . junit . Assert . assertThat ( allowed , org . hamcrest . core . Is . is ( true ) )
testEmptyPermissionsResultsInZeroBits ( ) { java . util . EnumSet < com . microsoft . windowsazure . services . media . models . AccessPolicyPermission > perms = java . util . EnumSet . noneOf ( com . microsoft . windowsazure . services . media . models . AccessPolicyPermission . class ) ; int bits = com . microsoft . windowsazure . services . media . models . AccessPolicyPermission . bitsFromPermissions ( perms ) ; "<AssertPlaceHolder>" ; } bitsFromPermissions ( java . util . EnumSet ) { int result = 0 ; for ( com . microsoft . windowsazure . services . media . models . AccessPolicyPermission p : perms ) { result |= p . getFlagValue ( ) ; } return result ; }
org . junit . Assert . assertEquals ( 0 , bits )
whenCallingForANewBuilderInstance_itShouldReturnBuilderCorrectly ( ) { com . graphhopper . jsprit . core . problem . job . Shipment . Builder builder = Shipment . Builder . newInstance ( "s" ) ; "<AssertPlaceHolder>" ; } newInstance ( java . lang . String ) { return new com . graphhopper . jsprit . core . problem . job . Shipment . Builder ( id ) ; }
org . junit . Assert . assertNotNull ( builder )
testQueueAppMetricsForMultipleFailures ( ) { java . lang . String queueName = "single" ; org . apache . hadoop . yarn . server . resourcemanager . scheduler . QueueMetrics metrics = org . apache . hadoop . yarn . server . resourcemanager . scheduler . QueueMetrics . forQueue ( ms , queueName , null , false , new org . apache . hadoop . conf . Configuration ( ) ) ; org . apache . hadoop . metrics2 . MetricsSource queueSource = org . apache . hadoop . yarn . server . resourcemanager . scheduler . TestQueueMetrics . queueSource ( ms , queueName ) ; org . apache . hadoop . yarn . server . resourcemanager . scheduler . AppSchedulingInfo app = org . apache . hadoop . yarn . server . resourcemanager . scheduler . TestQueueMetrics . mockApp ( org . apache . hadoop . yarn . server . resourcemanager . scheduler . TestQueueMetrics . USER ) ; metrics . submitApp ( org . apache . hadoop . yarn . server . resourcemanager . scheduler . TestQueueMetrics . USER ) ; org . apache . hadoop . metrics2 . MetricsSource userSource = org . apache . hadoop . yarn . server . resourcemanager . scheduler . TestQueueMetrics . userSource ( ms , queueName , org . apache . hadoop . yarn . server . resourcemanager . scheduler . TestQueueMetrics . USER ) ; org . apache . hadoop . yarn . server . resourcemanager . scheduler . AppMetricsChecker appMetricsChecker = org . apache . hadoop . yarn . server . resourcemanager . scheduler . AppMetricsChecker . create ( ) . counter ( AppMetricsKey . APPS_SUBMITTED , 1 ) . checkAgainst ( queueSource , true ) ; metrics . submitAppAttempt ( org . apache . hadoop . yarn . server . resourcemanager . scheduler . TestQueueMetrics . USER ) ; appMetricsChecker = org . apache . hadoop . yarn . server . resourcemanager . scheduler . AppMetricsChecker . createFromChecker ( appMetricsChecker ) . gaugeInt ( AppMetricsKey . APPS_PENDING , 1 ) . checkAgainst ( queueSource , true ) ; metrics . runAppAttempt ( app . getApplicationId ( ) , org . apache . hadoop . yarn . server . resourcemanager . scheduler . TestQueueMetrics . USER ) ; appMetricsChecker = org . apache . hadoop . yarn . server . resourcemanager . scheduler . AppMetricsChecker . createFromChecker ( appMetricsChecker ) . gaugeInt ( AppMetricsKey . APPS_PENDING , 0 ) . gaugeInt ( AppMetricsKey . APPS_RUNNING , 1 ) . checkAgainst ( queueSource , true ) ; metrics . finishAppAttempt ( app . getApplicationId ( ) , app . isPending ( ) , app . getUser ( ) ) ; appMetricsChecker = org . apache . hadoop . yarn . server . resourcemanager . scheduler . AppMetricsChecker . createFromChecker ( appMetricsChecker ) . gaugeInt ( AppMetricsKey . APPS_RUNNING , 0 ) . checkAgainst ( queueSource , true ) ; metrics . submitAppAttempt ( org . apache . hadoop . yarn . server . resourcemanager . scheduler . TestQueueMetrics . USER ) ; appMetricsChecker = org . apache . hadoop . yarn . server . resourcemanager . scheduler . AppMetricsChecker . createFromChecker ( appMetricsChecker ) . gaugeInt ( AppMetricsKey . APPS_PENDING , 1 ) . checkAgainst ( queueSource , true ) ; metrics . runAppAttempt ( app . getApplicationId ( ) , org . apache . hadoop . yarn . server . resourcemanager . scheduler . TestQueueMetrics . USER ) ; appMetricsChecker = org . apache . hadoop . yarn . server . resourcemanager . scheduler . AppMetricsChecker . createFromChecker ( appMetricsChecker ) . gaugeInt ( AppMetricsKey . APPS_PENDING , 0 ) . gaugeInt ( AppMetricsKey . APPS_RUNNING , 1 ) . checkAgainst ( queueSource , true ) ; metrics . finishAppAttempt ( app . getApplicationId ( ) , app . isPending ( ) , app . getUser ( ) ) ; appMetricsChecker = org . apache . hadoop . yarn . server . resourcemanager . scheduler . AppMetricsChecker . createFromChecker ( appMetricsChecker ) . gaugeInt ( AppMetricsKey . APPS_RUNNING , 0 ) . checkAgainst ( queueSource , true ) ; metrics . submitAppAttempt ( org . apache . hadoop . yarn . server . resourcemanager . scheduler . TestQueueMetrics . USER ) ; appMetricsChecker = org . apache . hadoop . yarn . server . resourcemanager . scheduler . AppMetricsChecker . createFromChecker ( appMetricsChecker ) . gaugeInt ( AppMetricsKey . APPS_PENDING , 1 ) . checkAgainst ( queueSource , true ) ; metrics . runAppAttempt ( app . getApplicationId ( ) , org . apache . hadoop . yarn . server . resourcemanager . scheduler . TestQueueMetrics . USER ) ; appMetricsChecker = org . apache . hadoop . yarn . server . resourcemanager . scheduler . AppMetricsChecker . createFromChecker ( appMetricsChecker ) . gaugeInt ( AppMetricsKey . APPS_PENDING , 0 ) . gaugeInt ( AppMetricsKey . APPS_RUNNING , 1 ) . checkAgainst ( queueSource , true ) ; metrics . finishAppAttempt ( app . getApplicationId ( ) , app . isPending ( ) , app . getUser ( ) ) ; appMetricsChecker = org . apache . hadoop . yarn . server . resourcemanager . scheduler . AppMetricsChecker . createFromChecker ( appMetricsChecker ) . gaugeInt ( AppMetricsKey . APPS_RUNNING , 0 ) . checkAgainst ( queueSource , true ) ; metrics . finishApp ( org . apache . hadoop . yarn . server . resourcemanager . scheduler . TestQueueMetrics . USER , RMAppState . FAILED ) ; org . apache . hadoop . yarn . server . resourcemanager . scheduler . AppMetricsChecker . createFromChecker ( appMetricsChecker ) . gaugeInt ( AppMetricsKey . APPS_RUNNING , 0 ) . counter ( AppMetricsKey . APPS_FAILED , 1 ) . checkAgainst ( queueSource , true ) ; "<AssertPlaceHolder>" ; } checkAgainst ( org . apache . hadoop . metrics2 . MetricsSource , boolean ) { if ( source == null ) { throw new java . lang . IllegalStateException ( "MetricsSource<sp>should<sp>not<sp>be<sp>null!" ) ; } org
org . junit . Assert . assertNull ( userSource )
testWorkingFrom ( ) { org . drools . workbench . models . datamodel . rule . FactPattern boundPattern = new org . drools . workbench . models . datamodel . rule . FactPattern ( "Person" ) ; boundPattern . setBoundName ( "person" ) ; boundPattern . addConstraint ( new org . drools . workbench . models . datamodel . rule . SingleFieldConstraint ( "addresses" ) ) ; org . drools . workbench . models . datamodel . rule . FactPattern pattern = new org . drools . workbench . models . datamodel . rule . FactPattern ( "Address" ) ; org . drools . workbench . models . datamodel . rule . SingleFieldConstraint constraint = new org . drools . workbench . models . datamodel . rule . SingleFieldConstraint ( "street" ) ; pattern . addConstraint ( constraint ) ; org . drools . workbench . models . datamodel . rule . FromCompositeFactPattern fromCompositeFactPattern = new org . drools . workbench . models . datamodel . rule . FromCompositeFactPattern ( ) ; fromCompositeFactPattern . setFactPattern ( pattern ) ; org . drools . workbench . models . datamodel . rule . ExpressionFormLine expression = new org . drools . workbench . models . datamodel . rule . ExpressionFormLine ( ) ; expression . setBinding ( "person.addresses" ) ; fromCompositeFactPattern . setExpression ( expression ) ; model . lhs = new org . drools . workbench . models . datamodel . rule . IPattern [ ] { boundPattern , fromCompositeFactPattern } ; "<AssertPlaceHolder>" ; } isValid ( ) { if ( ( model . lhs . length ) == 0 ) { return true ; } else { validateIPatterns ( model . lhs ) ; } return errors . isEmpty ( ) ; }
org . junit . Assert . assertTrue ( validator . isValid ( ) )
testGetPrivilegeNamesFromEmpty ( ) { java . util . Set < java . lang . String > names = bitsProvider . getPrivilegeNames ( PrivilegeBits . EMPTY ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return ( size ) == 0 ; }
org . junit . Assert . assertTrue ( names . isEmpty ( ) )
testEqualsDiffOneParameter ( ) { "<AssertPlaceHolder>" ; } }
org . junit . Assert . assertThat ( parameter , result , org . hamcrest . CoreMatchers . is ( false ) )
doTrimAll_A$String_notNull ( ) { org . junithelper . core . filter . TrimFilterManager target = new org . junithelper . core . filter . TrimFilterManager ( ) ; target . addFilter ( new org . junithelper . core . filter . impl . TrimCommentFilter ( ) , new org . junithelper . core . filter . impl . TrimInsideOfBraceFilter ( ) , new org . junithelper . core . filter . impl . TrimQuotationFilter ( ) ) ; java . lang . String src = "package<sp>foo.var;<sp>\r\n<sp>public<sp>class<sp>Sample<sp>{<sp>\r\n//<sp>hogehoge<sp>\r\n<sp>public<sp>void<sp>hoge()<sp>{<sp>System.out.println(\"aaa\");<sp>}<sp>}" ; java . lang . String actual = target . doTrimAll ( src ) ; java . lang . String expected = "package<sp>foo.var;<sp>public<sp>class<sp>Sample<sp>{<sp>public<sp>void<sp>hoge()<sp>{}<sp>}" ; "<AssertPlaceHolder>" ; } doTrimAll ( java . lang . String ) { java . lang . String dest = src ; for ( org . junithelper . core . filter . TrimFilter filter : filters ) { dest = filter . trimAll ( dest ) ; } return dest ; }
org . junit . Assert . assertEquals ( expected , actual )
testGetHasTypeRefs ( ) { final java . util . List < org . kie . workbench . common . dmn . api . definition . HasTypeRef > actualHasTypeRefs = outputClauseLiteralExpression . getHasTypeRefs ( ) ; final java . util . List < org . kie . workbench . common . dmn . api . definition . HasTypeRef > expectedHasTypeRefs = singletonList ( outputClauseLiteralExpression ) ; "<AssertPlaceHolder>" ; } getHasTypeRefs ( ) { final java . util . List < org . kie . workbench . common . dmn . api . definition . HasTypeRef > hasTypeRefs = super . getHasTypeRefs ( ) ; hasTypeRefs . addAll ( getFlatHasTypeRefs ( getColumn ( ) ) ) ; hasTypeRefs . addAll ( getFlatHasTypeRefs ( getRow ( ) ) ) ; return hasTypeRefs ; }
org . junit . Assert . assertEquals ( expectedHasTypeRefs , actualHasTypeRefs )
t04SendPowerPinSwitch ( ) { org . zu . ardulink . protocol . MessageInfo info = link . sendPowerPinSwitch ( 0 , IProtocol . HIGH , new org . zu . ardulink . protocol . LoggerReplyMessageCallback ( ) ) ; isLastTest = true ; "<AssertPlaceHolder>" ; } isSent ( ) { return sent ; }
org . junit . Assert . assertTrue ( info . isSent ( ) )
test_NamedDomain_MultiLocation ( ) { localRuntime = new org . apache . cayenne . configuration . server . ServerRuntimeBuilder ( "myd" ) . addConfigs ( CayenneProjects . TESTMAP_PROJECT , CayenneProjects . EMBEDDABLE_PROJECT ) . build ( ) ; "<AssertPlaceHolder>" ; } getDataDomain ( ) { return injector . getInstance ( org . apache . cayenne . access . DataDomain . class ) ; }
org . junit . Assert . assertEquals ( "myd" , localRuntime . getDataDomain ( ) . getName ( ) )
testNoRme ( ) { final org . pac4j . http . authorization . generator . RememberMeAuthorizationGenerator ag = new org . pac4j . http . authorization . generator . RememberMeAuthorizationGenerator ( ) ; ag . generate ( org . pac4j . core . context . MockWebContext . create ( ) , profile ) ; "<AssertPlaceHolder>" ; } isRemembered ( ) { return this . isRemembered ; }
org . junit . Assert . assertFalse ( profile . isRemembered ( ) )
getEntries ( ) { int [ ] keys = new int [ ] { 1 , 18 , 5 } ; java . lang . String [ ] values = new java . lang . String [ ] { "John" , "Jim" , "Jane" } ; java . util . Map < java . lang . Integer , java . lang . String > map = app . buildMap ( keys , values ) ; java . util . Set < java . util . Map . Entry < java . lang . Integer , java . lang . String > > entrySet = app . getEntries ( map ) ; "<AssertPlaceHolder>" ; } getEntries ( io . robusta . java . classic . Map ) { return null ; }
org . junit . Assert . assertTrue ( ( ( entrySet . size ( ) ) == 3 ) )
testCALLDATACOPY_7 ( ) { org . ethereum . vm . VM vm = new org . ethereum . vm . VM ( ) ; program = new org . ethereum . vm . Program ( org . spongycastle . util . encoders . Hex . decode ( "6020600073CC0929EB16730E7C14FEFC63006AC2D794C5795637" ) , invoke ) ; try { vm . step ( program ) ; vm . step ( program ) ; vm . step ( program ) ; vm . step ( program ) ; } finally { "<AssertPlaceHolder>" ; } } isStopped ( ) { return stopped ; }
org . junit . Assert . assertTrue ( program . isStopped ( ) )
unlockExclusiveAndTakeWriteLockMustInvalidateOptimisticReadLocks ( ) { pageList . unlockExclusiveAndTakeWriteLock ( pageRef ) ; long r = pageList . tryOptimisticReadLock ( pageRef ) ; "<AssertPlaceHolder>" ; } validateReadLock ( long , long ) { org . neo4j . unsafe . impl . internal . dragons . UnsafeUtil . loadFence ( ) ; return ( ( org . neo4j . io . pagecache . impl . muninn . OffHeapPageLock . getState ( address ) ) & ( org . neo4j . io . pagecache . impl . muninn . OffHeapPageLock . CHK_MASK ) ) == stamp ; }
org . junit . Assert . assertFalse ( pageList . validateReadLock ( pageRef , r ) )
testCreateQR2 ( ) { com . belerweb . social . bean . Result < com . belerweb . social . weixin . bean . QRTicket > result = weixin . createQR ( QRType . QR_LIMIT_SCENE , 0 ) ; "<AssertPlaceHolder>" ; com . belerweb . social . weixin . api . WeixinTest . logger . info ( result . getResult ( ) . getJsonObject ( ) . toString ( ) ) ; com . belerweb . social . weixin . api . WeixinTest . logger . info ( result . getResult ( ) . getQRUrl ( ) ) ; } success ( ) { return ( error ) == null ; }
org . junit . Assert . assertTrue ( result . success ( ) )
testMinKeyStrategy ( ) { io . datakernel . aggregation . ot . AggregationStructure structure = io . datakernel . aggregation . ot . AggregationStructure . create ( io . datakernel . aggregation . ChunkIdCodec . ofLong ( ) ) . withKey ( "key" , ofInt ( ) ) ; io . datakernel . aggregation . AggregationState state = new io . datakernel . aggregation . AggregationState ( structure ) ; io . datakernel . aggregation . Set < io . datakernel . aggregation . AggregationChunk > chunks1 = new io . datakernel . aggregation . HashSet ( ) ; chunks1 . add ( io . datakernel . aggregation . ConsolidationChunkSelectionTest . createTestChunk ( 1 , 1 , 2 ) ) ; chunks1 . add ( io . datakernel . aggregation . ConsolidationChunkSelectionTest . createTestChunk ( 2 , 1 , 2 ) ) ; chunks1 . add ( io . datakernel . aggregation . ConsolidationChunkSelectionTest . createTestChunk ( 3 , 1 , 4 ) ) ; chunks1 . add ( io . datakernel . aggregation . ConsolidationChunkSelectionTest . createTestChunk ( 4 , 3 , 4 ) ) ; io . datakernel . aggregation . Set < io . datakernel . aggregation . AggregationChunk > chunks2 = new io . datakernel . aggregation . HashSet ( ) ; chunks2 . add ( io . datakernel . aggregation . ConsolidationChunkSelectionTest . createTestChunk ( 9 , 9 , 10 ) ) ; chunks2 . add ( io . datakernel . aggregation . ConsolidationChunkSelectionTest . createTestChunk ( 10 , 9 , 10 ) ) ; chunks2 . add ( io . datakernel . aggregation . ConsolidationChunkSelectionTest . createTestChunk ( 11 , 10 , 11 ) ) ; chunks2 . add ( io . datakernel . aggregation . ConsolidationChunkSelectionTest . createTestChunk ( 12 , 10 , 13 ) ) ; chunks2 . add ( io . datakernel . aggregation . ConsolidationChunkSelectionTest . createTestChunk ( 13 , 12 , 13 ) ) ; state . apply ( io . datakernel . aggregation . ot . AggregationDiff . of ( concat ( chunks1 . stream ( ) , chunks2 . stream ( ) ) . collect ( toSet ( ) ) ) ) ; io . datakernel . aggregation . List < io . datakernel . aggregation . AggregationChunk > selectedChunks = state . findChunksForConsolidationMinKey ( 100 , 4000 ) ; "<AssertPlaceHolder>" ; } findChunksForConsolidationMinKey ( int , int ) { int partitioningKeyLength = aggregation . getPartitioningKey ( ) . size ( ) ; io . datakernel . aggregation . SortedMap < io . datakernel . aggregation . PrimaryKey , io . datakernel . aggregation . RangeTree < io . datakernel . aggregation . PrimaryKey , io . datakernel . aggregation . AggregationChunk > > partitioningKeyToTree = groupByPartition ( partitioningKeyLength ) ; if ( partitioningKeyToTree == null ) { io . datakernel . aggregation . List < io . datakernel . aggregation . AggregationChunk > chunks = findChunksForPartitioning ( partitioningKeyLength , maxChunks ) ; io . datakernel . aggregation . AggregationState . logChunksAndStrategy ( chunks , io . datakernel . aggregation . AggregationState . PickingStrategy . PARTITIONING ) ; return chunks ; } io . datakernel . aggregation . AggregationState . PickedChunks pickedChunks = io . datakernel . aggregation . AggregationState . findChunksWithMinKeyOrSizeFixStrategy ( partitioningKeyToTree , maxChunks , optimalChunkSize ) ; return io . datakernel . aggregation . AggregationState . processSelection ( pickedChunks . chunks , maxChunks , pickedChunks . partitionTree , pickedChunks . strategy ) ; }
org . junit . Assert . assertEquals ( chunks1 , new io . datakernel . aggregation . HashSet ( selectedChunks ) )
testReadObjectClassWithNullContextName ( ) { java . lang . Class < ? > clazz = getClass ( ) ; java . lang . String className = clazz . getName ( ) ; java . lang . String contextName = com . liferay . petra . string . StringPool . NULL ; java . nio . ByteBuffer byteBuffer = java . nio . ByteBuffer . allocate ( ( ( ( className . length ( ) ) + ( contextName . length ( ) ) ) + 11 ) ) ; byteBuffer . put ( SerializationConstants . TC_CLASS ) ; byteBuffer . put ( ( ( byte ) ( 1 ) ) ) ; byteBuffer . putInt ( contextName . length ( ) ) ; byteBuffer . put ( contextName . getBytes ( StringPool . UTF8 ) ) ; byteBuffer . put ( ( ( byte ) ( 1 ) ) ) ; byteBuffer . putInt ( className . length ( ) ) ; byteBuffer . put ( className . getBytes ( StringPool . UTF8 ) ) ; byteBuffer . flip ( ) ; com . liferay . petra . io . Deserializer deserializer = new com . liferay . petra . io . Deserializer ( byteBuffer ) ; java . lang . Class < ? > readClass = deserializer . readObject ( ) ; "<AssertPlaceHolder>" ; } readObject ( ) { byte tcByte = _buffer [ ( ( _index ) ++ ) ] ; if ( tcByte == ( SerializationConstants . TC_BOOLEAN ) ) { return ( ( T ) ( java . lang . Boolean . valueOf ( readBoolean ( ) ) ) ) ; } else if ( tcByte == ( SerializationConstants . TC_BYTE ) ) { return ( ( T ) ( java . lang . Byte . valueOf ( readByte ( ) ) ) ) ; } else if ( tcByte == ( SerializationConstants . TC_CHARACTER ) ) { return ( ( T ) ( java . lang . Character . valueOf ( readChar ( ) ) ) ) ; } else if ( tcByte == ( SerializationConstants . TC_CLASS ) ) { java . lang . String contextName = readString ( ) ; java . lang . String className = readString ( ) ; java . lang . ClassLoader classLoader = com . liferay . petra . lang . ClassLoaderPool . getClassLoader ( contextName ) ; return ( ( T ) ( com . liferay . petra . lang . ClassResolverUtil . resolve ( className , classLoader ) ) ) ; } else if ( tcByte == ( SerializationConstants . TC_DOUBLE ) ) { return ( ( T ) ( java . lang . Double . valueOf ( readDouble ( ) ) ) ) ; } else if ( tcByte == ( SerializationConstants . TC_FLOAT ) ) { return ( ( T ) ( java . lang . Float . valueOf ( readFloat ( ) ) ) ) ; } else if ( tcByte == ( SerializationConstants . TC_INTEGER ) ) { return ( ( T ) ( java . lang . Integer . valueOf ( readInt ( ) ) ) ) ; } else if ( tcByte == ( SerializationConstants . TC_LONG ) ) { return ( ( T ) ( java . lang . Long . valueOf ( readLong ( ) ) ) ) ; } else if ( tcByte == ( SerializationConstants . TC_NULL ) ) { return null ; } else if ( tcByte == ( SerializationConstants . TC_SHORT ) ) { return ( ( T ) ( java . lang . Short . valueOf ( readShort ( ) ) ) ) ; } else if ( tcByte == ( SerializationConstants . TC_STRING ) ) { return ( ( T ) ( readString ( ) ) ) ; } else if ( tcByte == ( SerializationConstants . TC_OBJECT ) ) { try { java . io . ObjectInputStream objectInputStream = new com . liferay . petra . io . ProtectedAnnotatedObjectInputStream ( new com . liferay . petra . io . Deserializer . BufferInputStream ( ) ) ; return ( ( T ) ( objectInputStream . readObject ( ) ) ) ; } catch ( java . io . IOException ioe ) { throw new java . lang . RuntimeException ( ioe ) ; } } throw new java . lang . IllegalStateException ( ( "Unkown<sp>TC<sp>code<sp>" + tcByte ) ) ; }
org . junit . Assert . assertSame ( clazz , readClass )
getName ( ) { "<AssertPlaceHolder>" ; } getName ( ) { org . junit . Assert . assertEquals ( "view" , this . mocker . getComponentUnderTest ( ) . getName ( ) ) ; }
org . junit . Assert . assertEquals ( "view" , this . mocker . getComponentUnderTest ( ) . getName ( ) )
givenUsingGuava_whenConvertingAnInputStreamToAString_thenCorrect ( ) { final java . lang . String originalString = org . apache . commons . lang3 . RandomStringUtils . randomAlphabetic ( org . baeldung . java . io . JavaInputStreamToXUnitTest . DEFAULT_SIZE ) ; final org . baeldung . java . io . InputStream inputStream = new org . baeldung . java . io . ByteArrayInputStream ( originalString . getBytes ( ) ) ; final com . google . common . io . ByteSource byteSource = new com . google . common . io . ByteSource ( ) { @ org . baeldung . java . io . Override public final java . io . InputStream openStream ( ) throws java . io . IOException { return inputStream ; } } ; final java . lang . String text = byteSource . asCharSource ( Charsets . UTF_8 ) . read ( ) ; "<AssertPlaceHolder>" ; } read ( ) { return "Reading<sp>file<sp>" + ( name ) ; }
org . junit . Assert . assertThat ( text , org . hamcrest . Matchers . equalTo ( originalString ) )
testInterruptSizeNoRetry ( ) { java . util . concurrent . atomic . AtomicInteger count = new java . util . concurrent . atomic . AtomicInteger ( ) ; com . ctrip . xpipe . api . utils . ControllableFile controllableFile = new com . ctrip . xpipe . utils . DefaultControllableFile ( file ) { @ com . ctrip . xpipe . utils . Override public long size ( ) { count . incrementAndGet ( ) ; return super . size ( ) ; } } ; java . lang . Thread . currentThread ( ) . interrupt ( ) ; try { controllableFile . size ( ) ; } catch ( java . lang . Exception e ) { } java . lang . Thread . interrupted ( ) ; "<AssertPlaceHolder>" ; } get ( ) { await ( ) ; java . lang . Throwable cause = cause ( ) ; if ( cause == null ) { return getNow ( ) ; } if ( cause instanceof com . ctrip . xpipe . command . CancellationException ) { throw ( ( com . ctrip . xpipe . command . CancellationException ) ( cause ) ) ; } throw new com . ctrip . xpipe . command . ExecutionException ( cause ) ; }
org . junit . Assert . assertEquals ( 1 , count . get ( ) )
deleteNonExistentEquipment ( ) { cern . c2mon . shared . client . configuration . api . equipment . Equipment equipment = buildDeleteEquipment ( 10L ) ; java . util . List < cern . c2mon . shared . client . configuration . api . equipment . Equipment > equipmentRemoveList = java . util . Arrays . asList ( equipment ) ; cern . c2mon . shared . client . configuration . api . Configuration config = new cern . c2mon . shared . client . configuration . api . Configuration ( 1L ) ; config . setEntities ( equipmentRemoveList ) ; org . easymock . EasyMock . expect ( equipmentCache . hasKey ( 10L ) ) . andReturn ( false ) ; org . easymock . EasyMock . replay ( equipmentCache ) ; "<AssertPlaceHolder>" ; org . easymock . EasyMock . verify ( equipmentCache ) ; } parse ( cern . c2mon . shared . client . configuration . api . Configuration ) { if ( ( ( configuration . getEntities ( ) ) != null ) && ( ! ( configuration . getEntities ( ) . isEmpty ( ) ) ) ) { return parseConfigurationList ( configuration . getEntities ( ) ) ; } else { throw new cern . c2mon . server . configuration . parser . exception . ConfigurationParseException ( "Empty<sp>configuration<sp>received!" ) ; } }
org . junit . Assert . assertEquals ( 0 , parser . parse ( config ) . size ( ) )
testFindAllPersonByIndexOnAnnotatedFieldWithAtIndexed ( ) { org . springframework . data . neo4j . support . Person person = org . springframework . data . neo4j . Person . persistedPerson ( org . springframework . data . neo4j . support . IndexTest . NAME_VALUE , 35 ) ; person . setNickname ( "Mike" ) ; final org . springframework . data . neo4j . support . Person found = personFinder . findByPropertyValue ( "nickname" , "Mike" ) ; "<AssertPlaceHolder>" ; } findByPropertyValue ( java . lang . String , java . lang . Object ) { return findByPropertyValue ( null , property , value ) ; }
org . junit . Assert . assertEquals ( person , found )
test_Filter14 ( ) { java . lang . String expected = "cbabc" ; java . lang . String actual ; org . osgi . framework . Bundle bundle = installBundle ( org . eclipse . equinox . http . servlet . tests . TEST_BUNDLE_1 ) ; try { bundle . start ( ) ; actual = requestAdvisor . request ( "something/a.TestFilter14" ) ; } finally { uninstallBundle ( bundle ) ; } "<AssertPlaceHolder>" ; } uninstallBundle ( org . osgi . framework . Bundle ) { org . eclipse . equinox . internal . security . tests . storage . WaitingRegistryListener listener = new org . eclipse . equinox . internal . security . tests . storage . WaitingRegistryListener ( ) ; listener . register ( "org.eclipse.equinox.security.secureStorage" ) ; try { bundle . uninstall ( ) ; bundle = null ; org . junit . Assert . assertTrue ( ( ( listener . waitFor ( 1 , ( 10 * ( org . eclipse . equinox . internal . security . tests . storage . DynamicPreferencesTest . MAX_TIME_PER_BUNDLE ) ) ) ) == 1 ) ) ; return true ; } finally { listener . unregister ( ) ; if ( bundle != null ) bundle . uninstall ( ) ; } }
org . junit . Assert . assertEquals ( expected , actual )
testBuscarLancamentoPorFuncionarioId ( ) { org . springframework . data . domain . Page < com . kazale . pontointeligente . api . entities . Lancamento > lancamento = this . lancamentoService . buscarPorFuncionarioId ( 1L , new org . springframework . data . domain . PageRequest ( 0 , 10 ) ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertNotNull ( lancamento )
testLongArrayWithIndex ( ) { java . lang . Long [ ] longArr = new java . lang . Long [ 4 ] ; longArr [ 0 ] = 1L ; longArr [ 1 ] = 2L ; longArr [ 2 ] = 4L ; longArr [ 3 ] = 5L ; org . apache . phoenix . schema . types . PhoenixArray arr = org . apache . phoenix . schema . types . PArrayDataType . instantiatePhoenixArray ( PLong . INSTANCE , longArr ) ; PLongArray . INSTANCE . toObject ( arr , PLongArray . INSTANCE ) ; byte [ ] bytes = PLongArray . INSTANCE . toBytes ( arr ) ; org . apache . hadoop . hbase . io . ImmutableBytesWritable ptr = new org . apache . hadoop . hbase . io . ImmutableBytesWritable ( bytes ) ; org . apache . phoenix . schema . types . PArrayDataType . positionAtArrayElement ( ptr , 2 , PLong . INSTANCE , PLong . INSTANCE . getByteSize ( ) ) ; int offset = ptr . getOffset ( ) ; int length = ptr . getLength ( ) ; byte [ ] bs = ptr . get ( ) ; byte [ ] res = new byte [ length ] ; java . lang . System . arraycopy ( bs , offset , res , 0 , length ) ; long result = ( ( java . lang . Long ) ( PLong . INSTANCE . toObject ( res ) ) ) ; "<AssertPlaceHolder>" ; } toObject ( java . lang . String ) { if ( ( value == null ) || ( ( value . length ( ) ) == 0 ) ) { return null ; } try { return java . lang . Long . parseLong ( value ) ; } catch ( java . lang . NumberFormatException e ) { throw newIllegalDataException ( e ) ; } }
org . junit . Assert . assertEquals ( 4L , result )
testNotEquals ( ) { java . util . List < java . security . Principal > principals = new java . util . ArrayList < java . security . Principal > ( ) ; principals . add ( new org . apache . jackrabbit . oak . spi . security . principal . PrincipalImpl ( "otherName" ) ) ; principals . add ( new java . security . Principal ( ) { @ org . apache . jackrabbit . oak . spi . security . principal . Override public java . lang . String getName ( ) { return "name" ; } } ) ; for ( java . security . Principal p : principals ) { "<AssertPlaceHolder>" ; } } equals ( java . lang . Object ) { if ( o instanceof org . apache . jackrabbit . api . security . principal . JackrabbitPrincipal ) { return getName ( ) . equals ( ( ( org . apache . jackrabbit . api . security . principal . JackrabbitPrincipal ) ( o ) ) . getName ( ) ) ; } return false ; }
org . junit . Assert . assertFalse ( principal . equals ( p ) )
documentInitNullPeer ( ) { "<AssertPlaceHolder>" ; } getPeer ( ) { if ( ( myContainer ) == null ) { throw new java . lang . IllegalStateException ( ) ; } return myContainer . getCellContainerPeer ( ) ; }
org . junit . Assert . assertNull ( container . getPeer ( ) )
message_does_not_exceed_max_length ( ) { for ( oakbot . chat . SplitStrategy strategy : oakbot . chat . SplitStrategy . values ( ) ) { java . util . List < java . lang . String > actual = strategy . split ( "message" , 50 ) ; java . util . List < java . lang . String > expected = java . util . Arrays . asList ( "message" ) ; "<AssertPlaceHolder>" ; } } split ( java . lang . String , int ) { if ( ( maxLength < 1 ) || ( ( message . length ( ) ) <= maxLength ) ) { return java . util . Arrays . asList ( message ) ; } return _split ( message , maxLength ) ; }
org . junit . Assert . assertEquals ( expected , actual )
testGetSettingsPanel ( ) { System . out . println ( "getSettingsPanel" ) ; kg . apc . jmeter . vizualizers . ResponseTimesPercentilesGui instance = new kg . apc . jmeter . vizualizers . ResponseTimesPercentilesGui ( ) ; kg . apc . jmeter . vizualizers . JSettingsPanel result = instance . createSettingsPanel ( ) ; "<AssertPlaceHolder>" ; } createSettingsPanel ( ) { return new kg . apc . jmeter . vizualizers . JSettingsPanel ( this , ( ( ( ( ( ( ( JSettingsPanel . TIMELINE_OPTION ) | ( JSettingsPanel . GRADIENT_OPTION ) ) | ( JSettingsPanel . FINAL_ZEROING_OPTION ) ) | ( JSettingsPanel . LIMIT_POINT_OPTION ) ) | ( JSettingsPanel . MAXY_OPTION ) ) | ( JSettingsPanel . RELATIVE_TIME_OPTION ) ) | ( JSettingsPanel . MARKERS_OPTION ) ) ) ; }
org . junit . Assert . assertNotNull ( result )
testZeroRateBaselineAndSubjectShouldNotBeDefined ( ) { org . drugis . addis . entities . RateMeasurement base = new org . drugis . addis . entities . BasicRateMeasurement ( 0 , 100 ) ; org . drugis . addis . entities . RateMeasurement subj = new org . drugis . addis . entities . BasicRateMeasurement ( 0 , 100 ) ; org . drugis . addis . entities . relativeeffect . CorrectedBasicOddsRatio or = new org . drugis . addis . entities . relativeeffect . CorrectedBasicOddsRatio ( base , subj ) ; "<AssertPlaceHolder>" ; } isDefined ( ) { return ( ( super . isDefined ( ) ) && ( isAdmissible ( d_subject ) ) ) && ( isAdmissible ( d_baseline ) ) ; }
org . junit . Assert . assertFalse ( or . isDefined ( ) )
testBuildWithDisabledStatusConstraint ( ) { unit . setActive ( false ) ; org . lnu . is . domain . specialty . Specialty context = new org . lnu . is . domain . specialty . Specialty ( ) ; java . lang . String expectedQuery = "SELECT<sp>e<sp>FROM<sp>Specialty<sp>e<sp>WHERE<sp>e.crtUserGroup<sp>IN<sp>(:userGroups)<sp>" ; org . lnu . is . pagination . MultiplePagedSearch < org . lnu . is . domain . specialty . Specialty > pagedSearch = new org . lnu . is . pagination . MultiplePagedSearch ( ) ; pagedSearch . setEntity ( context ) ; java . lang . String actualQuery = unit . build ( pagedSearch ) ; "<AssertPlaceHolder>" ; } setEntity ( T ) { this . entity = entity ; }
org . junit . Assert . assertEquals ( expectedQuery , actualQuery )
testGetParameters ( ) { java . lang . Long departmentId = 1L ; org . lnu . is . domain . department . Department department = new org . lnu . is . domain . department . Department ( ) ; department . setId ( departmentId ) ; java . lang . Long specialtyId = 2L ; org . lnu . is . domain . specialty . Specialty specialty = new org . lnu . is . domain . specialty . Specialty ( ) ; specialty . setId ( specialtyId ) ; org . lnu . is . domain . department . specialty . DepartmentSpecialty entity = new org . lnu . is . domain . department . specialty . DepartmentSpecialty ( ) ; entity . setDepartment ( department ) ; entity . setSpecialty ( specialty ) ; java . util . Map < java . lang . String , java . lang . Object > expected = new java . util . HashMap < java . lang . String , java . lang . Object > ( ) ; expected . put ( "status" , RowStatus . ACTIVE ) ; expected . put ( "userGroups" , groups ) ; expected . put ( "department" , department ) ; expected . put ( "specialty" , specialty ) ; when ( departmentDao . getEntityById ( anyLong ( ) ) ) . thenReturn ( department ) ; when ( specialtyDao . getEntityById ( anyLong ( ) ) ) . thenReturn ( specialty ) ; java . util . Map < java . lang . String , java . lang . Object > actual = unit . getParameters ( entity ) ; verify ( departmentDao ) . getEntityById ( departmentId ) ; verify ( specialtyDao ) . getEntityById ( specialtyId ) ; "<AssertPlaceHolder>" ; } getEntityById ( KEY ) { org . lnu . is . dao . dao . DefaultDao . LOG . info ( "Getting<sp>{}.entity<sp>wit<sp>id" , getEntityClass ( ) . getSimpleName ( ) , id ) ; return persistenceManager . findById ( getEntityClass ( ) , id ) ; }
org . junit . Assert . assertEquals ( expected , actual )
showTest ( ) { org . sickbeard . Show response = sick . show ( "71256" ) ; "<AssertPlaceHolder>" ; } show ( java . lang . String ) { java . lang . StringBuilder builder = new java . lang . StringBuilder ( "show" ) ; builder . append ( "&tvdbid=" ) ; builder . append ( tvdbid ) ; return new org . sickbeard . Show ( this . < org . sickbeard . ShowJson > commandData ( builder . toString ( ) , new com . google . gson . reflect . TypeToken < org . sickbeard . JsonResponse < org . sickbeard . ShowJson > > ( ) { } . getType ( ) ) ) ; }
org . junit . Assert . assertNotNull ( response )
testTooSmallPacketIsRejected ( ) { byte [ ] tooSmallPacket = new byte [ 12 ] ; "<AssertPlaceHolder>" ; } readFlexFecHeader ( byte [ ] , int , int ) { if ( length < ( org . jitsi . impl . neomedia . transform . fec . FlexFec03HeaderReader . HEADER_MIN_SIZE_BYTES ) ) { return null ; } boolean retransmissionBit = ( ( ( buffer [ flexFecOffset ] ) & 128 ) > > 7 ) == 1 ; if ( retransmissionBit ) { return null ; } int maskType = ( ( buffer [ flexFecOffset ] ) & 64 ) > > 6 ; if ( maskType != 0 ) { return null ; } int ssrcCount = ( buffer [ ( flexFecOffset + 8 ) ] ) & 255 ; if ( ssrcCount > 1 ) { return null ; } long protectedSsrc = org . jitsi . impl . neomedia . transform . fec . RTPUtils . readUint32AsLong ( buffer , ( flexFecOffset + 12 ) ) ; int seqNumBase = org . jitsi . impl . neomedia . transform . fec . RTPUtils . readUint16AsInt ( buffer , ( flexFecOffset + 16 ) ) ; org . jitsi . impl . neomedia . transform . fec . FlexFec03Mask mask ; try { mask = new org . jitsi . impl . neomedia . transform . fec . FlexFec03Mask ( buffer , ( flexFecOffset + ( org . jitsi . impl . neomedia . transform . fec . FlexFec03HeaderReader . MASK_START_OFFSET_BYTES ) ) , seqNumBase ) ; } catch ( org . jitsi . impl . neomedia . transform . fec . FlexFec03Mask e ) { return null ; } int flexFecHeaderSize = ( ( org . jitsi . impl . neomedia . transform . fec . FlexFec03HeaderReader . HEADER_MIN_SIZE_BYTES ) - 2 ) + ( mask . lengthBytes ( ) ) ; return new org . jitsi . impl . neomedia . transform . fec . FlexFec03Header ( protectedSsrc , seqNumBase , mask . getProtectedSeqNums ( ) , flexFecHeaderSize ) ; }
org . junit . Assert . assertNull ( org . jitsi . impl . neomedia . transform . fec . FlexFec03HeaderReader . readFlexFecHeader ( tooSmallPacket , 0 , tooSmallPacket . length ) )
write_SProfile ( ) { if ( ! ( isJp2KakDriverAvailable ) ) return ; it . geosolutions . imageio . plugins . jp2kakadu . JP2KWriteTest . LOGGER . info ( "PROFILE2" 0 ) ; final java . io . File inputFile = it . geosolutions . resources . TestData . file ( this , it . geosolutions . imageio . plugins . jp2kakadu . JP2KWriteTest . testFileName ) ; "<AssertPlaceHolder>" ; final java . io . File outputFile1 = it . geosolutions . resources . TestData . temp ( this , "SProfile1-.jp2" , it . geosolutions . imageio . plugins . jp2kakadu . JP2KWriteTest . deleteTempFilesOnExit ) ; final java . io . File outputFile2 = it . geosolutions . resources . TestData . temp ( this , "SProfile2-.jp2" , it . geosolutions . imageio . plugins . jp2kakadu . JP2KWriteTest . deleteTempFilesOnExit ) ; final javax . media . jai . ParameterBlockJAI pbjImageRead = new javax . media . jai . ParameterBlockJAI ( "PROFILE2" 1 ) ; pbjImageRead . setParameter ( "Input" , inputFile ) ; if ( it . geosolutions . imageio . plugins . jp2kakadu . JP2KWriteTest . ENABLE_SUBSAMPLING ) { javax . imageio . ImageReadParam readParam = new javax . imageio . ImageReadParam ( ) ; readParam . setSourceSubsampling ( 4 , 4 , 0 , 0 ) ; pbjImageRead . setParameter ( "readParam" , readParam ) ; } pbjImageRead . setParameter ( "Reader" , new it . geosolutions . imageio . plugins . jp2kakadu . JP2GDALKakaduImageReaderSpi ( ) . createReaderInstance ( ) ) ; javax . media . jai . RenderedOp image = javax . media . jai . JAI . create ( "PROFILE2" 1 , pbjImageRead ) ; final javax . media . jai . ParameterBlockJAI pbjImageWrite = new javax . media . jai . ParameterBlockJAI ( "ImageWrite" ) ; pbjImageWrite . setParameter ( "Output" , new it . geosolutions . imageio . stream . output . FileImageOutputStreamExtImpl ( outputFile1 ) ) ; javax . imageio . ImageWriter writer = new it . geosolutions . imageio . plugins . jp2kakadu . JP2GDALKakaduImageWriterSpi ( ) . createWriterInstance ( ) ; pbjImageWrite . setParameter ( "Writer" , writer ) ; pbjImageWrite . addSource ( image ) ; javax . imageio . ImageWriteParam param = writer . getDefaultWriteParam ( ) ; ( ( it . geosolutions . imageio . plugins . jp2kakadu . JP2GDALKakaduImageWriteParam ) ( param ) ) . setSProfile ( 1 ) ; pbjImageWrite . setParameter ( "writeParam" , param ) ; final javax . media . jai . RenderedOp op = javax . media . jai . JAI . create ( "ImageWrite" , pbjImageWrite ) ; final javax . media . jai . ParameterBlockJAI pbjImageWrite2 = new javax . media . jai . ParameterBlockJAI ( "ImageWrite" ) ; pbjImageWrite2 . setParameter ( "Output" , new it . geosolutions . imageio . stream . output . FileImageOutputStreamExtImpl ( outputFile2 ) ) ; javax . imageio . ImageWriter writer2 = new it . geosolutions . imageio . plugins . jp2kakadu . JP2GDALKakaduImageWriterSpi ( ) . createWriterInstance ( ) ; pbjImageWrite2 . setParameter ( "Writer" , writer2 ) ; pbjImageWrite2 . addSource ( image ) ; javax . imageio . ImageWriteParam param2 = writer2 . getDefaultWriteParam ( ) ; ( ( it . geosolutions . imageio . plugins . jp2kakadu . JP2GDALKakaduImageWriteParam ) ( param2 ) ) . setSProfile ( "PROFILE2" ) ; pbjImageWrite2 . setParameter ( "writeParam" , param2 ) ; final javax . media . jai . RenderedOp op2 = javax . media . jai . JAI . create ( "ImageWrite" , pbjImageWrite2 ) ; } file ( java . lang . Object , java . lang . String ) { final java . net . URL url = it . geosolutions . resources . TestData . url ( caller , path ) ; final java . io . File file = it . geosolutions . imageio . utilities . Utilities . urlToFile ( url ) ; if ( ! ( file . exists ( ) ) ) { throw new java . io . FileNotFoundException ( ( "Could<sp>not<sp>locate<sp>test-data:<sp>" + path ) ) ; } return file ; }
org . junit . Assert . assertTrue ( inputFile . exists ( ) )
testDisassociateTrustBundleFromDomain_noEntityManager_assertException ( ) { final org . nhindirect . config . store . dao . impl . TrustBundleDaoImpl dao = new org . nhindirect . config . store . dao . impl . TrustBundleDaoImpl ( ) ; boolean exceptionOccured = false ; try { dao . disassociateTrustBundleFromDomain ( 1234 , 5678 ) ; } catch ( java . lang . IllegalStateException ex ) { exceptionOccured = true ; } "<AssertPlaceHolder>" ; }
org . junit . Assert . assertTrue ( exceptionOccured )
testStringTemplateExecutesWithStringConn ( ) { org . junit . Assume . assumeTrue ( ( ( redisTemplate ) instanceof org . springframework . data . redis . core . StringRedisTemplate ) ) ; java . lang . String value = redisTemplate . execute ( ( ( org . springframework . data . redis . core . RedisCallback < java . lang . String > ) ( ( connection ) -> { org . springframework . data . redis . connection . StringRedisConnection stringConn = ( ( org . springframework . data . redis . connection . StringRedisConnection ) ( connection ) ) ; stringConn . set ( "test" , "it" ) ; return stringConn . get ( "test" ) ; } ) ) ) ; "<AssertPlaceHolder>" ; } get ( java . lang . Object ) { return delegate . get ( key ) ; }
org . junit . Assert . assertEquals ( value , "it" )
testSubClassStoresSuperInterfaceValues ( ) { org . apache . beam . sdk . options . ProxyInvocationHandler handler = new org . apache . beam . sdk . options . ProxyInvocationHandler ( org . apache . beam . vendor . guava . v20_0 . com . google . common . collect . Maps . newHashMap ( ) ) ; org . apache . beam . sdk . options . ProxyInvocationHandlerTest . SubClass extended = handler . as ( org . apache . beam . sdk . options . ProxyInvocationHandlerTest . SubClass . class ) ; extended . setString ( "parentValue" ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( "parentValue" , extended . getString ( ) )
testBuildWithDisabledStatusCOnstsintWithOrderBy ( ) { unit . setActive ( false ) ; org . lnu . is . domain . course . type . CourseType context = new org . lnu . is . domain . course . type . CourseType ( ) ; org . lnu . is . pagination . OrderBy orderBy1 = new org . lnu . is . pagination . OrderBy ( "abbrName" , org . lnu . is . pagination . OrderByType . ASC ) ; org . lnu . is . pagination . OrderBy orderBy2 = new org . lnu . is . pagination . OrderBy ( "name" , org . lnu . is . pagination . OrderByType . DESC ) ; java . util . List < org . lnu . is . pagination . OrderBy > orders = java . util . Arrays . asList ( orderBy1 , orderBy2 ) ; java . lang . String expected = "SELECT<sp>e<sp>FROM<sp>CourseType<sp>e<sp>WHERE<sp>e.crtUserGroup<sp>IN<sp>(:userGroups)<sp>ORDER<sp>BY<sp>e.abbrName<sp>ASC,<sp>e.name<sp>DESC" ; org . lnu . is . pagination . MultiplePagedSearch < org . lnu . is . domain . course . type . CourseType > pagedSearch = new org . lnu . is . pagination . MultiplePagedSearch ( ) ; pagedSearch . setEntity ( context ) ; pagedSearch . setOrders ( orders ) ; java . lang . String actualQuery = unit . build ( pagedSearch ) ; "<AssertPlaceHolder>" ; } setOrders ( java . util . List ) { this . orders = orders ; }
org . junit . Assert . assertEquals ( expected , actualQuery )
givenTypeAndNameAndAnnotationWhenNameAndAnnotationMatchButTypeNotFoundThenFieldNotFoundExceptionIsThrown ( ) { org . awaitility . FakeRepositoryWithAnnotation repository = new org . awaitility . FakeRepositoryWithAnnotation ( ) ; new org . awaitility . Asynch ( repository ) . perform ( ) ; byte one = ( ( byte ) ( 1 ) ) ; org . awaitility . Awaitility . await ( ) . until ( org . awaitility . Awaitility . fieldIn ( repository ) . ofType ( byte . class ) . andWithName ( "value" ) . andAnnotatedWith ( org . awaitility . ExampleAnnotation . class ) , org . hamcrest . Matchers . equalTo ( one ) ) ; "<AssertPlaceHolder>" ; } getValue ( ) { return currentConditionValue ; }
org . junit . Assert . assertEquals ( 1 , repository . getValue ( ) )
url_type_request_returns_instance_of_url ( ) { java . lang . Object result = this . urlBuilder . create ( java . net . URL . class , null ) ; "<AssertPlaceHolder>" ; } create ( java . lang . Object , com . flextrade . jfixture . SpecimenContext ) { if ( ! ( request instanceof com . flextrade . jfixture . utility . SpecimenType ) ) { return new com . flextrade . jfixture . NoSpecimen ( ) ; } com . flextrade . jfixture . utility . SpecimenType type = ( ( com . flextrade . jfixture . utility . SpecimenType ) ( request ) ) ; if ( ! ( org . joda . time . base . BaseDateTime . class . isAssignableFrom ( type . getRawType ( ) ) ) ) { return new com . flextrade . jfixture . NoSpecimen ( ) ; } try { java . util . Date date = ( ( java . util . Date ) ( context . resolve ( java . util . Date . class ) ) ) ; long instant = date . getTime ( ) ; org . joda . time . DateTimeZone timeZone = ( ( org . joda . time . DateTimeZone ) ( context . resolve ( org . joda . time . DateTimeZone . class ) ) ) ; return type . getRawType ( ) . getDeclaredConstructor ( long . class , org . joda . time . DateTimeZone . class ) . newInstance ( instant , timeZone ) ; } catch ( java . lang . Exception e ) { e . printStackTrace ( ) ; return new com . flextrade . jfixture . NoSpecimen ( ) ; } }
org . junit . Assert . assertTrue ( ( result instanceof java . net . URL ) )
test_skip ( java . lang . String ) { org . joda . beans . ser . json . JsonInput input = new org . joda . beans . ser . json . JsonInput ( new java . io . StringReader ( ( text + ',' ) ) ) ; input . skipData ( ) ; "<AssertPlaceHolder>" ; } readEvent ( ) { char next = readNext ( ) ; while ( ( ( ( next == '<sp>' ) || ( next == '\t' ) ) || ( next == '\n' ) ) || ( next == '\r' ) ) { next = readNext ( ) ; } switch ( next ) { case '{' : return JsonEvent . OBJECT ; case '}' : return JsonEvent . OBJECT_END ; case '[' : return JsonEvent . ARRAY ; case ']' : return JsonEvent . ARRAY_END ; case '"' : return JsonEvent . STRING ; case '-' : case '0' : case '1' : case '2' : case '3' : case '4' : case '5' : case '6' : case '7' : case '8' : case '9' : return acceptNumber ( next ) ; case 'n' : return acceptNull ( ) ; case 't' : return acceptTrue ( ) ; case 'f' : return acceptFalse ( ) ; case ',' : return JsonEvent . COMMA ; case ':' : return JsonEvent . COLON ; default : throw new java . lang . IllegalArgumentException ( ( ( "Invalid<sp>JSON<sp>data:<sp>Expected<sp>JSON<sp>character<sp>but<sp>found<sp>'" + next ) + "'" ) ) ; } }
org . junit . Assert . assertEquals ( input . readEvent ( ) , JsonEvent . COMMA )
testQueue ( ) { org . skyscreamer . nevado . jms . destination . NevadoQueue queue = createTempQueue ( createSession ( ) ) ; org . skyscreamer . nevado . jms . destination . NevadoQueue testQueue = ( ( org . skyscreamer . nevado . jms . destination . NevadoQueue ) ( org . skyscreamer . nevado . jms . util . SerializeUtil . deserialize ( org . skyscreamer . nevado . jms . util . SerializeUtil . serialize ( queue ) ) ) ) ; "<AssertPlaceHolder>" ; } serialize ( java . io . Serializable ) { java . io . ByteArrayOutputStream byteArrayOutputStream = new java . io . ByteArrayOutputStream ( ) ; com . caucho . hessian . io . Hessian2Output hessian2Output = new com . caucho . hessian . io . Hessian2Output ( byteArrayOutputStream ) ; hessian2Output . setSerializerFactory ( new com . caucho . hessian . io . SerializerFactory ( org . skyscreamer . nevado . jms . util . SerializeUtil . class . getClassLoader ( ) ) ) ; hessian2Output . startMessage ( ) ; if ( serializable instanceof java . lang . Character ) { serializable = new org . skyscreamer . nevado . jms . util . CharWrapper ( ( ( java . lang . Character ) ( serializable ) ) ) ; } hessian2Output . writeObject ( serializable ) ; hessian2Output . completeMessage ( ) ; hessian2Output . close ( ) ; return byteArrayOutputStream . toByteArray ( ) ; }
org . junit . Assert . assertEquals ( queue , testQueue )
testOverlijdenOpZelfdeDag00uur ( ) { final java . util . List < nl . bzk . brp . model . basis . BerichtEntiteit > resultaat = brby0907 . voerRegelUit ( maakOverledenHuidigePersoon ( null , "20110831<sp>00:00:00.00" ) , maakOverledenNieuwePersoon ( 20110831 ) , null , null ) ; "<AssertPlaceHolder>" ; } size ( ) { return elementen . size ( ) ; }
org . junit . Assert . assertEquals ( 0 , resultaat . size ( ) )
getClientAddressWithXRemoteAddr ( ) { org . mockito . Mockito . when ( request . getHeader ( org . mockito . Mockito . eq ( "Remote_Addr" ) ) ) . thenReturn ( "192.168.1.1" ) ; "<AssertPlaceHolder>" ; } getClientAddress ( javax . servlet . http . HttpServletRequest ) { for ( final java . lang . String header : com . cloud . api . ApiServlet . s_clientAddressHeaders ) { final java . lang . String ip = com . cloud . api . ApiServlet . getCorrectIPAddress ( request . getHeader ( header ) ) ; if ( ip != null ) { return ip ; } } return request . getRemoteAddr ( ) ; }
org . junit . Assert . assertEquals ( "192.168.1.1" , com . cloud . api . ApiServlet . getClientAddress ( request ) )
testAddString005 ( ) { javax . naming . ldap . LdapName ln = new javax . naming . ldap . LdapName ( "" ) ; java . lang . String x = "" ; ln . add ( x ) ; "<AssertPlaceHolder>" ; } toString ( ) { return this . toString ( "" ) ; }
org . junit . Assert . assertEquals ( "" , ln . toString ( ) )
TSDLClockTest ( ) { org . eclipse . tracecompass . ctf . core . tests . trace . IOstructgenTest . createDummyTrace ( org . eclipse . tracecompass . ctf . core . tests . trace . IOstructgenTest . clockTSDL ) ; trace = new org . eclipse . tracecompass . ctf . core . trace . CTFTrace ( org . eclipse . tracecompass . ctf . core . tests . trace . IOstructgenTest . tempTraceDir ) ; "<AssertPlaceHolder>" ; } createDummyTrace ( java . lang . String ) { java . io . File dir = new java . io . File ( org . eclipse . tracecompass . ctf . core . tests . trace . IOstructgenTest . tempTraceDir ) ; if ( dir . exists ( ) ) { org . eclipse . tracecompass . ctf . core . tests . trace . IOstructgenTest . deltree ( dir ) ; } dir . mkdirs ( ) ; java . io . File metadataFile = new java . io . File ( ( ( org . eclipse . tracecompass . ctf . core . tests . trace . IOstructgenTest . tempTraceDir ) + "/metadata" ) ) ; try ( java . io . FileWriter fw = new java . io . FileWriter ( metadataFile ) ) { fw . write ( metadata ) ; } catch ( java . io . IOException e ) { e . printStackTrace ( ) ; } byte [ ] magicLE = new byte [ ] { ( ( byte ) ( 193 ) ) , ( ( byte ) ( 31 ) ) , ( ( byte ) ( 252 ) ) , ( ( byte ) ( 193 ) ) } ; byte [ ] uuid = new byte [ ] { ( ( byte ) ( 176 ) ) , 77 , 57 , 27 , ( ( byte ) ( 231 ) ) , 54 , 68 , ( ( byte ) ( 193 ) ) , ( ( byte ) ( 141 ) ) , ( ( byte ) ( 137 ) ) , 75 , ( ( byte ) ( 180 ) ) , 56 , ( ( byte ) ( 133 ) ) , 127 , ( ( byte ) ( 141 ) ) } ; org . eclipse . tracecompass . ctf . core . tests . trace . IOstructgenTest . Event ev = new org . eclipse . tracecompass . ctf . core . tests . trace . IOstructgenTest . Event ( 2 , 2 ) ; final int nbEvents = ( ( org . eclipse . tracecompass . ctf . core . tests . trace . IOstructgenTest . DATA_SIZE ) / ( ev . getSize ( ) ) ) - 1 ; final int contentSize = ( ( nbEvents * ( ev . getSize ( ) ) ) + ( org . eclipse . tracecompass . ctf . core . tests . trace . IOstructgenTest . HEADER_SIZE ) ) * 8 ; java . nio . ByteBuffer data = java . nio . ByteBuffer . allocate ( org . eclipse . tracecompass . ctf . core . tests . trace . IOstructgenTest . PACKET_SIZE ) ; data . order ( ByteOrder . LITTLE_ENDIAN ) ; data . clear ( ) ; data . put ( magicLE ) ; data . put ( uuid ) ; data . putInt ( 0 ) ; data . putLong ( 42240 ) ; data . putLong ( ( ( nbEvents * 65536 ) + 42406 ) ) ; data . putLong ( contentSize ) ; data . putLong ( ( ( org . eclipse . tracecompass . ctf . core . tests . trace . IOstructgenTest . PACKET_SIZE ) * 8 ) ) ; data . putLong ( 0 ) ; data . putInt ( 0 ) ; for ( int i = 0 ; i < nbEvents ; i ++ ) { ev . setEventTimestamp ( ( ( i * 65536 ) + 42405 ) ) ; ev . setEventContent ( i ) ; ev . writeEvent ( data ) ; } data . flip ( ) ; java . io . File dummyFile = new java . io . File ( ( ( org . eclipse . tracecompass . ctf . core . tests . trace . IOstructgenTest . tempTraceDir ) + "/dummyChan" ) ) ; try ( java . io . FileOutputStream fos = new java . io . FileOutputStream ( dummyFile ) ) { fos . getChannel ( ) . write ( data ) ; } catch ( java . io . IOException e ) { e . printStackTrace ( ) ; } }
org . junit . Assert . assertNotNull ( trace )
testFailBadVersion ( ) { final org . apache . druid . indexing . common . task . Task task = new org . apache . druid . indexing . common . task . NoopTask ( null , null , 0 , 0 , null , null , null ) ; final org . apache . druid . indexing . common . actions . SegmentTransactionalInsertAction action = new org . apache . druid . indexing . common . actions . SegmentTransactionalInsertAction ( com . google . common . collect . ImmutableSet . of ( org . apache . druid . indexing . common . actions . SegmentTransactionalInsertActionTest . SEGMENT3 ) ) ; actionTestKit . getTaskLockbox ( ) . add ( task ) ; actionTestKit . getTaskLockbox ( ) . lock ( TaskLockType . EXCLUSIVE , task , org . apache . druid . indexing . common . actions . SegmentTransactionalInsertActionTest . INTERVAL , 5000 ) ; thrown . expect ( org . apache . druid . indexing . common . actions . IllegalStateException . class ) ; thrown . expectMessage ( org . hamcrest . CoreMatchers . containsString ( "are<sp>not<sp>covered<sp>by<sp>locks" ) ) ; org . apache . druid . indexing . overlord . SegmentPublishResult result = action . perform ( task , actionTestKit . getTaskActionToolbox ( ) ) ; "<AssertPlaceHolder>" ; } ok ( java . util . Set ) { return new org . apache . druid . indexing . overlord . SegmentPublishResult ( segments , true , null ) ; }
org . junit . Assert . assertEquals ( org . apache . druid . indexing . overlord . SegmentPublishResult . ok ( com . google . common . collect . ImmutableSet . of ( org . apache . druid . indexing . common . actions . SegmentTransactionalInsertActionTest . SEGMENT3 ) ) , result )
getByID ( ) { net . billforward . model . Invoice invoice = net . billforward . model . Invoice . getByID ( "95CE489C-0867-4EB0-9086-C09E444B1249" ) ; "<AssertPlaceHolder>" ; System . out . println ( invoice . toString ( ) ) ; } getID ( ) { return id ; }
org . junit . Assert . assertEquals ( "95CE489C-0867-4EB0-9086-C09E444B1249" , invoice . getID ( ) )
whenClosingSession_tryingToUseTransactionThrowsException ( ) { grakn . core . server . session . SessionImpl localSession = grakn . core . server . session . SessionIT . server . sessionFactory ( ) . session ( grakn . core . server . keyspace . KeyspaceImpl . of ( "test" ) ) ; grakn . core . server . session . TransactionOLTP tx1 = localSession . transaction ( ) . write ( ) ; "<AssertPlaceHolder>" ; localSession . close ( ) ; expectedException . expect ( grakn . core . server . exception . TransactionException . class ) ; expectedException . expectMessage ( "The<sp>session<sp>for<sp>graph<sp>[test]<sp>is<sp>closed.<sp>Create<sp>a<sp>new<sp>session<sp>to<sp>interact<sp>with<sp>the<sp>graph." ) ; grakn . core . concept . type . SchemaConcept thing = tx1 . getSchemaConcept ( grakn . core . concept . Label . of ( "thing" ) ) ; } isClosed ( ) { return ! ( isTxOpen ) ; }
org . junit . Assert . assertFalse ( tx1 . isClosed ( ) )
testGettingANonExistingRootCategoryReturnsNull ( ) { org . pentaho . marketplace . domain . model . factories . interfaces . ICategoryFactory factory = this . createFactory ( ) ; org . pentaho . marketplace . domain . model . entities . interfaces . ICategory category = factory . get ( "IDoNotExist" ) ; "<AssertPlaceHolder>" ; } get ( java . lang . String ) { return org . pentaho . marketplace . domain . model . factories . CategoryFactory . categories . get ( name ) ; }
org . junit . Assert . assertNull ( category )
testSetSameValueDoesNotFireEvent ( ) { point . setValue ( new java . lang . Double ( val1 ) ) ; "<AssertPlaceHolder>" ; verify ( eventHelper , times ( 0 ) ) . addEventPayload ( eq ( continuousMapping ) , any ( org . cytoscape . view . vizmap . events . VisualMappingFunctionChangeRecord . class ) , eq ( org . cytoscape . view . vizmap . events . VisualMappingFunctionChangedEvent . class ) ) ; } getValue ( ) { return nodelist ; }
org . junit . Assert . assertEquals ( val1 , point . getValue ( ) )
testConstructor ( ) { new org . jtrfp . trcl . pool . IndexList < java . lang . Object > ( delegate ) ; "<AssertPlaceHolder>" ; } size ( ) { final javax . swing . DefaultComboBoxModel < java . lang . String > model = ( ( javax . swing . DefaultComboBoxModel ) ( comboBox . getModel ( ) ) ) ; return model . getSize ( ) ; }
org . junit . Assert . assertEquals ( 0 , delegate . size ( ) )
testMissingMaxCacheSize ( ) { java . util . Map < java . lang . String , java . lang . Object > cacheConfig = com . google . common . collect . ImmutableMap . of ( CachingStellarProcessor . MAX_TIME_RETAIN_PARAM , 10 ) ; cache = org . apache . metron . stellar . common . CachingStellarProcessor . createCache ( cacheConfig ) ; "<AssertPlaceHolder>" ; } createCache ( java . util . Map ) { if ( config == null ) { org . apache . metron . stellar . common . CachingStellarProcessor . LOG . debug ( "Cannot<sp>create<sp>cache;<sp>missing<sp>cache<sp>configuration" ) ; return null ; } java . lang . Long maxSize = org . apache . metron . stellar . common . CachingStellarProcessor . getParam ( config , org . apache . metron . stellar . common . CachingStellarProcessor . MAX_CACHE_SIZE_PARAM , null , org . apache . metron . stellar . common . Long . class ) ; if ( ( maxSize == null ) || ( maxSize <= 0 ) ) { org . apache . metron . stellar . common . CachingStellarProcessor . LOG . error ( "Cannot<sp>create<sp>cache;<sp>missing<sp>or<sp>invalid<sp>configuration;<sp>{}<sp>=<sp>{}" , org . apache . metron . stellar . common . CachingStellarProcessor . MAX_CACHE_SIZE_PARAM , maxSize ) ; return null ; } java . lang . Integer maxTimeRetain = org . apache . metron . stellar . common . CachingStellarProcessor . getParam ( config , org . apache . metron . stellar . common . CachingStellarProcessor . MAX_TIME_RETAIN_PARAM , null , java . lang . Integer . class ) ; if ( ( maxTimeRetain == null ) || ( maxTimeRetain <= 0 ) ) { org . apache . metron . stellar . common . CachingStellarProcessor . LOG . error ( "Cannot<sp>create<sp>cache;<sp>missing<sp>or<sp>invalid<sp>configuration;<sp>{}<sp>=<sp>{}" , org . apache . metron . stellar . common . CachingStellarProcessor . MAX_TIME_RETAIN_PARAM , maxTimeRetain ) ; return null ; } com . github . benmanes . caffeine . cache . Caffeine < java . lang . Object , java . lang . Object > cache = com . github . benmanes . caffeine . cache . Caffeine . newBuilder ( ) . maximumSize ( maxSize ) . expireAfterWrite ( maxTimeRetain , TimeUnit . MINUTES ) ; java . lang . Boolean recordStats = org . apache . metron . stellar . common . CachingStellarProcessor . getParam ( config , org . apache . metron . stellar . common . CachingStellarProcessor . RECORD_STATS , false , java . lang . Boolean . class ) ; if ( recordStats ) { cache . recordStats ( ) ; } return cache . build ( ) ; }
org . junit . Assert . assertNull ( cache )
validate ( ) { com . behase . relumin . config . ReluminConfig config = new com . behase . relumin . config . ReluminConfig ( ) ; config . getServer ( ) . setPort ( "hoge" ) ; config . getServer ( ) . setMonitorPort ( "hoge" ) ; config . getAuth ( ) . setEnabled ( "hoge" ) ; config . getAuth ( ) . setAllowAnonymous ( "hoge" ) ; config . getRedis ( ) . setHost ( "" ) ; config . getRedis ( ) . setPort ( "hoge" ) ; config . getScheduler ( ) . setRefreshClustersIntervalMillis ( "hoge" ) ; config . getScheduler ( ) . setCollectStaticsInfoIntervalMillis ( "hoge" ) ; config . getScheduler ( ) . setCollectStaticsInfoMaxCount ( "hoge" ) ; config . getNotice ( ) . getMail ( ) . setPort ( "hoge" ) ; config . getOutputMetrics ( ) . getFluentd ( ) . setEnabled ( "hoge" ) ; config . getOutputMetrics ( ) . getFluentd ( ) . setPort ( "hoge" ) ; config . getOutputMetrics ( ) . getFluentd ( ) . setTimeout ( "hoge" ) ; config . getOutputMetrics ( ) . getFluentd ( ) . setBufferCapacity ( "hoge" ) ; try { config . validate ( ) ; } catch ( java . lang . Exception e ) { } "<AssertPlaceHolder>" ; } validate ( ) { com . behase . relumin . config . ReluminConfig config = new com . behase . relumin . config . ReluminConfig ( ) ; config . getServer ( ) . setPort ( "hoge" ) ; config . getServer ( ) . setMonitorPort ( "hoge" ) ; config . getAuth ( ) . setEnabled ( "hoge" ) ; config . getAuth ( ) . setAllowAnonymous ( "hoge" ) ; config . getRedis ( ) . setHost ( "" ) ; config . getRedis ( ) . setPort ( "hoge" ) ; config . getScheduler ( ) . setRefreshClustersIntervalMillis ( "hoge" ) ; config . getScheduler ( ) . setCollectStaticsInfoIntervalMillis ( "hoge" ) ; config . getScheduler ( ) . setCollectStaticsInfoMaxCount ( "hoge" ) ; config . getNotice ( ) . getMail ( ) . setPort ( "hoge" ) ; config . getOutputMetrics ( ) . getFluentd ( ) . setEnabled ( "hoge" ) ; config . getOutputMetrics ( ) . getFluentd ( ) . setPort ( "hoge" ) ; config . getOutputMetrics ( ) . getFluentd ( ) . setTimeout ( "hoge" ) ; config . getOutputMetrics ( ) . getFluentd ( ) . setBufferCapacity ( "hoge" ) ; try { config . validate ( ) ; } catch ( java . lang . Exception e ) { } org . junit . Assert . assertThat ( config . getErrors ( ) . size ( ) , org . hamcrest . CoreMatchers . is ( 14 ) ) ; }
org . junit . Assert . assertThat ( config . getErrors ( ) . size ( ) , org . hamcrest . CoreMatchers . is ( 14 ) )
testForString ( ) { java . net . InetAddress loopbackAddress = java . net . InetAddress . getLoopbackAddress ( ) ; java . lang . String uri = io . ebeaninternal . server . type . ConvertInetAddresses . toUriString ( loopbackAddress ) ; java . net . InetAddress inetAddress = io . ebeaninternal . server . type . ConvertInetAddresses . forString ( uri ) ; "<AssertPlaceHolder>" ; } forString ( java . lang . String ) { byte [ ] addr = io . ebeaninternal . server . type . ConvertInetAddresses . textToNumericFormatV4 ( ipString ) ; if ( addr == null ) { addr = io . ebeaninternal . server . type . ConvertInetAddresses . textToNumericFormatV6 ( ipString ) ; } if ( addr == null ) { throw new java . lang . IllegalArgumentException ( java . lang . String . format ( "'%s'<sp>is<sp>not<sp>an<sp>IP<sp>string<sp>literal." , ipString ) ) ; } try { return java . net . InetAddress . getByAddress ( addr ) ; } catch ( java . net . UnknownHostException e ) { throw new java . lang . IllegalArgumentException ( java . lang . String . format ( "'%s'<sp>is<sp>extremely<sp>broken." , ipString ) , e ) ; } }
org . junit . Assert . assertEquals ( loopbackAddress , inetAddress )
testSetGetMessage ( ) { fr . gouv . vitam . common . error . VitamErrorTest . vitamError . setMessage ( "2" ) ; "<AssertPlaceHolder>" ; } getMessage ( ) { return message ; }
org . junit . Assert . assertEquals ( "2" , fr . gouv . vitam . common . error . VitamErrorTest . vitamError . getMessage ( ) )
testResolveDataPolicyWhenPersistentAndRegionAttributesPersistentPartitionDataPolicy ( ) { org . apache . geode . cache . RegionFactory mockRegionFactory = createMockRegionFactory ( ) ; factoryBean . setAttributes ( createMockRegionAttributes ( DataPolicy . PERSISTENT_PARTITION ) ) ; factoryBean . setDataPolicy ( ( ( org . apache . geode . cache . DataPolicy ) ( null ) ) ) ; factoryBean . setPersistent ( true ) ; factoryBean . resolveDataPolicy ( mockRegionFactory , true , ( ( java . lang . String ) ( null ) ) ) ; verify ( mockRegionFactory , times ( 1 ) ) . setDataPolicy ( eq ( DataPolicy . PERSISTENT_PARTITION ) ) ; "<AssertPlaceHolder>" ; } getDataPolicy ( ) { return this . dataPolicy ; }
org . junit . Assert . assertEquals ( DataPolicy . PERSISTENT_PARTITION , factoryBean . getDataPolicy ( ) )
testBetrokkenheidActueelIndicatieOuderUitWieKindIsGeboren ( ) { final java . util . Set < nl . bzk . brp . domain . algemeen . ZoekCriterium > zoekCriteria = new java . util . HashSet ( ) ; nl . bzk . brp . domain . algemeen . ZoekCriterium zoekCriteria1 = new nl . bzk . brp . domain . algemeen . ZoekCriterium ( getAttribuutElement ( Element . PERSOON_IDENTIFICATIENUMMERS_BURGERSERVICENUMMER ) , nl . bzk . algemeenbrp . dal . domein . brp . enums . Zoekoptie . EXACT , "402533930" ) ; nl . bzk . brp . domain . algemeen . ZoekCriterium zoekCriteria2 = new nl . bzk . brp . domain . algemeen . ZoekCriterium ( getAttribuutElement ( Element . PERSOON_OUDER_OUDERSCHAP_INDICATIEOUDERUITWIEKINDISGEBOREN ) , nl . bzk . algemeenbrp . dal . domein . brp . enums . Zoekoptie . EXACT , true ) ; zoekCriteria . add ( zoekCriteria1 ) ; zoekCriteria . add ( zoekCriteria2 ) ; nl . bzk . brp . delivery . dataaccess . bevraging . SqlStamementZoekPersoon sql = new nl . bzk . brp . delivery . dataaccess . bevraging . SqlBepaler ( zoekCriteria , 10 , false , null , false ) . maakSql ( ) ; final java . util . List < java . lang . Long > ids = zoekPersoonRepository . zoekPersonen ( sql , postgres ) ; "<AssertPlaceHolder>" ; } size ( ) { return elementen . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , ids . size ( ) )
testSerialization ( ) { org . jfree . data . time . ohlc . OHLC i1 = new org . jfree . data . time . ohlc . OHLC ( 2.0 , 4.0 , 1.0 , 3.0 ) ; java . io . ByteArrayOutputStream buffer = new java . io . ByteArrayOutputStream ( ) ; java . io . ObjectOutput out = new java . io . ObjectOutputStream ( buffer ) ; out . writeObject ( i1 ) ; out . close ( ) ; java . io . ObjectInput in = new java . io . ObjectInputStream ( new java . io . ByteArrayInputStream ( buffer . toByteArray ( ) ) ) ; org . jfree . data . time . ohlc . OHLC i2 = ( ( org . jfree . data . time . ohlc . OHLC ) ( in . readObject ( ) ) ) ; in . close ( ) ; "<AssertPlaceHolder>" ; } close ( ) { try { this . connection . close ( ) ; } catch ( java . lang . Exception e ) { System . err . println ( "JdbcXYDataset:<sp>swallowing<sp>exception." ) ; } }
org . junit . Assert . assertEquals ( i1 , i2 )
shouldNotCancelKick ( ) { given ( settings . getProperty ( RestrictionSettings . FORCE_SINGLE_SESSION ) ) . willReturn ( false ) ; java . lang . String name = "Bobby" ; org . bukkit . entity . Player player = fr . xephi . authme . listener . PlayerListenerTest . mockPlayerWithName ( name ) ; fr . xephi . authme . listener . PlayerKickEvent event = new fr . xephi . authme . listener . PlayerKickEvent ( player , "You<sp>logged<sp>in<sp>from<sp>another<sp>location" , "" ) ; given ( antiBotService . wasPlayerKicked ( name ) ) . willReturn ( false ) ; listener . onPlayerKick ( event ) ; "<AssertPlaceHolder>" ; verify ( antiBotService ) . wasPlayerKicked ( name ) ; verify ( management ) . performQuit ( player ) ; } isCancelled ( ) { return isCancelled ; }
org . junit . Assert . assertThat ( event . isCancelled ( ) , org . hamcrest . Matchers . equalTo ( false ) )
testIsNamedTrueWhenNameSet ( ) { com . ocpsoft . pretty . faces . config . mapping . PathParameter parameter = new com . ocpsoft . pretty . faces . config . mapping . PathParameter ( ) ; parameter . setName ( "name" ) ; "<AssertPlaceHolder>" ; } isNamed ( ) { return ( null != ( super . getName ( ) ) ) && ( ! ( "" . equals ( super . getName ( ) . trim ( ) ) ) ) ; }
org . junit . Assert . assertTrue ( parameter . isNamed ( ) )
testDoOnNth ( ) { java . util . concurrent . atomic . AtomicInteger item = new java . util . concurrent . atomic . AtomicInteger ( ) ; rx . Observable . just ( 1 , 2 , 3 ) . compose ( com . github . davidmoten . rx . Transformers . doOnNext ( 2 , com . github . davidmoten . rx . Actions . setAtomic ( item ) ) ) . subscribe ( ) ; "<AssertPlaceHolder>" ; } get ( ) { if ( present ) return value ; else throw new com . github . davidmoten . util . Optional . NotPresentException ( ) ; }
org . junit . Assert . assertEquals ( 2 , item . get ( ) )
testExplicitType ( ) { factoryBean . setLocation ( SSLTestConstants . KEYSTORE_JKS_RESOURCE ) ; factoryBean . setType ( SSL . DEFAULT_KEYSTORE_TYPE ) ; "<AssertPlaceHolder>" ; } createKeyStore ( ) { if ( ( getLocation ( ) ) == null ) { throw new java . lang . IllegalArgumentException ( "location<sp>is<sp>required" ) ; } java . io . InputStream inputStream = null ; try { java . net . URL url = ch . qos . logback . core . util . LocationUtil . urlForResource ( getLocation ( ) ) ; inputStream = url . openStream ( ) ; java . security . KeyStore keyStore = newKeyStore ( ) ; keyStore . load ( inputStream , getPassword ( ) . toCharArray ( ) ) ; return keyStore ; } catch ( java . security . NoSuchProviderException ex ) { throw new java . security . NoSuchProviderException ( ( "no<sp>such<sp>keystore<sp>provider:<sp>" + ( getProvider ( ) ) ) ) ; } catch ( java . security . NoSuchAlgorithmException ex ) { throw new java . security . NoSuchAlgorithmException ( ( "no<sp>such<sp>keystore<sp>type:<sp>" + ( getType ( ) ) ) ) ; } catch ( java . io . FileNotFoundException ex ) { throw new java . security . KeyStoreException ( ( ( getLocation ( ) ) + ":<sp>file<sp>not<sp>found" ) ) ; } catch ( java . lang . Exception ex ) { throw new java . security . KeyStoreException ( ( ( ( getLocation ( ) ) + ":<sp>" ) + ( ex . getMessage ( ) ) ) , ex ) ; } finally { try { if ( inputStream != null ) { inputStream . close ( ) ; } } catch ( java . io . IOException ex ) { ex . printStackTrace ( System . err ) ; } } }
org . junit . Assert . assertNotNull ( factoryBean . createKeyStore ( ) )