input
stringlengths
28
18.7k
output
stringlengths
39
1.69k
testFindByPrimaryKeyExisting ( ) { com . liferay . changeset . model . ChangesetCollection newChangesetCollection = addChangesetCollection ( ) ; com . liferay . changeset . model . ChangesetCollection existingChangesetCollection = _persistence . findByPrimaryKey ( newChangesetCollection . getPrimaryKey ( ) ) ; "<AssertPlaceHolder>" ; } getPrimaryKey ( ) { return _amImageEntryId ; }
org . junit . Assert . assertEquals ( existingChangesetCollection , newChangesetCollection )
testResolveMethod3 ( ) { java . lang . Class < ? > declareClass = java . lang . Class . forName ( "java.lang.Math" ) ; java . lang . String methodName = "abd" ; java . lang . Class < ? > [ ] paramTypes = new java . lang . Class < ? > [ ] { com . huawei . streaming . expression . Integer . class } ; java . lang . reflect . Method actualMethod = com . huawei . streaming . expression . MethodResolver . resolveMethod ( declareClass , methodName , paramTypes ) ; java . lang . reflect . Method expectedMethod = null ; "<AssertPlaceHolder>" ; } resolveMethod ( java . lang . Class , java . lang . String , com . huawei . streaming . expression . Class [ ] ) { java . lang . reflect . Method [ ] declaredMethods = declareClass . getDeclaredMethods ( ) ; java . lang . reflect . Method bestMatchMethod = null ; int bestConversionCount = - 1 ; for ( java . lang . reflect . Method method : declaredMethods ) { if ( ! ( com . huawei . streaming . expression . MethodResolver . isPublic ( method ) ) ) { continue ; } if ( ! ( method . getName ( ) . equals ( methodName ) ) ) { continue ; } int conversionCount = com . huawei . streaming . expression . MethodResolver . compareParamTypes ( method . getParameterTypes ( ) , paramTypes ) ; if ( conversionCount == ( - 1 ) ) { continue ; } if ( conversionCount == 0 ) { bestMatchMethod = method ; break ; } if ( null == bestMatchMethod ) { bestMatchMethod = method ; bestConversionCount = conversionCount ; } else { if ( conversionCount < bestConversionCount ) { bestMatchMethod = method ; bestConversionCount = conversionCount ; } } } return bestMatchMethod ; }
org . junit . Assert . assertEquals ( expectedMethod , actualMethod )
trunkatedXref ( ) { try ( org . sejda . sambox . pdmodel . PDDocument doc = org . sejda . sambox . input . PDFParser . parse ( org . sejda . io . SeekableSources . inMemorySeekableSourceFrom ( getClass ( ) . getResourceAsStream ( "/sambox/test_trunkated_xref_table.pdf" ) ) ) ) { "<AssertPlaceHolder>" ; } } getPage ( int ) { return getDocumentCatalog ( ) . getPages ( ) . get ( pageIndex ) ; }
org . junit . Assert . assertNotNull ( doc . getPage ( 0 ) )
testLucaSize ( ) { System . out . println ( "testLucaSize" ) ; org . roaringbitmap . RoaringBitmap rb = org . roaringbitmap . RoaringBitmap . bitmapOf ( 2946000 , 2997491 , 10478289 , 10490227 , 10502444 , 19866827 ) ; System . out . println ( ( "cardinality<sp>=<sp>" + ( rb . getCardinality ( ) ) ) ) ; System . out . println ( ( "total<sp>size<sp>in<sp>bytes<sp>=<sp>" + ( rb . getSizeInBytes ( ) ) ) ) ; "<AssertPlaceHolder>" ; } getSizeInBytes ( ) { return ( ( int ) ( getLongSizeInBytes ( ) ) ) ; }
org . junit . Assert . assertTrue ( ( ( rb . getSizeInBytes ( ) ) <= 50 ) )
testSetAlpha ( ) { gc . setAlpha ( 123 ) ; "<AssertPlaceHolder>" ; } getAlpha ( ) { checkWidget ( ) ; return alpha ; }
org . junit . Assert . assertEquals ( 123 , gc . getAlpha ( ) )
appendExtensionPostAssignSourceCode_A$StringBuilder$String$StringArray$String_StringIsEmpty ( ) { org . junithelper . core . generator . LineBreakProvider lineBreakProvider = new org . junithelper . core . generator . LineBreakProvider ( config , currentLineBreak ) ; org . junithelper . core . generator . SourceCodeAppender target = new org . junithelper . core . generator . SourceCodeAppender ( lineBreakProvider , indentationProvider ) ; java . lang . StringBuilder buf = new java . lang . StringBuilder ( ) ; java . lang . String code = "" ; java . lang . String [ ] fromList = new java . lang . String [ ] { } ; java . lang . String to = "" ; target . appendExtensionPostAssignSourceCode ( buf , code , fromList , to ) ; "<AssertPlaceHolder>" ; } toString ( ) { return this . name ; }
org . junit . Assert . assertThat ( buf . toString ( ) , org . hamcrest . CoreMatchers . is ( org . hamcrest . CoreMatchers . equalTo ( "" ) ) )
testGetStencilsetRef ( ) { org . oryxeditor . server . diagram . StencilSetReference ssr = new org . oryxeditor . server . diagram . StencilSetReference ( "aNamespace" ) ; diagram2 . setStencilsetRef ( ssr ) ; "<AssertPlaceHolder>" ; } getStencilsetRef ( ) { return stencilsetRef ; }
org . junit . Assert . assertEquals ( ssr , diagram2 . getStencilsetRef ( ) )
testGeefAttribuutElementen ( ) { java . util . Collection < nl . bzk . brp . domain . element . AttribuutElement > alleAttribuutElementen = modelAanwijzer . geefAttribuutElementen ( ) ; "<AssertPlaceHolder>" ; } isEmpty ( ) { return elementen . isEmpty ( ) ; }
org . junit . Assert . assertFalse ( alleAttribuutElementen . isEmpty ( ) )
testOr ( ) { org . apache . druid . collections . bitmap . WrappedBitSetBitmap bitSet = org . apache . druid . collections . bitmap . WrappedBitSetBitmapBitSetTest . defaultBitSet ( ) ; org . apache . druid . collections . bitmap . WrappedBitSetBitmap bitSet2 = org . apache . druid . collections . bitmap . WrappedBitSetBitmapBitSetTest . defaultBitSet ( ) ; java . util . Set < java . lang . Integer > defaultBitSet = org . apache . druid . collections . IntSetTestUtility . getSetBits ( ) ; bitSet . remove ( 1 ) ; bitSet2 . remove ( 2 ) ; bitSet . or ( bitSet2 ) ; "<AssertPlaceHolder>" ; } equalSets ( java . util . Set , org . apache . druid . collections . bitmap . ImmutableBitmap ) { java . util . Set < java . lang . Integer > s3 = new java . util . HashSet ( ) ; for ( java . lang . Integer i : new org . apache . druid . collections . IntSetTestUtility . IntIt ( s2 . iterator ( ) ) ) { s3 . add ( i ) ; } return com . google . common . collect . Sets . difference ( s1 , s3 ) . isEmpty ( ) ; }
org . junit . Assert . assertTrue ( org . apache . druid . collections . IntSetTestUtility . equalSets ( defaultBitSet , bitSet ) )
testCreateCollectionValueSuccessfulThree ( ) { final com . orientechnologies . orient . core . index . OCompositeIndexDefinition compositeIndexDefinition = new com . orientechnologies . orient . core . index . OCompositeIndexDefinition ( "testCollectionClass" ) ; compositeIndexDefinition . addIndex ( new com . orientechnologies . orient . core . index . OPropertyIndexDefinition ( "testCollectionClass" , "fOne" , com . orientechnologies . orient . core . metadata . schema . OType . INTEGER ) ) ; compositeIndexDefinition . addIndex ( new com . orientechnologies . orient . core . index . OPropertyListIndexDefinition ( "testCollectionClass" , "fTwo" , com . orientechnologies . orient . core . metadata . schema . OType . INTEGER ) ) ; compositeIndexDefinition . addIndex ( new com . orientechnologies . orient . core . index . OPropertyIndexDefinition ( "testCollectionClass" , "fThree" , com . orientechnologies . orient . core . metadata . schema . OType . STRING ) ) ; final java . lang . Object result = compositeIndexDefinition . createValue ( 12 , com . orientechnologies . orient . core . index . Arrays . asList ( 1 , 2 ) , "test" ) ; final com . orientechnologies . orient . core . index . ArrayList < com . orientechnologies . orient . core . index . OCompositeKey > expectedResult = new com . orientechnologies . orient . core . index . ArrayList < com . orientechnologies . orient . core . index . OCompositeKey > ( ) ; expectedResult . add ( new com . orientechnologies . orient . core . index . OCompositeKey ( 12 , 1 , "test" ) ) ; expectedResult . add ( new com . orientechnologies . orient . core . index . OCompositeKey ( 12 , 2 , "test" ) ) ; "<AssertPlaceHolder>" ; } add ( long , int , com . orientechnologies . orient . core . storage . impl . local . paginated . atomicoperations . OAtomicOperation ) { final long lastPage = ( getFilledUpTo ( atomicOperation , fileId ) ) - 1 ; com . orientechnologies . orient . core . storage . cache . OCacheEntry cacheEntry ; boolean clear = false ; if ( lastPage < 0 ) { cacheEntry = addPage ( atomicOperation , fileId ) ; clear = true ; } else { cacheEntry = loadPageForWrite ( atomicOperation , fileId , lastPage , false , true ) ; } try { com . orientechnologies . orient . core . storage . cluster . OClusterPositionMapBucket bucket = new com . orientechnologies . orient . core . storage . cluster . OClusterPositionMapBucket ( cacheEntry , clear ) ; if ( bucket . isFull ( ) ) { releasePageFromWrite ( atomicOperation , cacheEntry ) ; cacheEntry = addPage ( atomicOperation , fileId ) ; bucket = new com . orientechnologies . orient . core . storage . cluster . OClusterPositionMapBucket ( cacheEntry , true ) ; } final long index = bucket . add ( pageIndex , recordPosition ) ; return index + ( ( cacheEntry . getPageIndex ( ) ) * ( com . orientechnologies . orient . core . storage . cluster . OClusterPositionMapBucket . MAX_ENTRIES ) ) ; } finally { releasePageFromWrite ( atomicOperation , cacheEntry ) ; } }
org . junit . Assert . assertEquals ( result , expectedResult )
testAutocastLong ( ) { org . openl . meta . BigIntegerValue expectedResult = new org . openl . meta . BigIntegerValue ( "1234" ) ; org . openl . meta . BigIntegerValue result = org . openl . meta . BigIntegerValue . autocast ( ( ( long ) ( 1234 ) ) , null ) ; "<AssertPlaceHolder>" ; } autocast ( byte , org . openl . meta . BigIntegerValue ) { return new org . openl . meta . BigIntegerValue ( java . lang . String . valueOf ( x ) ) ; }
org . junit . Assert . assertEquals ( expectedResult , result )
testLogViewNotNull ( ) { com . aliyun . odps . LogView log = odps . logview ( ) ; System . out . println ( log . getLogViewHost ( ) ) ; "<AssertPlaceHolder>" ; } getLogViewHost ( ) { return this . logViewHost ; }
org . junit . Assert . assertNotNull ( log . getLogViewHost ( ) )
testAddConnection ( ) { org . pentaho . database . model . DatabaseConnection connection = new org . pentaho . database . model . DatabaseConnection ( ) ; connection . setName ( org . pentaho . platform . dataaccess . datasource . wizard . service . impl . ConnectionServiceImplIT . SECOND_CONNECTION ) ; org . pentaho . platform . dataaccess . datasource . wizard . service . impl . ConnectionServiceImplIT . connectionServiceImpl . addConnection ( connection ) ; org . pentaho . database . model . IDatabaseConnection actualCnnection = org . pentaho . platform . dataaccess . datasource . wizard . service . impl . ConnectionServiceImplIT . connectionServiceImpl . getConnectionByName ( org . pentaho . platform . dataaccess . datasource . wizard . service . impl . ConnectionServiceImplIT . SECOND_CONNECTION ) ; "<AssertPlaceHolder>" ; } getConnectionByName ( java . lang . String ) { for ( org . pentaho . database . model . IDatabaseConnection connection : connectionList ) { if ( connection . getName ( ) . equals ( name ) ) { return connection ; } } org . pentaho . platform . dataaccess . datasource . wizard . service . impl . InMemoryConnectionServiceImpl . logger . error ( org . pentaho . platform . dataaccess . datasource . wizard . service . messages . Messages . getErrorString ( "ConnectionServiceInMemoryDelegate.ERROR_0003_UNABLE_TO_GET_CONNECTION" , name , null ) ) ; throw new org . pentaho . platform . dataaccess . datasource . wizard . service . ConnectionServiceException ( org . pentaho . platform . dataaccess . datasource . wizard . service . messages . Messages . getErrorString ( "ConnectionServiceInMemoryDelegate.ERROR_0003_UNABLE_TO_GET_CONNECTION" , name , null ) ) ; }
org . junit . Assert . assertEquals ( connection , actualCnnection )
testFieldByte ( ) { @ org . simpleflatmapper . reflect . test . asm . SuppressWarnings ( "unchecked" ) org . simpleflatmapper . reflect . primitive . ByteSetter < org . simpleflatmapper . test . beans . DbPrimitiveObjectFields > setter = ( ( org . simpleflatmapper . reflect . primitive . ByteSetter < org . simpleflatmapper . test . beans . DbPrimitiveObjectFields > ) ( factory . createSetter ( org . simpleflatmapper . test . beans . DbPrimitiveObjectFields . class . getDeclaredField ( "pByte" ) ) ) ) ; setter . setByte ( objectField , ( ( byte ) ( 195 ) ) ) ; "<AssertPlaceHolder>" ; } getpByte ( ) { return pByte ; }
org . junit . Assert . assertEquals ( ( ( byte ) ( 195 ) ) , objectField . getpByte ( ) )
getFileItems_noItems ( ) { "<AssertPlaceHolder>" ; } getFileItems ( ) { java . util . List < org . apache . commons . fileupload . FileItem > items = createLinkedList ( ) ; for ( java . lang . Object value : values ) { if ( value instanceof org . apache . commons . fileupload . FileItem ) { items . add ( ( ( org . apache . commons . fileupload . FileItem ) ( value ) ) ) ; } } return items . toArray ( new org . apache . commons . fileupload . FileItem [ items . size ( ) ] ) ; }
org . junit . Assert . assertArrayEquals ( new org . apache . commons . fileupload . FileItem [ 0 ] , values . getFileItems ( ) )
testGetUniprotName ( ) { uk . ac . ebi . ep . ebeye . protein . model . EntryTest . LOGGER . info ( "getUniprotName" ) ; java . lang . String expResult = "ABC" ; java . lang . String result = instance . getUniprotName ( ) ; "<AssertPlaceHolder>" ; } getUniprotName ( ) { return uniprotName . substring ( 0 , uniprotName . indexOf ( "_" ) ) ; }
org . junit . Assert . assertEquals ( expResult , result )
testSourceTableRowExists ( ) { org . apache . hadoop . hbase . client . Result result = mock ( org . apache . hadoop . hbase . client . Result . class ) ; when ( result . isEmpty ( ) ) . thenReturn ( false ) ; org . pentaho . hbase . factory . HBaseTable table = mock ( org . pentaho . hbase . factory . HBaseTable . class ) ; when ( table . get ( ( ( org . apache . hadoop . hbase . client . Get ) ( any ( ) ) ) ) ) . thenReturn ( result ) ; commonHBaseConnection . m_sourceTable = table ; boolean exists = commonHBaseConnection . sourceTableRowExists ( new byte [ ] { 1 , 2 , 3 } ) ; "<AssertPlaceHolder>" ; } get ( org . apache . hadoop . hbase . client . Get ) { return tab . get ( toGet ) ; }
org . junit . Assert . assertTrue ( exists )
test01WindowSimple ( ) { final com . jogamp . opengl . test . junit . newt . GLCapabilities caps = new com . jogamp . opengl . test . junit . newt . GLCapabilities ( com . jogamp . opengl . test . junit . newt . TestGLWindows01NEWT . glp ) ; "<AssertPlaceHolder>" ; final com . jogamp . opengl . test . junit . newt . GLWindow window = com . jogamp . opengl . test . junit . newt . TestGLWindows01NEWT . createWindow ( null , caps , com . jogamp . opengl . test . junit . newt . TestGLWindows01NEWT . width , com . jogamp . opengl . test . junit . newt . TestGLWindows01NEWT . height , true , false ) ; System . out . println ( ( "Created:<sp>" + window ) ) ; int state ; for ( state = 0 ; ( state * 100 ) < ( com . jogamp . opengl . test . junit . newt . TestGLWindows01NEWT . durationPerTest ) ; state ++ ) { java . lang . Thread . sleep ( 100 ) ; } System . out . println ( ( "duration:<sp>" + ( window . getTotalFPSDuration ( ) ) ) ) ; com . jogamp . opengl . test . junit . newt . TestGLWindows01NEWT . destroyWindow ( window ) ; }
org . junit . Assert . assertNotNull ( caps )
setValueMessage ( ) { org . openhab . binding . zwave . internal . protocol . commandclass . ZWaveBasicCommandClass cls = ( ( org . openhab . binding . zwave . internal . protocol . commandclass . ZWaveBasicCommandClass ) ( getCommandClass ( CommandClass . COMMAND_CLASS_BASIC ) ) ) ; org . openhab . binding . zwave . internal . protocol . transaction . ZWaveCommandClassTransactionPayload msg ; byte [ ] expectedResponseV1 = new byte [ ] { 32 , 1 , 34 } ; cls . setVersion ( 1 ) ; msg = cls . setValueMessage ( 34 ) ; "<AssertPlaceHolder>" ; } getPayloadBuffer ( ) { return payload ; }
org . junit . Assert . assertTrue ( java . util . Arrays . equals ( msg . getPayloadBuffer ( ) , expectedResponseV1 ) )
testCreateBizkeeperCommand ( ) { com . netflix . hystrix . strategy . HystrixPlugins . reset ( ) ; org . apache . servicecomb . bizkeeper . ConsumerBizkeeperHandler consumerBizkeeperHandler = new org . apache . servicecomb . bizkeeper . ConsumerBizkeeperHandler ( ) ; org . apache . servicecomb . core . Invocation invocation = org . mockito . Mockito . mock ( org . apache . servicecomb . core . Invocation . class ) ; org . mockito . Mockito . when ( invocation . getOperationMeta ( ) ) . thenReturn ( org . mockito . Mockito . mock ( org . apache . servicecomb . core . definition . OperationMeta . class ) ) ; org . mockito . Mockito . when ( invocation . getOperationMeta ( ) . getMicroserviceQualifiedName ( ) ) . thenReturn ( "test1" ) ; org . apache . servicecomb . bizkeeper . CommandKey . toHystrixCommandGroupKey ( "groupname" , invocation ) ; org . apache . servicecomb . bizkeeper . CommandKey . toHystrixCommandKey ( "groupname" , invocation ) ; org . apache . servicecomb . bizkeeper . BizkeeperCommand command = consumerBizkeeperHandler . createBizkeeperCommand ( invocation ) ; "<AssertPlaceHolder>" ; } createBizkeeperCommand ( org . apache . servicecomb . core . Invocation ) { if ( Configuration . INSTANCE . isFallbackForce ( handler . groupname , invocation . getMicroserviceName ( ) , invocation . getOperationMeta ( ) . getMicroserviceQualifiedName ( ) ) ) { return forceFallbackCommand ( invocation ) ; } return handler . createBizkeeperCommand ( invocation ) ; }
org . junit . Assert . assertNotNull ( command )
testWorldCreate ( ) { com . turt2live . antishare . object . AWorld world = mock ( com . turt2live . antishare . object . AWorld . class ) ; com . turt2live . antishare . object . ASLocation location = new com . turt2live . antishare . object . ASLocation ( world , 0 , 0 , 0 ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertEquals ( world , location . world )
refinementStringEquals ( ) { generateDrugHierarchy ( ) ; final com . b2international . index . query . Expression actual = eval ( java . lang . String . format ( "<%s:<sp>%s<sp>=<sp>'PANADOL'" , com . b2international . snowowl . snomed . core . ecl . SnomedEclEvaluationRequestTest . DRUG_ROOT , com . b2international . snowowl . snomed . core . ecl . SnomedEclEvaluationRequestTest . HAS_TRADE_NAME ) ) ; final com . b2international . index . query . Expression expected = com . b2international . snowowl . snomed . core . ecl . SnomedEclEvaluationRequestTest . and ( descendantsOf ( com . b2international . snowowl . snomed . core . ecl . SnomedEclEvaluationRequestTest . DRUG_ROOT ) , ids ( com . google . common . collect . ImmutableSet . of ( com . b2international . snowowl . snomed . core . ecl . SnomedEclEvaluationRequestTest . PANADOL_TABLET ) ) ) ; "<AssertPlaceHolder>" ; } of ( com . b2international . commons . exceptions . ApiError ) { return new com . b2international . snowowl . snomed . api . rest . domain . RestApiError . Builder ( error ) ; }
org . junit . Assert . assertEquals ( expected , actual )
staticFromStringHelper ( ) { org . apache . accumulo . core . replication . ReplicationTarget expected = new org . apache . accumulo . core . replication . ReplicationTarget ( "foo" , "bar" , org . apache . accumulo . core . data . TableId . of ( "1" ) ) ; org . apache . hadoop . io . DataOutputBuffer buffer = new org . apache . hadoop . io . DataOutputBuffer ( ) ; expected . write ( buffer ) ; org . apache . hadoop . io . Text t = new org . apache . hadoop . io . Text ( ) ; t . set ( buffer . getData ( ) , 0 , buffer . getLength ( ) ) ; "<AssertPlaceHolder>" ; } from ( org . apache . hadoop . io . Text ) { org . apache . accumulo . core . replication . ReplicationTarget target = new org . apache . accumulo . core . replication . ReplicationTarget ( ) ; org . apache . hadoop . io . DataInputBuffer buffer = new org . apache . hadoop . io . DataInputBuffer ( ) ; buffer . reset ( t . getBytes ( ) , t . getLength ( ) ) ; try { target . readFields ( buffer ) ; } catch ( java . io . IOException e ) { throw new java . lang . RuntimeException ( e ) ; } return target ; }
org . junit . Assert . assertEquals ( expected , org . apache . accumulo . core . replication . ReplicationTarget . from ( t . toString ( ) ) )
testMark ( ) { rtpPacket . wrap ( true , 8 , 27244 , 3249239181L , 3001189225L , p , 12 , ( ( p . length ) - 12 ) ) ; "<AssertPlaceHolder>" ; } getMarker ( ) { return ( ( ( buffer . get ( 1 ) ) & 255 ) & 128 ) == 128 ; }
org . junit . Assert . assertEquals ( true , rtpPacket . getMarker ( ) )
givenRegexWithDefaultMatcher_whenMatchFailsOnDifferentCases_thenCorrect ( ) { int matches = com . baeldung . java . regex . RegexUnitTest . runTest ( "dog" , "This<sp>is<sp>a<sp>Dog" ) ; "<AssertPlaceHolder>" ; } runTest ( java . lang . String , java . lang . String ) { com . baeldung . java . regex . RegexUnitTest . pattern = java . util . regex . Pattern . compile ( regex ) ; com . baeldung . java . regex . RegexUnitTest . matcher = com . baeldung . java . regex . RegexUnitTest . pattern . matcher ( text ) ; int matches = 0 ; while ( com . baeldung . java . regex . RegexUnitTest . matcher . find ( ) ) matches ++ ; return matches ; }
org . junit . Assert . assertFalse ( ( matches > 0 ) )
testOverrideMinDelayBetweenRequests ( java . time . Duration ) { com . orbitz . consul . config . CacheConfig config = com . orbitz . consul . config . CacheConfig . builder ( ) . withMinDelayBetweenRequests ( delayBetweenRequests ) . build ( ) ; "<AssertPlaceHolder>" ; } getMinimumDurationBetweenRequests ( ) { return minDelayBetweenRequests ; }
org . junit . Assert . assertEquals ( delayBetweenRequests , config . getMinimumDurationBetweenRequests ( ) )
unicodeFormFieldTest ( ) { java . lang . String filename = ( com . itextpdf . forms . PdfFormFieldTest . sourceFolder ) + "unicodeFormFieldFile.pdf" ; com . itextpdf . kernel . pdf . PdfDocument pdfDoc = new com . itextpdf . kernel . pdf . PdfDocument ( new com . itextpdf . kernel . pdf . PdfReader ( filename ) ) ; com . itextpdf . forms . PdfAcroForm form = com . itextpdf . forms . PdfAcroForm . getAcroForm ( pdfDoc , true ) ; java . util . Map < java . lang . String , com . itextpdf . forms . fields . PdfFormField > formFields = form . getFormFields ( ) ; java . lang . String fieldName = "帐号1" ; "<AssertPlaceHolder>" ; } keySet ( ) { return map . keySet ( ) ; }
org . junit . Assert . assertEquals ( fieldName , formFields . keySet ( ) . toArray ( new java . lang . String [ 1 ] ) [ 0 ] )
testInvalidParameter ( ) { final net . violet . platform . api . actions . Action theAction = new net . violet . platform . api . actions . stores . Update ( ) ; final net . violet . platform . api . callers . APICaller caller = getPublicApplicationAPICaller ( ) ; final java . util . Map < java . lang . String , java . lang . Object > theParams = new java . util . HashMap < java . lang . String , java . lang . Object > ( ) ; theParams . put ( "language" , "JPFR" ) ; final net . violet . platform . api . actions . ActionParam theActionParam = new net . violet . platform . api . actions . ActionParam ( caller , theParams ) ; final java . lang . Object theResult = theAction . processRequest ( theActionParam ) ; "<AssertPlaceHolder>" ; } put ( K , T ) { this . mMap . put ( theRef , new net . violet . db . cache . CacheReference < K , T > ( theRef , theRecord , this . mReferenceQueue ) ) ; this . mLinkedMap . put ( theRef , theRecord ) ; }
org . junit . Assert . assertNotNull ( theResult )
testGetRepubBirthCertOnModemDetectionPropNotBoolean ( ) { java . util . Map < java . lang . String , java . lang . Object > properties = new java . util . HashMap ( ) ; properties . put ( org . eclipse . kura . core . cloud . CloudServiceOptionsTest . REPUB_BIRTH_ON_MODEM_DETECT , "invalid" ) ; org . eclipse . kura . core . cloud . CloudServiceOptions options = new org . eclipse . kura . core . cloud . CloudServiceOptions ( properties , org . eclipse . kura . core . cloud . CloudServiceOptionsTest . systemService ) ; boolean republish = options . getRepubBirthCertOnModemDetection ( ) ; "<AssertPlaceHolder>" ; } getRepubBirthCertOnModemDetection ( ) { boolean repubBirth = false ; if ( ( ( this . properties ) != null ) && ( ( this . properties . get ( org . eclipse . kura . core . cloud . CloudServiceOptions . REPUB_BIRTH_ON_MODEM_DETECT ) ) instanceof java . lang . Boolean ) ) { repubBirth = ( ( java . lang . Boolean ) ( this . properties . get ( org . eclipse . kura . core . cloud . CloudServiceOptions . REPUB_BIRTH_ON_MODEM_DETECT ) ) ) ; } return repubBirth ; }
org . junit . Assert . assertFalse ( republish )
testNoCompressionCodec ( ) { com . streamsets . pipeline . stage . destination . hdfs . HdfsDTarget dTarget = new com . streamsets . pipeline . stage . destination . hdfs . BaseHdfsTargetIT . ForTestHdfsTarget ( ) ; configure ( dTarget ) ; com . streamsets . pipeline . stage . destination . hdfs . HdfsTarget target = ( ( com . streamsets . pipeline . stage . destination . hdfs . HdfsTarget ) ( dTarget . createTarget ( ) ) ) ; try { com . streamsets . pipeline . api . Target . Context context = com . streamsets . pipeline . sdk . ContextInfoCreator . createTargetContext ( com . streamsets . pipeline . stage . destination . hdfs . HdfsDTarget . class , "n" , false , OnRecordError . TO_ERROR , null ) ; target . init ( null , context ) ; "<AssertPlaceHolder>" ; } finally { target . destroy ( ) ; } } getCompressionCodec ( ) { return compressionCodec ; }
org . junit . Assert . assertNull ( target . getCompressionCodec ( ) )
testGetSymbolFirstAtom ( ) { org . openscience . smsd . helper . BondEnergy instance = new org . openscience . smsd . helper . BondEnergy ( "H" , "I" , org . openscience . cdk . interfaces . IBond . Order . SINGLE , 295 ) ; java . lang . String expResult = "H" ; java . lang . String result = instance . getSymbolFirstAtom ( ) ; "<AssertPlaceHolder>" ; } getSymbolFirstAtom ( ) { return symbol1 ; }
org . junit . Assert . assertEquals ( expResult , result )
testSync ( ) { try { org . apache . zookeeper . test . LOG . info ( ( "Starting<sp>ZK:" + ( new java . util . Date ( ) . toString ( ) ) ) ) ; opsCount = new java . util . concurrent . CountDownLatch ( limit ) ; org . apache . zookeeper . ZooKeeper zk = createClient ( ) ; org . apache . zookeeper . test . LOG . info ( ( "Beginning<sp>test:" + ( new java . util . Date ( ) . toString ( ) ) ) ) ; for ( int i = 0 ; i < 100 ; i ++ ) zk . create ( ( "/test" + i ) , new byte [ 0 ] , Ids . OPEN_ACL_UNSAFE , CreateMode . PERSISTENT , this , results ) ; zk . sync ( "/test" , this , results ) ; for ( int i = 0 ; i < 100 ; i ++ ) zk . delete ( ( "/test" + i ) , 0 , this , results ) ; for ( int i = 0 ; i < 100 ; i ++ ) zk . getChildren ( "/" , new org . apache . zookeeper . test . NullWatcher ( ) , ( ( org . apache . zookeeper . AsyncCallback . ChildrenCallback ) ( this ) ) , results ) ; for ( int i = 0 ; i < 100 ; i ++ ) zk . getChildren ( "/" , new org . apache . zookeeper . test . NullWatcher ( ) , ( ( org . apache . zookeeper . AsyncCallback . Children2Callback ) ( this ) ) , results ) ; org . apache . zookeeper . test . LOG . info ( ( "Submitted<sp>all<sp>operations:" + ( new java . util . Date ( ) . toString ( ) ) ) ) ; if ( ! ( opsCount . await ( 10000 , TimeUnit . MILLISECONDS ) ) ) org . junit . Assert . fail ( ( "Haven't<sp>received<sp>all<sp>confirmations" + ( opsCount . getCount ( ) ) ) ) ; for ( int i = 0 ; i < ( limit ) ; i ++ ) { "<AssertPlaceHolder>" ; } } catch ( java . io . IOException e ) { System . out . println ( e . toString ( ) ) ; } } get ( int ) { return org . apache . zookeeper . KeeperException . Code . lookup . get ( code ) ; }
org . junit . Assert . assertEquals ( 0 , ( ( int ) ( results . get ( i ) ) ) )
testPDFBox3964 ( ) { org . sejda . sambox . pdmodel . PDDocument doc = org . sejda . sambox . pdmodel . PDDocument . load ( new java . io . File ( org . sejda . sambox . input . PDFParserTest . TARGETPDFDIR , "PDFBOX-3964-c687766d68ac766be3f02aaec5e0d713_2.pdf" ) ) ; "<AssertPlaceHolder>" ; doc . close ( ) ; } getNumberOfPages ( ) { return getDocumentCatalog ( ) . getPages ( ) . getCount ( ) ; }
org . junit . Assert . assertEquals ( 10 , doc . getNumberOfPages ( ) )
testGetVideoFolder ( ) { net . holmes . core . backend . manager . BackendManager backendManager = createMock ( net . holmes . core . backend . manager . BackendManager . class ) ; expect ( backendManager . getFolder ( "videosTest" , net . holmes . core . backend . handler . VIDEO ) ) . andReturn ( new net . holmes . core . backend . response . ConfigurationFolder ( "videosTest" , "videosTest" , java . lang . System . getProperty ( "java.io.tmpdir" ) ) ) . atLeastOnce ( ) ; replay ( backendManager ) ; net . holmes . core . backend . handler . VideoFoldersHandler videoFoldersHandler = new net . holmes . core . backend . handler . VideoFoldersHandler ( backendManager ) ; "<AssertPlaceHolder>" ; verify ( backendManager ) ; } getVideoFolder ( java . lang . String ) { return getFolder ( id ) ; }
org . junit . Assert . assertNotNull ( videoFoldersHandler . getVideoFolder ( "videosTest" ) )
split_string_whitespace_using_guava ( ) { java . util . List < java . lang . String > elementsInString = com . google . common . collect . Lists . newArrayList ( com . google . common . base . Splitter . on ( "<sp>" ) . split ( com . levelup . java . string . SplitStrings . CONSTANT_STRING ) ) ; com . levelup . java . string . SplitStrings . logger . info ( elementsInString ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertTrue ( ( ( elementsInString . size ( ) ) == 16 ) )
addDuppedProxy ( ) { org . openqa . grid . internal . Registry registry = org . openqa . grid . internal . Registry . newInstance ( ) ; org . openqa . grid . internal . RemoteProxy p1 = org . openqa . grid . internal . RemoteProxyFactory . getNewBasicRemoteProxy ( "app1" , "http://machine1:4444/" , registry ) ; org . openqa . grid . internal . RemoteProxy p2 = org . openqa . grid . internal . RemoteProxyFactory . getNewBasicRemoteProxy ( "app1" , "http://machine2:4444/" , registry ) ; org . openqa . grid . internal . RemoteProxy p3 = org . openqa . grid . internal . RemoteProxyFactory . getNewBasicRemoteProxy ( "app1" , "http://machine3:4444/" , registry ) ; org . openqa . grid . internal . RemoteProxy p4 = org . openqa . grid . internal . RemoteProxyFactory . getNewBasicRemoteProxy ( "app1" , "http://machine4:4444/" , registry ) ; try { registry . add ( p1 ) ; registry . add ( p2 ) ; registry . add ( p3 ) ; registry . add ( p4 ) ; registry . add ( p4 ) ; "<AssertPlaceHolder>" ; } finally { registry . stop ( ) ; } } getAllProxies ( ) { return proxies ; }
org . junit . Assert . assertTrue ( ( ( registry . getAllProxies ( ) . size ( ) ) == 4 ) )
copyDoesntDeleteOriginalFile ( ) { java . io . File source = temporaryFolder . newFile ( "test.txt" ) ; java . io . File destination = temporaryFolder . newFolder ( ) ; com . hotels . bdp . circustrain . avro . util . SchemaCopier copier = new com . hotels . bdp . circustrain . avro . util . SchemaCopier ( new org . apache . hadoop . hive . conf . HiveConf ( ) , new org . apache . hadoop . hive . conf . HiveConf ( ) ) ; copier . copy ( source . toString ( ) , destination . toString ( ) ) ; org . apache . hadoop . fs . FileSystem fs = new org . apache . hadoop . fs . Path ( destination . toString ( ) ) . getFileSystem ( new org . apache . hadoop . hive . conf . HiveConf ( ) ) ; "<AssertPlaceHolder>" ; } toString ( ) { return com . google . common . base . Objects . toStringHelper ( this ) . add ( "message" , message ) . add ( "left" , left ) . add ( "right" , right ) . toString ( ) ; }
org . junit . Assert . assertTrue ( fs . exists ( new org . apache . hadoop . fs . Path ( source . toString ( ) ) ) )
testGet ( ) { org . apache . druid . server . lookup . cache . polling . PollingCache mockPollingCache = org . easymock . EasyMock . createStrictMock ( org . apache . druid . server . lookup . cache . polling . PollingCache . class ) ; org . apache . druid . server . lookup . PollingLookup . CacheRefKeeper cacheRefKeeper = new org . apache . druid . server . lookup . PollingLookup . CacheRefKeeper ( mockPollingCache ) ; "<AssertPlaceHolder>" ; } getAndIncrementRef ( ) { synchronized ( refCounts ) { if ( ( refCounts . get ( ) ) < 0 ) { return null ; } refCounts . incrementAndGet ( ) ; return pollingCache ; } }
org . junit . Assert . assertEquals ( mockPollingCache , cacheRefKeeper . getAndIncrementRef ( ) )
testSingleSave ( ) { org . geoserver . config . GeoServer gs = createMock ( org . geoserver . config . GeoServer . class ) ; java . util . List < org . geoserver . config . ConfigurationListener > listeners = new java . util . ArrayList ( ) ; gs . addListener ( capture ( listeners ) ) ; expectLastCall ( ) . atLeastOnce ( ) ; java . util . List < org . geoserver . wps . ProcessGroupInfo > procGroups = new java . util . ArrayList ( ) ; org . geoserver . wps . WPSInfo wps = createNiceMock ( org . geoserver . wps . WPSInfo . class ) ; expect ( wps . getProcessGroups ( ) ) . andReturn ( procGroups ) . anyTimes ( ) ; replay ( wps ) ; expect ( gs . getService ( org . geoserver . wps . WPSInfo . class ) ) . andReturn ( wps ) . anyTimes ( ) ; gs . save ( wps ) ; expectLastCall ( ) . once ( ) ; replay ( gs ) ; initer . initialize ( gs ) ; "<AssertPlaceHolder>" ; org . geoserver . config . ConfigurationListener l = listeners . get ( 0 ) ; l . handleGlobalChange ( null , null , null , null ) ; l . handlePostGlobalChange ( null ) ; verify ( gs ) ; } size ( ) { if ( ( featureCache ) != null ) { return featureCache . size ( ) ; } if ( ( org . geoserver . wfs . FeatureSizeFeatureCollection . FEATURE_CACHE_LIMIT ) > 0 ) { try { int count = featureSource . getCount ( query ) ; if ( count == 0 ) { featureCache = new java . util . ArrayList ( ) ; return count ; } if ( count > 0 ) { return count ; } java . util . List < org . opengis . feature . simple . SimpleFeature > tempFeatureCache = new java . util . ArrayList ( ) ; try ( org . geotools . data . simple . SimpleFeatureIterator it = featureSource . getFeatures ( query ) . features ( ) ) { count = 0 ; while ( it . hasNext ( ) ) { org . opengis . feature . simple . SimpleFeature feature = it . next ( ) ; if ( ( tempFeatureCache . size ( ) ) < ( org . geoserver . wfs . FeatureSizeFeatureCollection . FEATURE_CACHE_LIMIT ) ) { tempFeatureCache . add ( feature ) ; } count ++ ; } if ( count <= ( org . geoserver . wfs . FeatureSizeFeatureCollection . FEATURE_CACHE_LIMIT ) ) { featureCache = tempFeatureCache ; } else { tempFeatureCache . clear ( ) ; } return count ; } } catch ( java . io . IOException e ) { throw new java . lang . RuntimeException ( e ) ; } } return super . size ( ) ; }
org . junit . Assert . assertEquals ( 1 , listeners . size ( ) )
testParseConfig61CustomFields ( ) { final java . lang . String CONFIG_61_noCustomFields = "1,2017/08/11<sp>12:37:58,008900008659,CONFIG,0,1,2017/08/11<sp>11:37:58,192.168.14.162,vsys1,edit,admin,Web,Succeeded,<sp>config<sp>shared<sp>log-settings<sp>config,1354,0x0,/FatherNode/KidNode/GrandsonNode1,/FatherNode/KidNode/GrandsonNode2" ; org . json . simple . JSONObject actual = parser . parse ( CONFIG_61_noCustomFields . getBytes ( ) ) . get ( 0 ) ; org . json . simple . JSONObject expected = new org . json . simple . JSONObject ( ) ; expected . put ( BasicPaloAltoFirewallParser . PaloAltoDomain , "admin" 0 ) ; expected . put ( BasicPaloAltoFirewallParser . ReceiveTime , "2017/08/11<sp>12:37:58" ) ; expected . put ( BasicPaloAltoFirewallParser . SerialNum , "admin" 6 ) ; expected . put ( BasicPaloAltoFirewallParser . Type , "CONFIG" ) ; expected . put ( BasicPaloAltoFirewallParser . ThreatContentType , "0" ) ; expected . put ( BasicPaloAltoFirewallParser . ConfigVersion , "admin" 0 ) ; expected . put ( BasicPaloAltoFirewallParser . GenerateTime , "2017/08/11<sp>11:37:58" ) ; expected . put ( BasicPaloAltoFirewallParser . HOST , "admin" 5 ) ; expected . put ( BasicPaloAltoFirewallParser . VirtualSystem , "vsys1" ) ; expected . put ( BasicPaloAltoFirewallParser . Command , "edit" ) ; expected . put ( BasicPaloAltoFirewallParser . Admin , "admin" ) ; expected . put ( BasicPaloAltoFirewallParser . Client , "Web" ) ; expected . put ( BasicPaloAltoFirewallParser . Result , "admin" 2 ) ; expected . put ( BasicPaloAltoFirewallParser . ConfigurationPath , "config<sp>shared<sp>log-settings<sp>config" ) ; expected . put ( BasicPaloAltoFirewallParser . Seqno , "admin" 9 ) ; expected . put ( BasicPaloAltoFirewallParser . ActionFlags , "admin" 8 ) ; expected . put ( BasicPaloAltoFirewallParser . BeforeChangeDetail , "admin" 7 ) ; expected . put ( BasicPaloAltoFirewallParser . AfterChangeDetail , "admin" 4 ) ; expected . put ( BasicPaloAltoFirewallParser . ParserVersion , 61 ) ; expected . put ( "admin" 1 , CONFIG_61_noCustomFields ) ; expected . put ( "admin" 3 , actual . get ( "admin" 3 ) ) ; "<AssertPlaceHolder>" ; } get ( org . apache . metron . performance . load . CommandLine ) { return cli . getOptionValue ( shortCode ) ; }
org . junit . Assert . assertEquals ( expected , actual )
testDefineCommand4 ( ) { try { com . gabstudios . cmdline . CmdLine . defineCommand ( "file,<sp>!fileName,<sp>:file\\d.txt,<sp>#Load<sp>files<sp>into<sp>the<sp>system,<sp>#Load<sp>files<sp>into<sp>the<sp>system" ) ; org . junit . Assert . fail ( ) ; } catch ( com . gabstudios . cmdline . DuplicateException e ) { "<AssertPlaceHolder>" ; } } defineCommand ( java . lang . String [ ] ) { com . gabstudios . validate . Validate . defineBoolean ( ( ( ( nameArgs != null ) && ( ( nameArgs . length ) > 0 ) ) && ( ( nameArgs . length ) <= ( com . gabstudios . cmdline . CmdLine . MAX_LENGTH ) ) ) ) . testTrue ( ) . throwValidationExceptionOnFail ( ) . validate ( ) ; final java . util . List < com . gabstudios . cmdline . Token > tokens = com . gabstudios . cmdline . CmdLine . DEFINED_COMMAND_TOKENIZER . tokenize ( nameArgs ) ; final com . gabstudios . cmdline . CommandDefinition command = com . gabstudios . cmdline . CmdLine . createCommandDefinition ( tokens ) ; final java . util . List < java . lang . String > names = command . getNames ( ) ; for ( final java . lang . String name : names ) { final com . gabstudios . cmdline . CommandDefinition existingCommand = com . gabstudios . cmdline . CmdLine . COMMAND_DEFINITION_MAP . put ( name , command ) ; if ( existingCommand != null ) { throw new com . gabstudios . cmdline . DuplicateException ( ( ( "Error:<sp>The<sp>command<sp>'" + name ) + "'<sp>has<sp>already<sp>been<sp>defined.<sp>Define<sp>a<sp>new<sp>command<sp>name." ) ) ; } } return com . gabstudios . cmdline . CmdLine . INSTANCE ; }
org . junit . Assert . assertTrue ( true )
testPathMissingLength ( ) { byte [ ] data = new byte [ 26 ] ; java . nio . ByteBuffer buffer = java . nio . ByteBuffer . wrap ( data ) ; buffer . putShort ( ( ( short ) ( AdobePathSegment . CLOSED_SUBPATH_BEZIER_LINKED ) ) ) ; java . awt . geom . Path2D path = new com . twelvemonkeys . imageio . path . AdobePathBuilder ( data ) . path ( ) ; "<AssertPlaceHolder>" ; } path ( ) { java . util . List < java . util . List < com . twelvemonkeys . imageio . path . AdobePathSegment > > subPaths = new java . util . ArrayList < java . util . List < com . twelvemonkeys . imageio . path . AdobePathSegment > > ( ) ; java . util . List < com . twelvemonkeys . imageio . path . AdobePathSegment > currentPath = null ; int currentPathLength = 0 ; com . twelvemonkeys . imageio . path . AdobePathSegment segment ; while ( ( segment = nextSegment ( ) ) != null ) { if ( com . twelvemonkeys . imageio . path . AdobePathBuilder . DEBUG ) { System . out . println ( segment ) ; } if ( ( ( segment . selector ) == ( AdobePathSegment . OPEN_SUBPATH_LENGTH_RECORD ) ) || ( ( segment . selector ) == ( AdobePathSegment . CLOSED_SUBPATH_LENGTH_RECORD ) ) ) { if ( currentPath != null ) { if ( currentPathLength != ( currentPath . size ( ) ) ) { throw new javax . imageio . IIOException ( java . lang . String . format ( "Bad<sp>path,<sp>expected<sp>%d<sp>segments,<sp>found<sp>only<sp>%d" , currentPathLength , currentPath . size ( ) ) ) ; } subPaths . add ( currentPath ) ; } currentPath = new java . util . ArrayList < com . twelvemonkeys . imageio . path . AdobePathSegment > ( segment . length ) ; currentPathLength = segment . length ; } else if ( ( ( ( ( segment . selector ) == ( AdobePathSegment . OPEN_SUBPATH_BEZIER_LINKED ) ) || ( ( segment . selector ) == ( AdobePathSegment . OPEN_SUBPATH_BEZIER_UNLINKED ) ) ) || ( ( segment . selector ) == ( AdobePathSegment . CLOSED_SUBPATH_BEZIER_LINKED ) ) ) || ( ( segment . selector ) == ( AdobePathSegment . CLOSED_SUBPATH_BEZIER_UNLINKED ) ) ) { if ( currentPath == null ) { throw new javax . imageio . IIOException ( "Bad<sp>path,<sp>missing<sp>subpath<sp>length<sp>record" ) ; } if ( ( currentPath . size ( ) ) >= currentPathLength ) { throw new javax . imageio . IIOException ( java . lang . String . format ( "Bad<sp>path,<sp>expected<sp>%d<sp>segments,<sp>found%d" , currentPathLength , currentPath . size ( ) ) ) ; } currentPath . add ( segment ) ; } } if ( currentPath != null ) { if ( currentPathLength != ( currentPath . size ( ) ) ) { throw new javax . imageio . IIOException ( java . lang . String . format ( "Bad<sp>path,<sp>expected<sp>%d<sp>segments,<sp>found<sp>only<sp>%d" , currentPathLength , currentPath . size ( ) ) ) ; } subPaths . add ( currentPath ) ; } return pathToShape ( subPaths ) ; }
org . junit . Assert . assertNotNull ( path )
shouldGetConnectionWithEmptyCredentials ( ) { com . arjuna . ats . internal . jdbc . DirectRecoverableConnection connection = new com . arjuna . ats . internal . jdbc . DirectRecoverableConnection ( "testDb" , "" , "" , com . arjuna . ats . internal . jdbc . DirectRecoverableConnectionUnitTest . TestDynamicClass . class . getName ( ) , null ) ; javax . sql . XAConnection xaConnection = connection . getConnection ( ) ; "<AssertPlaceHolder>" ; verify ( com . arjuna . ats . internal . jdbc . DirectRecoverableConnectionUnitTest . TestDynamicClass . lastInstance . getDataSource ( "testDb" ) ) . getXAConnection ( ) ; } getXaConnection ( ) { return xaConnection ; }
org . junit . Assert . assertEquals ( com . arjuna . ats . internal . jdbc . DirectRecoverableConnectionUnitTest . TestDynamicClass . lastInstance . getXaConnection ( ) , xaConnection )
testOk ( ) { cn . com . warlock . wisp . core . plugin . processor . support . transform . IEntryTransform iEntryTransform = cn . com . warlock . wisp . core . plugin . processor . support . transform . EntryTransformFactory . getInsertTransform ( ) ; cn . com . warlock . wisp . core . dto . MysqlEntry mysqlEntry = new cn . com . warlock . wisp . core . dto . MysqlEntry ( ) ; java . util . List < cn . com . warlock . wisp . core . dto . MysqlColumn > mysqlColumns = new java . util . ArrayList ( ) ; cn . com . warlock . wisp . core . dto . MysqlColumn mysqlColumn = new cn . com . warlock . wisp . core . dto . MysqlColumn ( ) ; mysqlColumn . setName ( "id" ) ; mysqlColumn . setValue ( java . lang . String . valueOf ( 1L ) ) ; mysqlColumns . add ( mysqlColumn ) ; mysqlEntry . setColumns ( mysqlColumns ) ; cn . com . warlock . wisp . core . plugin . processor . support . transform . TransformResult transformResult = iEntryTransform . entry2Json ( mysqlEntry , "id" ) ; "<AssertPlaceHolder>" ; } entry2Json ( cn . com . warlock . wisp . core . dto . MysqlEntry , java . lang . String ) { java . util . Map < java . lang . String , java . lang . String > map = new java . util . HashMap ( ) ; java . lang . String currentKeyValue = "" ; java . util . List < cn . com . warlock . wisp . core . dto . MysqlColumn > columns = entry . getColumns ( ) ; for ( cn . com . warlock . wisp . core . dto . MysqlColumn mysqlColumn : columns ) { boolean isNull = mysqlColumn . isNull ( ) ; java . lang . String key = mysqlColumn . getName ( ) ; java . lang . String value = mysqlColumn . getValue ( ) ; if ( value == null ) { isNull = true ; } if ( isNull ) { map . put ( key , "" ) ; } else { map . put ( key , value ) ; if ( key . equals ( tableKey ) ) { currentKeyValue = value ; } } } if ( currentKeyValue . equals ( "" ) ) { cn . com . warlock . wisp . core . plugin . processor . support . transform . impl . UpdateEntryTransformImpl . LOGGER . error ( "cannot<sp>find<sp>key<sp>{}<sp>for<sp>table<sp>{}<sp>with<sp>entry<sp>{}" , tableKey , entry . getTable ( ) , entry ) ; return null ; } else { java . lang . String value ; try { value = new com . fasterxml . jackson . databind . ObjectMapper ( ) . writeValueAsString ( map ) ; } catch ( com . fasterxml . jackson . core . JsonProcessingException e ) { cn . com . warlock . wisp . core . plugin . processor . support . transform . impl . UpdateEntryTransformImpl . LOGGER . error ( "json<sp>dump<sp>error,<sp>key<sp>{}<sp>for<sp>table<sp>{}<sp>with<sp>entry<sp>{}" , tableKey , entry . getTable ( ) , entry ) ; return null ; } return new cn . com . warlock . wisp . core . plugin . processor . support . transform . TransformResult ( currentKeyValue , value ) ; } }
org . junit . Assert . assertEquals ( transformResult . getKey ( ) , java . lang . String . valueOf ( 1L ) )
whenCriteriaSortingStringNullsLastAsc_thenNullsLast ( ) { final org . hibernate . Criteria criteria = session . createCriteria ( com . baeldung . persistence . model . Foo . class , "FOO" ) ; criteria . addOrder ( org . hibernate . criterion . Order . asc ( "name" ) . nulls ( NullPrecedence . LAST ) ) ; final java . util . List < com . baeldung . persistence . model . Foo > fooList = criteria . list ( ) ; "<AssertPlaceHolder>" ; for ( final com . baeldung . persistence . model . Foo foo : fooList ) { System . out . println ( ( ( ( "Id:<sp>" + ( foo . getId ( ) ) ) + ",<sp>FirstName:<sp>" ) + ( foo . getName ( ) ) ) ) ; } } get ( java . lang . Integer ) { return emf . unwrap ( org . hibernate . SessionFactory . class ) . getCurrentSession ( ) . get ( org . baeldung . demo . model . Foo . class , id ) ; }
org . junit . Assert . assertNull ( fooList . get ( ( ( fooList . toArray ( ) . length ) - 1 ) ) . getName ( ) )
getByUrlTest ( ) { System . out . print ( "-><sp>getByUrl()<sp>-<sp>" ) ; ua . com . alexcoffee . model . product . Product product = ua . com . alexcoffee . service . impl . ProductServiceImplTest . productService . getByUrl ( ua . com . alexcoffee . service . impl . URL ) ; "<AssertPlaceHolder>" ; System . out . println ( "OK!" ) ; } getByUrl ( java . lang . String ) { if ( isEmpty ( url ) ) { throw new java . lang . IllegalArgumentException ( "No<sp>product<sp>URL!" ) ; } final ua . com . alexcoffee . model . product . Product product = this . repository . findByUrl ( url ) ; if ( isNull ( product ) ) { throw new java . lang . NullPointerException ( ( ( "Can't<sp>find<sp>product<sp>by<sp>url<sp>" + url ) + "!" ) ) ; } return product ; }
org . junit . Assert . assertNotNull ( product )
testInvalidLanguageTokeniseJSON ( ) { java . lang . String tokMode = "TOKENISE" ; java . io . File tokenOutJSON = folder . newFile ( "tokens-testTweetTokeniseJSON.json" ) ; java . lang . String commandArgs = java . lang . String . format ( commandFormat , jsonTwitterInputFile , tokenOutJSON , tokMode , "APPEND" ) ; commandArgs += "<sp>-m<sp>LANG_ID" ; java . lang . String [ ] commandArgsArr = commandArgs . split ( "<sp>" ) ; org . openimaj . tools . twitter . TwitterPreprocessingTool . main ( commandArgsArr ) ; org . openimaj . twitter . collection . TwitterStatusList < org . openimaj . twitter . USMFStatus > toktweets = org . openimaj . twitter . collection . FileTwitterStatusList . readUSMF ( tokenOutJSON , "UTF-8" ) ; org . openimaj . tools . twitter . modes . preprocessing . LanguageDetectionMode langDet = new org . openimaj . tools . twitter . modes . preprocessing . LanguageDetectionMode ( ) ; org . openimaj . tools . twitter . modes . preprocessing . TokeniseMode tokModeInst = new org . openimaj . tools . twitter . modes . preprocessing . TokeniseMode ( ) ; for ( org . openimaj . twitter . USMFStatus twitterStatus : toktweets ) { java . util . Map < java . lang . String , java . lang . Object > a = org . openimaj . tools . twitter . modes . preprocessing . LanguageDetectionMode . results ( twitterStatus , langDet ) ; boolean validLanguage = org . openimaj . text . nlp . TweetTokeniser . isValid ( ( ( java . lang . String ) ( a . get ( "language" ) ) ) ) ; java . util . Map < java . lang . String , java . util . List < java . lang . String > > tokens = org . openimaj . tools . twitter . modes . preprocessing . TokeniseMode . results ( twitterStatus , tokModeInst ) ; boolean containsTokens = ( tokens . size ( ) ) != 0 ; boolean valid = ( containsTokens && validLanguage ) || ( ( ! containsTokens ) && ( ! validLanguage ) ) ; if ( ! valid ) { System . out . println ( ( "Language<sp>was:<sp>" + ( a . get ( "language" ) ) ) ) ; System . out . println ( ( "Tokens<sp>were:<sp>" + tokens ) ) ; } "<AssertPlaceHolder>" ; } tokenOutJSON . delete ( ) ; } get ( int ) { int done = 0 ; for ( final org . openimaj . tools . clusterquantiser . FeatureFileFeature fff : this ) { if ( ( done ++ ) == index ) return fff ; } return null ; }
org . junit . Assert . assertTrue ( valid )
testInt ( ) { int [ ] values = new int [ ] { 0 , 53 , 253 , 62323 } ; for ( int in : values ) { int out = org . fusesource . amqp . codec . TestSupport . writeRead ( new org . fusesource . amqp . codec . AMQPInt ( in ) ) . getValue ( ) . intValue ( ) ; "<AssertPlaceHolder>" ; } } writeRead ( T extends org . fusesource . amqp . codec . AMQPType ) { return org . fusesource . amqp . codec . TestSupport . writeRead ( value , true ) ; }
org . junit . Assert . assertEquals ( in , out )
source_notfound_2 ( ) { org . apache . jena . graph . Graph graph = org . apache . jena . sparql . graph . GraphFactory . createGraphMem ( ) ; org . apache . jena . riot . RDFParser . create ( ) . source ( ( ( org . apache . jena . riot . TestRDFParser . DIR ) + "data.nosuchfile.ttl" ) ) . parse ( graph ) ; "<AssertPlaceHolder>" ; } size ( ) { return pmap . size ( ) ; }
org . junit . Assert . assertEquals ( 3 , graph . size ( ) )
testTransformReturnsNullIfSpecValidationFails ( ) { java . util . Map < java . lang . String , org . apache . beam . runners . dataflow . worker . DataflowExecutionContext . DataflowStepContext > stepContextMapping = new java . util . HashMap ( ) ; org . apache . beam . runners . dataflow . worker . fn . control . UserDistributionMonitoringInfoToCounterUpdateTransformer testObject = new org . apache . beam . runners . dataflow . worker . fn . control . UserDistributionMonitoringInfoToCounterUpdateTransformer ( mockSpecValidator , stepContextMapping ) ; java . util . Optional < java . lang . String > error = java . util . Optional . of ( "Error<sp>text" ) ; when ( mockSpecValidator . validate ( any ( ) ) ) . thenReturn ( error ) ; "<AssertPlaceHolder>" ; } transform ( org . apache . beam . model . pipeline . v1 . MetricsApi . MonitoringInfo ) { java . util . Optional < java . lang . String > validationResult = validate ( monitoringInfo ) ; if ( validationResult . isPresent ( ) ) { org . apache . beam . runners . dataflow . worker . fn . control . ElementCountMonitoringInfoToCounterUpdateTransformer . LOG . info ( validationResult . get ( ) ) ; return null ; } long value = monitoringInfo . getMetric ( ) . getCounterData ( ) . getInt64Value ( ) ; final java . lang . String pcollectionId = monitoringInfo . getLabelsMap ( ) . get ( MonitoringInfoConstants . Labels . PCOLLECTION ) ; final java . lang . String pcollectionName = pcollectionIdToNameContext . get ( pcollectionId ) . userName ( ) ; java . lang . String counterName = pcollectionName + "-ElementCount" ; com . google . api . services . dataflow . model . NameAndKind name = new com . google . api . services . dataflow . model . NameAndKind ( ) ; name . setName ( counterName ) . setKind ( "SUM" ) ; return new com . google . api . services . dataflow . model . CounterUpdate ( ) . setNameAndKind ( name ) . setCumulative ( true ) . setInteger ( org . apache . beam . runners . dataflow . worker . counters . DataflowCounterUpdateExtractor . longToSplitInt ( value ) ) ; }
org . junit . Assert . assertEquals ( null , testObject . transform ( null ) )
testCreateTableTableAlreadyExistsCorrectKeySchema ( ) { final java . util . Collection < com . amazonaws . services . dynamodbv2 . model . AttributeDefinition > ads = java . util . Arrays . asList ( new com . amazonaws . services . dynamodbv2 . model . AttributeDefinition ( "Hash" , com . amazonaws . services . dynamodbv2 . model . ScalarAttributeType . S ) ) ; final java . util . Collection < com . amazonaws . services . dynamodbv2 . model . KeySchemaElement > kses = java . util . Arrays . asList ( new com . amazonaws . services . dynamodbv2 . model . KeySchemaElement ( "Hash" , com . amazonaws . services . dynamodbv2 . model . KeyType . HASH ) ) ; final com . amazonaws . services . dynamodbv2 . model . TableDescription description = new com . amazonaws . services . dynamodbv2 . model . TableDescription ( ) . withAttributeDefinitions ( ads ) . withKeySchema ( kses ) . withTableName ( com . amazonaws . services . dynamodbv2 . json . demo . mars . util . DynamoDBManagerTest . tableName ) ; final com . amazonaws . services . dynamodbv2 . model . DescribeTableResult result = new com . amazonaws . services . dynamodbv2 . model . DescribeTableResult ( ) . withTable ( description ) ; org . easymock . EasyMock . expect ( dynamoDB . describeTable ( com . amazonaws . services . dynamodbv2 . json . demo . mars . util . DynamoDBManagerTest . tableName ) ) . andReturn ( result ) ; final com . amazonaws . services . dynamodbv2 . model . CreateTableRequest request = new com . amazonaws . services . dynamodbv2 . model . CreateTableRequest ( ) . withAttributeDefinitions ( ads ) . withKeySchema ( kses ) . withTableName ( com . amazonaws . services . dynamodbv2 . json . demo . mars . util . DynamoDBManagerTest . tableName ) ; org . powermock . api . easymock . PowerMock . replayAll ( ) ; "<AssertPlaceHolder>" ; org . powermock . api . easymock . PowerMock . verifyAll ( ) ; } createTable ( com . amazonaws . services . dynamodbv2 . AmazonDynamoDB , com . amazonaws . services . dynamodbv2 . model . CreateTableRequest ) { try { final com . amazonaws . services . dynamodbv2 . model . DescribeTableResult result = dynamoDB . describeTable ( request . getTableName ( ) ) ; if ( ! ( request . getKeySchema ( ) . equals ( result . getTable ( ) . getKeySchema ( ) ) ) ) { throw new java . lang . IllegalStateException ( ( ( "Table<sp>" + ( request . getTableName ( ) ) ) + "<sp>already<sp>exists<sp>and<sp>has<sp>an<sp>invalid<sp>schema" ) ) ; } com . amazonaws . services . dynamodbv2 . json . demo . mars . util . DynamoDBManager . LOGGER . warning ( ( ( "Table<sp>" + ( request . getTableName ( ) ) ) + "<sp>already<sp>exists" ) ) ; return result . getTable ( ) ; } catch ( final com . amazonaws . services . dynamodbv2 . model . ResourceNotFoundException e ) { return dynamoDB . createTable ( request ) . getTableDescription ( ) ; } }
org . junit . Assert . assertEquals ( description , com . amazonaws . services . dynamodbv2 . json . demo . mars . util . DynamoDBManager . createTable ( dynamoDB , request ) )
differentSizedPages ( ) { final int BATCH_SIZE = 5 ; org . apache . hadoop . dynamodb . preader . PageResultMultiplexer < java . lang . Integer > mux = new org . apache . hadoop . dynamodb . preader . PageResultMultiplexer ( BATCH_SIZE , org . apache . hadoop . dynamodb . read . PageResultMultiplexerTest . DEFAULT_CAPACITY ) ; mux . addPageResults ( new org . apache . hadoop . dynamodb . preader . PageResults ( java . util . Arrays . asList ( 1 , 1 , 1 ) , null ) ) ; mux . addPageResults ( new org . apache . hadoop . dynamodb . preader . PageResults ( java . util . Arrays . asList ( 2 ) , null ) ) ; mux . addPageResults ( new org . apache . hadoop . dynamodb . preader . PageResults ( java . util . Arrays . asList ( 3 , 3 , 3 , 3 ) , null ) ) ; mux . addPageResults ( new org . apache . hadoop . dynamodb . preader . PageResults ( java . util . Arrays . asList ( 4 , 4 ) , null ) ) ; mux . setDraining ( true ) ; java . lang . Integer [ ] expectedOrder = new org . apache . hadoop . dynamodb . read . Integer [ ] { 1 , 2 , 3 , 4 , 1 , 3 , 4 , 1 , 3 , 3 } ; "<AssertPlaceHolder>" ; } muxToList ( org . apache . hadoop . dynamodb . preader . PageResultMultiplexer ) { java . lang . Integer n ; java . util . List < java . lang . Integer > list = new java . util . ArrayList ( ) ; while ( true ) { if ( ( n = mux . next ( ) ) == null ) { break ; } list . add ( n ) ; } return list ; }
org . junit . Assert . assertArrayEquals ( expectedOrder , muxToList ( mux ) . toArray ( ) )
shouldReturnTrueWhenClassPresent ( ) { java . lang . String className = "java.io.IOException" ; boolean isPresent = com . navercorp . volleyextensions . volleyer . util . ClassUtils . isPresent ( className ) ; "<AssertPlaceHolder>" ; } isPresent ( java . lang . String ) { if ( com . navercorp . volleyextensions . volleyer . util . StringUtils . isEmpty ( clazzName ) ) { return false ; } boolean isPresent ; try { @ com . navercorp . volleyextensions . volleyer . util . SuppressWarnings ( "unused" ) java . lang . Class < ? > clazz = java . lang . Class . forName ( clazzName ) ; isPresent = true ; } catch ( java . lang . ClassNotFoundException e ) { isPresent = false ; } return isPresent ; }
org . junit . Assert . assertThat ( isPresent , org . hamcrest . CoreMatchers . is ( true ) )
checkTraverseByCollectionNode ( ) { org . neo4j . graphdb . traversal . Traverser heroesTraverserRest = getHeroesViaRest ( ) ; org . neo4j . graphdb . traversal . Traverser heroesTraverserByCollection = getHeroesByCollectionNodeViaRest ( ) ; "<AssertPlaceHolder>" ; } nodes ( ) { return new org . neo4j . helpers . collection . IterableWrapper < org . neo4j . graphdb . Node , org . neo4j . graphdb . Path > ( paths ) { @ org . neo4j . rest . graphdb . traversal . Override protected org . neo4j . graphdb . Node underlyingObjectToObject ( org . neo4j . graphdb . Path path ) { return path . endNode ( ) ; } } ; }
org . junit . Assert . assertEquals ( heroesTraverserRest . nodes ( ) . iterator ( ) . next ( ) , heroesTraverserByCollection . nodes ( ) . iterator ( ) . next ( ) )
canRetrieveSubtype ( ) { org . mongolink . domain . session . UnitOfWork unitOfWork = new org . mongolink . domain . session . UnitOfWork ( mock ( org . mongolink . domain . session . MongoSessionImpl . class ) ) ; org . mongolink . domain . session . TestsUnitOfWork . Child entity = new org . mongolink . domain . session . TestsUnitOfWork . Child ( ) ; unitOfWork . registerDirty ( 1 , entity , new org . bson . Document ( ) ) ; java . lang . Object entityFound = unitOfWork . getEntity ( org . mongolink . domain . session . TestsUnitOfWork . Parent . class , 1 ) ; "<AssertPlaceHolder>" . isNotNull ( ) ; } getEntity ( java . lang . Class , java . lang . Object ) { return ( ( T ) ( getValue ( type , dbId ) . entity ) ) ; }
org . junit . Assert . assertThat ( entityFound )
testGetSubjectMatterExpert ( ) { org . finra . herd . model . api . xml . SubjectMatterExpertKey originalSubjectMatterExpertKey = new org . finra . herd . model . api . xml . SubjectMatterExpertKey ( USER_ID ) ; org . finra . herd . model . api . xml . SubjectMatterExpertKey updatedSubjectMatterExpertKey = new org . finra . herd . model . api . xml . SubjectMatterExpertKey ( USER_ID_2 ) ; org . finra . herd . model . api . xml . SubjectMatterExpertContactDetails subjectMatterExpertContactDetails = new org . finra . herd . model . api . xml . SubjectMatterExpertContactDetails ( USER_FULL_NAME , USER_JOB_TITLE , USER_EMAIL_ADDRESS , USER_TELEPHONE_NUMBER ) ; when ( alternateKeyHelper . validateStringParameter ( "user<sp>id" , org . finra . herd . service . impl . USER_ID ) ) . thenReturn ( org . finra . herd . service . impl . USER_ID_2 ) ; when ( subjectMatterExpertDao . getSubjectMatterExpertByKey ( updatedSubjectMatterExpertKey ) ) . thenReturn ( subjectMatterExpertContactDetails ) ; org . finra . herd . model . api . xml . SubjectMatterExpert result = subjectMatterExpertServiceImpl . getSubjectMatterExpert ( originalSubjectMatterExpertKey ) ; "<AssertPlaceHolder>" ; verify ( alternateKeyHelper ) . validateStringParameter ( "user<sp>id" , org . finra . herd . service . impl . USER_ID ) ; verify ( subjectMatterExpertDao ) . getSubjectMatterExpertByKey ( updatedSubjectMatterExpertKey ) ; verifyNoMoreInteractionsHelper ( ) ; } getSubjectMatterExpert ( java . lang . String ) { return subjectMatterExpertService . getSubjectMatterExpert ( new org . finra . herd . model . api . xml . SubjectMatterExpertKey ( userId ) ) ; }
org . junit . Assert . assertEquals ( new org . finra . herd . model . api . xml . SubjectMatterExpert ( updatedSubjectMatterExpertKey , subjectMatterExpertContactDetails ) , result )
testGetValue ( ) { picker . setValue ( org . kie . workbench . common . dmn . client . editors . types . DataTypePickerWidgetTest . VALUE ) ; "<AssertPlaceHolder>" ; } getValue ( ) { final org . uberfire . client . views . pfly . widgets . Moment currentDate = view . getDate ( ) ; if ( java . util . Objects . isNull ( currentDate ) ) { return "" ; } return currentDate . format ( org . kie . workbench . common . dmn . client . editors . types . listview . constraint . common . typed . time . picker . TimePicker . TIME_FORMAT ) ; }
org . junit . Assert . assertEquals ( org . kie . workbench . common . dmn . client . editors . types . DataTypePickerWidgetTest . VALUE , picker . getValue ( ) )
delete ( ) { com . ebay . cloud . cms . entmgr . entity . EntityContext context = com . ebay . cloud . cms . entmgr . entity . impl . EntityServiceTest . newEntityContext ( ) ; com . ebay . cloud . cms . dal . entity . JsonEntity newInst = newServiceInstance ( IBranch . DEFAULT_BRANCH ) ; java . lang . String newId = com . ebay . cloud . cms . entmgr . entity . impl . EntityServiceTest . entityService . create ( newInst , context ) ; com . ebay . cloud . cms . dal . entity . IEntity qEntity = buildQueryEntity ( com . ebay . cloud . cms . entmgr . entity . impl . EntityServiceTest . RAPTOR_REPO , IBranch . DEFAULT_BRANCH , com . ebay . cloud . cms . entmgr . entity . impl . EntityServiceTest . SERVICE_INSTANCE , newId ) ; com . ebay . cloud . cms . entmgr . entity . impl . EntityServiceTest . entityService . delete ( qEntity , context ) ; com . ebay . cloud . cms . dal . entity . IEntity entityGet = com . ebay . cloud . cms . entmgr . entity . impl . EntityServiceTest . entityService . get ( qEntity , context ) ; "<AssertPlaceHolder>" ; } get ( com . ebay . cloud . cms . dal . entity . IEntity , com . ebay . cloud . cms . entmgr . entity . EntityContext ) { return entityGetManager . get ( queryEntity , context ) ; }
org . junit . Assert . assertTrue ( ( entityGet == null ) )
shouldEncryptAndDecryptDatabase ( ) { java . lang . String password = "abcdefg" ; byte [ ] passwordBytes = password . getBytes ( "UTF-8" ) ; byte [ ] hashedPassword = de . slackspace . openkeepass . crypto . Sha256 . hash ( passwordBytes ) ; de . slackspace . openkeepass . domain . KeePassHeader keepassHeader = new de . slackspace . openkeepass . domain . KeePassHeader ( ) ; keepassHeader . setValue ( KeePassHeader . MASTER_SEED , de . slackspace . openkeepass . util . ByteUtils . hexStringToByteArray ( "35ac8b529bc4f6e44194bccd0537fcb433a30bcb847e63156262c4df99c528ca" ) ) ; keepassHeader . setValue ( KeePassHeader . TRANSFORM_SEED , de . slackspace . openkeepass . util . ByteUtils . hexStringToByteArray ( "0d52d93efc5493ae6623f0d5d69bb76bd976bb717f4ee67abbe43528ebfbb646" ) ) ; keepassHeader . setValue ( KeePassHeader . TRANSFORM_ROUNDS , de . slackspace . openkeepass . util . ByteUtils . hexStringToByteArray ( "401F000000000000" ) ) ; keepassHeader . setValue ( KeePassHeader . ENCRYPTION_IV , de . slackspace . openkeepass . util . ByteUtils . hexStringToByteArray ( "2c605455f181fbc9462aefb817852b37" ) ) ; keepassHeader . setCompression ( CompressionAlgorithm . Gzip ) ; keepassHeader . setCrsAlgorithm ( CrsAlgorithm . Salsa20 ) ; byte [ ] data = de . slackspace . openkeepass . util . ByteUtils . hexStringToByteArray ( "69d788d9b01ea1facd1c0bf0187e7d74e4aa07b20d464f3d23d0b2dc2f059ff8000000008e0c5564df5e72b78c7b009b3e4a494e05c0d0387957d246d6b8cc8489e4ba19ad0600001f8b0800000000000400ed5adb6ee248107d66a5fd07c4be6d94d8e616888c4704cce40664b824b3f3b26a4c010ec626dd4d08339a7fdff615dbb48164273b62365114e1ea53d5edf63955d50ef287e799917e024c74cbac64a41331930653b386ba39ae641674745ccaa40945e61019960995cc0a48e683f2fb6ff235c02d22a4a11bc02e53721328b23fa4e48f600246d4c28a879185b5c941d411450344a08566a0f48050df200b91a10d706d82cc310c95ac28e58f25e9389bed89e5b342f92c57fe12f5f591911075201ad6e794ddab72b520348dd204cde606a48783b57b1895e4bedf42380e6e4018a18541fb04b0bdd6b4c03527cc513a2b14ed39f860275013e9260513991a5ce8846dfbaa8e5644c9150bb29030e6f8d52cc3c2fe6a9a8850c0d7b0dabe8e0d18cfb9039a722c6d806d330fdeb0b0063c0777c07581195bfb2db62868c1a34ac99ea1a753c6cb06326c4e456c1198bf7b3164608e806d2a2f2d3c547a78b1c606d668e0ce4d3c26b344202df697c440aecdb93f817b8332dbb29566c0b96eaa261a18e0af66d31e83f7fb9775a5cafda954c2fe0e30e6bc9d029b38c75d35295ef580e90bb1bbfa88adc57ccb0278e8a430db97b3c5c109e8d1be899e2f29cc882289b210b7c5805dfd2b28c56c59ca178a61b06377b0378ca85d30d8a382a1bbf8c6e353f3aab47a30faddaa4acbed59abfb90756e75131b84e859f33b9de8ec01ba03067e5ae256365750a19d3f6ab7c70df5a1b3f483c4d14e18f61810d681f83aaeb14467cdec94e45a6c66b9a95aee5816f57276e09f921d02244f1cf023253b1aa191fc1d928d43666f1529f952b34ce6982fcb82f7d1b5f7f49947796f0b9aacf88c740dd99cb707139e3217ea85a961d8e51e818466af6a1a10c27364bf396fde10c873559fe73a231d773ec9760b01c22eeb0ce05f7aa37d82c650b31626554a6ccfd757fe5a2df7b6b74b218e72b75c08edb97c49d8ccacbe0799246470115ea5a92ea8d55bcda10b8f0bd62240f064dd84e30f2be6c2306c11468c61641710d626acc18840d756171be5b7a3e92dd98387f6e70c3efbd4d607c39bfb5249eb17cf9ff0ed27e3e9fab1598d523be0ab18a56b4a662508c64c2de6305c2d6ddd216dca1d68b3f60aeb436015606deca131095d8544b09f0abc4e638b0ab6cbc0f3e7c860b70ef285241dbc46083b94109182c491c24e2df85bc5d342540c29b94b71404176c91a0fc54963b2607ff4cd77c858803fe05ef8d1220178e1fc7e811331ed157b185632b616334a3f57a6d74f77aa59562b2f9ec9e977780befba3dafa38c1747652ce6c59c503a3f138429c09cdde0896e8e2ce1e5b1bdbe8b37813d967607b7458d669c20e7f8c92ddc1ca5bcd6be8791494680db83d182b81cb1659f34e4bb5609b1343d624bc9f7ba39b4964a0fe131d0b47b250b9e3540b1bb23145bd320972adffa5db5d3aa36d5efdf7ad5f3efdf6eabddee7dbb53f72ed5564fed7c77b625e6e8af46d85c0eb34537c36f6882cc2d6c4f9417a7b9d1d7bb1be1a2bf58156ba3ac06bdc7cbf17ba2dc3f5196fe4f89f24599edbe51ec5fa1e51b65b5f41fd9b74c6c1330e6c274c07ec5b298178b7fdb3df0c8c2b393099d193f34ed35756d82c0c865a583c87b6c1fd24cf8b3f471da7bf9f3e71ee92fd0f78fc863a1a38c97c7924f62e13ce69c5cdc175546f81c133bc8ac4f32a578c63bace4947c9a79ebe4947bebe4b4f34cb3f350f38253cd8b8e35fc73cde339993db5af6735f1b1b3babc6aa3ac5eed6c3fd744ebb547f4e43821a22756ece4929d5cb3f9453b56b563cad82d0d493c13c5b3acb8431abbb451282468639b3862736f8ae355ead8258fddc57b0f8138f7cb17484c21f1329474d2f1ab50f465b9c970d16a142f47fcf8dc0621b14380c9d174727ff757eb34f7b1599586a57151fa5cea4f8ab569e535b36f360e418d5d718e42fb058d370db1ae616c5963034e346bf6aae0bcbe60bd687b7857d8785bb0a32ff8578d41bc4c6fd4e958a116c22f1d37ab36fe4c4feb476a53101fef253cbe3dcf8f0be6925bb5dd0683ec53b573875eb5a55ff648716855fb356f237770bebe1c56d1b599a7eaea73c3e8178a42b67ffe89cbf91650964ba77b71feb0291fccfc9f535e7ca7fc9b53bed61ef7f12d2ddc8ac5d3621d565f4c012df894bf34296013e83e9c970e9bf385eccfe27cf69df36fce79e1f97ed41e99b5e2543f2d92d6047fe99c627e6b034da4eff53a422a1f36e3dfb3fcafccf8cef08166a52b0389476ae1535935ae2e96d2c6bf121c965f5833182073ca16b95773737ad8bcff7999febda1ffb1bc8f7cae83011486edc1036894f8dfbcf1be70631fe8d7df9efc076926aea08429000001000000000000000000000000000000000000000000000000000000000000000000000000000000" ) ; de . slackspace . openkeepass . crypto . CryptoInformation cryptoInformation = new de . slackspace . openkeepass . crypto . CryptoInformation ( de . slackspace . openkeepass . domain . KeePassHeader . VERSION_SIGNATURE_LENGTH , keepassHeader . getMasterSeed ( ) , keepassHeader . getTransformSeed ( ) , keepassHeader . getEncryptionIV ( ) , keepassHeader . getTransformRounds ( ) , keepassHeader . getHeaderSize ( ) ) ; byte [ ] encryptedDatabase = new de . slackspace . openkeepass . crypto . Decrypter ( ) . encryptDatabase ( hashedPassword , cryptoInformation , data ) ; byte [ ] decryptedDatabase = new de . slackspace . openkeepass . crypto . Decrypter ( ) . decryptDatabase ( hashedPassword , cryptoInformation , encryptedDatabase ) ; "<AssertPlaceHolder>" ; } decryptDatabase ( byte [ ] , de . slackspace . openkeepass . crypto . CryptoInformation , byte [ ] ) { byte [ ] aesKey = createAesKey ( password , cryptoInformation ) ; return processDatabaseEncryption ( false , database , cryptoInformation , aesKey ) ; }
org . junit . Assert . assertArrayEquals ( data , decryptedDatabase )
testGetSimpleText ( ) { new org . apache . poi . xdgf . extractor . XDGFVisioExtractor ( xml ) . close ( ) ; new org . apache . poi . xdgf . extractor . XDGFVisioExtractor ( pkg ) . close ( ) ; org . apache . poi . xdgf . extractor . XDGFVisioExtractor extractor = new org . apache . poi . xdgf . extractor . XDGFVisioExtractor ( xml ) ; extractor . getText ( ) ; java . lang . String text = extractor . getText ( ) ; java . lang . String expected = "Text<sp>here\nText<sp>there\nText,<sp>text,<sp>everywhere!\nRouter<sp>here\n" ; "<AssertPlaceHolder>" ; extractor . close ( ) ; } getText ( ) { try { org . apache . poi . hwpf . converter . WordToTextConverter wordToTextConverter = new org . apache . poi . hwpf . converter . WordToTextConverter ( ) ; org . apache . poi . hwpf . usermodel . HeaderStories hs = new org . apache . poi . hwpf . usermodel . HeaderStories ( doc ) ; if ( ( hs . getFirstHeaderSubrange ( ) ) != null ) wordToTextConverter . processDocumentPart ( doc , hs . getFirstHeaderSubrange ( ) ) ; if ( ( hs . getEvenHeaderSubrange ( ) ) != null ) wordToTextConverter . processDocumentPart ( doc , hs . getEvenHeaderSubrange ( ) ) ; if ( ( hs . getOddHeaderSubrange ( ) ) != null ) wordToTextConverter . processDocumentPart ( doc , hs . getOddHeaderSubrange ( ) ) ; wordToTextConverter . processDocument ( doc ) ; wordToTextConverter . processDocumentPart ( doc , doc . getMainTextboxRange ( ) ) ; if ( ( hs . getFirstFooterSubrange ( ) ) != null ) wordToTextConverter . processDocumentPart ( doc , hs . getFirstFooterSubrange ( ) ) ; if ( ( hs . getEvenFooterSubrange ( ) ) != null ) wordToTextConverter . processDocumentPart ( doc , hs . getEvenFooterSubrange ( ) ) ; if ( ( hs . getOddFooterSubrange ( ) ) != null ) wordToTextConverter . processDocumentPart ( doc , hs . getOddFooterSubrange ( ) ) ; return wordToTextConverter . getText ( ) ; } catch ( java . lang . RuntimeException e ) { throw e ; } catch ( java . lang . Exception exc ) { throw new java . lang . RuntimeException ( exc ) ; } }
org . junit . Assert . assertEquals ( expected , text )
spongeSystemEnables ( ) { try { system . enable ( ) ; "<AssertPlaceHolder>" ; } finally { system . disable ( ) ; } } isEnabled ( ) { return enabled ; }
org . junit . Assert . assertTrue ( system . isEnabled ( ) )
testHashCode_null ( ) { fixture . setStream ( ( ( org . eclipse . tracecompass . internal . ctf . core . trace . CTFStream ) ( null ) ) ) ; fixture . setName ( ( ( java . lang . String ) ( null ) ) ) ; int result = fixture . hashCode ( ) ; "<AssertPlaceHolder>" ; } hashCode ( ) { return java . util . Objects . hash ( super . hashCode ( ) , fBaseAddress ) ; }
org . junit . Assert . assertTrue ( ( 0 != result ) )
testThatBuilderRejectsNullKeys ( ) { boolean gotException = false ; try { ninja . Cookie . builder ( null , "" ) ; } catch ( java . lang . NullPointerException nullPointerException ) { gotException = true ; } "<AssertPlaceHolder>" ; } builder ( java . lang . String , java . lang . String ) { com . google . common . base . Preconditions . checkNotNull ( name ) ; com . google . common . base . Preconditions . checkNotNull ( value ) ; return new ninja . Cookie . Builder ( name , value ) ; }
org . junit . Assert . assertTrue ( gotException )
testRewriteBasicPomWithScmExpression ( ) { java . util . List < org . apache . maven . project . MavenProject > reactorProjects = createReactorProjects ( "basic-pom-with-scm-expression" ) ; org . apache . maven . shared . release . config . ReleaseDescriptorBuilder builder = createDescriptorFromProjects ( reactorProjects , "basic-pom-with-scm-expression" ) ; mapNextVersion ( builder , "groupId:artifactId" ) ; phase . execute ( org . apache . maven . shared . release . config . ReleaseUtils . buildReleaseDescriptor ( builder ) , new org . apache . maven . shared . release . env . DefaultReleaseEnvironment ( ) , reactorProjects ) ; "<AssertPlaceHolder>" ; } comparePomFiles ( java . util . List ) { return comparePomFiles ( reactorProjects , true ) ; }
org . junit . Assert . assertTrue ( comparePomFiles ( reactorProjects ) )
cacheTreeShouldBeGreaterThanWorkingTree ( ) { org . eclipse . jface . viewers . Viewer viewer = mock ( org . eclipse . jface . viewers . Viewer . class ) ; org . eclipse . egit . ui . internal . synchronize . mapping . GitChangeSetSorter sorter = new org . eclipse . egit . ui . internal . synchronize . mapping . GitChangeSetSorter ( ) ; org . eclipse . egit . ui . internal . synchronize . model . GitModelCache cache = mock ( org . eclipse . egit . ui . internal . synchronize . model . GitModelCache . class ) ; org . eclipse . egit . ui . internal . synchronize . model . GitModelWorkingTree workingTree = mock ( org . eclipse . egit . ui . internal . synchronize . model . GitModelWorkingTree . class ) ; int actual = sorter . compare ( viewer , cache , workingTree ) ; "<AssertPlaceHolder>" ; } compare ( org . eclipse . compare . ITypedElement , org . eclipse . compare . ITypedElement , org . eclipse . compare . ITypedElement ) { if ( actLeft . getType ( ) . equals ( ITypedElement . FOLDER_TYPE ) ) { org . eclipse . compare . structuremergeviewer . DiffNode diffNode = new org . eclipse . compare . structuremergeviewer . DiffNode ( null , org . eclipse . compare . structuremergeviewer . Differencer . CHANGE , actAncestor , actLeft , actRight ) ; org . eclipse . compare . ITypedElement [ ] lc = ( ( org . eclipse . compare . ITypedElement [ ] ) ( ( ( org . eclipse . compare . structuremergeviewer . IStructureComparator ) ( actLeft ) ) . getChildren ( ) ) ) ; org . eclipse . compare . ITypedElement [ ] rc = ( ( org . eclipse . compare . ITypedElement [ ] ) ( ( ( org . eclipse . compare . structuremergeviewer . IStructureComparator ) ( actRight ) ) . getChildren ( ) ) ) ; org . eclipse . compare . ITypedElement [ ] ac = null ; if ( actAncestor != null ) ac = ( ( org . eclipse . compare . ITypedElement [ ] ) ( ( ( org . eclipse . compare . structuremergeviewer . IStructureComparator ) ( actAncestor ) ) . getChildren ( ) ) ) ; int li = 0 ; int ri = 0 ; while ( ( li < ( lc . length ) ) && ( ri < ( rc . length ) ) ) { org . eclipse . compare . ITypedElement ln = lc [ li ] ; org . eclipse . compare . ITypedElement rn = rc [ ri ] ; org . eclipse . compare . ITypedElement an = null ; if ( ac != null ) an = ac [ ri ] ; int compareTo = ln . getName ( ) . compareTo ( rn . getName ( ) ) ; if ( compareTo == 0 ) { if ( ! ( ln . equals ( rn ) ) ) diffNode . add ( compare ( ln , rn , an ) ) ; ++ li ; ++ ri ; } else if ( compareTo < 0 ) { org . eclipse . compare . structuremergeviewer . DiffNode childDiffNode = new org . eclipse . compare . structuremergeviewer . DiffNode ( org . eclipse . compare . structuremergeviewer . Differencer . ADDITION , an , ln , null ) ; diffNode . add ( childDiffNode ) ; if ( ln . getType ( ) . equals ( ITypedElement . FOLDER_TYPE ) ) { org . eclipse . compare . ITypedElement [ ] children = ( ( org . eclipse . compare . ITypedElement [ ] ) ( ( ( org . eclipse . compare . structuremergeviewer . IStructureComparator ) ( ln ) ) . getChildren ( ) ) ) ; if ( ( children != null ) && ( ( children . length ) > 0 ) ) { for ( org . eclipse . compare . ITypedElement child : children ) { childDiffNode . add ( addDirectoryFiles ( child , Differencer . ADDITION ) ) ; } } } ++ li ; } else { org . eclipse . compare . structuremergeviewer . DiffNode childDiffNode = new org . eclipse . compare . structuremergeviewer . DiffNode ( org . eclipse . compare . structuremergeviewer . Differencer . DELETION , an , null , rn ) ; diffNode . add ( childDiffNode ) ; if ( rn . getType ( ) . equals ( ITypedElement . FOLDER_TYPE ) ) { org . eclipse . compare . ITypedElement [ ] children = ( ( org . eclipse . compare . ITypedElement [ ] ) ( ( ( org . eclipse . compare . structuremergeviewer . IStructureComparator ) ( rn ) ) . getChildren ( ) ) ) ; if ( ( children != null ) && ( ( children . length ) > 0 ) ) { for ( org . eclipse . compare . ITypedElement child : children ) { childDiffNode . add ( addDirectoryFiles ( child , Differencer . DELETION ) ) ; } } } ++ ri ; } } while ( li < ( lc . length ) ) { org . eclipse . compare . ITypedElement ln = lc [ li ] ; org . eclipse . compare . ITypedElement an = null ; if ( ac != null ) an = ac [ li ] ; org . eclipse . compare . structuremergeviewer . DiffNode childDiffNode = new org . eclipse . compare . structuremergeviewer . DiffNode ( org . eclipse . compare . structuremergeviewer . Differencer . ADDITION , an , ln , null ) ; diffNode . add ( childDiffNode ) ; if ( ln . getType ( ) . equals ( ITypedElement . FOLDER_TYPE ) ) { org . eclipse . compare . ITypedElement
org . junit . Assert . assertTrue ( ( actual > 0 ) )
getValueMessage ( ) { org . openhab . binding . zwave . internal . protocol . commandclass . ZWaveBasicCommandClass cls = ( ( org . openhab . binding . zwave . internal . protocol . commandclass . ZWaveBasicCommandClass ) ( getCommandClass ( CommandClass . COMMAND_CLASS_BASIC ) ) ) ; org . openhab . binding . zwave . internal . protocol . transaction . ZWaveCommandClassTransactionPayload msg ; byte [ ] expectedResponseV1 = new byte [ ] { 32 , 2 } ; cls . setVersion ( 1 ) ; msg = cls . getValueMessage ( ) ; "<AssertPlaceHolder>" ; } getPayloadBuffer ( ) { return payload ; }
org . junit . Assert . assertTrue ( java . util . Arrays . equals ( msg . getPayloadBuffer ( ) , expectedResponseV1 ) )
testGetXml ( ) { java . lang . String xml = "xml" ; when ( delegate . getXML ( ) ) . thenReturn ( xml ) ; "<AssertPlaceHolder>" ; } getXML ( ) { return delegate . getXML ( ) ; }
org . junit . Assert . assertEquals ( xml , mapping . getXML ( ) )
matchInvalidXPath ( ) { java . lang . String xmlString = "<xml><doc><id>US123456789</id></doc></xml>" ; gov . uspto . bulkdata . find . PatternXpathValueRegex pattern = new gov . uspto . bulkdata . find . PatternXpathValueRegex ( "//BAD/LOCATION" , "^ABCBAD.+" ) ; gov . uspto . bulkdata . find . PatternMatcher matcher = new gov . uspto . bulkdata . find . PatternMatcher ( ) ; matcher . add ( pattern ) ; "<AssertPlaceHolder>" ; } match ( java . lang . String ) { if ( xmlDocStr == null ) { return false ; } javax . xml . parsers . DocumentBuilderFactory dbfact = javax . xml . parsers . DocumentBuilderFactory . newInstance ( ) ; dbfact . setAttribute ( "http://apache.org/xml/features/nonvalidating/load-external-dtd" , false ) ; try { javax . xml . parsers . DocumentBuilder builder = dbfact . newDocumentBuilder ( ) ; org . xml . sax . InputSource inputSource = new org . xml . sax . InputSource ( new java . io . StringReader ( xmlDocStr ) ) ; org . w3c . dom . Document document = builder . parse ( inputSource ) ; javax . xml . xpath . XPathFactory fact = javax . xml . xpath . XPathFactory . newInstance ( ) ; javax . xml . xpath . XPath xpath = fact . newXPath ( ) ; java . lang . String value = ( ( java . lang . String ) ( xpath . evaluate ( xPathExpression , document , XPathConstants . STRING ) ) ) ; java . util . regex . Matcher matcher = regexPattern . matcher ( value ) ; if ( matcher . matches ( ) ) { return true ; } } catch ( javax . xml . xpath . XPathExpressionException e ) { gov . uspto . bulkdata . find . MatchValueRegex . LOGGER . error ( "XML<sp>XPathExpressionException<sp>on<sp>doc:{}" , xmlDocStr , e ) ; return false ; } catch ( javax . xml . parsers . ParserConfigurationException e ) { gov . uspto . bulkdata . find . MatchValueRegex . LOGGER . error ( "XML<sp>ParserConfigurationException" , e ) ; return false ; } catch ( org . xml . sax . SAXException e ) { gov . uspto . bulkdata . find . MatchValueRegex . LOGGER . error ( "XML<sp>SAXException<sp>on<sp>doc:{}" , xmlDocStr , e ) ; return false ; } catch ( java . io . IOException e ) { gov . uspto . bulkdata . find . MatchValueRegex . LOGGER . error ( "IOException" , e ) ; return false ; } return false ; }
org . junit . Assert . assertFalse ( matcher . match ( xmlString ) )
shouldReturnNoValueOnMissingProperty ( ) { long nodeId = 10 ; org . neo4j . internal . kernel . api . helpers . StubNodeCursor nodeCursor = new org . neo4j . internal . kernel . api . helpers . StubNodeCursor ( ) . withNode ( nodeId , new long [ ] { } , genericMap ( 999 , org . neo4j . values . storable . Values . of ( 12345 ) ) ) ; org . neo4j . kernel . impl . newapi . CursorPropertyAccessor accessor = new org . neo4j . kernel . impl . newapi . CursorPropertyAccessor ( nodeCursor , new org . neo4j . internal . kernel . api . helpers . StubPropertyCursor ( ) , new org . neo4j . internal . kernel . api . helpers . StubRead ( ) ) ; org . neo4j . values . storable . Value readValue = accessor . getNodePropertyValue ( nodeId , 0 ) ; "<AssertPlaceHolder>" ; } getNodePropertyValue ( long , int ) { org . neo4j . kernel . impl . store . record . NodeRecord nodeRecord = nodeStore . newRecord ( ) ; if ( nodeStore . getRecord ( nodeId , nodeRecord , org . neo4j . consistency . checking . full . FORCE ) . inUse ( ) ) { org . neo4j . consistency . checking . full . PropertyReader . SpecificValueVisitor visitor = new org . neo4j . consistency . checking . full . PropertyReader . SpecificValueVisitor ( propertyKeyId ) ; try { if ( visitPropertyRecordChain ( nodeRecord . getNextProp ( ) , visitor ) ) { return visitor . foundPropertyValue ; } } catch ( org . neo4j . consistency . checking . full . PropertyReader . CircularPropertyRecordChainException e ) { } } return org . neo4j . values . storable . Values . NO_VALUE ; }
org . junit . Assert . assertEquals ( org . neo4j . kernel . impl . newapi . NO_VALUE , readValue )
testEmptyString ( ) { javax . validation . Path path = org . apache . bval . jsr . util . PathImpl . createPathFromString ( "" ) ; "<AssertPlaceHolder>" ; } iterator ( ) { @ org . apache . bval . jsr . util . SuppressWarnings ( { "unchecked" , "rawtypes" } ) final java . util . Iterator < javax . validation . Path . Node > result = ( ( java . util . List ) ( nodeList ) ) . iterator ( ) ; return result ; }
org . junit . Assert . assertEquals ( null , path . iterator ( ) . next ( ) . getName ( ) )
duplicateStatuses ( ) { org . apache . accumulo . server . replication . proto . Replication . Status newFile = org . apache . accumulo . server . replication . StatusUtil . fileCreated ( 100 ) ; org . apache . accumulo . server . replication . proto . Replication . Status update1 = org . apache . accumulo . server . replication . StatusUtil . ingestedUntil ( builder , 100 ) ; org . apache . accumulo . server . replication . proto . Replication . Status update2 = org . apache . accumulo . server . replication . StatusUtil . ingestedUntil ( builder , 200 ) ; org . apache . accumulo . server . replication . proto . Replication . Status repl1 = org . apache . accumulo . server . replication . StatusUtil . replicated ( builder , 50 ) ; org . apache . accumulo . server . replication . proto . Replication . Status repl2 = org . apache . accumulo . server . replication . StatusUtil . replicated ( builder , 150 ) ; org . apache . accumulo . server . replication . proto . Replication . Status order1 = combiner . typedReduce ( key , java . util . Arrays . asList ( newFile , update1 , repl1 , update2 , repl2 ) . iterator ( ) ) ; org . apache . accumulo . server . replication . proto . Replication . Status permutation = combiner . typedReduce ( key , java . util . Arrays . asList ( newFile , repl1 , update1 , update1 , repl2 , update2 , update2 ) . iterator ( ) ) ; "<AssertPlaceHolder>" ; } iterator ( ) { if ( ( iflag ) != null ) ( ( org . apache . accumulo . core . iterators . system . InterruptibleIterator ) ( iter ) ) . setInterruptFlag ( iflag ) ; return iter ; }
org . junit . Assert . assertEquals ( order1 , permutation )
testKeyOverwrite ( ) { java . lang . String userName = "user" + ( org . apache . commons . lang3 . RandomStringUtils . randomNumeric ( 5 ) ) ; java . lang . String adminName = "admin" + ( org . apache . commons . lang3 . RandomStringUtils . randomNumeric ( 5 ) ) ; java . lang . String volumeName = "volume" + ( org . apache . commons . lang3 . RandomStringUtils . randomNumeric ( 5 ) ) ; java . lang . String bucketName = "bucket" + ( org . apache . commons . lang3 . RandomStringUtils . randomNumeric ( 5 ) ) ; java . lang . String keyName = "key" + ( org . apache . commons . lang3 . RandomStringUtils . randomNumeric ( 5 ) ) ; long numKeyAllocateFails = omMetrics . getNumKeyAllocateFails ( ) ; org . apache . hadoop . ozone . web . handlers . VolumeArgs createVolumeArgs = new org . apache . hadoop . ozone . web . handlers . VolumeArgs ( volumeName , userArgs ) ; createVolumeArgs . setUserName ( userName ) ; createVolumeArgs . setAdminName ( adminName ) ; storageHandler . createVolume ( createVolumeArgs ) ; org . apache . hadoop . ozone . web . handlers . BucketArgs bucketArgs = new org . apache . hadoop . ozone . web . handlers . BucketArgs ( bucketName , createVolumeArgs ) ; bucketArgs . setAddAcls ( new java . util . LinkedList ( ) ) ; bucketArgs . setRemoveAcls ( new java . util . LinkedList ( ) ) ; bucketArgs . setStorageType ( StorageType . DISK ) ; storageHandler . createBucket ( bucketArgs ) ; org . apache . hadoop . ozone . web . handlers . KeyArgs keyArgs = new org . apache . hadoop . ozone . web . handlers . KeyArgs ( volumeName , bucketName , keyName , userArgs ) ; keyArgs . setSize ( 100 ) ; java . lang . String dataString = org . apache . commons . lang3 . RandomStringUtils . randomAscii ( 100 ) ; try ( java . io . OutputStream stream = storageHandler . newKeyWriter ( keyArgs ) ) { stream . write ( dataString . getBytes ( ) ) ; } org . apache . hadoop . ozone . web . handlers . KeyArgs keyArgs2 = new org . apache . hadoop . ozone . web . handlers . KeyArgs ( volumeName , bucketName , keyName , userArgs ) ; storageHandler . newKeyWriter ( keyArgs2 ) ; "<AssertPlaceHolder>" ; } getNumKeyAllocateFails ( ) { return numKeyAllocateFails . value ( ) ; }
org . junit . Assert . assertEquals ( numKeyAllocateFails , omMetrics . getNumKeyAllocateFails ( ) )
testIsValidStartupPageWithInvalidHtml ( ) { org . eclipse . rap . rwt . cluster . testfixture . test . TestHttpUrlConnection connection = new org . eclipse . rap . rwt . cluster . testfixture . test . TestHttpUrlConnection ( 200 , org . eclipse . rap . rwt . cluster . testfixture . client . Response_Test . TYPE_HTML , "no<sp>html" ) ; org . eclipse . rap . rwt . cluster . testfixture . client . Response response = new org . eclipse . rap . rwt . cluster . testfixture . client . Response ( connection ) ; "<AssertPlaceHolder>" ; } isValidStartupPage ( ) { return ( ( ( ( responseCode ) == 200 ) && ( contentType . toLowerCase ( ) . startsWith ( "text/html" ) ) ) && ( contentType . toLowerCase ( ) . endsWith ( "charset=utf-8" ) ) ) && ( getContentText ( ) . startsWith ( org . eclipse . rap . rwt . cluster . testfixture . client . Response . HTML_PROLOGUE ) ) ; }
org . junit . Assert . assertFalse ( response . isValidStartupPage ( ) )
testGetHash ( ) { java . lang . String s = "EzFY" ; wsattacker . plugin . dos . dosExtension . attackClasses . hashDos . CollisionDJBX31A instance = new wsattacker . plugin . dos . dosExtension . attackClasses . hashDos . CollisionDJBX31A ( ) ; int expResult = 2175080 ; int result = instance . getHash ( s ) ; "<AssertPlaceHolder>" ; } getHash ( java . lang . String ) { byte [ ] byteArray = s . getBytes ( java . nio . charset . Charset . defaultCharset ( ) ) ; int n = byteArray . length ; int key = 0 ; for ( int i = 0 ; i < n ; i ++ ) { int h = key ; key = key << 10 ; for ( int ab = 0 ; ab < 6 ; ab ++ ) { key = ( key << 1 ) + h ; } key = key + ( byteArray [ i ] ) ; } key = key + ( key > > 5 ) ; return key ; }
org . junit . Assert . assertEquals ( expResult , result )
testAMQPValue ( ) { org . fusesource . amqp . codec . AMQPValue in = new org . fusesource . amqp . codec . AMQPValue ( ) ; in . setValue ( new org . fusesource . amqp . codec . AMQPString ( "Hello<sp>world!" ) ) ; org . fusesource . amqp . codec . AMQPValue out = org . fusesource . amqp . codec . TestSupport . writeRead ( in ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( ( "Wrapper{" + "node=" ) + ( node ) ) + ",<sp>weight=" ) + ( weight ) ) + '}' ; }
org . junit . Assert . assertEquals ( in . toString ( ) , out . toString ( ) )
testMatchGraphTargetAtLeastPluginWithWSC01 ( ) { es . usc . citius . composit . core . knowledge . Concept c = es . usc . citius . composit . test . matcher . graph . WSCHashMatchGraphTest . dataset . getKb ( ) . getConcept ( "con1233457844" ) ; java . util . Map < es . usc . citius . composit . core . knowledge . Concept , es . usc . citius . composit . core . matcher . logic . LogicMatchType > expected = new java . util . HashMap < es . usc . citius . composit . core . knowledge . Concept , es . usc . citius . composit . core . matcher . logic . LogicMatchType > ( ) ; expected . put ( es . usc . citius . composit . test . matcher . graph . WSCHashMatchGraphTest . dataset . getKb ( ) . getConcept ( "con1653328292" ) , LogicMatchType . PLUGIN ) ; expected . put ( es . usc . citius . composit . test . matcher . graph . WSCHashMatchGraphTest . dataset . getKb ( ) . getConcept ( "con1988815758" ) , LogicMatchType . PLUGIN ) ; expected . put ( es . usc . citius . composit . test . matcher . graph . WSCHashMatchGraphTest . dataset . getKb ( ) . getConcept ( "con1233457844" ) , LogicMatchType . EXACT ) ; expected . put ( es . usc . citius . composit . test . matcher . graph . WSCHashMatchGraphTest . dataset . getKb ( ) . getConcept ( "con1226699739" ) , LogicMatchType . PLUGIN ) ; java . util . Map < es . usc . citius . composit . core . knowledge . Concept , es . usc . citius . composit . core . matcher . logic . LogicMatchType > result = es . usc . citius . composit . test . matcher . graph . WSCHashMatchGraphTest . matchGraph . getTargetElementsMatchedBy ( c , LogicMatchType . PLUGIN , MatchGraph . TypeSelector . AT_LEAST ) ; "<AssertPlaceHolder>" ; } getTargetElementsMatchedBy ( E , T extends es . usc . citius . composit . core . matcher . graph . Comparable , es . usc . citius . composit . core . matcher . graph . TypeSelector ) { return filter ( getTargetElementsMatchedBy ( source ) , type , selector ) ; }
org . junit . Assert . assertEquals ( expected , result )
testDriverShouldBeAbleToFindElementsAfterLoadingMoreThanOnePageAtATime ( ) { driver . get ( pages . formPage ) ; driver . get ( pages . xhtmlTestPage ) ; org . openqa . selenium . WebElement link = driver . findElement ( org . openqa . selenium . By . linkText ( "click<sp>me" ) ) ; "<AssertPlaceHolder>" ; } getText ( ) { return stb . getText ( ) ; }
org . junit . Assert . assertThat ( link . getText ( ) , org . hamcrest . Matchers . is ( "click<sp>me" ) )
createValid ( ) { org . geotools . grid . PolygonElement oblong = new org . geotools . grid . oblong . OblongImpl ( 0 , 0 , org . geotools . grid . oblong . OblongTest . WIDTH , org . geotools . grid . oblong . OblongTest . HEIGHT , null ) ; "<AssertPlaceHolder>" ; }
org . junit . Assert . assertNotNull ( oblong )
testCopyFormatAttributesWithContributorNameAsString ( ) { org . sonatype . nexus . repository . npm . internal . NpmFormatAttributesExtractor underTest = new org . sonatype . nexus . repository . npm . internal . NpmFormatAttributesExtractor ( new ImmutableMap . Builder < java . lang . String , java . lang . Object > ( ) . put ( "contributors" , "Foo<sp><foo@example.com><sp>(http://www.example.com/foo)" ) . build ( ) ) ; underTest . copyFormatAttributes ( asset ) ; "<AssertPlaceHolder>" ; } formatAttributes ( ) { return component . formatAttributes ( ) ; }
org . junit . Assert . assertThat ( asset . formatAttributes ( ) . get ( "contributors" ) , org . hamcrest . Matchers . is ( singletonList ( "Foo<sp><foo@example.com><sp>(http://www.example.com/foo)" ) ) )
testCopy ( ) { java . lang . String data = "testCopy\nasdf" ; java . io . File file = org . eclipse . aether . internal . test . util . TestFileUtils . createTempFile ( data ) ; java . io . File target = new java . io . File ( targetDir , "testCopy.txt" ) ; fileProcessor . copy ( file , target ) ; "<AssertPlaceHolder>" ; file . delete ( ) ; } readString ( java . io . File ) { byte [ ] content = org . eclipse . aether . internal . test . util . TestFileUtils . readBytes ( file ) ; return new java . lang . String ( content , "UTF-8" ) ; }
org . junit . Assert . assertEquals ( data , org . eclipse . aether . internal . test . util . TestFileUtils . readString ( file ) )
testBuildSqlLockError ( ) { java . lang . String logFileName = constructor . addSQlDEADLOCK ( ) . addSQlTimeout ( ) . build ( LogFileTypes . FILE ) ; org . v8LogScanner . LocalTCPLogScanner . ClientsManager manager = new org . v8LogScanner . LocalTCPLogScanner . ClientsManager ( ) ; org . v8LogScanner . LocalTCPLogScanner . V8LogScannerClient localClient = manager . localClient ( ) ; org . v8LogScanner . rgx . ScanProfile profile = localClient . getProfile ( ) ; org . v8LogScanner . rgx . ScanProfile . buildSqlDeadlLockError ( profile ) ; profile . addLogPath ( logFileName ) ; manager . startRgxOp ( ) ; java . util . List < org . v8LogScanner . rgx . SelectorEntry > logs = localClient . select ( 100 , SelectDirections . FORWARD ) ; "<AssertPlaceHolder>" ; org . v8LogScanner . testV8LogScanner . V8LogFileConstructor . deleteLogFile ( logFileName ) ; } size ( ) { return elements . size ( ) ; }
org . junit . Assert . assertEquals ( 2 , logs . size ( ) )
hydrogensAppearBeforeWhenBondIsFromRight ( ) { org . openscience . cdk . interfaces . IAtom atom1 = mock ( org . openscience . cdk . interfaces . IAtom . class ) ; org . openscience . cdk . interfaces . IAtom atom2 = mock ( org . openscience . cdk . interfaces . IAtom . class ) ; when ( atom1 . getPoint2d ( ) ) . thenReturn ( new javax . vecmath . Point2d ( 0 , 0 ) ) ; when ( atom2 . getPoint2d ( ) ) . thenReturn ( new javax . vecmath . Point2d ( 1 , 0 ) ) ; "<AssertPlaceHolder>" ; } position ( org . openscience . cdk . interfaces . IAtom , java . util . List ) { final java . util . List < javax . vecmath . Vector2d > vectors = org . openscience . cdk . renderer . generators . standard . VecmathUtil . newUnitVectors ( atom , neighbors ) ; if ( ( neighbors . size ( ) ) > 2 ) { return org . openscience . cdk . renderer . generators . standard . HydrogenPosition . usingAngularExtent ( vectors ) ; } else if ( ( neighbors . size ( ) ) > 1 ) { return org . openscience . cdk . renderer . generators . standard . HydrogenPosition . usingCardinalDirection ( org . openscience . cdk . renderer . generators . standard . VecmathUtil . average ( vectors ) ) ; } else if ( ( neighbors . size ( ) ) == 1 ) { return ( vectors . get ( 0 ) . x ) > ( org . openscience . cdk . renderer . generators . standard . HydrogenPosition . VERTICAL_THRESHOLD ) ? org . openscience . cdk . renderer . generators . standard . HydrogenPosition . Left : org . openscience . cdk . renderer . generators . standard . HydrogenPosition . Right ; } else { return org . openscience . cdk . renderer . generators . standard . HydrogenPosition . usingDefaultPlacement ( atom ) ; } }
org . junit . Assert . assertThat ( org . openscience . cdk . renderer . generators . standard . HydrogenPosition . position ( atom1 , java . util . Arrays . asList ( atom2 ) ) , org . hamcrest . CoreMatchers . is ( org . openscience . cdk . renderer . generators . standard . HydrogenPosition . Left ) )
testListsPrevWhenItReturnsEmpty ( ) { final elemental2 . dom . Element dataTypeRow1 = fakeDataTypeRow ( "123" ) ; final elemental2 . dom . Element dataTypeRow2 = fakeDataTypeRow ( "456" ) ; final elemental2 . dom . Element dataTypeRow3 = fakeDataTypeRow ( "789" ) ; final java . util . List < elemental2 . dom . Element > elements = asList ( dataTypeRow1 , dataTypeRow2 , dataTypeRow3 ) ; final java . util . Optional < elemental2 . dom . Element > actual = view . utils . prev ( elements , dataTypeRow1 ) ; "<AssertPlaceHolder>" ; } isPresent ( ) { return isPresent ; }
org . junit . Assert . assertFalse ( actual . isPresent ( ) )
testDeletingLogWhileReading ( ) { java . lang . String name = testName . getMethodName ( ) ; org . apache . distributedlog . api . DistributedLogManager dlm = createNewDLM ( conf , name ) ; org . apache . distributedlog . BKSyncLogWriter out = ( ( org . apache . distributedlog . BKSyncLogWriter ) ( dlm . startLogSegmentNonPartitioned ( ) ) ) ; for ( long i = 1 ; i < 10 ; i ++ ) { org . apache . distributedlog . LogRecord op = org . apache . distributedlog . DLMTestUtil . getLogRecordInstance ( i ) ; out . write ( op ) ; } out . closeAndComplete ( ) ; org . apache . distributedlog . api . LogReader reader = dlm . getInputStream ( 1L ) ; for ( int i = 1 ; i < 10 ; i ++ ) { org . apache . distributedlog . LogRecord record = waitForNextRecord ( reader ) ; "<AssertPlaceHolder>" ; } org . apache . distributedlog . api . DistributedLogManager deleteDLM = createNewDLM ( conf , name ) ; deleteDLM . delete ( ) ; org . apache . distributedlog . LogRecord record ; try { record = reader . readNext ( false ) ; while ( null == record ) { record = reader . readNext ( false ) ; } org . junit . Assert . fail ( "Should<sp>fail<sp>reading<sp>next<sp>with<sp>LogNotFound" ) ; } catch ( org . apache . distributedlog . exceptions . LogNotFoundException lnfe ) { } } getTransactionId ( ) { return txid ; }
org . junit . Assert . assertEquals ( ( ( long ) ( i ) ) , record . getTransactionId ( ) )
shouldReturnTheSameCVectorIfThetaNotNull ( ) { eu . smartenit . sbox . ntm . dtm . sender . ThetaCoefficientHandler tch = new eu . smartenit . sbox . ntm . dtm . sender . ThetaCoefficientHandler ( ) ; eu . smartenit . sbox . db . dto . CVector cVector = new eu . smartenit . sbox . db . dto . CVector ( null , 1 ) ; eu . smartenit . sbox . db . dto . RVector rVector = new eu . smartenit . sbox . db . dto . RVector ( null , 1 , new java . util . ArrayList < eu . smartenit . sbox . db . dto . ThetaCoefficient > ( ) ) ; eu . smartenit . sbox . db . dto . CVector normalizedCVector = tch . normalizeCVector ( cVector , rVector ) ; "<AssertPlaceHolder>" ; } normalizeCVector ( eu . smartenit . sbox . db . dto . CVector , eu . smartenit . sbox . db . dto . RVector ) { if ( ( rVector . getThetaCollection ( ) ) == null ) { eu . smartenit . sbox . ntm . dtm . sender . ThetaCoefficientHandler . logger . debug ( "Theta<sp>collection<sp>in<sp>received<sp>reference<sp>vector<sp>is<sp>not<sp>set." ) ; eu . smartenit . sbox . ntm . dtm . sender . ThetaCoefficientHandler . logger . debug ( "Compensation<sp>vector<sp>normalization<sp>will<sp>not<sp>be<sp>performed." ) ; return cVector ; } return cVector ; }
org . junit . Assert . assertEquals ( cVector , normalizedCVector )
testGetViolationsWithPolicy3 ( ) { eu . atos . sla . datamodel . IPolicy policy ; System . out . println ( "---testEvaluateWithPolicy3()---" ) ; java . util . List < eu . atos . sla . monitoring . IMonitoringMetric > metrics = setUpWithPolicies ( ) ; policy = newPolicy ( "LATENCY" , 3 , new java . util . Date ( 7000 ) ) ; term . setPolicies ( java . util . Arrays . asList ( new eu . atos . sla . datamodel . IPolicy [ ] { policy } ) ) ; java . util . List < eu . atos . sla . datamodel . IViolation > violations = evaluator . evaluate ( contract , term , metrics , now ) ; "<AssertPlaceHolder>" ; } size ( ) { return ( list1 . size ( ) ) + ( list2 . size ( ) ) ; }
org . junit . Assert . assertEquals ( 0 , violations . size ( ) )
testDriverGood ( ) { final java . lang . String result = testWithDummyJADT ( com . pogofish . jadt . JADTTest . GOOD_ARGS , com . pogofish . jadt . JADTTest . NO_SYNTAX_ERRORS , com . pogofish . jadt . JADTTest . NO_SEMANTIC_ERRORS ) ; "<AssertPlaceHolder>" ; } testWithDummyJADT ( java . lang . String [ ] , java . util . List , java . util . List ) { final com . pogofish . jadt . sink . StringSinkFactoryFactory factory = new com . pogofish . jadt . sink . StringSinkFactoryFactory ( ) ; com . pogofish . jadt . JADT . createDummyJADT ( syntaxErrors , semanticErrors , JADT . TEST_SRC_INFO , factory ) . parseAndEmit ( args ) ; return factory . results ( ) . get ( JADT . TEST_DIR ) . get ( 0 ) . getResults ( ) . get ( JADT . TEST_CLASS_NAME ) ; }
org . junit . Assert . assertEquals ( JADT . TEST_SRC_INFO , result )
testAbortedTransactionPutAndGet ( ) { transactionContext . start ( ) ; org . apache . hadoop . hbase . client . Put put = new org . apache . hadoop . hbase . client . Put ( co . cask . tephra . hbase10 . TransactionAwareHTableTest . TestBytes . row ) ; put . add ( co . cask . tephra . hbase10 . TransactionAwareHTableTest . TestBytes . family , co . cask . tephra . hbase10 . TransactionAwareHTableTest . TestBytes . qualifier , co . cask . tephra . hbase10 . TransactionAwareHTableTest . TestBytes . value ) ; transactionAwareHTable . put ( put ) ; transactionContext . abort ( ) ; transactionContext . start ( ) ; org . apache . hadoop . hbase . client . Result result = transactionAwareHTable . get ( new org . apache . hadoop . hbase . client . Get ( co . cask . tephra . hbase10 . TransactionAwareHTableTest . TestBytes . row ) ) ; transactionContext . finish ( ) ; byte [ ] value = result . getValue ( co . cask . tephra . hbase10 . TransactionAwareHTableTest . TestBytes . family , co . cask . tephra . hbase10 . TransactionAwareHTableTest . TestBytes . qualifier ) ; "<AssertPlaceHolder>" ; } finish ( ) { com . google . common . base . Preconditions . checkState ( ( ( currentTx ) != null ) , "Cannot<sp>finish<sp>tx<sp>that<sp>has<sp>not<sp>been<sp>started" ) ; checkForConflicts ( ) ; persist ( ) ; commit ( ) ; postCommit ( ) ; currentTx = null ; }
org . junit . Assert . assertArrayEquals ( value , null )
testFindWithFilterEqualOrderBy ( ) { createIndexIfNotExists ( org . springframework . data . aerospike . sample . Person . class , "age_index" , "age_index" 5 , IndexType . NUMERIC ) ; createIndexIfNotExists ( org . springframework . data . aerospike . sample . Person . class , "dave-004" 9 , "age_index" 1 , IndexType . STRING ) ; client . put ( policy , new com . aerospike . client . Key ( getNameSpace ( ) , org . springframework . data . aerospike . core . AerospikeTemplateIntegrationTests . SET_NAME_PERSON , "dave-001" ) , new com . aerospike . client . Bin ( "firstname" , "dave-004" 7 ) , new com . aerospike . client . Bin ( "age_index" 1 , "age_index" 2 ) , new com . aerospike . client . Bin ( "age_index" 5 , 21 ) ) ; client . put ( policy , new com . aerospike . client . Key ( getNameSpace ( ) , org . springframework . data . aerospike . core . AerospikeTemplateIntegrationTests . SET_NAME_PERSON , "dave-002" ) , new com . aerospike . client . Bin ( "firstname" , "dave-004" 5 ) , new com . aerospike . client . Bin ( "age_index" 1 , "age_index" 2 ) , new com . aerospike . client . Bin ( "age_index" 5 , 22 ) ) ; client . put ( policy , new com . aerospike . client . Key ( getNameSpace ( ) , org . springframework . data . aerospike . core . AerospikeTemplateIntegrationTests . SET_NAME_PERSON , "dave-004" 3 ) , new com . aerospike . client . Bin ( "firstname" , "age_index" 3 ) , new com . aerospike . client . Bin ( "age_index" 1 , "age_index" 2 ) , new com . aerospike . client . Bin ( "age_index" 5 , 23 ) ) ; client . put ( policy , new com . aerospike . client . Key ( getNameSpace ( ) , org . springframework . data . aerospike . core . AerospikeTemplateIntegrationTests . SET_NAME_PERSON , "dave-004" ) , new com . aerospike . client . Bin ( "firstname" , "Dave" ) , new com . aerospike . client . Bin ( "age_index" 1 , "age_index" 2 ) , new com . aerospike . client . Bin ( "age_index" 5 , 24 ) ) ; client . put ( policy , new com . aerospike . client . Key ( getNameSpace ( ) , org . springframework . data . aerospike . core . AerospikeTemplateIntegrationTests . SET_NAME_PERSON , "dave-004" 1 ) , new com . aerospike . client . Bin ( "firstname" , "dave-004" 0 ) , new com . aerospike . client . Bin ( "age_index" 1 , "age_index" 2 ) , new com . aerospike . client . Bin ( "age_index" 5 , 25 ) ) ; client . put ( policy , new com . aerospike . client . Key ( getNameSpace ( ) , org . springframework . data . aerospike . core . AerospikeTemplateIntegrationTests . SET_NAME_PERSON , "age_index" 0 ) , new com . aerospike . client . Bin ( "firstname" , "dave-004" 8 ) , new com . aerospike . client . Bin ( "age_index" 1 , "age_index" 2 ) , new com . aerospike . client . Bin ( "age_index" 5 , 26 ) ) ; client . put ( policy , new com . aerospike . client . Key ( getNameSpace ( ) , org . springframework . data . aerospike . core . AerospikeTemplateIntegrationTests . SET_NAME_PERSON , "dave-007" ) , new com . aerospike . client . Bin ( "firstname" , "dave-004" 6 ) , new com . aerospike . client . Bin ( "age_index" 1 , "age_index" 2 ) , new com . aerospike . client . Bin ( "age_index" 5 , 27 ) ) ; client . put ( policy , new com . aerospike . client . Key ( getNameSpace ( ) , org . springframework . data . aerospike . core . AerospikeTemplateIntegrationTests . SET_NAME_PERSON , "dave-008" ) , new com . aerospike . client . Bin ( "firstname" , "dave-004" 4 ) , new com . aerospike . client . Bin ( "age_index" 1 , "age_index" 2 ) , new com . aerospike . client . Bin ( "age_index" 5 , 28 ) ) ; client . put ( policy , new com . aerospike . client . Key ( getNameSpace ( ) , org . springframework . data . aerospike . core . AerospikeTemplateIntegrationTests . SET_NAME_PERSON , "age_index" 6 ) , new com . aerospike . client . Bin ( "firstname" , "Alister" ) , new com . aerospike . client . Bin ( "age_index" 1 , "age_index" 2 ) , new com . aerospike . client . Bin ( "age_index" 5 , 29 ) ) ; client . put ( policy , new com . aerospike . client . Key ( getNameSpace ( ) , org . springframework . data . aerospike . core . AerospikeTemplateIntegrationTests . SET_NAME_PERSON , "dave-010" ) , new com . aerospike . client . Bin ( "firstname" , "age_index" 4 ) , new com . aerospike . client . Bin ( "age_index" 1 , "age_index" 2 ) , new com . aerospike . client . Bin ( "age_index" 5 , 30 ) ) ; org . springframework . data . aerospike . repository . query . Query query = createQueryForMethodWithArgs ( "dave-004" 2 , "age_index" 2 ) ; java . lang . Iterable < org . springframework . data . aerospike . sample . Person > it = template . find ( query , org . springframework . data . aerospike . sample . Person . class ) ; int count = 0 ; for ( org . springframework . data . aerospike . sample . Person person : it ) { count ++ ; } "<AssertPlaceHolder>" ; } find ( org . springframework . data . keyvalue . core . query . KeyValueQuery ,
org . junit . Assert . assertEquals ( 10 , count )
booleanArrayTest ( ) { com . jfireframework . baseutil . collection . buffer . ByteBuf < ? > buf = com . jfireframework . baseutil . collection . buffer . HeapByteBuf . allocate ( 256 ) ; boolean [ ] array = new boolean [ ] { true , false , false , true , true , true } ; com . jfireframework . licp . Licp lbse = new com . jfireframework . licp . Licp ( ) ; lbse . serialize ( array , buf . clear ( ) ) ; byte [ ] src = buf . toArray ( ) ; java . nio . ByteBuffer buffer = java . nio . ByteBuffer . allocate ( src . length ) ; buffer . put ( src ) . flip ( ) ; boolean [ ] result = ( ( boolean [ ] ) ( lbse . deserialize ( buffer ) ) ) ; for ( int i = 0 ; i < ( array . length ) ; i ++ ) { "<AssertPlaceHolder>" ; } } deserialize ( java . nio . ByteBuffer ) { collect . clear ( ) ; register . clear ( ) ; if ( ( buffer . get ( ) ) != ( com . jfireframework . licp . Licp . version ) ) { throw new com . jfireframework . baseutil . exception . UnSupportException ( "licp" ) ; } return ( ( T ) ( _deserialize ( buffer ) ) ) ; }
org . junit . Assert . assertEquals ( array [ i ] , result [ i ] )
testMapToKeyPartial ( ) { one . util . streamex . Map < java . lang . Integer , java . lang . Integer > original = new one . util . streamex . HashMap ( ) ; original . put ( 1 , 1 ) ; original . put ( 2 , 5 ) ; original . put ( 3 , 3 ) ; original . put ( 4 , 4 ) ; one . util . streamex . Map < java . lang . Integer , java . lang . Integer > expected = new one . util . streamex . HashMap ( ) ; expected . put ( 1 , 1 ) ; expected . put ( 9 , 3 ) ; expected . put ( 16 , 4 ) ; one . util . streamex . Map < java . lang . Integer , java . lang . Integer > actual = one . util . streamex . EntryStream . of ( original ) . mapToKeyPartial ( ( key , value ) -> { if ( key . equals ( value ) ) { return one . util . streamex . Optional . of ( ( key * value ) ) ; } return one . util . streamex . Optional . empty ( ) ; } ) . toMap ( ) ; "<AssertPlaceHolder>" ; } toMap ( ) { one . util . streamex . Map < K , V > map = ( one . util . streamex . EntryStream . isParallel ( ) ) ? new java . util . concurrent . ConcurrentHashMap ( ) : new one . util . streamex . HashMap ( ) ; forEach ( one . util . streamex . EntryStream . toMapConsumer ( map ) ) ; return map ; }
org . junit . Assert . assertEquals ( expected , actual )
UpdateEmployeeName ( ) { newSession ( ) ; com . mysema . rdfbean . object . UpdateTest . Employee employee = getEmployee ( ) ; employee . name = "Jane<sp>Doe" ; session . save ( employee ) ; newSession ( ) ; employee = getEmployee ( ) ; "<AssertPlaceHolder>" ; } getEmployee ( ) { return session . get ( com . mysema . rdfbean . object . UpdateTest . Employee . class , ids . get ( 0 ) ) ; }
org . junit . Assert . assertEquals ( "Jane<sp>Doe" , employee . name )
testCreateTable ( ) { com . nearinfinity . blur . thrift . generated . Blur . Iface client = getClient ( ) ; com . nearinfinity . blur . thrift . generated . TableDescriptor tableDescriptor = new com . nearinfinity . blur . thrift . generated . TableDescriptor ( ) ; tableDescriptor . setName ( "test" ) ; tableDescriptor . setShardCount ( 5 ) ; tableDescriptor . setTableUri ( ( ( com . nearinfinity . blur . MiniCluster . getFileSystemUri ( ) . toString ( ) ) + "/blur/test" ) ) ; client . createTable ( tableDescriptor ) ; java . util . List < java . lang . String > tableList = client . tableList ( ) ; "<AssertPlaceHolder>" ; } tableList ( ) { try { return _clusterStatus . getTableList ( true ) ; } catch ( java . lang . Exception e ) { com . nearinfinity . blur . thrift . TableAdmin . LOG . error ( "Unknown<sp>error<sp>while<sp>trying<sp>to<sp>get<sp>a<sp>table<sp>list." , e ) ; throw new com . nearinfinity . blur . thrift . BException ( "Unknown<sp>error<sp>while<sp>trying<sp>to<sp>get<sp>a<sp>table<sp>list." , e ) ; } }
org . junit . Assert . assertEquals ( java . util . Arrays . asList ( "test" ) , tableList )
testSetsSaveToAlbum ( ) { com . eclipsesource . tabris . camera . CameraOptions cameraOptions = new com . eclipsesource . tabris . camera . CameraOptions ( ) ; cameraOptions . setSaveToAlbum ( true ) ; "<AssertPlaceHolder>" ; } savesToAlbum ( ) { return saveToAlbum ; }
org . junit . Assert . assertTrue ( cameraOptions . savesToAlbum ( ) )
find_A$Function1_Nil ( ) { com . m3 . scalaflavor4j . Seq < java . lang . String > seq = com . m3 . scalaflavor4j . Seq . apply ( ) ; com . m3 . scalaflavor4j . F1 < java . lang . String , java . lang . Boolean > p = new com . m3 . scalaflavor4j . F1 < java . lang . String , java . lang . Boolean > ( ) { public com . m3 . scalaflavor4j . Boolean apply ( java . lang . String v1 ) { return v1 . equals ( "c" ) ; } } ; com . m3 . scalaflavor4j . Option < java . lang . String > actual = seq . find ( p ) ; "<AssertPlaceHolder>" ; } isDefined ( ) { return false ; }
org . junit . Assert . assertThat ( actual . isDefined ( ) , org . hamcrest . CoreMatchers . is ( false ) )
testConvertAll ( ) { java . lang . String abbrName = "abbrName" ; java . lang . String name = "name" ; java . lang . String abbrName1 = "abbrName1" ; java . lang . String name1 = "name1" ; java . lang . Long parentId = 2L ; org . lnu . is . domain . order . type . OrderType parent = new org . lnu . is . domain . order . type . OrderType ( ) ; parent . setId ( parentId ) ; org . lnu . is . domain . order . type . OrderType source = new org . lnu . is . domain . order . type . OrderType ( ) ; source . setAbbrName ( abbrName ) ; source . setName ( name ) ; source . setParent ( parent ) ; org . lnu . is . resource . order . type . OrderTypeResource expected = new org . lnu . is . resource . order . type . OrderTypeResource ( ) ; expected . setAbbrName ( abbrName ) ; expected . setName ( name ) ; expected . setParentId ( parentId ) ; org . lnu . is . domain . order . type . OrderType source1 = new org . lnu . is . domain . order . type . OrderType ( ) ; source1 . setAbbrName ( abbrName1 ) ; source1 . setName ( name1 ) ; org . lnu . is . resource . order . type . OrderTypeResource expected1 = new org . lnu . is . resource . order . type . OrderTypeResource ( ) ; expected1 . setAbbrName ( abbrName1 ) ; expected1 . setName ( name1 ) ; java . util . List < org . lnu . is . resource . order . type . OrderTypeResource > expecteds = java . util . Arrays . asList ( expected , expected1 ) ; java . util . List < org . lnu . is . domain . order . type . OrderType > sources = java . util . Arrays . asList ( source , source1 ) ; java . util . List < org . lnu . is . resource . order . type . OrderTypeResource > actuals = unit . convertAll ( sources ) ; "<AssertPlaceHolder>" ; } convertAll ( java . util . List ) { return convertAll ( sources , new java . util . ArrayList < TARGET > ( sources . size ( ) ) ) ; }
org . junit . Assert . assertEquals ( expecteds , actuals )
testRangeDelete ( ) { org . apache . poi . hwpf . HWPFDocument doc = org . apache . poi . hwpf . HWPFTestDataSamples . openSampleFile ( "Bug28627.doc" ) ; org . apache . poi . hwpf . usermodel . Range range = doc . getRange ( ) ; int numParagraphs = range . numParagraphs ( ) ; int totalLength = 0 ; int deletedLength = 0 ; for ( int i = 0 ; i < numParagraphs ; i ++ ) { org . apache . poi . hwpf . usermodel . Paragraph para = range . getParagraph ( i ) ; java . lang . String text = para . text ( ) ; totalLength += text . length ( ) ; if ( text . contains ( "{delete<sp>me}" ) ) { para . delete ( ) ; deletedLength = text . length ( ) ; } } int newLength = 0 ; range = doc . getRange ( ) ; numParagraphs = range . numParagraphs ( ) ; for ( int i = 0 ; i < numParagraphs ; i ++ ) { org . apache . poi . hwpf . usermodel . Paragraph para = range . getParagraph ( i ) ; java . lang . String text = para . text ( ) ; newLength += text . length ( ) ; } "<AssertPlaceHolder>" ; doc . close ( ) ; } length ( ) { return getString ( ) . length ( ) ; }
org . junit . Assert . assertEquals ( newLength , ( totalLength - deletedLength ) )
testBoundsOfOneRectangle ( ) { java . util . ArrayList < technology . tabula . Rectangle > shapes = new java . util . ArrayList ( ) ; shapes . add ( new technology . tabula . Rectangle ( 0 , 0 , 20 , 40 ) ) ; technology . tabula . Rectangle r = technology . tabula . Utils . bounds ( shapes ) ; "<AssertPlaceHolder>" ; } get ( int ) { return begin + index ; }
org . junit . Assert . assertEquals ( r , shapes . get ( 0 ) )
testSetNetworkType ( ) { defaultIsisInterface . setNetworkType ( IsisNetworkType . BROADCAST ) ; resultNwType = defaultIsisInterface . networkType ( ) ; "<AssertPlaceHolder>" ; } is ( java . lang . Class ) { return true ; }
org . junit . Assert . assertThat ( resultNwType , org . hamcrest . CoreMatchers . is ( IsisNetworkType . BROADCAST ) )
testSetVersion ( ) { java . lang . String content = "<extension><version>OLD_VERSION</version></extension>" ; org . jdom . Element extensionElm = builder . build ( new java . io . StringReader ( content ) ) . getRootElement ( ) ; new org . apache . maven . shared . release . transform . jdom . JDomExtension ( extensionElm ) . setVersion ( "NEW_VERSION" ) ; "<AssertPlaceHolder>" ; } getVersion ( org . jdom . Element ) { return extensionElm . getChildTextTrim ( "version" , extensionElm . getNamespace ( ) ) ; }
org . junit . Assert . assertEquals ( "NEW_VERSION" , getVersion ( extensionElm ) )