input
stringlengths 35
67.8k
| output
stringlengths 39
1.53k
|
|---|---|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . METHOD_1 ( ) ; org . apache . hadoop . hdfs . IDENT_0 cluster = new org . apache . hadoop . hdfs . IDENT_0 . Builder ( conf ) . format ( true ) . build ( ) ; org . apache . hadoop . fs . IDENT_1 fs = cluster . METHOD_2 ( ) ; org . apache . hadoop . hdfs . IDENT_2 IDENT_3 = ( ( org . apache . hadoop . hdfs . IDENT_2 ) ( org . apache . hadoop . fs . IDENT_1 . newInstance ( fs . METHOD_3 ( ) , fs . METHOD_4 ( ) ) ) ) ; try { fs . METHOD_5 ( new org . apache . hadoop . fs . Path ( STRING_0 ) ) ; fs . METHOD_5 ( new org . apache . hadoop . fs . Path ( STRING_1 ) ) ; fs . METHOD_5 ( new org . apache . hadoop . fs . Path ( STRING_2 ) ) ; "<AssertPlaceHolder>" ; } finally { cluster . METHOD_6 ( ) ; } } delete ( org . apache . hadoop . fs . Path , boolean ) { return false ; }
|
org . junit . Assert . assertTrue ( fs . delete ( new org . apache . hadoop . fs . Path ( STRING_0 ) , true ) )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . METHOD_1 ( ) ; final int IDENT_0 = 1 ; final int IDENT_1 = IDENT_0 ; final int IDENT_2 = 1 * IDENT_1 ; conf . METHOD_2 ( IDENT_3 . IDENT_4 , IDENT_0 ) ; conf . METHOD_3 ( IDENT_3 . IDENT_5 , IDENT_2 ) ; conf . METHOD_2 ( IDENT_3 . IDENT_6 , IDENT_1 ) ; org . apache . hadoop . hdfs . IDENT_7 cluster = new org . apache . hadoop . hdfs . IDENT_7 . Builder ( conf ) . METHOD_4 ( 1 ) . format ( true ) . build ( ) ; org . apache . hadoop . fs . IDENT_8 fs = cluster . METHOD_5 ( ) ; org . apache . hadoop . hdfs . IDENT_9 IDENT_10 = ( ( org . apache . hadoop . hdfs . IDENT_9 ) ( org . apache . hadoop . fs . IDENT_8 . newInstance ( fs . METHOD_6 ( ) , fs . METHOD_7 ( ) ) ) ) ; try { final int IDENT_11 = 10 ; final int IDENT_12 = 2 ; org . apache . hadoop . fs . Path base = new org . apache . hadoop . fs . Path ( STRING_0 ) ; IDENT_10 . METHOD_8 ( base ) ; for ( int f = 0 ; f < IDENT_11 ; f ++ ) { org . apache . hadoop . hdfs . IDENT_13 out = IDENT_10 . create ( new org . apache . hadoop . fs . Path ( base , ( "test" + f ) ) ) ; for ( int k = 0 ; k < IDENT_12 ; k ++ ) { out . write ( k ) ; } out . close ( ) ; } for ( int f = 0 ; f < IDENT_11 ; f ++ ) { org . apache . hadoop . hdfs . IDENT_14 in = IDENT_10 . METHOD_9 ( new org . apache . hadoop . fs . Path ( base , ( "test" + f ) ) ) ; for ( int k = 0 ; k < IDENT_12 ; k ++ ) { "<AssertPlaceHolder>" ; } in . close ( ) ; } } catch ( java . lang . Exception e ) { e . printStackTrace ( ) ; } finally { cluster . METHOD_10 ( ) ; } } read ( ) { if ( ( IDENT_15 ) > 0 ) { try { java . lang . Thread . sleep ( IDENT_16 ) ; } catch ( java . lang . IDENT_17 e ) { } ( IDENT_15 ) OP_0 ; return 1 ; } return - 1 ; }
|
org . junit . Assert . assertTrue ( ( ( in . read ( ) ) == k ) )
|
METHOD_0 ( ) { final int IDENT_0 = 3 ; org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . METHOD_1 ( ) ; conf . METHOD_2 ( IDENT_1 . IDENT_2 , 1 ) ; conf . METHOD_2 ( IDENT_1 . IDENT_3 , IDENT_0 ) ; org . apache . hadoop . hdfs . IDENT_4 cluster = new org . apache . hadoop . hdfs . IDENT_4 . Builder ( conf ) . METHOD_3 ( IDENT_0 ) . build ( ) ; try { final org . apache . hadoop . fs . IDENT_5 fs = cluster . METHOD_4 ( ) ; final org . apache . hadoop . fs . Path IDENT_6 = new org . apache . hadoop . fs . Path ( STRING_0 ) ; org . apache . hadoop . hdfs . IDENT_7 . IDENT_8 [ ] IDENT_9 = new org . apache . hadoop . hdfs . IDENT_7 . IDENT_8 [ 10 ] ; for ( int i = 0 ; i < ( IDENT_9 . length ) ; i ++ ) { IDENT_9 [ i ] = new org . apache . hadoop . hdfs . IDENT_7 . METHOD_5 ( fs , new org . apache . hadoop . fs . Path ( IDENT_6 , ( STRING_1 + i ) ) ) ; } try { for ( org . apache . hadoop . hdfs . IDENT_7 . METHOD_5 IDENT_10 : IDENT_9 ) { IDENT_10 . start ( ) ; } java . lang . Thread . sleep ( 1000 ) ; cluster . METHOD_6 ( org . apache . hadoop . hdfs . IDENT_11 . METHOD_7 ( IDENT_0 ) ) ; System . out . println ( STRING_2 ) ; java . lang . Thread . sleep ( INT_0 ) ; } finally { for ( org . apache . hadoop . hdfs . IDENT_7 . METHOD_5 IDENT_12 : IDENT_9 ) { if ( IDENT_12 != null ) { IDENT_12 . IDENT_13 = false ; IDENT_12 . METHOD_8 ( ) ; } } for ( org . apache . hadoop . hdfs . IDENT_7 . METHOD_5 IDENT_10 : IDENT_9 ) { if ( IDENT_10 != null ) { IDENT_10 . join ( ) ; } } } System . out . println ( STRING_3 ) ; for ( org . apache . hadoop . hdfs . IDENT_7 . METHOD_5 IDENT_10 : IDENT_9 ) { System . out . println ( ( ( ( IDENT_10 . IDENT_14 ) + STRING_4 ) + ( fs . METHOD_9 ( IDENT_10 . IDENT_14 ) . METHOD_10 ( ) ) ) ) ; org . apache . hadoop . fs . IDENT_15 in = null ; try { in = fs . METHOD_11 ( IDENT_10 . IDENT_14 ) ; for ( int j = 0 , x ; ( x = in . read ( ) ) != ( - 1 ) ; j ++ ) { "<AssertPlaceHolder>" ; } } finally { org . apache . hadoop . io . IDENT_16 . METHOD_12 ( in ) ; } } } finally { if ( cluster != null ) { cluster . METHOD_13 ( ) ; } } } read ( ) { if ( ( IDENT_17 ) > 0 ) { try { java . lang . Thread . sleep ( IDENT_18 ) ; } catch ( java . lang . IDENT_19 e ) { } ( IDENT_17 ) OP_0 ; return 1 ; } return - 1 ; }
|
org . junit . Assert . assertEquals ( j , x )
|
METHOD_0 ( ) { final java . lang . Thread . IDENT_0 IDENT_1 = java . lang . Thread . METHOD_1 ( ) ; java . lang . Thread . METHOD_2 ( new java . lang . Thread . METHOD_3 ( ) { @ org . apache . hadoop . hdfs . Override public void METHOD_4 ( java . lang . Thread t , java . lang . Throwable e ) { if ( e instanceof java . util . IDENT_2 ) { IDENT_3 . LOG . error ( ( STRING_0 + t ) , e ) ; IDENT_4 = true ; } } } ) ; System . out . println ( STRING_1 ) ; final long IDENT_5 = 1000 ; final int IDENT_6 = 3 ; final org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . METHOD_5 ( ) ; conf . METHOD_6 ( IDENT_7 . IDENT_8 , 1000 ) ; conf . METHOD_6 ( IDENT_7 . IDENT_9 , 1 ) ; org . apache . hadoop . hdfs . IDENT_10 cluster = new org . apache . hadoop . hdfs . IDENT_10 . Builder ( conf ) . METHOD_7 ( IDENT_6 ) . build ( ) ; try { cluster . METHOD_8 ( ) ; org . apache . hadoop . hdfs . IDENT_11 IDENT_12 = ( ( org . apache . hadoop . hdfs . IDENT_11 ) ( cluster . METHOD_9 ( ) ) ) ; org . apache . hadoop . hdfs . IDENT_13 . METHOD_10 ( IDENT_12 , new org . apache . hadoop . fs . Path ( STRING_2 ) , IDENT_6 ) ; org . apache . hadoop . hdfs . IDENT_13 . METHOD_10 ( IDENT_12 , new org . apache . hadoop . fs . Path ( STRING_3 ) , IDENT_6 ) ; org . apache . hadoop . hdfs . IDENT_13 . METHOD_10 ( IDENT_12 , new org . apache . hadoop . fs . Path ( STRING_4 ) , IDENT_6 ) ; cluster . METHOD_11 ( IDENT_5 , IDENT_5 ) ; try { java . lang . Thread . sleep ( ( 5 * IDENT_5 ) ) ; } catch ( java . lang . IDENT_14 e ) { } "<AssertPlaceHolder>" ; } finally { java . lang . Thread . METHOD_2 ( IDENT_1 ) ; cluster . METHOD_12 ( ) ; } } sleep ( long ) { org . apache . hadoop . IDENT_15 . IDENT_16 . LOG . info ( ( ( STRING_5 + IDENT_17 ) + STRING_6 ) ) ; try { java . lang . Thread . sleep ( IDENT_17 ) ; } catch ( java . lang . IDENT_14 e ) { org . apache . hadoop . IDENT_15 . IDENT_16 . LOG . info ( STRING_7 , e ) ; return false ; } return true ; }
|
org . junit . Assert . assertFalse ( IDENT_4 )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . METHOD_1 ( ) ; org . apache . hadoop . hdfs . IDENT_0 cluster = new org . apache . hadoop . hdfs . IDENT_0 . Builder ( conf ) . METHOD_2 ( 2 ) . build ( ) ; org . apache . hadoop . hdfs . IDENT_1 IDENT_2 = ( ( org . apache . hadoop . hdfs . IDENT_1 ) ( cluster . METHOD_3 ( ) ) ) ; try { org . apache . hadoop . fs . Path path = new org . apache . hadoop . fs . Path ( IDENT_3 ) ; org . apache . hadoop . fs . IDENT_4 IDENT_5 = IDENT_2 . create ( path , true , INT_0 , ( ( short ) ( 2 ) ) , IDENT_6 . IDENT_7 ) ; System . out . println ( ( STRING_0 + ( path . toString ( ) ) ) ) ; ( ( org . apache . hadoop . hdfs . IDENT_8 ) ( IDENT_5 . METHOD_4 ( ) ) ) . METHOD_5 ( java . util . IDENT_9 . of ( IDENT_10 . IDENT_11 ) ) ; long IDENT_12 = IDENT_2 . METHOD_6 ( path ) . METHOD_7 ( ) ; "<AssertPlaceHolder>" ; IDENT_5 . close ( ) ; } finally { IDENT_2 . close ( ) ; cluster . METHOD_8 ( ) ; } } METHOD_7 ( ) { return length ; }
|
org . junit . Assert . assertEquals ( 0L , IDENT_12 )
|
METHOD_0 ( ) { org . apache . hadoop . hdfs . IDENT_0 cluster = new org . apache . hadoop . hdfs . IDENT_0 . Builder ( conf ) . METHOD_1 ( 2 ) . build ( ) ; try { cluster . METHOD_2 ( ) ; org . apache . hadoop . hdfs . server . protocol . IDENT_1 IDENT_2 = cluster . METHOD_3 ( ) ; org . apache . hadoop . hdfs . server . protocol . IDENT_1 IDENT_3 = METHOD_4 ( IDENT_2 ) ; org . apache . hadoop . hdfs . IDENT_4 IDENT_5 = new org . apache . hadoop . hdfs . METHOD_5 ( null , IDENT_3 , conf , null ) ; byte [ ] buf = new byte [ INT_0 ] ; org . apache . hadoop . fs . IDENT_6 IDENT_7 = METHOD_6 ( IDENT_5 , ( ( org . apache . hadoop . hdfs . IDENT_8 . IDENT_9 ) + STRING_0 ) ) ; IDENT_7 . write ( buf , 0 , INT_0 ) ; IDENT_7 . close ( ) ; org . apache . hadoop . hdfs . IDENT_10 IDENT_11 = IDENT_5 . METHOD_7 ( ( ( org . apache . hadoop . hdfs . IDENT_8 . IDENT_9 ) + STRING_0 ) ) ; org . apache . hadoop . fs . IDENT_6 IDENT_12 = METHOD_6 ( IDENT_5 , ( ( org . apache . hadoop . hdfs . IDENT_8 . IDENT_9 ) + STRING_1 ) ) ; METHOD_8 ( new org . apache . hadoop . IDENT_13 . METHOD_9 ( org . apache . hadoop . security . token . IDENT_14 . IDENT_15 . class . getName ( ) , STRING_2 ) ) . when ( IDENT_3 ) . METHOD_10 ( anyString ( ) ) ; org . apache . hadoop . hdfs . IDENT_16 IDENT_17 = IDENT_5 . METHOD_11 ( ) ; IDENT_5 . IDENT_18 = ( ( org . apache . hadoop . util . IDENT_19 . now ( ) ) - ( org . apache . hadoop . hdfs . protocol . IDENT_20 . IDENT_21 ) ) - 1000 ; try { IDENT_5 . METHOD_10 ( ) ; } catch ( java . io . IOException e ) { } try { IDENT_12 . write ( buf , 0 , INT_0 ) ; org . apache . hadoop . hdfs . IDENT_8 . LOG . info ( STRING_0 1 ) ; } catch ( java . io . IOException e ) { org . junit . Assert . fail ( STRING_3 ) ; } IDENT_5 . IDENT_18 = ( ( org . apache . hadoop . util . IDENT_19 . now ( ) ) - ( org . apache . hadoop . hdfs . protocol . IDENT_20 . IDENT_22 ) ) - 1000 ; IDENT_5 . METHOD_10 ( ) ; try { IDENT_12 . write ( buf , 0 , INT_0 ) ; IDENT_12 . close ( ) ; org . junit . Assert . fail ( STRING_4 ) ; } catch ( java . io . IOException e ) { org . apache . hadoop . hdfs . IDENT_8 . LOG . info ( STRING_5 , e ) ; } java . lang . Thread . sleep ( 1000 ) ; "<AssertPlaceHolder>" ; METHOD_12 ( ) . when ( IDENT_3 ) . METHOD_10 ( anyString ( ) ) ; try { int IDENT_23 = IDENT_11 . read ( buf , 0 , 1 ) ; if ( IDENT_23 != 1 ) { org . junit . Assert . fail ( STRING_6 ) ; } IDENT_11 . close ( ) ; } catch ( java . io . IOException e ) { org . apache . hadoop . hdfs . IDENT_8 . LOG . error ( STRING_0 0 , e ) ; org . junit . Assert . fail ( STRING_7 ) ; } try { IDENT_7 = METHOD_6 ( IDENT_5 , ( ( org . apache . hadoop . hdfs . IDENT_8 . IDENT_9 ) + STRING_0 ) ) ; IDENT_7 . write ( buf , 0 , INT_0 ) ; IDENT_7 . close ( ) ; } catch ( java . io . IOException e ) { org . apache . hadoop . hdfs . IDENT_8 . LOG . error ( STRING_8 , e ) ; org . junit . Assert . fail ( STRING_9 ) ; } } finally { cluster . METHOD_13 ( ) ; } } isEmpty ( ) { return ( IDENT_24 . METHOD_14 ( 0 ) ) == ( - 1 ) ; }
|
org . junit . Assert . assertTrue ( IDENT_17 . isEmpty ( ) )
|
METHOD_0 ( ) { final java . lang . String [ ] IDENT_0 = new java . lang . String [ ] { STRING_0 , STRING_1 } ; org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . METHOD_1 ( ) ; for ( final java . lang . String IDENT_1 : IDENT_0 ) { conf . set ( IDENT_2 . IDENT_3 , IDENT_1 ) ; org . apache . hadoop . hdfs . IDENT_4 cluster = new org . apache . hadoop . hdfs . IDENT_4 . Builder ( conf ) . build ( ) ; org . apache . hadoop . fs . IDENT_5 IDENT_6 = cluster . METHOD_2 ( ) ; try { org . apache . hadoop . fs . Path IDENT_7 = IDENT_6 . METHOD_3 ( new org . apache . hadoop . fs . Path ( ( ( IDENT_1 + "/" ) + ( org . apache . hadoop . hdfs . IDENT_8 . METHOD_4 ( IDENT_6 ) ) ) ) ) ; org . apache . hadoop . fs . Path IDENT_9 = IDENT_6 . METHOD_5 ( ) ; "<AssertPlaceHolder>" ; } finally { IDENT_6 . close ( ) ; cluster . METHOD_6 ( ) ; } } } METHOD_5 ( ) { return METHOD_3 ( new org . apache . hadoop . fs . Path ( org . apache . hadoop . hdfs . web . IDENT_10 . METHOD_7 ( IDENT_11 ) ) ) ; }
|
org . junit . Assert . assertEquals ( IDENT_7 , IDENT_9 )
|
METHOD_0 ( ) { cluster . METHOD_1 ( 0 ) . METHOD_2 ( IDENT_0 . IDENT_1 , INT_0 ) ; cluster . METHOD_3 ( ) ; java . lang . String status = cluster . METHOD_4 ( ) . METHOD_5 ( ) . METHOD_6 ( ) ; "<AssertPlaceHolder>" ; } METHOD_6 ( ) { if ( ! ( this . METHOD_7 ( ) ) ) { return "" ; } return STRING_0 + ( this . METHOD_8 ( ) ) ; }
|
org . junit . Assert . assertEquals ( "" , status )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . METHOD_1 ( ) ; org . apache . hadoop . hdfs . IDENT_0 cluster = new org . apache . hadoop . hdfs . IDENT_0 . Builder ( conf ) . build ( ) ; org . apache . hadoop . fs . IDENT_1 fs = cluster . METHOD_2 ( ) ; try { org . apache . hadoop . fs . Path IDENT_2 = new org . apache . hadoop . fs . Path ( STRING_0 ) ; org . apache . hadoop . hdfs . IDENT_3 . METHOD_3 ( fs , IDENT_2 , org . apache . hadoop . hdfs . IDENT_4 . IDENT_5 , fs . METHOD_4 ( IDENT_2 ) , org . apache . hadoop . hdfs . IDENT_4 . IDENT_6 ) ; org . apache . hadoop . fs . IDENT_7 stream = fs . METHOD_5 ( IDENT_2 ) ; stream . METHOD_6 ( INT_0 ) ; "<AssertPlaceHolder>" ; stream . METHOD_6 ( ( - INT_1 ) ) ; } finally { fs . close ( ) ; cluster . METHOD_7 ( ) ; } } METHOD_8 ( ) { return IDENT_8 ; }
|
org . junit . Assert . assertEquals ( INT_0 , stream . METHOD_8 ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . METHOD_1 ( ) ; org . apache . hadoop . hdfs . IDENT_0 cluster = new org . apache . hadoop . hdfs . IDENT_0 . Builder ( conf ) . build ( ) ; org . apache . hadoop . fs . IDENT_1 fs = cluster . METHOD_2 ( ) ; try { org . apache . hadoop . fs . Path IDENT_2 = new org . apache . hadoop . fs . Path ( STRING_0 ) ; org . apache . hadoop . hdfs . IDENT_3 . METHOD_3 ( fs , IDENT_2 , org . apache . hadoop . hdfs . IDENT_4 . IDENT_5 , fs . METHOD_4 ( IDENT_2 ) , org . apache . hadoop . hdfs . IDENT_4 . IDENT_6 ) ; org . apache . hadoop . fs . IDENT_7 stream = fs . METHOD_5 ( IDENT_2 ) ; stream . METHOD_6 ( INT_0 ) ; "<AssertPlaceHolder>" ; stream . METHOD_6 ( ( ( ( org . apache . hadoop . hdfs . IDENT_4 . IDENT_5 ) + ( org . apache . hadoop . hdfs . IDENT_4 . IDENT_5 ) ) + ( org . apache . hadoop . hdfs . IDENT_4 . IDENT_5 ) ) ) ; } finally { fs . close ( ) ; cluster . METHOD_7 ( ) ; } } METHOD_8 ( ) { return IDENT_8 ; }
|
org . junit . Assert . assertEquals ( INT_0 , stream . METHOD_8 ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . METHOD_1 ( ) ; conf . METHOD_2 ( IDENT_0 . IDENT_1 , true ) ; conf . METHOD_3 ( IDENT_0 . IDENT_2 , ( INT_0 * INT_0 ) ) ; conf . METHOD_3 ( IDENT_0 . IDENT_3 , ( INT_0 * INT_0 ) ) ; org . apache . hadoop . hdfs . IDENT_4 cluster = new org . apache . hadoop . hdfs . IDENT_4 . Builder ( conf ) . build ( ) ; org . apache . hadoop . fs . IDENT_5 fs = cluster . METHOD_4 ( ) ; try { org . apache . hadoop . fs . Path IDENT_6 = new org . apache . hadoop . fs . Path ( STRING_0 ) ; org . apache . hadoop . hdfs . IDENT_7 . METHOD_5 ( fs , IDENT_6 , org . apache . hadoop . hdfs . IDENT_8 . IDENT_9 , fs . METHOD_6 ( IDENT_6 ) , org . apache . hadoop . hdfs . IDENT_8 . IDENT_10 ) ; org . apache . hadoop . fs . IDENT_11 stream = fs . METHOD_7 ( IDENT_6 ) ; stream . METHOD_8 ( INT_1 ) ; "<AssertPlaceHolder>" ; stream . METHOD_8 ( ( - INT_2 ) ) ; } finally { fs . close ( ) ; cluster . METHOD_9 ( ) ; } } METHOD_10 ( ) { return IDENT_12 ; }
|
org . junit . Assert . assertEquals ( INT_1 , stream . METHOD_10 ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . METHOD_1 ( ) ; conf . METHOD_2 ( IDENT_0 . IDENT_1 , true ) ; conf . METHOD_3 ( IDENT_0 . IDENT_2 , ( INT_0 * INT_0 ) ) ; conf . METHOD_3 ( IDENT_0 . IDENT_3 , ( INT_0 * INT_0 ) ) ; org . apache . hadoop . hdfs . IDENT_4 cluster = new org . apache . hadoop . hdfs . IDENT_4 . Builder ( conf ) . build ( ) ; org . apache . hadoop . fs . IDENT_5 fs = cluster . METHOD_4 ( ) ; try { org . apache . hadoop . fs . Path IDENT_6 = new org . apache . hadoop . fs . Path ( STRING_0 ) ; org . apache . hadoop . hdfs . IDENT_7 . METHOD_5 ( fs , IDENT_6 , org . apache . hadoop . hdfs . IDENT_8 . IDENT_9 , fs . METHOD_6 ( IDENT_6 ) , org . apache . hadoop . hdfs . IDENT_8 . IDENT_10 ) ; org . apache . hadoop . fs . IDENT_11 stream = fs . METHOD_7 ( IDENT_6 ) ; stream . METHOD_8 ( INT_1 ) ; "<AssertPlaceHolder>" ; stream . METHOD_8 ( ( ( ( org . apache . hadoop . hdfs . IDENT_8 . IDENT_9 ) + ( org . apache . hadoop . hdfs . IDENT_8 . IDENT_9 ) ) + ( org . apache . hadoop . hdfs . IDENT_8 . IDENT_9 ) ) ) ; } finally { fs . close ( ) ; cluster . METHOD_9 ( ) ; } } METHOD_10 ( ) { return IDENT_12 ; }
|
org . junit . Assert . assertEquals ( INT_1 , stream . METHOD_10 ( ) )
|
METHOD_0 ( ) { IDENT_0 = false ; IDENT_1 = false ; java . lang . String IDENT_2 = IDENT_3 ; long IDENT_4 = 0 ; int IDENT_5 = METHOD_1 ( IDENT_2 , org . apache . hadoop . hdfs . IDENT_6 . IDENT_7 , org . apache . hadoop . hdfs . IDENT_6 . IDENT_8 , IDENT_4 ) ; org . apache . hadoop . hdfs . IDENT_6 . LOG . info ( ( STRING_0 + IDENT_5 ) ) ; "<AssertPlaceHolder>" ; } info ( java . lang . String ) { if ( ( LOG ) != null ) { LOG . info ( msg ) ; } else if ( ( LOGGER ) != null ) { LOGGER . info ( msg ) ; } }
|
org . junit . Assert . assertEquals ( 0 , IDENT_5 )
|
METHOD_0 ( ) { final org . apache . camel . component . IDENT_0 . api . dto . IDENT_1 . IDENT_2 tree = new org . apache . camel . component . IDENT_0 . api . dto . IDENT_1 . METHOD_1 ( ) ; tree . METHOD_2 ( new org . apache . camel . component . IDENT_0 . dto . IDENT_3 . METHOD_3 ( ) ) . METHOD_4 ( new org . apache . camel . component . IDENT_0 . dto . IDENT_3 . METHOD_5 ( ) ) . METHOD_4 ( STRING_0 , new org . apache . camel . component . IDENT_0 . dto . IDENT_3 . METHOD_6 ( ) ) ; tree . METHOD_2 ( new org . apache . camel . component . IDENT_0 . dto . IDENT_3 . METHOD_3 ( ) ) ; final java . lang . Class [ ] types = tree . METHOD_7 ( ) ; java . util . Arrays . METHOD_8 ( types , ( final java . lang . Class l , final java . lang . Class r ) -> l . getName ( ) . compareTo ( r . getName ( ) ) ) ; "<AssertPlaceHolder>" ; } getName ( ) { return name ; }
|
org . junit . Assert . assertArrayEquals ( new java . lang . Class [ ] { org . apache . camel . component . IDENT_0 . dto . IDENT_3 . METHOD_3 . class , org . apache . camel . component . IDENT_0 . dto . IDENT_3 . METHOD_6 . class , org . apache . camel . component . IDENT_0 . dto . IDENT_3 . METHOD_5 . class } , types )
|
METHOD_0 ( ) { java . lang . String IDENT_0 = IDENT_1 ; IDENT_2 = true ; int IDENT_3 = ( ( int ) ( org . apache . hadoop . hdfs . IDENT_4 . IDENT_5 ) ) + ( ( int ) ( ( org . apache . hadoop . hdfs . IDENT_4 . IDENT_5 ) / 2 ) ) ; long IDENT_6 = ( org . apache . hadoop . hdfs . IDENT_4 . IDENT_5 ) + 1 ; int IDENT_7 = 5 ; int IDENT_8 = METHOD_1 ( IDENT_0 , IDENT_7 , IDENT_3 , IDENT_6 ) ; "<AssertPlaceHolder>" ; } METHOD_1 ( java . lang . String , int , int , long ) { int IDENT_9 = 0 ; long IDENT_10 = 0 ; org . apache . hadoop . fs . IDENT_11 out = null ; byte [ ] IDENT_12 = new byte [ org . apache . hadoop . hdfs . IDENT_4 . IDENT_13 ] ; byte [ ] IDENT_14 = new byte [ org . apache . hadoop . hdfs . IDENT_4 . IDENT_13 ] ; for ( int i = 0 ; i < ( org . apache . hadoop . hdfs . IDENT_4 . IDENT_13 ) ; i ++ ) { IDENT_12 [ i ] = ( ( byte ) ( i & 255 ) ) ; } try { org . apache . hadoop . fs . Path path = METHOD_2 ( IDENT_0 ) ; long IDENT_15 = 0 ; if ( METHOD_3 ( path ) ) { if ( IDENT_16 ) { out = ( IDENT_17 ) ? IDENT_18 . create ( path , java . util . IDENT_19 . of ( IDENT_20 . IDENT_21 ) ) : IDENT_22 . create ( path , IDENT_16 ) ; org . apache . hadoop . hdfs . IDENT_4 . LOG . info ( ( STRING_0 7 + path ) ) ; } else { out = ( IDENT_17 ) ? IDENT_18 . create ( path , java . util . IDENT_19 . of ( IDENT_20 . IDENT_23 ) ) : IDENT_22 . append ( path ) ; IDENT_15 = METHOD_4 ( path ) ; org . apache . hadoop . hdfs . IDENT_4 . LOG . info ( ( ( ( STRING_0 5 + IDENT_15 ) + STRING_1 ) + path ) ) ; } } else { out = ( IDENT_17 ) ? IDENT_18 . create ( path , java . util . IDENT_19 . of ( IDENT_20 . IDENT_24 ) ) : IDENT_22 . create ( path ) ; } long IDENT_25 = IDENT_15 ; long IDENT_26 = IDENT_15 ; long IDENT_27 = 0 ; boolean IDENT_28 ; for ( int i = 0 ; i < IDENT_29 ; i ++ ) { IDENT_28 = ( i OP_0 2 ) == 0 ; METHOD_5 ( out , IDENT_12 , IDENT_30 ) ; IDENT_25 += IDENT_30 ; if ( IDENT_28 ) { out . METHOD_6 ( ) ; IDENT_26 += IDENT_30 + IDENT_27 ; IDENT_27 = 0 ; } else { IDENT_27 += IDENT_30 ; } if ( IDENT_31 ) { org . apache . hadoop . hdfs . IDENT_4 . LOG . info ( ( ( ( ( ( ( ( STRING_0 2 + IDENT_30 ) + STRING_2 ) + IDENT_25 ) + STRING_3 ) + IDENT_26 ) + STRING_0 0 ) + IDENT_0 ) ) ; } IDENT_10 = METHOD_7 ( IDENT_0 , IDENT_14 , IDENT_26 , IDENT_32 ) ; java . lang . String IDENT_33 = ( ( ( ( ( ( STRING_0 3 + IDENT_25 ) + STRING_0 6 ) + IDENT_26 ) + STRING_4 ) + IDENT_10 ) + STRING_0 9 ) + IDENT_0 ; if ( ( IDENT_10 >= IDENT_26 ) && ( IDENT_10 <= IDENT_25 ) ) { IDENT_33 = ( STRING_0 8 + IDENT_33 ) + STRING_0 4 ; } else { IDENT_9 ++ ; IDENT_33 = ( STRING_5 + IDENT_33 ) + STRING_6 ; if ( IDENT_34 ) { throw new java . io . IOException ( IDENT_33 ) ; } } org . apache . hadoop . hdfs . IDENT_4 . LOG . info ( IDENT_33 ) ; } METHOD_5 ( out , IDENT_12 , IDENT_30 ) ; IDENT_25 += IDENT_30 ; IDENT_26 += IDENT_30 + IDENT_27 ; IDENT_27 += 0 ; out . close ( ) ; IDENT_10 = METHOD_7 ( IDENT_0 , IDENT_14 , IDENT_26 , IDENT_32 ) ; java . lang . String IDENT_35 = ( ( ( ( ( ( STRING_0 3 + IDENT_25 ) + STRING_0 6 ) + IDENT_26 ) + STRING_4 ) + IDENT_10 ) + STRING_0 9 ) + IDENT_0 ; java . lang . String IDENT_33 ; if ( ( IDENT_10 >= IDENT_26 ) && ( IDENT_10 <= IDENT_25 ) ) { IDENT_33 = ( STRING_7 + IDENT_35 ) + STRING_0 4 ; } else { IDENT_9 ++ ; IDENT_33 = ( STRING_8 + IDENT_35 ) + STRING_6 ; org . apache . hadoop . hdfs . IDENT_4 . LOG . info ( IDENT_33 ) ; if ( IDENT_34 ) { throw new java . io . IOException ( IDENT_33 ) ; } } long IDENT_36 = METHOD_4 ( path ) ; if ( IDENT_36 != IDENT_10 ) { IDENT_33 = ( STRING_0 1 + IDENT_35 ) + STRING_6 ; throw new java . io . IOException ( IDENT_33 ) ; } } catch ( java . io . IOException e ) { throw new java . io . IOException ( ( ( STRING_0 + STRING_9 ) + IDENT_10 ) , e ) ; } finally { if ( out != null ) { out . close ( ) ; } } return - IDENT_9 ; }
|
org . junit . Assert . assertEquals ( 0 , IDENT_8 )
|
METHOD_0 ( ) { org . apache . hadoop . hdfs . server . protocol . IDENT_0 . IDENT_1 [ ] list = new org . apache . hadoop . hdfs . server . protocol . IDENT_0 . IDENT_1 [ ] { org . apache . hadoop . hdfs . IDENT_2 . IDENT_3 . METHOD_1 ( 1 ) , org . apache . hadoop . hdfs . IDENT_2 . IDENT_3 . METHOD_1 ( 2 ) } ; org . apache . hadoop . hdfs . server . protocol . IDENT_0 IDENT_4 = new org . apache . hadoop . hdfs . server . protocol . METHOD_2 ( list ) ; org . apache . hadoop . hdfs . protocol . IDENT_5 . IDENT_6 . IDENT_7 IDENT_8 = org . apache . hadoop . hdfs . IDENT_2 . IDENT_9 . convert ( IDENT_4 ) ; org . apache . hadoop . hdfs . server . protocol . METHOD_2 IDENT_10 = org . apache . hadoop . hdfs . IDENT_2 . IDENT_9 . convert ( IDENT_8 ) ; org . apache . hadoop . hdfs . server . protocol . METHOD_2 . IDENT_1 [ ] IDENT_11 = IDENT_4 . METHOD_3 ( ) ; org . apache . hadoop . hdfs . server . protocol . METHOD_2 . IDENT_1 [ ] IDENT_12 = IDENT_10 . METHOD_3 ( ) ; "<AssertPlaceHolder>" ; for ( int i = 0 ; i < ( IDENT_11 . length ) ; i ++ ) { compare ( IDENT_11 [ i ] , IDENT_12 [ i ] ) ; } } METHOD_3 ( ) { return IDENT_11 ; }
|
org . junit . Assert . assertEquals ( IDENT_11 . length , IDENT_12 . length )
|
METHOD_0 ( ) { org . apache . hadoop . io . IDENT_0 t = new org . apache . hadoop . io . METHOD_1 ( STRING_0 . getBytes ( ) ) ; java . lang . String s = t . toString ( ) ; org . apache . hadoop . io . METHOD_1 IDENT_1 = new org . apache . hadoop . io . METHOD_1 ( s ) ; "<AssertPlaceHolder>" ; } toString ( ) { java . lang . StringBuilder sb = new java . lang . StringBuilder ( ) ; sb . append ( ( ( org . apache . hadoop . yarn . api . records . IDENT_2 . IDENT_3 . IDENT_4 ) + STRING_1 ) ) ; org . apache . hadoop . yarn . api . records . IDENT_5 IDENT_6 = METHOD_2 ( ) ; sb . append ( IDENT_6 . METHOD_3 ( ) ) . append ( STRING_1 ) ; sb . append ( IDENT_6 . getId ( ) ) . append ( STRING_1 ) ; sb . append ( METHOD_4 ( ) ) ; return sb . toString ( ) ; }
|
org . junit . Assert . assertEquals ( t , IDENT_1 )
|
METHOD_0 ( ) { java . util . ArrayList < org . apache . hadoop . hdfs . protocol . IDENT_0 > IDENT_1 = new java . util . ArrayList < org . apache . hadoop . hdfs . protocol . IDENT_0 > ( ) ; for ( int i = 0 ; i < 3 ; i ++ ) { IDENT_1 . add ( METHOD_1 ( ) ) ; } java . util . List < org . apache . hadoop . hdfs . protocol . IDENT_2 . IDENT_3 . IDENT_4 > IDENT_5 = org . apache . hadoop . hdfs . IDENT_6 . IDENT_7 . METHOD_2 ( IDENT_1 ) ; java . util . List < org . apache . hadoop . hdfs . protocol . IDENT_0 > IDENT_8 = org . apache . hadoop . hdfs . IDENT_6 . IDENT_7 . METHOD_3 ( IDENT_5 ) ; "<AssertPlaceHolder>" ; for ( org . apache . hadoop . hdfs . protocol . IDENT_0 IDENT_9 : IDENT_1 ) { compare ( IDENT_9 , IDENT_8 . get ( 2 ) ) ; } } size ( ) { return ( ( java . lang . Integer ) ( new io . IDENT_10 . transaction . handler . METHOD_4 ( io . IDENT_10 . transaction . handler . IDENT_11 . IDENT_12 ) { @ io . IDENT_10 . metadata . IDENT_13 . Override public java . lang . Object METHOD_5 ( ) throws java . io . IOException { io . IDENT_10 . metadata . hdfs . dal . IDENT_14 IDENT_15 = ( ( io . IDENT_10 . metadata . hdfs . dal . IDENT_14 ) ( io . IDENT_10 . metadata . IDENT_16 . METHOD_6 ( io . IDENT_10 . metadata . hdfs . dal . IDENT_14 . class ) ) ) ; return IDENT_15 . METHOD_7 ( ) ; } } . METHOD_8 ( ) ) ) ; }
|
org . junit . Assert . assertEquals ( IDENT_1 . size ( ) , IDENT_8 . size ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . hdfs . protocol . IDENT_0 [ ] IDENT_1 = new org . apache . hadoop . hdfs . protocol . IDENT_0 [ 3 ] ; for ( int i = 0 ; i < 3 ; i ++ ) { IDENT_1 [ i ] = METHOD_1 ( ) ; } org . apache . hadoop . hdfs . protocol . IDENT_2 . IDENT_3 . IDENT_4 [ ] IDENT_5 = org . apache . hadoop . hdfs . IDENT_6 . IDENT_7 . METHOD_2 ( IDENT_1 ) ; org . apache . hadoop . hdfs . protocol . IDENT_0 [ ] IDENT_8 = org . apache . hadoop . hdfs . IDENT_6 . IDENT_7 . METHOD_2 ( IDENT_5 ) ; "<AssertPlaceHolder>" ; for ( int i = 0 ; i < ( IDENT_1 . length ) ; i ++ ) { compare ( IDENT_1 [ i ] , IDENT_8 [ i ] ) ; } } METHOD_2 ( org . apache . hadoop . hdfs . protocol . IDENT_0 [ ] ) { if ( IDENT_9 == null ) { return null ; } return org . apache . hadoop . hdfs . IDENT_6 . IDENT_7 . METHOD_3 ( java . util . Arrays . asList ( IDENT_9 ) ) . toArray ( new org . apache . hadoop . hdfs . protocol . IDENT_2 . IDENT_3 . IDENT_4 [ IDENT_9 . length ] ) ; }
|
org . junit . Assert . assertEquals ( IDENT_1 . length , IDENT_8 . length )
|
METHOD_0 ( ) { org . apache . hadoop . hdfs . protocol . IDENT_0 IDENT_1 = org . apache . hadoop . hdfs . IDENT_2 . METHOD_1 ( ) ; org . apache . hadoop . hdfs . security . token . block . IDENT_3 [ ] keys = new org . apache . hadoop . hdfs . security . token . block . IDENT_3 [ ] { org . apache . hadoop . hdfs . IDENT_4 . IDENT_5 . METHOD_2 ( 2 ) , org . apache . hadoop . hdfs . IDENT_4 . IDENT_5 . METHOD_2 ( 3 ) } ; org . apache . hadoop . hdfs . security . token . block . IDENT_6 IDENT_7 = new org . apache . hadoop . hdfs . security . token . block . METHOD_3 ( true , 9 , 10 , org . apache . hadoop . hdfs . IDENT_4 . IDENT_5 . METHOD_2 ( 1 ) , keys ) ; org . apache . hadoop . hdfs . server . protocol . IDENT_8 IDENT_9 = new org . apache . hadoop . hdfs . server . protocol . METHOD_4 ( IDENT_1 , new org . apache . hadoop . hdfs . server . common . METHOD_5 ( org . apache . hadoop . hdfs . server . common . IDENT_10 . IDENT_11 . IDENT_12 ) , IDENT_7 , STRING_0 ) ; org . apache . hadoop . hdfs . protocol . IDENT_13 . IDENT_14 . IDENT_15 IDENT_13 = org . apache . hadoop . hdfs . IDENT_4 . IDENT_16 . convert ( IDENT_9 ) ; org . apache . hadoop . hdfs . server . protocol . METHOD_4 IDENT_17 = org . apache . hadoop . hdfs . IDENT_4 . IDENT_16 . convert ( IDENT_13 ) ; compare ( IDENT_9 . METHOD_6 ( ) , IDENT_17 . METHOD_6 ( ) ) ; compare ( IDENT_9 . METHOD_7 ( ) , IDENT_17 . METHOD_7 ( ) ) ; compare ( IDENT_9 , IDENT_17 ) ; "<AssertPlaceHolder>" ; } METHOD_8 ( ) { return IDENT_18 ; }
|
org . junit . Assert . assertEquals ( IDENT_9 . METHOD_8 ( ) , IDENT_17 . METHOD_8 ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . security . token . IDENT_0 < org . apache . hadoop . hdfs . security . token . IDENT_1 . IDENT_2 > token = METHOD_1 ( STRING_0 , STRING_1 ) ; try { IDENT_3 . METHOD_2 ( token , STRING_2 ) ; org . junit . Assert . fail ( STRING_3 ) ; } catch ( org . apache . hadoop . security . IDENT_4 IDENT_5 ) { } IDENT_3 . METHOD_2 ( token , STRING_1 ) ; org . apache . hadoop . hdfs . security . token . IDENT_1 . IDENT_2 IDENT_6 = new org . apache . hadoop . hdfs . security . token . IDENT_1 . METHOD_3 ( ) ; byte [ ] IDENT_7 = token . METHOD_4 ( ) ; IDENT_6 . METHOD_5 ( new java . io . METHOD_6 ( new java . io . ByteArrayInputStream ( IDENT_7 ) ) ) ; "<AssertPlaceHolder>" ; org . apache . hadoop . hdfs . security . IDENT_8 . LOG . info ( STRING_4 ) ; java . lang . Thread . sleep ( INT_0 ) ; try { IDENT_3 . METHOD_7 ( IDENT_6 ) ; org . junit . Assert . fail ( STRING_5 ) ; } catch ( org . apache . hadoop . security . token . IDENT_9 . IDENT_10 e ) { } IDENT_3 . METHOD_2 ( token , STRING_1 ) ; org . apache . hadoop . hdfs . security . IDENT_8 . LOG . info ( STRING_6 ) ; java . lang . Thread . sleep ( INT_1 ) ; try { IDENT_3 . METHOD_2 ( token , STRING_1 ) ; org . junit . Assert . fail ( STRING_7 ) ; } catch ( org . apache . hadoop . security . token . IDENT_9 . IDENT_10 it ) { } } METHOD_7 ( org . apache . hadoop . hdfs . security . token . block . IDENT_11 ) { if ( METHOD_8 ( IDENT_6 . METHOD_9 ( ) ) ) { throw new io . IDENT_12 . metadata . security . token . block . METHOD_10 ( ( ( STRING_8 + ( IDENT_6 . toString ( ) ) ) + STRING_9 ) ) ; } org . apache . hadoop . hdfs . security . token . block . IDENT_13 key = null ; try { key = METHOD_11 ( IDENT_6 . METHOD_12 ( ) ) ; } catch ( java . io . IOException ex ) { } if ( key == null ) { throw new io . IDENT_12 . metadata . security . token . block . METHOD_10 ( ( ( ( ( STRING_10 + ( IDENT_6 . toString ( ) ) ) + STRING_11 ) + ( IDENT_6 . METHOD_12 ( ) ) ) + STRING_12 ) ) ; } return METHOD_13 ( IDENT_6 . getBytes ( ) , key . getKey ( ) ) ; }
|
org . junit . Assert . assertTrue ( ( null != ( IDENT_3 . METHOD_7 ( IDENT_6 ) ) ) )
|
METHOD_0 ( ) { final org . apache . hadoop . hdfs . IDENT_0 IDENT_1 = cluster . METHOD_1 ( ) ; final org . apache . hadoop . security . IDENT_2 IDENT_3 = new org . apache . hadoop . security . METHOD_2 ( ) ; final org . apache . hadoop . security . token . IDENT_4 < ? > [ ] tokens = IDENT_1 . METHOD_3 ( STRING_0 , IDENT_3 ) ; "<AssertPlaceHolder>" ; @ org . apache . hadoop . hdfs . security . METHOD_4 ( STRING_1 ) final org . apache . hadoop . security . token . IDENT_4 < org . apache . hadoop . hdfs . security . token . IDENT_5 . IDENT_6 > token = ( ( org . apache . hadoop . security . token . IDENT_4 < org . apache . hadoop . hdfs . security . token . IDENT_5 . IDENT_6 > ) ( tokens [ 0 ] ) ) ; final org . apache . hadoop . security . IDENT_7 IDENT_8 = org . apache . hadoop . security . IDENT_7 . METHOD_5 ( STRING_2 ) ; final org . apache . hadoop . security . IDENT_7 IDENT_9 = org . apache . hadoop . security . IDENT_7 . METHOD_5 ( STRING_0 ) ; IDENT_8 . METHOD_6 ( new java . security . IDENT_10 < java . lang . Object > ( ) { @ org . apache . hadoop . hdfs . security . Override public java . lang . Object run ( ) throws java . io . IOException { try { token . METHOD_7 ( config ) ; } catch ( java . lang . Exception e ) { org . junit . Assert . fail ( ( STRING_3 + IDENT_8 ) ) ; } return null ; } } ) ; IDENT_9 . METHOD_6 ( new java . security . IDENT_10 < java . lang . Object > ( ) { @ org . apache . hadoop . hdfs . security . Override public java . lang . Object run ( ) throws org . apache . hadoop . hdfs . security . Exception { token . METHOD_7 ( config ) ; return null ; } } ) ; IDENT_8 . METHOD_6 ( new java . security . IDENT_10 < java . lang . Object > ( ) { @ org . apache . hadoop . hdfs . security . Override public java . lang . Object run ( ) throws java . io . IOException { try { token . METHOD_8 ( config ) ; } catch ( java . lang . Exception e ) { org . junit . Assert . fail ( ( STRING_4 + IDENT_8 ) ) ; } return null ; } } ) ; } METHOD_3 ( java . lang . String , org . apache . hadoop . security . METHOD_2 ) { org . apache . hadoop . security . token . IDENT_4 < ? > [ ] tokens = null ; org . apache . hadoop . io . IDENT_11 IDENT_12 = METHOD_9 ( ) ; org . apache . hadoop . security . token . IDENT_4 < ? > token = IDENT_13 . METHOD_10 ( IDENT_12 ) ; if ( token == null ) { final java . net . URL url = METHOD_11 ( null , null , null , null ) ; final org . apache . hadoop . security . token . IDENT_5 . web . IDENT_14 IDENT_15 = new org . apache . hadoop . security . token . IDENT_5 . web . METHOD_12 ( IDENT_16 ) ; try { final java . lang . String IDENT_17 = METHOD_13 ( ) ; token = METHOD_14 ( ) . METHOD_6 ( new java . security . IDENT_10 < org . apache . hadoop . security . token . IDENT_4 < ? > > ( ) { @ org . apache . hadoop . crypto . key . IDENT_18 . Override public org . apache . hadoop . security . token . IDENT_4 < ? > run ( ) throws org . apache . hadoop . crypto . key . IDENT_18 . Exception { return IDENT_15 . METHOD_15 ( url , new org . apache . hadoop . security . token . IDENT_5 . web . METHOD_12 . METHOD_16 ( ) , IDENT_19 , IDENT_17 ) ; } } ) ; if ( token != null ) { IDENT_13 . METHOD_17 ( token . METHOD_18 ( ) , token ) ; tokens = new org . apache . hadoop . security . token . METHOD_16 < ? > [ ] { token } ; } else { throw new java . io . IOException ( STRING_5 ) ; } } catch ( java . lang . IDENT_20 e ) { java . lang . Thread . METHOD_19 ( ) . METHOD_20 ( ) ; } catch ( java . lang . Exception e ) { throw new java . io . IOException ( e ) ; } } return tokens ; }
|
org . junit . Assert . assertEquals ( 1 , tokens . length )
|
METHOD_0 ( ) { final org . apache . camel . component . IDENT_0 . api . dto . IDENT_1 . IDENT_2 tree = new org . apache . camel . component . IDENT_0 . api . dto . IDENT_1 . METHOD_1 ( ) ; tree . METHOD_2 ( new org . apache . camel . component . IDENT_0 . dto . IDENT_3 . METHOD_3 ( ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return results . size ( ) ; }
|
org . junit . Assert . assertEquals ( 1 , tree . size ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = METHOD_1 ( ) ; short IDENT_0 = 2 ; final org . apache . hadoop . fs . Path IDENT_1 = new org . apache . hadoop . fs . Path ( STRING_0 ) ; java . lang . String [ ] IDENT_2 = new java . lang . String [ ] { STRING_1 , STRING_1 , STRING_2 } ; org . apache . hadoop . hdfs . IDENT_3 cluster = new org . apache . hadoop . hdfs . IDENT_3 . Builder ( conf ) . METHOD_2 ( IDENT_2 . length ) . METHOD_3 ( METHOD_3 ) . build ( ) ; final org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_5 IDENT_6 = cluster . METHOD_4 ( ) . METHOD_5 ( ) ; final org . apache . hadoop . hdfs . server . IDENT_7 . IDENT_8 IDENT_9 = IDENT_6 . METHOD_6 ( ) . METHOD_7 ( ) ; try { final org . apache . hadoop . fs . IDENT_10 fs = cluster . METHOD_8 ( ) ; org . apache . hadoop . hdfs . IDENT_11 . METHOD_9 ( fs , IDENT_1 , 1L , IDENT_0 , 1L ) ; org . apache . hadoop . hdfs . protocol . IDENT_12 b = org . apache . hadoop . hdfs . IDENT_11 . METHOD_10 ( fs , IDENT_1 ) ; org . apache . hadoop . hdfs . IDENT_11 . METHOD_11 ( cluster , b , 2 , IDENT_0 , 0 ) ; java . util . ArrayList < org . apache . hadoop . hdfs . server . IDENT_13 . IDENT_14 > IDENT_15 = cluster . METHOD_12 ( ) ; "<AssertPlaceHolder>" ; org . apache . hadoop . hdfs . server . IDENT_13 . IDENT_14 IDENT_16 = IDENT_15 . get ( 2 ) ; org . apache . hadoop . hdfs . protocol . IDENT_17 IDENT_18 = IDENT_16 . METHOD_13 ( ) ; cluster . METHOD_14 ( 2 ) ; IDENT_9 . METHOD_15 ( IDENT_18 , false ) ; org . apache . hadoop . hdfs . IDENT_11 . METHOD_11 ( cluster , b , 1 , IDENT_0 , 1 ) ; java . lang . String [ ] IDENT_19 = new java . lang . String [ ] { STRING_2 } ; cluster . METHOD_16 ( conf , 1 , true , null , IDENT_19 ) ; cluster . METHOD_17 ( ) ; org . apache . hadoop . hdfs . IDENT_11 . METHOD_11 ( cluster , b , 2 , IDENT_0 , 0 ) ; } finally { cluster . METHOD_18 ( ) ; } } size ( ) { return ( ( java . lang . Integer ) ( new io . IDENT_20 . transaction . handler . METHOD_19 ( io . IDENT_20 . transaction . handler . IDENT_21 . IDENT_22 ) { @ io . IDENT_20 . metadata . IDENT_7 . Override public java . lang . Object METHOD_20 ( ) throws java . io . IOException { io . IDENT_20 . metadata . hdfs . dal . IDENT_23 IDENT_24 = ( ( io . IDENT_20 . metadata . hdfs . dal . IDENT_23 ) ( io . IDENT_20 . metadata . IDENT_25 . METHOD_21 ( io . IDENT_20 . metadata . hdfs . dal . IDENT_23 . class ) ) ) ; return IDENT_24 . METHOD_22 ( ) ; } } . METHOD_23 ( ) ) ) ; }
|
org . junit . Assert . assertEquals ( 3 , IDENT_15 . size ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = METHOD_1 ( ) ; short IDENT_0 = 2 ; final org . apache . hadoop . fs . Path IDENT_1 = new org . apache . hadoop . fs . Path ( STRING_0 ) ; org . apache . hadoop . fs . IDENT_2 IDENT_3 = org . apache . hadoop . fs . IDENT_2 . METHOD_2 ( conf ) ; org . apache . hadoop . fs . Path IDENT_4 = IDENT_3 . METHOD_3 ( ) ; org . apache . hadoop . fs . Path IDENT_5 = new org . apache . hadoop . fs . Path ( IDENT_4 , STRING_1 ) ; org . apache . hadoop . fs . Path IDENT_6 = new org . apache . hadoop . fs . Path ( IDENT_5 , STRING_2 ) ; org . apache . hadoop . fs . Path IDENT_7 = new org . apache . hadoop . fs . Path ( IDENT_5 , STRING_3 ) ; "<AssertPlaceHolder>" ; org . apache . hadoop . hdfs . IDENT_8 . METHOD_4 ( IDENT_3 , IDENT_6 , "" ) ; org . apache . hadoop . hdfs . IDENT_8 . METHOD_4 ( IDENT_3 , IDENT_7 , "" ) ; conf . set ( IDENT_9 . IDENT_10 , IDENT_6 . METHOD_5 ( ) . getPath ( ) ) ; conf . set ( IDENT_9 . IDENT_11 , IDENT_7 . METHOD_5 ( ) . getPath ( ) ) ; java . lang . String [ ] IDENT_12 = new java . lang . String [ ] { STRING_4 , STRING_4 , STRING_5 , STRING_5 } ; org . apache . hadoop . hdfs . IDENT_13 cluster = new org . apache . hadoop . hdfs . IDENT_13 . Builder ( conf ) . METHOD_6 ( IDENT_12 . length ) . METHOD_7 ( METHOD_7 ) . build ( ) ; final org . apache . hadoop . hdfs . server . IDENT_14 . IDENT_15 IDENT_16 = cluster . METHOD_8 ( ) . METHOD_9 ( ) ; try { final org . apache . hadoop . fs . IDENT_2 fs = cluster . METHOD_10 ( ) ; org . apache . hadoop . hdfs . IDENT_8 . METHOD_11 ( fs , IDENT_1 , 1L , IDENT_0 , 1L ) ; org . apache . hadoop . hdfs . protocol . IDENT_17 b = org . apache . hadoop . hdfs . IDENT_8 . METHOD_12 ( fs , IDENT_1 ) ; org . apache . hadoop . hdfs . IDENT_8 . METHOD_13 ( cluster , b , 2 , IDENT_0 , 0 ) ; org . apache . hadoop . fs . IDENT_18 [ ] IDENT_19 = fs . METHOD_14 ( fs . METHOD_15 ( IDENT_1 ) , 0 , Long . MAX_VALUE ) ; java . lang . String name = IDENT_19 [ 0 ] . METHOD_16 ( ) [ 0 ] ; org . apache . hadoop . hdfs . IDENT_8 . METHOD_4 ( IDENT_3 , IDENT_6 , name ) ; IDENT_16 . METHOD_17 ( ) . METHOD_18 ( ) . METHOD_19 ( conf ) ; org . apache . hadoop . hdfs . IDENT_8 . METHOD_20 ( fs , name ) ; org . apache . hadoop . hdfs . IDENT_8 . METHOD_13 ( cluster , b , 2 , IDENT_0 , 0 ) ; } finally { cluster . METHOD_21 ( ) ; } } METHOD_22 ( java . lang . String ) { return METHOD_22 ( IDENT_20 , null , true ) ; }
|
org . junit . Assert . assertTrue ( IDENT_3 . METHOD_22 ( IDENT_5 ) )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = METHOD_1 ( ) ; short IDENT_0 = 5 ; final org . apache . hadoop . fs . Path IDENT_1 = new org . apache . hadoop . fs . Path ( STRING_0 ) ; org . apache . hadoop . fs . IDENT_2 IDENT_3 = org . apache . hadoop . fs . IDENT_2 . METHOD_2 ( conf ) ; org . apache . hadoop . fs . Path IDENT_4 = IDENT_3 . METHOD_3 ( ) ; org . apache . hadoop . fs . Path IDENT_5 = new org . apache . hadoop . fs . Path ( IDENT_4 , STRING_1 ) ; org . apache . hadoop . fs . Path IDENT_6 = new org . apache . hadoop . fs . Path ( IDENT_5 , STRING_2 ) ; org . apache . hadoop . fs . Path IDENT_7 = new org . apache . hadoop . fs . Path ( IDENT_5 , STRING_3 ) ; "<AssertPlaceHolder>" ; org . apache . hadoop . hdfs . IDENT_8 . METHOD_4 ( IDENT_3 , IDENT_6 , "" ) ; org . apache . hadoop . hdfs . IDENT_8 . METHOD_4 ( IDENT_3 , IDENT_7 , "" ) ; conf . set ( IDENT_9 . IDENT_10 , IDENT_7 . METHOD_5 ( ) . getPath ( ) ) ; conf . set ( IDENT_9 . IDENT_11 , IDENT_6 . METHOD_5 ( ) . getPath ( ) ) ; java . lang . String [ ] IDENT_12 = new java . lang . String [ ] { STRING_4 , STRING_5 , STRING_4 , STRING_4 , STRING_4 } ; org . apache . hadoop . hdfs . IDENT_13 cluster = new org . apache . hadoop . hdfs . IDENT_13 . Builder ( conf ) . METHOD_6 ( IDENT_12 . length ) . METHOD_7 ( METHOD_7 ) . build ( ) ; final org . apache . hadoop . hdfs . server . IDENT_14 . IDENT_15 IDENT_16 = cluster . METHOD_8 ( ) . METHOD_9 ( ) ; try { final org . apache . hadoop . fs . IDENT_2 fs = cluster . METHOD_10 ( ) ; org . apache . hadoop . hdfs . IDENT_8 . METHOD_11 ( fs , IDENT_1 , 1L , IDENT_0 , 1L ) ; org . apache . hadoop . hdfs . protocol . IDENT_17 b = org . apache . hadoop . hdfs . IDENT_8 . METHOD_12 ( fs , IDENT_1 ) ; org . apache . hadoop . hdfs . IDENT_8 . METHOD_13 ( cluster , b , 2 , IDENT_0 , 0 ) ; IDENT_0 = 2 ; fs . METHOD_14 ( IDENT_1 , IDENT_0 ) ; org . apache . hadoop . fs . IDENT_18 [ ] IDENT_19 = fs . METHOD_15 ( fs . METHOD_16 ( IDENT_1 ) , 0 , Long . MAX_VALUE ) ; for ( java . lang . String IDENT_20 : IDENT_19 [ 0 ] . METHOD_17 ( ) ) { if ( ! ( IDENT_20 . startsWith ( STRING_5 ) ) ) { java . lang . String name = IDENT_20 . substring ( ( ( STRING_4 . length ( ) ) + 1 ) ) ; org . apache . hadoop . hdfs . IDENT_8 . METHOD_4 ( IDENT_3 , IDENT_6 , name ) ; IDENT_16 . METHOD_18 ( ) . METHOD_19 ( ) . METHOD_20 ( conf ) ; org . apache . hadoop . hdfs . IDENT_8 . METHOD_21 ( fs , name ) ; break ; } } org . apache . hadoop . hdfs . IDENT_8 . METHOD_13 ( cluster , b , 2 , IDENT_0 , 0 ) ; } finally { cluster . METHOD_22 ( ) ; } } METHOD_23 ( java . lang . String ) { return METHOD_23 ( IDENT_21 , null , true ) ; }
|
org . junit . Assert . assertTrue ( IDENT_3 . METHOD_23 ( IDENT_5 ) )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . METHOD_1 ( ) ; conf . METHOD_2 ( IDENT_0 . IDENT_1 , IDENT_2 . IDENT_3 . class , org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_5 . class ) ; conf . METHOD_3 ( IDENT_0 . IDENT_6 , 1 ) ; java . util . List < org . apache . hadoop . fs . Path > IDENT_7 = com . google . common . collect . Lists . newArrayList ( ) ; for ( int i = 0 ; i < 10 ; i ++ ) { IDENT_7 . add ( new org . apache . hadoop . fs . Path ( ( STRING_0 + i ) ) ) ; } org . apache . hadoop . hdfs . IDENT_8 cluster = new org . apache . hadoop . hdfs . IDENT_8 . Builder ( conf ) . METHOD_4 ( 2 ) . build ( ) ; try { java . util . List < org . apache . hadoop . fs . IDENT_9 > streams = com . google . common . collect . Lists . newArrayList ( ) ; try { for ( org . apache . hadoop . fs . Path path : IDENT_7 ) { org . apache . hadoop . fs . IDENT_9 out = cluster . METHOD_5 ( ) . create ( path , ( ( short ) ( 2 ) ) ) ; streams . add ( out ) ; out . METHOD_6 ( STRING_1 ) ; out . METHOD_7 ( ) ; } org . apache . hadoop . hdfs . IDENT_8 . IDENT_10 IDENT_11 = cluster . METHOD_8 ( 0 ) ; for ( int i = 0 ; i < ( streams . size ( ) ) ; i ++ ) { org . apache . hadoop . fs . Path path = IDENT_7 . get ( i ) ; org . apache . hadoop . fs . IDENT_9 out = streams . get ( i ) ; out . METHOD_6 ( STRING_2 ) ; out . METHOD_7 ( ) ; cluster . METHOD_5 ( ) . METHOD_9 ( path , ( ( short ) ( 1 ) ) ) ; out . close ( ) ; } org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_12 . LOG . info ( STRING_3 ) ; org . apache . hadoop . hdfs . IDENT_8 . IDENT_10 IDENT_13 = cluster . METHOD_8 ( 0 ) ; cluster . METHOD_10 ( false ) ; cluster . METHOD_11 ( IDENT_11 ) ; cluster . METHOD_12 ( ) ; cluster . METHOD_11 ( IDENT_13 ) ; cluster . METHOD_12 ( ) ; cluster . METHOD_13 ( ) . METHOD_14 ( ) . METHOD_15 ( ) . METHOD_16 ( 2 ) ; cluster . METHOD_17 ( ) ; org . apache . hadoop . hdfs . server . IDENT_14 . IDENT_15 . IDENT_16 . METHOD_18 ( cluster ) ; cluster . METHOD_19 ( ) ; for ( org . apache . hadoop . fs . Path path : IDENT_7 ) { java . lang . String ret = org . apache . hadoop . hdfs . IDENT_17 . METHOD_20 ( cluster . METHOD_5 ( ) , path ) ; "<AssertPlaceHolder>" ; } } finally { org . apache . hadoop . io . IDENT_18 . METHOD_21 ( org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_12 . LOG , streams . toArray ( new java . io . IDENT_19 [ 0 ] ) ) ; } } finally { cluster . METHOD_22 ( ) ; } } METHOD_5 ( ) { METHOD_23 ( ) ; return METHOD_5 ( 0 ) ; }
|
org . junit . Assert . assertEquals ( ( STRING_1 + STRING_2 ) , ret )
|
METHOD_0 ( ) { java . net . URI IDENT_0 = org . apache . hadoop . hdfs . server . common . IDENT_1 . METHOD_1 ( org . apache . hadoop . hdfs . server . common . IDENT_2 . IDENT_3 ) ; org . apache . hadoop . hdfs . server . common . IDENT_2 . LOG . info ( ( STRING_0 + IDENT_0 ) ) ; "<AssertPlaceHolder>" ; } info ( java . lang . String ) { if ( ( LOG ) != null ) { LOG . info ( msg ) ; } else if ( ( LOGGER ) != null ) { LOGGER . info ( msg ) ; } }
|
org . junit . Assert . assertNotNull ( IDENT_0 )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . METHOD_1 ( ) ; org . apache . hadoop . fs . IDENT_0 . METHOD_2 ( conf , STRING_0 ) ; long IDENT_1 = INT_0 * INT_1 ; final long IDENT_2 = 6 * IDENT_1 ; long IDENT_3 = IDENT_2 ; long start = org . apache . hadoop . util . IDENT_4 . now ( ) ; org . apache . hadoop . hdfs . util . IDENT_5 IDENT_6 = new org . apache . hadoop . hdfs . util . METHOD_3 ( IDENT_1 ) ; long IDENT_7 = 0L ; long IDENT_8 = INT_0 * INT_2 ; IDENT_6 . METHOD_4 ( IDENT_8 ) ; IDENT_3 OP_0 IDENT_8 ; IDENT_8 = INT_0 * INT_3 ; IDENT_6 . METHOD_4 ( IDENT_8 ) ; IDENT_3 OP_0 IDENT_8 ; try { java . lang . Thread . sleep ( 1000 ) ; } catch ( java . lang . IDENT_9 IDENT_10 ) { } IDENT_6 . METHOD_4 ( IDENT_3 ) ; long end = org . apache . hadoop . util . IDENT_4 . now ( ) ; "<AssertPlaceHolder>" ; } now ( ) { return java . lang . System . currentTimeMillis ( ) ; }
|
org . junit . Assert . assertTrue ( ( ( ( IDENT_7 * 1000 ) / ( end - start ) ) <= IDENT_1 ) )
|
METHOD_0 ( ) { try { METHOD_1 ( ) ; org . apache . hadoop . fs . Path IDENT_0 = new org . apache . hadoop . fs . Path ( ( ( "/" + ( org . apache . hadoop . test . IDENT_1 . METHOD_2 ( ) ) ) + STRING_0 ) ) ; org . apache . hadoop . fs . Path IDENT_2 = new org . apache . hadoop . fs . Path ( ( ( "/" + ( org . apache . hadoop . test . IDENT_1 . METHOD_2 ( ) ) ) + STRING_1 ) ) ; org . apache . hadoop . hdfs . IDENT_3 . METHOD_3 ( fs , IDENT_0 , org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_5 . IDENT_6 , org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_5 . IDENT_7 , org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_5 . IDENT_8 ) ; org . apache . hadoop . hdfs . IDENT_3 . METHOD_3 ( fs , IDENT_2 , org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_5 . IDENT_6 , org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_5 . IDENT_7 , org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_5 . IDENT_8 ) ; METHOD_4 ( ) ; java . io . File IDENT_9 = METHOD_5 ( IDENT_2 , true ) ; java . io . File IDENT_10 = METHOD_6 ( IDENT_9 , false ) ; METHOD_7 ( IDENT_2 , IDENT_9 , IDENT_10 ) ; METHOD_8 ( ) ; "<AssertPlaceHolder>" ; assert ! ( fs . exists ( IDENT_2 ) ) ; assert fs . exists ( IDENT_0 ) ; } finally { METHOD_9 ( ) ; } } METHOD_10 ( ) { org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_11 IDENT_12 = IDENT_13 . METHOD_11 ( ) . METHOD_12 ( IDENT_14 ) ; return IDENT_12 . METHOD_13 ( ) ; }
|
org . junit . Assert . assertFalse ( METHOD_10 ( ) )
|
METHOD_0 ( ) { "<AssertPlaceHolder>" ; doReturn ( ( ( org . apache . hadoop . hdfs . protocol . IDENT_0 . IDENT_1 ) * 1000 ) ) . when ( IDENT_2 ) . METHOD_1 ( ) ; try { IDENT_3 . METHOD_2 ( ) ; } catch ( java . io . IOException e ) { org . junit . Assert . fail ( STRING_0 ) ; } } METHOD_2 ( ) { org . apache . hadoop . hdfs . server . protocol . IDENT_4 IDENT_5 = null ; while ( METHOD_3 ( ) ) { try { IDENT_5 = IDENT_6 . METHOD_4 ( ) ; org . apache . hadoop . hdfs . server . IDENT_7 . IDENT_8 . LOG . debug ( ( ( ( this ) + STRING_1 ) + IDENT_5 ) ) ; break ; } catch ( java . net . IDENT_9 e ) { org . apache . hadoop . hdfs . server . IDENT_7 . IDENT_8 . LOG . warn ( ( STRING_2 + ( IDENT_10 ) ) ) ; } catch ( java . io . IOException e ) { org . apache . hadoop . hdfs . server . IDENT_7 . IDENT_8 . LOG . warn ( ( STRING_2 + ( IDENT_10 ) ) ) ; } METHOD_5 ( INT_0 , STRING_3 ) ; } if ( IDENT_5 != null ) { METHOD_6 ( IDENT_5 ) ; } else { throw new java . io . IOException ( STRING_4 ) ; } return IDENT_5 ; }
|
org . junit . Assert . assertEquals ( IDENT_0 . IDENT_1 , IDENT_3 . METHOD_2 ( ) . METHOD_1 ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . hdfs . IDENT_0 . METHOD_1 ( ) ; org . apache . hadoop . fs . IDENT_1 stream = fs . create ( new org . apache . hadoop . fs . Path ( STRING_0 ) , ( ( short ) ( 3 ) ) ) ; stream . write ( org . apache . hadoop . hdfs . IDENT_2 . METHOD_2 ( 0 , 100 ) ) ; stream . close ( ) ; final java . lang . String IDENT_3 = java . util . UUID . METHOD_3 ( ) . toString ( ) ; final org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_5 IDENT_6 = new org . apache . hadoop . hdfs . server . IDENT_4 . METHOD_4 ( IDENT_3 ) ; org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_7 [ ] IDENT_8 = org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_9 . METHOD_5 ( org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_9 . METHOD_6 ( ) , IDENT_6 ) ; cluster . METHOD_7 ( ) . METHOD_8 ( IDENT_10 , IDENT_11 , IDENT_8 ) ; org . apache . hadoop . hdfs . server . IDENT_12 . IDENT_13 IDENT_14 = cluster . METHOD_9 ( ) . METHOD_10 ( ) . METHOD_11 ( ) . METHOD_12 ( ) . METHOD_13 ( IDENT_15 . METHOD_14 ( ) ) . METHOD_15 ( IDENT_3 ) ; "<AssertPlaceHolder>" ; } METHOD_15 ( org . apache . hadoop . hdfs . server . common . IDENT_16 . IDENT_17 ) { return new org . apache . hadoop . hdfs . server . common . METHOD_16 ( 1 , 2 , STRING_1 , 3 , type , STRING_2 ) ; }
|
org . junit . Assert . assertNotNull ( IDENT_14 )
|
METHOD_0 ( ) { client . METHOD_1 ( new org . apache . hadoop . hdfs . protocol . IDENT_0 [ ] { new org . apache . hadoop . hdfs . protocol . METHOD_2 ( IDENT_1 , new org . apache . hadoop . hdfs . protocol . IDENT_2 [ ] { IDENT_3 } ) } ) ; METHOD_3 ( 1 ) ; org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_5 IDENT_6 = org . apache . hadoop . hdfs . server . IDENT_7 . IDENT_8 . METHOD_4 ( IDENT_9 , block ) ; "<AssertPlaceHolder>" ; } METHOD_5 ( ) { return METHOD_5 ; }
|
org . junit . Assert . assertThat ( IDENT_6 . METHOD_5 ( ) , org . hamcrest . CoreMatchers . is ( 0 ) )
|
METHOD_0 ( ) { final org . apache . camel . component . IDENT_0 . api . dto . IDENT_1 . IDENT_2 tree = new org . apache . camel . component . IDENT_0 . api . dto . IDENT_1 . METHOD_1 ( ) ; tree . METHOD_2 ( new org . apache . camel . component . IDENT_0 . dto . IDENT_3 . METHOD_3 ( ) ) ; tree . METHOD_2 ( new org . apache . camel . component . IDENT_0 . dto . IDENT_3 . METHOD_3 ( ) ) ; "<AssertPlaceHolder>" ; } size ( ) { return results . size ( ) ; }
|
org . junit . Assert . assertEquals ( 2 , tree . size ( ) )
|
METHOD_0 ( ) { final org . apache . camel . component . IDENT_0 . IDENT_1 config = new org . apache . camel . component . IDENT_0 . METHOD_1 ( STRING_0 , STRING_1 , STRING_2 , STRING_3 , STRING_4 , true ) ; config . METHOD_2 ( STRING_5 ) ; final org . apache . camel . component . IDENT_0 . internal . IDENT_2 session = org . apache . camel . component . IDENT_0 . internal . IDENT_3 . METHOD_3 ( config ) ; "<AssertPlaceHolder>" ; } METHOD_4 ( ) { return IDENT_4 ; }
|
org . junit . Assert . assertEquals ( STRING_5 , session . METHOD_4 ( ) )
|
METHOD_0 ( ) { for ( int i = 0 ; i < ( org . apache . hadoop . hdfs . server . IDENT_0 . IDENT_1 . IDENT_2 ) ; i ++ ) { java . util . IDENT_3 IDENT_4 = new java . util . METHOD_1 ( ( ( java . lang . System . currentTimeMillis ( ) ) + i ) ) ; java . net . IDENT_5 [ ] IDENT_6 = METHOD_2 ( IDENT_4 ) ; org . apache . hadoop . fs . Path p = new org . apache . hadoop . fs . Path ( ( STRING_0 + i ) ) ; org . apache . hadoop . fs . IDENT_7 out = org . apache . hadoop . hdfs . server . IDENT_0 . IDENT_1 . IDENT_8 . create ( p , org . apache . hadoop . fs . IDENT_9 . IDENT_10 . METHOD_3 ( ) , true , INT_0 , ( ( short ) ( 3 ) ) , INT_1 , null , IDENT_6 , new io . IDENT_11 . metadata . hdfs . entity . METHOD_4 ( STRING_1 , ( ( short ) ( 1 ) ) ) ) ; out . write ( org . apache . hadoop . hdfs . server . IDENT_0 . IDENT_1 . IDENT_12 ) ; out . close ( ) ; org . apache . hadoop . fs . IDENT_13 [ ] IDENT_14 = METHOD_5 ( p ) ; for ( org . apache . hadoop . fs . IDENT_13 IDENT_15 : IDENT_14 ) { java . lang . String [ ] IDENT_16 = IDENT_15 . METHOD_6 ( ) ; java . lang . String [ ] IDENT_17 = METHOD_7 ( IDENT_6 ) ; "<AssertPlaceHolder>" ; } } } METHOD_8 ( java . lang . String [ ] , java . lang . String [ ] ) { for ( int i = 0 ; i < ( IDENT_18 . length ) ; i ++ ) { boolean IDENT_19 = false ; for ( int j = 0 ; j < ( IDENT_20 . length ) ; j ++ ) { if ( IDENT_18 [ i ] . equals ( IDENT_20 [ j ] ) ) { IDENT_19 = true ; break ; } } if ( IDENT_19 == false ) { org . junit . Assert . fail ( ( ( IDENT_18 [ i ] ) + STRING_2 ) ) ; } } return true ; }
|
org . junit . Assert . assertTrue ( METHOD_8 ( IDENT_16 , IDENT_17 ) )
|
METHOD_0 ( ) { org . apache . hadoop . fs . Path IDENT_0 = new org . apache . hadoop . fs . Path ( STRING_0 ) ; org . apache . hadoop . hdfs . server . IDENT_1 . IDENT_2 . fs . METHOD_1 ( IDENT_0 ) ; final org . apache . hadoop . fs . Path p = new org . apache . hadoop . fs . Path ( IDENT_0 , STRING_1 ) ; final byte [ ] data = new byte [ 2 * ( org . apache . hadoop . hdfs . server . IDENT_1 . IDENT_2 . IDENT_3 ) ] ; org . apache . hadoop . hdfs . IDENT_4 . METHOD_2 ( ) . METHOD_3 ( data ) ; org . apache . hadoop . hdfs . server . IDENT_1 . IDENT_2 . METHOD_4 ( data , data . length , p ) ; final int IDENT_5 = ( data . length ) - 1 ; boolean IDENT_6 = org . apache . hadoop . hdfs . server . IDENT_1 . IDENT_2 . fs . METHOD_5 ( p , IDENT_5 ) ; "<AssertPlaceHolder>" ; org . apache . hadoop . hdfs . server . IDENT_1 . IDENT_2 . fs . METHOD_6 ( p , ( ( short ) ( ( org . apache . hadoop . hdfs . server . IDENT_1 . IDENT_2 . IDENT_7 ) - 1 ) ) ) ; org . apache . hadoop . hdfs . server . IDENT_1 . IDENT_2 . fs . METHOD_7 ( p , org . apache . hadoop . fs . IDENT_8 . IDENT_9 . METHOD_8 ( ( ( short ) ( INT_0 ) ) ) ) ; final org . apache . hadoop . fs . Path q = new org . apache . hadoop . fs . Path ( IDENT_0 , STRING_2 ) ; org . apache . hadoop . hdfs . server . IDENT_1 . IDENT_2 . fs . METHOD_9 ( p , q ) ; org . apache . hadoop . hdfs . server . IDENT_1 . IDENT_2 . METHOD_10 ( q ) ; org . apache . hadoop . hdfs . server . IDENT_1 . IDENT_2 . METHOD_11 ( q , IDENT_5 , data ) ; org . apache . hadoop . hdfs . server . IDENT_1 . IDENT_2 . cluster . METHOD_12 ( ) ; org . apache . hadoop . hdfs . server . IDENT_1 . IDENT_2 . METHOD_11 ( q , IDENT_5 , data ) ; org . apache . hadoop . hdfs . server . IDENT_1 . IDENT_2 . fs . delete ( IDENT_0 , true ) ; } METHOD_5 ( java . lang . String , long ) { METHOD_13 ( ) ; if ( IDENT_5 < 0 ) { throw new org . apache . hadoop . METHOD_14 ( ( ( STRING_3 + IDENT_5 ) + STRING_4 ) ) ; } try { return IDENT_1 . METHOD_5 ( IDENT_10 , IDENT_5 , IDENT_11 ) ; } catch ( org . apache . hadoop . IDENT_12 . IDENT_13 IDENT_14 ) { throw IDENT_14 . METHOD_15 ( org . apache . hadoop . security . IDENT_15 . class , org . apache . hadoop . hdfs . protocol . IDENT_16 . class ) ; } }
|
org . junit . Assert . assertFalse ( IDENT_6 )
|
METHOD_0 ( ) { for ( byte i = 1 ; i < 16 ; i ++ ) { final org . apache . hadoop . hdfs . server . IDENT_0 . IDENT_1 f = org . apache . hadoop . hdfs . server . IDENT_0 . IDENT_2 . METHOD_1 ( i ) ; "<AssertPlaceHolder>" ; } } METHOD_2 ( ) { byte id = METHOD_3 ( ) ; if ( id == ( org . apache . hadoop . hdfs . server . IDENT_3 . IDENT_4 . IDENT_5 ) ) { return ( this . METHOD_4 ( ) ) != null ? this . METHOD_4 ( ) . METHOD_2 ( ) : id ; } return id ; }
|
org . junit . Assert . assertEquals ( i , f . METHOD_2 ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . conf . Configuration ( ) ; org . apache . hadoop . hdfs . IDENT_0 cluster = new org . apache . hadoop . hdfs . IDENT_0 . Builder ( conf ) . METHOD_1 ( 1 ) . build ( ) ; cluster . METHOD_2 ( ) ; org . apache . hadoop . fs . IDENT_1 fs = cluster . METHOD_3 ( ) ; org . apache . hadoop . fs . Path path = new org . apache . hadoop . fs . Path ( STRING_0 ) ; "<AssertPlaceHolder>" ; int size = conf . METHOD_4 ( IDENT_2 . IDENT_3 , INT_0 ) ; byte [ ] data = new byte [ size ] ; org . apache . hadoop . fs . Path IDENT_4 = new org . apache . hadoop . fs . Path ( STRING_1 ) ; org . apache . hadoop . fs . IDENT_5 IDENT_6 = fs . create ( IDENT_4 ) ; fs . delete ( IDENT_4 , false ) ; try { IDENT_6 . write ( data , 0 , data . length ) ; IDENT_6 . METHOD_5 ( ) ; org . junit . Assert . fail ( STRING_2 ) ; } catch ( java . lang . Exception e ) { } finally { if ( cluster != null ) { cluster . METHOD_6 ( ) ; } } } METHOD_7 ( org . apache . hadoop . fs . Path ) { return false ; }
|
org . junit . Assert . assertTrue ( fs . METHOD_7 ( path ) )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . conf . Configuration ( ) ; org . apache . hadoop . hdfs . IDENT_0 cluster = null ; try { cluster = new org . apache . hadoop . hdfs . IDENT_0 . Builder ( conf ) . METHOD_1 ( 1 ) . build ( ) ; cluster . METHOD_2 ( ) ; org . apache . hadoop . fs . IDENT_1 fs = cluster . METHOD_3 ( ) ; METHOD_4 ( fs , STRING_0 , false ) ; METHOD_4 ( fs , STRING_0 , false ) ; org . apache . hadoop . fs . Path IDENT_2 = new org . apache . hadoop . fs . Path ( STRING_0 ) ; IDENT_3 . IDENT_4 = false ; METHOD_5 ( cluster ) ; fs . delete ( IDENT_2 , true ) ; org . apache . hadoop . hdfs . IDENT_5 . METHOD_6 ( fs , IDENT_2 , 10 , ( ( short ) ( 1 ) ) , 0L ) ; "<AssertPlaceHolder>" ; METHOD_7 ( cluster ) ; } finally { if ( cluster != null ) { cluster . METHOD_8 ( ) ; } } } METHOD_9 ( org . apache . hadoop . fs . Path ) { return false ; }
|
org . junit . Assert . assertTrue ( ( ! ( fs . METHOD_9 ( IDENT_2 ) ) ) )
|
METHOD_0 ( ) { io . IDENT_0 . metadata . IDENT_1 . METHOD_1 ( conf ) ; io . IDENT_0 . metadata . IDENT_1 . METHOD_2 ( ) ; org . apache . hadoop . hdfs . server . IDENT_2 . IDENT_3 IDENT_4 = org . mockito . Mockito . mock ( org . apache . hadoop . hdfs . server . IDENT_2 . IDENT_3 . class ) ; org . apache . hadoop . hdfs . server . IDENT_2 . IDENT_5 IDENT_6 = org . mockito . Mockito . mock ( org . apache . hadoop . hdfs . server . IDENT_2 . IDENT_5 . class ) ; org . mockito . Mockito . when ( IDENT_6 . METHOD_3 ( ) ) . thenReturn ( true ) ; org . apache . hadoop . hdfs . server . IDENT_2 . IDENT_7 IDENT_2 = org . mockito . Mockito . mock ( org . apache . hadoop . hdfs . server . IDENT_2 . IDENT_7 . class ) ; org . mockito . Mockito . when ( IDENT_6 . METHOD_4 ( ) ) . thenReturn ( IDENT_2 ) ; org . mockito . Mockito . when ( IDENT_2 . getId ( ) ) . thenReturn ( 0L ) ; io . IDENT_0 . IDENT_8 . node . IDENT_9 IDENT_10 = org . mockito . Mockito . mock ( io . IDENT_0 . IDENT_8 . node . IDENT_9 . class ) ; org . mockito . Mockito . when ( IDENT_2 . METHOD_5 ( ) ) . thenReturn ( IDENT_10 ) ; org . mockito . Mockito . when ( IDENT_10 . METHOD_6 ( ) ) . thenReturn ( new java . util . ArrayList < io . IDENT_0 . IDENT_8 . node . IDENT_11 > ( ) ) ; org . mockito . Mockito . when ( IDENT_6 . METHOD_7 ( ) ) . thenReturn ( IDENT_4 ) ; org . apache . hadoop . hdfs . server . IDENT_2 . IDENT_12 IDENT_13 = new org . apache . hadoop . hdfs . server . IDENT_2 . METHOD_8 ( IDENT_6 ) ; IDENT_13 . METHOD_9 ( 0 , 0 ) ; METHOD_10 ( IDENT_13 , STRING_0 , STRING_1 ) ; METHOD_10 ( IDENT_13 , STRING_2 , STRING_3 ) ; METHOD_10 ( IDENT_13 , STRING_4 , STRING_5 ) ; "<AssertPlaceHolder>" ; IDENT_13 . METHOD_11 ( ) ; } METHOD_12 ( ) { io . IDENT_0 . transaction . handler . IDENT_14 IDENT_15 = new io . IDENT_0 . transaction . handler . METHOD_13 ( io . IDENT_0 . transaction . handler . IDENT_16 . IDENT_17 ) { @ org . apache . hadoop . hdfs . server . IDENT_2 . Override public void METHOD_14 ( io . IDENT_0 . transaction . IDENT_18 . IDENT_19 IDENT_20 ) throws java . io . IOException { } @ org . apache . hadoop . hdfs . server . IDENT_2 . Override public java . lang . Object METHOD_15 ( ) throws io . IDENT_0 . exception . IDENT_21 , java . io . IOException { io . IDENT_0 . metadata . hdfs . dal . IDENT_22 < org . apache . hadoop . hdfs . server . IDENT_2 . IDENT_23 > IDENT_24 = ( ( io . IDENT_0 . metadata . hdfs . dal . IDENT_22 ) ( io . IDENT_0 . metadata . IDENT_1 . METHOD_16 ( io . IDENT_0 . metadata . hdfs . dal . IDENT_22 . class ) ) ) ; return IDENT_24 . METHOD_17 ( ) ; } } ; return ( ( int ) ( IDENT_15 . METHOD_18 ( IDENT_25 ) ) ) ; }
|
org . junit . Assert . assertEquals ( IDENT_13 . METHOD_12 ( ) , 3 )
|
METHOD_0 ( ) { com . IDENT_0 . IDENT_1 . IDENT_2 . IDENT_3 . log ( STRING_0 ) ; com . IDENT_0 . IDENT_1 . IDENT_2 . IDENT_4 instance = new com . IDENT_0 . IDENT_1 . IDENT_2 . METHOD_1 ( new com . IDENT_0 . IDENT_1 . IDENT_2 . IDENT_3 . METHOD_2 ( ) ) ; instance . METHOD_3 ( ) ; "<AssertPlaceHolder>" ; } METHOD_4 ( ) { return this . IDENT_5 . METHOD_4 ( ) ; }
|
org . junit . Assert . assertFalse ( instance . METHOD_4 ( ) )
|
METHOD_0 ( ) { final org . eclipse . IDENT_0 . client . api . IDENT_1 request = mock ( org . eclipse . IDENT_0 . client . api . IDENT_1 . class ) ; final org . mockito . IDENT_2 < org . eclipse . IDENT_0 . client . api . Response . IDENT_3 > listener = org . mockito . IDENT_2 . METHOD_1 ( org . eclipse . IDENT_0 . client . api . Response . IDENT_3 . class ) ; METHOD_2 ( ) . when ( request ) . METHOD_3 ( listener . METHOD_4 ( ) ) ; client . METHOD_5 ( request , ( response , headers , exception ) -> { } ) ; final org . eclipse . IDENT_0 . client . api . Result result = mock ( org . eclipse . IDENT_0 . client . api . Result . class ) ; final org . eclipse . IDENT_0 . client . api . Response response = mock ( org . eclipse . IDENT_0 . client . api . Response . class ) ; when ( result . METHOD_6 ( ) ) . thenReturn ( response ) ; when ( response . METHOD_7 ( ) ) . thenReturn ( new org . eclipse . IDENT_0 . http . METHOD_8 ( ) ) ; final org . apache . camel . component . IDENT_4 . internal . client . IDENT_5 IDENT_6 = mock ( org . apache . camel . component . IDENT_4 . internal . client . IDENT_5 . class ) ; when ( result . METHOD_9 ( ) ) . thenReturn ( IDENT_6 ) ; final org . eclipse . IDENT_0 . client . IDENT_7 IDENT_8 = mock ( org . eclipse . IDENT_0 . client . IDENT_7 . class ) ; when ( IDENT_6 . METHOD_10 ( ) ) . thenReturn ( IDENT_8 ) ; when ( IDENT_8 . METHOD_11 ( IDENT_9 . IDENT_10 ) ) . thenReturn ( IDENT_6 ) ; listener . getValue ( ) . METHOD_12 ( result ) ; final long IDENT_11 = java . lang . System . currentTimeMillis ( ) ; client . METHOD_13 ( ) ; final long IDENT_12 = ( java . lang . System . currentTimeMillis ( ) ) - IDENT_11 ; "<AssertPlaceHolder>" ; } METHOD_13 ( ) { if ( ( IDENT_13 ) != null ) { IDENT_13 . METHOD_14 ( ) ; if ( ! ( IDENT_13 . METHOD_15 ( ) ) ) { try { IDENT_13 . METHOD_16 ( 0 , IDENT_14 , TimeUnit . IDENT_15 ) ; } catch ( java . lang . IDENT_16 | java . util . concurrent . IDENT_17 IDENT_18 ) { } } } session . METHOD_17 ( this ) ; }
|
org . junit . Assert . assertTrue ( ( IDENT_12 < 10 ) )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . hdfs . server . IDENT_0 . METHOD_1 ( ) ; final int IDENT_1 = 100 ; final java . lang . String IDENT_2 = STRING_0 ; java . lang . String IDENT_3 = org . apache . hadoop . security . IDENT_4 . METHOD_2 ( ) . METHOD_3 ( ) ; conf . set ( java . lang . String . format ( STRING_1 , IDENT_3 ) , STRING_2 ) ; conf . set ( java . lang . String . format ( STRING_3 , IDENT_3 ) , STRING_2 ) ; conf . set ( java . lang . String . format ( STRING_4 , IDENT_3 ) , STRING_2 ) ; org . apache . hadoop . security . IDENT_5 . IDENT_6 . METHOD_4 ( conf ) ; final org . apache . hadoop . hdfs . server . IDENT_0 . IDENT_7 cluster = new org . apache . hadoop . hdfs . server . IDENT_0 . IDENT_7 . Builder ( conf ) . format ( true ) . build ( ) ; cluster . METHOD_5 ( ) ; org . apache . hadoop . fs . IDENT_8 IDENT_9 = cluster . METHOD_6 ( ) ; java . util . List < org . apache . hadoop . security . IDENT_4 > IDENT_10 = new java . util . ArrayList ( ) ; java . util . List < org . apache . hadoop . fs . IDENT_8 > IDENT_11 = new java . util . ArrayList ( ) ; for ( int IDENT_12 = 0 ; IDENT_12 < IDENT_1 ; IDENT_12 ++ ) { org . apache . hadoop . security . IDENT_4 IDENT_13 = org . apache . hadoop . security . IDENT_4 . METHOD_7 ( ( IDENT_2 + IDENT_12 ) , org . apache . hadoop . security . IDENT_4 . METHOD_8 ( ) , new java . lang . String [ ] { IDENT_2 + IDENT_12 } ) ; org . apache . hadoop . fs . IDENT_8 fs = IDENT_13 . METHOD_9 ( new java . security . IDENT_14 < org . apache . hadoop . fs . IDENT_8 > ( ) { @ org . apache . hadoop . hdfs . server . IDENT_0 . Override public org . apache . hadoop . fs . IDENT_8 run ( ) throws java . lang . Exception { return cluster . METHOD_6 ( ) ; } } ) ; IDENT_10 . add ( IDENT_13 ) ; IDENT_11 . add ( fs ) ; } try { IDENT_9 . METHOD_10 ( new org . apache . hadoop . fs . Path ( STRING_5 ) ) ; IDENT_9 . METHOD_11 ( new org . apache . hadoop . fs . Path ( STRING_5 ) , new org . apache . hadoop . fs . IDENT_15 . METHOD_12 ( org . apache . hadoop . fs . IDENT_15 . IDENT_16 . IDENT_17 , org . apache . hadoop . fs . IDENT_15 . IDENT_16 . IDENT_17 , org . apache . hadoop . fs . IDENT_15 . IDENT_16 . IDENT_17 ) ) ; for ( int IDENT_12 = 0 ; IDENT_12 < ( IDENT_11 . size ( ) ) ; IDENT_12 ++ ) { org . apache . hadoop . fs . IDENT_8 fs = IDENT_11 . get ( IDENT_12 ) ; org . apache . hadoop . fs . Path root = new org . apache . hadoop . fs . Path ( java . lang . String . format ( STRING_6 , IDENT_12 ) ) ; fs . METHOD_10 ( root ) ; fs . METHOD_13 ( root , ( IDENT_2 + IDENT_12 ) , ( IDENT_2 + IDENT_12 ) ) ; fs . METHOD_10 ( new org . apache . hadoop . fs . Path ( root , ( "b" + IDENT_12 ) ) ) ; fs . METHOD_10 ( new org . apache . hadoop . fs . Path ( root , ( STRING_7 + IDENT_12 ) ) ) ; fs . create ( new org . apache . hadoop . fs . Path ( root , ( ( "b" + IDENT_12 ) + STRING_8 ) ) ) . close ( ) ; fs . create ( new org . apache . hadoop . fs . Path ( root , ( ( STRING_7 + IDENT_12 ) + STRING_8 ) ) ) . close ( ) ; } for ( int IDENT_12 = 0 ; IDENT_12 < ( IDENT_11 . size ( ) ) ; IDENT_12 ++ ) { org . apache . hadoop . fs . IDENT_8 fs = IDENT_11 . get ( IDENT_12 ) ; "<AssertPlaceHolder>" ; org . apache . hadoop . fs . IDENT_8 . METHOD_14 ( IDENT_10 . get ( IDENT_12 ) ) ; } } finally { cluster . METHOD_15 ( ) ; } } delete ( org . apache . commons . net . IDENT_18 . IDENT_19 , org . apache . hadoop . fs . Path , boolean ) { org . apache . hadoop . fs . Path IDENT_20 = new org . apache . hadoop . fs . Path ( client . METHOD_16 ( ) ) ; org . apache . hadoop . fs . Path IDENT_21 = METHOD_17 ( IDENT_20 , file ) ; java . lang . String IDENT_22 = IDENT_21 . METHOD_18 ( ) . getPath ( ) ; try { org . apache . hadoop . fs . IDENT_23 IDENT_24 = METHOD_19 ( client , IDENT_21 ) ; if ( IDENT_24 . METHOD_20 ( ) ) { return client . METHOD_21 ( IDENT_22 ) ; } } catch ( java . io . IDENT_25 e ) { return false ; } org . apache . hadoop . fs . IDENT_23 [ ] IDENT_26 = METHOD_22 ( client , IDENT_21 ) ; if ( ( ( IDENT_26 != null ) && ( ( IDENT_26 . length ) > 0 ) ) && ( ! IDENT_27 ) ) { throw new java . io . IOException ( ( ( STRING_9 + file ) + STRING_10 ) ) ; } for ( org . apache . hadoop . fs . IDENT_23 IDENT_28 : IDENT_26 ) { delete ( client , new org . apache . hadoop . fs . Path ( IDENT_21 , IDENT_28 . getPath ( ) ) , IDENT_27 ) ; } return client . METHOD_23 ( IDENT_22 ) ; }
|
org . junit . Assert . assertTrue ( fs . delete ( new org . apache . hadoop . fs . Path ( java . lang . String . format ( STRING_6 , IDENT_12 ) ) , true ) )
|
METHOD_0 ( ) { final org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . conf . Configuration ( ) ; conf . METHOD_1 ( IDENT_0 . IDENT_1 , false ) ; conf . METHOD_2 ( IDENT_0 . IDENT_2 , org . apache . hadoop . hdfs . server . IDENT_3 . IDENT_4 . IDENT_5 . IDENT_6 ) ; org . apache . hadoop . fs . IDENT_7 IDENT_8 = null ; final org . apache . hadoop . hdfs . IDENT_9 cluster = new org . apache . hadoop . hdfs . IDENT_9 . Builder ( conf ) . METHOD_3 ( org . apache . hadoop . hdfs . IDENT_10 . METHOD_4 ( 2 ) ) . METHOD_5 ( 3 ) . build ( ) ; try { cluster . METHOD_6 ( ) ; java . lang . Thread . sleep ( INT_0 ) ; org . apache . hadoop . hdfs . server . IDENT_3 . IDENT_4 . IDENT_5 . LOG . info ( STRING_0 ) ; org . apache . hadoop . fs . IDENT_11 fs = cluster . METHOD_7 ( 0 ) ; IDENT_8 = fs . create ( org . apache . hadoop . hdfs . server . IDENT_3 . IDENT_4 . IDENT_5 . IDENT_12 ) ; org . apache . hadoop . hdfs . IDENT_13 . write ( IDENT_8 , 0 , org . apache . hadoop . hdfs . server . IDENT_3 . IDENT_4 . IDENT_5 . IDENT_14 ) ; IDENT_8 . METHOD_8 ( ) ; org . apache . hadoop . hdfs . server . IDENT_3 . IDENT_4 . IDENT_5 . LOG . info ( STRING_1 ) ; cluster . METHOD_9 ( 0 ) ; cluster . METHOD_6 ( ) ; "<AssertPlaceHolder>" ; org . apache . hadoop . fs . IDENT_11 IDENT_15 = METHOD_10 ( cluster , conf ) ; org . apache . hadoop . hdfs . server . IDENT_3 . IDENT_4 . IDENT_5 . METHOD_11 ( IDENT_15 , org . apache . hadoop . hdfs . server . IDENT_3 . IDENT_4 . IDENT_5 . IDENT_12 ) ; org . apache . hadoop . hdfs . IDENT_13 . METHOD_12 ( fs , org . apache . hadoop . hdfs . server . IDENT_3 . IDENT_4 . IDENT_5 . IDENT_12 , org . apache . hadoop . hdfs . server . IDENT_3 . IDENT_4 . IDENT_5 . IDENT_14 ) ; } finally { org . apache . hadoop . io . IDENT_16 . METHOD_13 ( IDENT_8 ) ; cluster . METHOD_14 ( ) ; } } exists ( org . apache . hadoop . fs . Path ) { return fs . exists ( p ) ; }
|
org . junit . Assert . assertTrue ( fs . exists ( org . apache . hadoop . hdfs . server . IDENT_3 . IDENT_4 . IDENT_5 . IDENT_12 ) )
|
METHOD_0 ( ) { org . apache . hadoop . hdfs . IDENT_0 conf = new org . apache . hadoop . hdfs . METHOD_1 ( ) ; IDENT_1 = org . apache . hadoop . fs . IDENT_2 . METHOD_2 ( conf ) ; org . apache . hadoop . fs . Path IDENT_3 = IDENT_1 . METHOD_3 ( ) ; org . apache . hadoop . fs . Path IDENT_4 = new org . apache . hadoop . fs . Path ( IDENT_3 , ( ( java . lang . System . getProperty ( STRING_0 , STRING_1 ) ) + STRING_2 ) ) ; org . apache . hadoop . fs . Path IDENT_5 = new org . apache . hadoop . fs . Path ( IDENT_4 , STRING_3 ) ; org . apache . hadoop . fs . Path IDENT_6 = new org . apache . hadoop . fs . Path ( IDENT_4 , STRING_4 ) ; conf . set ( IDENT_7 . IDENT_8 , IDENT_5 . METHOD_4 ( ) . getPath ( ) ) ; conf . set ( IDENT_7 . IDENT_9 , IDENT_6 . METHOD_4 ( ) . getPath ( ) ) ; METHOD_5 ( IDENT_5 , null ) ; METHOD_5 ( IDENT_6 , null ) ; java . lang . String [ ] args = new java . lang . String [ ] { STRING_5 } ; java . lang . String ret = METHOD_6 ( conf , args , true ) ; "<AssertPlaceHolder>" ; METHOD_7 ( IDENT_1 , IDENT_6 . METHOD_8 ( ) ) ; } METHOD_4 ( ) { return uri ; }
|
org . junit . Assert . assertEquals ( IDENT_6 . METHOD_4 ( ) . getPath ( ) , ret . trim ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . hdfs . IDENT_0 conf = new org . apache . hadoop . hdfs . METHOD_1 ( ) ; IDENT_1 = org . apache . hadoop . fs . IDENT_2 . METHOD_2 ( conf ) ; org . apache . hadoop . fs . Path IDENT_3 = IDENT_1 . METHOD_3 ( ) ; org . apache . hadoop . fs . Path IDENT_4 = new org . apache . hadoop . fs . Path ( IDENT_3 , ( ( java . lang . System . getProperty ( STRING_0 , STRING_1 ) ) + STRING_2 ) ) ; org . apache . hadoop . fs . Path IDENT_5 = new org . apache . hadoop . fs . Path ( IDENT_4 , STRING_3 ) ; org . apache . hadoop . fs . Path IDENT_6 = new org . apache . hadoop . fs . Path ( IDENT_4 , STRING_4 ) ; conf . set ( IDENT_7 . IDENT_8 , IDENT_5 . METHOD_4 ( ) . getPath ( ) ) ; conf . set ( IDENT_7 . IDENT_9 , IDENT_6 . METHOD_4 ( ) . getPath ( ) ) ; METHOD_5 ( IDENT_5 , null ) ; METHOD_5 ( IDENT_6 , null ) ; java . lang . String [ ] args = new java . lang . String [ ] { STRING_5 } ; java . lang . String ret = METHOD_6 ( conf , args , true ) ; "<AssertPlaceHolder>" ; METHOD_7 ( IDENT_1 , IDENT_6 . METHOD_8 ( ) ) ; } METHOD_4 ( ) { return uri ; }
|
org . junit . Assert . assertEquals ( IDENT_5 . METHOD_4 ( ) . getPath ( ) , ret . trim ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . hdfs . util . IDENT_0 < java . lang . Integer > l = new org . apache . hadoop . hdfs . util . IDENT_0 < java . lang . Integer > ( ) ; for ( int i = 0 ; i < INT_0 ; i ++ ) { l . add ( i ) ; } int i = 0 ; for ( int IDENT_1 : l ) { "<AssertPlaceHolder>" ; i ++ ; } } add ( org . apache . hadoop . yarn . server . IDENT_2 . IDENT_3 . dao . IDENT_4 ) { this . IDENT_5 . add ( IDENT_5 ) ; }
|
org . junit . Assert . assertEquals ( i , IDENT_1 )
|
METHOD_0 ( ) { org . apache . hadoop . hdfs . util . IDENT_0 s = new org . apache . hadoop . hdfs . util . METHOD_1 ( org . apache . hadoop . hdfs . util . IDENT_1 . METHOD_2 ( STRING_0 ) , 5 ) ; "<AssertPlaceHolder>" ; try { s . METHOD_3 ( 1 ) ; org . junit . Assert . fail ( STRING_1 ) ; } catch ( java . io . IDENT_2 e ) { } } METHOD_3 ( long ) { METHOD_4 ( ) ; if ( n > 0 ) { if ( ( n + ( pos ) ) > ( length ) ) { n = ( length ) - ( pos ) ; } pos += n ; return n ; } return n < 0 ? - 1 : 0 ; }
|
org . junit . Assert . assertEquals ( 2 , s . METHOD_3 ( 3 ) )
|
METHOD_0 ( ) { org . apache . hadoop . hdfs . util . IDENT_0 s = new org . apache . hadoop . hdfs . util . METHOD_1 ( org . apache . hadoop . hdfs . util . IDENT_1 . METHOD_2 ( STRING_0 ) , 5 ) ; byte [ ] buf = new byte [ 10 ] ; "<AssertPlaceHolder>" ; try { s . read ( buf , 2 , 3 ) ; org . junit . Assert . fail ( STRING_1 ) ; } catch ( java . io . IDENT_2 e ) { } } read ( byte [ ] , int , int ) { if ( b == null ) { throw new java . lang . METHOD_3 ( ) ; } else if ( ( ( IDENT_3 < 0 ) || ( len < 0 ) ) || ( len > ( ( b . length ) - IDENT_3 ) ) ) { throw new java . lang . METHOD_4 ( ) ; } else if ( len == 0 ) { return 0 ; } METHOD_5 ( ) ; if ( ( pos ) < ( length ) ) { int n = ( ( int ) ( java . lang . Math . min ( len , ( ( length ) - ( pos ) ) ) ) ) ; java . lang . System . METHOD_6 ( data , pos , b , IDENT_3 , n ) ; pos += n ; return n ; } return - 1 ; }
|
org . junit . Assert . assertEquals ( 2 , s . read ( buf , 0 , 5 ) )
|
METHOD_0 ( ) { org . apache . hadoop . hdfs . util . IDENT_0 s = new org . apache . hadoop . hdfs . util . METHOD_1 ( org . apache . hadoop . hdfs . util . IDENT_1 . METHOD_2 ( STRING_0 ) , 5 ) ; "<AssertPlaceHolder>" ; try { s . METHOD_3 ( 1 ) ; org . junit . Assert . fail ( STRING_1 ) ; } catch ( java . lang . IDENT_2 IDENT_3 ) { } } METHOD_4 ( ) { return false ; }
|
org . junit . Assert . assertFalse ( s . METHOD_4 ( ) )
|
METHOD_0 ( ) { final int IDENT_0 = 0 ; final org . apache . camel . maven . IDENT_1 IDENT_2 = new org . apache . camel . maven . METHOD_1 ( ) ; IDENT_2 . IDENT_3 = org . apache . camel . maven . IDENT_4 . METHOD_2 ( ) ; IDENT_2 . IDENT_5 = org . apache . camel . maven . IDENT_4 . METHOD_3 ( ) ; int IDENT_6 = 0 ; try { IDENT_2 . METHOD_4 ( ) ; } catch ( final java . lang . IllegalArgumentException e ) { IDENT_6 ++ ; } try { IDENT_2 . METHOD_5 ( ) ; } catch ( final java . lang . IllegalArgumentException e ) { IDENT_6 ++ ; } "<AssertPlaceHolder>" ; } METHOD_5 ( ) { org . apache . camel . maven . METHOD_1 . METHOD_6 ( IDENT_3 , IDENT_7 ) ; }
|
org . junit . Assert . assertEquals ( IDENT_0 , IDENT_6 )
|
METHOD_0 ( ) { final long now = org . apache . hadoop . util . IDENT_0 . now ( ) ; final java . lang . String parent = STRING_0 ; final org . apache . hadoop . hdfs . protocol . IDENT_1 status = new org . apache . hadoop . hdfs . protocol . METHOD_1 ( INT_0 , false , 3 , ( 1L << INT_1 ) , now , ( now + 10 ) , new org . apache . hadoop . fs . IDENT_2 . METHOD_2 ( ( ( short ) ( INT_2 ) ) ) , STRING_1 , STRING_2 , org . apache . hadoop . hdfs . IDENT_3 . METHOD_3 ( "bar" ) , org . apache . hadoop . hdfs . IDENT_3 . METHOD_3 ( "foo" ) , ( - 1 ) , 0 , false , ( ( byte ) ( 0 ) ) ) ; final org . apache . hadoop . fs . IDENT_4 IDENT_5 = org . apache . hadoop . hdfs . web . IDENT_6 . METHOD_4 ( status , parent ) ; System . out . println ( ( STRING_3 + status ) ) ; System . out . println ( ( STRING_4 + IDENT_5 ) ) ; final java . lang . String json = org . apache . hadoop . hdfs . web . IDENT_7 . METHOD_5 ( status , true ) ; System . out . println ( ( STRING_5 + ( json . replace ( "bar" 1 , STRING_6 ) ) ) ) ; org . IDENT_8 . jackson . map . IDENT_9 reader = new org . IDENT_8 . jackson . map . METHOD_6 ( ) . reader ( java . util . Map . class ) ; final org . apache . hadoop . hdfs . protocol . METHOD_1 IDENT_10 = org . apache . hadoop . hdfs . web . IDENT_7 . METHOD_4 ( ( ( java . util . Map < ? , ? > ) ( reader . METHOD_7 ( json ) ) ) , true ) ; final org . apache . hadoop . fs . IDENT_4 IDENT_11 = org . apache . hadoop . hdfs . web . IDENT_6 . METHOD_4 ( IDENT_10 , parent ) ; System . out . println ( ( STRING_7 + IDENT_10 ) ) ; System . out . println ( ( "bar" 0 + IDENT_11 ) ) ; "<AssertPlaceHolder>" ; } println ( java . lang . Object ) { System . out . println ( s ) ; }
|
org . junit . Assert . assertEquals ( IDENT_5 , IDENT_11 )
|
METHOD_0 ( ) { java . util . concurrent . IDENT_0 < java . lang . String > future = METHOD_1 ( org . apache . hadoop . hdfs . web . IDENT_1 . IDENT_2 ) ; try { org . apache . hadoop . hdfs . web . IDENT_1 . fs . METHOD_2 ( org . apache . hadoop . hdfs . web . IDENT_1 . p ) ; org . junit . Assert . fail ( ) ; } catch ( java . io . IOException IDENT_3 ) { } "<AssertPlaceHolder>" ; } METHOD_3 ( java . util . concurrent . IDENT_0 ) { java . lang . String request = null ; try { request = future . get ( 2 , TimeUnit . IDENT_4 ) ; } catch ( java . lang . Exception e ) { org . junit . Assert . fail ( e . toString ( ) ) ; } java . util . regex . IDENT_5 matcher = org . apache . hadoop . hdfs . web . IDENT_1 . IDENT_6 . matcher ( request ) ; return matcher . find ( ) ? matcher . group ( 2 ) : null ; }
|
org . junit . Assert . assertEquals ( null , METHOD_3 ( future ) )
|
METHOD_0 ( ) { java . util . concurrent . IDENT_0 < java . lang . String > future = METHOD_1 ( org . apache . hadoop . hdfs . web . IDENT_1 . IDENT_2 ) ; try { org . apache . hadoop . hdfs . web . IDENT_1 . fs . delete ( org . apache . hadoop . hdfs . web . IDENT_1 . p , false ) ; org . junit . Assert . fail ( ) ; } catch ( java . io . IOException IDENT_3 ) { } "<AssertPlaceHolder>" ; } METHOD_2 ( java . util . concurrent . IDENT_0 ) { java . lang . String request = null ; try { request = future . get ( 2 , TimeUnit . IDENT_4 ) ; } catch ( java . lang . Exception e ) { org . junit . Assert . fail ( e . toString ( ) ) ; } java . util . regex . IDENT_5 matcher = org . apache . hadoop . hdfs . web . IDENT_1 . IDENT_6 . matcher ( request ) ; return matcher . find ( ) ? matcher . group ( 2 ) : null ; }
|
org . junit . Assert . assertEquals ( null , METHOD_2 ( future ) )
|
METHOD_0 ( ) { final org . apache . hadoop . hdfs . web . resources . IDENT_0 p = new org . apache . hadoop . hdfs . web . resources . METHOD_1 ( METHOD_1 . IDENT_1 ) ; "<AssertPlaceHolder>" ; } getValue ( ) { return this . value ; }
|
org . junit . Assert . assertEquals ( null , p . getValue ( ) )
|
METHOD_0 ( ) { final org . apache . hadoop . hdfs . web . resources . IDENT_0 p = new org . apache . hadoop . hdfs . web . resources . METHOD_1 ( METHOD_1 . IDENT_1 ) ; "<AssertPlaceHolder>" ; new org . apache . hadoop . hdfs . web . resources . METHOD_1 ( STRING_0 ) ; try { new org . apache . hadoop . hdfs . web . resources . METHOD_1 ( STRING_1 ) ; org . junit . Assert . fail ( ) ; } catch ( java . lang . IllegalArgumentException e ) { org . apache . hadoop . hdfs . web . resources . IDENT_2 . LOG . info ( ( STRING_2 + e ) ) ; } } getValue ( ) { return this . value ; }
|
org . junit . Assert . assertEquals ( false , p . getValue ( ) )
|
METHOD_0 ( ) { java . lang . String IDENT_0 = STRING_0 + ( org . apache . hadoop . net . IDENT_1 . getPath ( IDENT_2 [ 0 ] ) ) ; java . util . Map < org . apache . hadoop . net . Node , java . lang . Integer > IDENT_3 = METHOD_1 ( 100 , IDENT_0 ) ; for ( org . apache . hadoop . net . Node key : IDENT_2 ) { "<AssertPlaceHolder>" ; } } get ( org . apache . hadoop . hdfs . server . IDENT_4 . IDENT_5 ) { java . util . Collection < io . IDENT_6 . metadata . hdfs . entity . IDENT_7 > IDENT_8 = METHOD_2 ( IDENT_9 ) ; if ( IDENT_8 == null ) { return null ; } java . util . IDENT_10 < java . lang . String > IDENT_11 = new java . util . METHOD_3 ( ) ; for ( io . IDENT_6 . metadata . hdfs . entity . IDENT_7 IDENT_12 : IDENT_8 ) { IDENT_11 . add ( IDENT_13 . METHOD_4 ( IDENT_12 . METHOD_5 ( ) ) . METHOD_6 ( ) ) ; } return IDENT_11 ; }
|
org . junit . Assert . assertTrue ( ( ( ( IDENT_3 . get ( key ) ) > 0 ) || ( key == ( IDENT_2 [ 0 ] ) ) ) )
|
METHOD_0 ( ) { org . apache . hadoop . IDENT_0 . IDENT_1 IDENT_2 = new org . apache . hadoop . IDENT_0 . METHOD_1 ( "test" , 1 , true , 2 ) ; org . apache . hadoop . IDENT_0 . METHOD_1 IDENT_3 = org . apache . hadoop . IDENT_0 . METHOD_1 . METHOD_2 ( STRING_0 ) ; "<AssertPlaceHolder>" ; } compareTo ( org . apache . hadoop . fs . IDENT_4 ) { if ( IDENT_5 == null ) { return 1 ; } return ( hashCode ( ) ) - ( IDENT_5 . hashCode ( ) ) ; }
|
org . junit . Assert . assertEquals ( 0 , IDENT_2 . compareTo ( IDENT_3 ) )
|
METHOD_0 ( ) { org . apache . hadoop . mapreduce . v2 . app . job . impl . IDENT_0 . IDENT_1 IDENT_2 = new org . apache . hadoop . mapreduce . v2 . app . job . impl . IDENT_0 . METHOD_1 ( false ) ; org . apache . hadoop . yarn . event . IDENT_3 IDENT_4 = mock ( org . apache . hadoop . yarn . event . IDENT_3 . class ) ; java . lang . String [ ] IDENT_5 = new java . lang . String [ 3 ] ; IDENT_5 [ 0 ] = STRING_0 ; IDENT_5 [ 1 ] = STRING_1 ; IDENT_5 [ 2 ] = STRING_2 ; org . apache . hadoop . mapreduce . split . IDENT_6 . IDENT_7 IDENT_8 = new org . apache . hadoop . mapreduce . split . IDENT_6 . METHOD_2 ( IDENT_5 , 0 , ( ( INT_0 * INT_1 ) * INT_2 ) ) ; org . apache . hadoop . mapreduce . v2 . app . job . impl . IDENT_0 IDENT_9 = METHOD_3 ( IDENT_4 , IDENT_8 ) ; org . apache . hadoop . mapreduce . v2 . app . job . impl . IDENT_0 IDENT_10 = METHOD_4 ( IDENT_9 ) ; when ( IDENT_10 . METHOD_5 ( IDENT_5 [ 0 ] ) ) . thenReturn ( STRING_3 ) ; IDENT_10 . IDENT_11 = IDENT_10 . METHOD_6 ( IDENT_8 . METHOD_7 ( ) ) ; org . apache . hadoop . mapreduce . v2 . app . job . event . IDENT_12 IDENT_13 = mock ( org . apache . hadoop . mapreduce . v2 . app . job . event . IDENT_12 . class ) ; IDENT_2 . METHOD_8 ( IDENT_10 , IDENT_13 ) ; verify ( IDENT_10 ) . METHOD_5 ( IDENT_5 [ 0 ] ) ; org . mockito . IDENT_14 < org . apache . hadoop . yarn . event . IDENT_15 > IDENT_16 = org . mockito . IDENT_14 . METHOD_9 ( org . apache . hadoop . yarn . event . IDENT_15 . class ) ; verify ( IDENT_4 , times ( 2 ) ) . METHOD_10 ( IDENT_16 . METHOD_11 ( ) ) ; if ( ! ( ( IDENT_16 . METHOD_12 ( ) . get ( 1 ) ) instanceof org . apache . hadoop . mapreduce . v2 . app . IDENT_17 . IDENT_18 ) ) { org . junit . Assert . fail ( STRING_4 ) ; } java . util . Map < java . lang . String , java . lang . Boolean > expected = new java . util . HashMap < java . lang . String , java . lang . Boolean > ( ) ; expected . put ( STRING_3 , true ) ; expected . put ( STRING_1 , true ) ; expected . put ( STRING_2 , true ) ; org . apache . hadoop . mapreduce . v2 . app . IDENT_17 . IDENT_18 IDENT_19 = ( ( org . apache . hadoop . mapreduce . v2 . app . IDENT_17 . IDENT_18 ) ( IDENT_16 . METHOD_12 ( ) . get ( 1 ) ) ) ; java . lang . String [ ] IDENT_20 = IDENT_19 . METHOD_13 ( ) ; for ( java . lang . String IDENT_21 : IDENT_20 ) { expected . remove ( IDENT_21 ) ; } "<AssertPlaceHolder>" ; } size ( ) { return ( ( java . lang . Integer ) ( new io . IDENT_22 . transaction . handler . METHOD_14 ( io . IDENT_22 . transaction . handler . IDENT_23 . IDENT_24 ) { @ io . IDENT_22 . metadata . IDENT_25 . Override public java . lang . Object METHOD_15 ( ) throws java . io . IOException { io . IDENT_22 . metadata . hdfs . dal . IDENT_26 IDENT_27 = ( ( io . IDENT_22 . metadata . hdfs . dal . IDENT_26 ) ( io . IDENT_22 . metadata . IDENT_28 . METHOD_16 ( io . IDENT_22 . metadata . hdfs . dal . IDENT_26 . class ) ) ) ; return IDENT_27 . METHOD_17 ( ) ; } } . METHOD_10 ( ) ) ) ; }
|
org . junit . Assert . assertEquals ( 0 , expected . size ( ) )
|
METHOD_0 ( ) { final org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . conf . Configuration ( ) ; conf . METHOD_1 ( IDENT_0 . IDENT_1 , FLOAT_0 ) ; final org . apache . hadoop . mapreduce . v2 . app . IDENT_2 . IDENT_3 . IDENT_4 IDENT_2 = new org . apache . hadoop . mapreduce . v2 . app . IDENT_2 . IDENT_3 . METHOD_2 ( conf ) ; IDENT_2 . start ( ) ; final org . apache . hadoop . yarn . event . IDENT_5 IDENT_6 = ( ( org . apache . hadoop . yarn . event . IDENT_5 ) ( IDENT_2 . METHOD_3 ( ) . METHOD_4 ( ) ) ) ; final org . apache . hadoop . yarn . server . IDENT_7 . IDENT_8 . IDENT_9 app = IDENT_2 . METHOD_5 ( INT_0 ) ; IDENT_6 . METHOD_6 ( ) ; final java . lang . String host = STRING_0 ; final org . apache . hadoop . yarn . server . IDENT_7 . IDENT_10 IDENT_11 = IDENT_2 . METHOD_7 ( java . lang . String . format ( STRING_1 , host ) , INT_1 ) ; IDENT_11 . METHOD_8 ( true ) ; IDENT_6 . METHOD_6 ( ) ; final org . apache . hadoop . yarn . api . records . IDENT_12 IDENT_13 = app . METHOD_9 ( ) . METHOD_10 ( ) ; IDENT_2 . METHOD_11 ( IDENT_13 ) ; IDENT_6 . METHOD_6 ( ) ; final org . apache . hadoop . mapreduce . v2 . api . records . IDENT_14 IDENT_15 = org . apache . hadoop . mapreduce . v2 . util . IDENT_16 . METHOD_12 ( IDENT_13 . METHOD_13 ( ) , 0 ) ; final org . apache . hadoop . mapreduce . v2 . app . job . IDENT_17 IDENT_18 = mock ( org . apache . hadoop . mapreduce . v2 . app . job . IDENT_17 . class ) ; when ( IDENT_18 . METHOD_14 ( ) ) . thenReturn ( org . apache . hadoop . mapreduce . v2 . util . IDENT_16 . METHOD_15 ( IDENT_15 , STRING_2 , STRING_3 , IDENT_19 . IDENT_20 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , STRING_4 , null , false , "" ) ) ; final org . apache . hadoop . mapreduce . v2 . app . IDENT_2 . IDENT_3 . IDENT_21 IDENT_22 = new org . apache . hadoop . mapreduce . v2 . app . IDENT_2 . IDENT_3 . METHOD_16 ( IDENT_2 , conf , IDENT_13 , IDENT_18 , new org . apache . hadoop . yarn . util . METHOD_17 ( ) ) ; final java . lang . String [ ] IDENT_23 = new java . lang . String [ ] { host } ; IDENT_22 . METHOD_18 ( METHOD_19 ( IDENT_15 , 0 , INT_2 , IDENT_23 , false , true ) ) ; IDENT_22 . METHOD_20 ( ) ; IDENT_22 . METHOD_21 ( ) ; IDENT_11 . METHOD_8 ( true ) ; IDENT_6 . METHOD_6 ( ) ; IDENT_22 . METHOD_18 ( METHOD_19 ( IDENT_15 , 1 , INT_2 , IDENT_23 , false , false ) ) ; int IDENT_24 ; for ( IDENT_24 = 0 ; IDENT_24 < 1 ; ) { IDENT_24 += IDENT_22 . METHOD_22 ( ) . size ( ) ; IDENT_11 . METHOD_8 ( true ) ; IDENT_6 . METHOD_6 ( ) ; } "<AssertPlaceHolder>" ; } METHOD_6 ( ) { while ( ( this . val ) == null ) { this . IDENT_25 . METHOD_6 ( ) ; } return this . val ; }
|
org . junit . Assert . assertEquals ( IDENT_24 , 1 )
|
METHOD_0 ( ) { org . apache . hadoop . IDENT_0 . IDENT_1 IDENT_2 = new org . apache . hadoop . IDENT_0 . METHOD_1 ( ) ; IDENT_2 . METHOD_2 ( STRING_0 , IDENT_3 . IDENT_4 ) . METHOD_3 ( 1 ) ; IDENT_2 . METHOD_2 ( STRING_1 , IDENT_3 . IDENT_4 ) . METHOD_3 ( 1 ) ; IDENT_2 . METHOD_4 ( STRING_2 , STRING_3 , 1 ) ; java . util . HashSet < java . lang . String > IDENT_5 = new java . util . HashSet < java . lang . String > ( IDENT_2 . METHOD_5 ( ) ) ; java . util . HashSet < java . lang . String > IDENT_6 = new java . util . HashSet < java . lang . String > ( ) ; IDENT_6 . add ( STRING_2 ) ; IDENT_6 . add ( STRING_4 ) ; IDENT_6 . add ( STRING_5 ) ; "<AssertPlaceHolder>" ; } add ( java . util . Collection ) { for ( io . IDENT_7 . transaction . IDENT_8 . IDENT_9 IDENT_8 : IDENT_10 ) { add ( IDENT_8 ) ; } return this ; }
|
org . junit . Assert . assertEquals ( IDENT_6 , IDENT_5 )
|
METHOD_0 ( ) { final java . lang . String IDENT_0 = STRING_0 ; org . apache . hadoop . yarn . conf . IDENT_1 conf = new org . apache . hadoop . yarn . conf . METHOD_1 ( ) ; conf . set ( IDENT_2 . IDENT_3 , IDENT_0 ) ; org . apache . hadoop . mapreduce . v2 . IDENT_4 . IDENT_5 . IDENT_6 plugin = new org . apache . hadoop . mapreduce . v2 . IDENT_4 . IDENT_5 . METHOD_2 ( ) ; plugin . METHOD_3 ( conf ) ; org . apache . hadoop . yarn . api . records . IDENT_7 id = org . apache . hadoop . yarn . api . records . IDENT_7 . newInstance ( INT_0 , 5 ) ; java . lang . String IDENT_8 = id . toString ( ) . METHOD_4 ( STRING_1 , STRING_2 ) ; java . net . URI expected = new java . net . URI ( ( ( ( STRING_3 + IDENT_0 ) + STRING_4 ) + IDENT_8 ) ) ; java . net . URI actual = plugin . METHOD_5 ( id ) ; "<AssertPlaceHolder>" ; } METHOD_5 ( org . apache . hadoop . yarn . api . records . IDENT_7 ) { java . lang . String IDENT_8 = id . toString ( ) . METHOD_4 ( STRING_1 , STRING_2 ) ; java . lang . String IDENT_9 = org . apache . hadoop . mapreduce . v2 . util . IDENT_10 . METHOD_6 ( METHOD_7 ( ) ) ; return new java . net . URI ( ( ( IDENT_9 + STRING_4 ) + IDENT_8 ) ) ; }
|
org . junit . Assert . assertEquals ( expected , actual )
|
METHOD_0 ( ) { org . apache . hadoop . mapreduce . v2 . api . records . IDENT_0 IDENT_1 = mock ( org . apache . hadoop . mapreduce . v2 . api . records . IDENT_0 . class ) ; org . apache . hadoop . mapreduce . IDENT_2 . IDENT_3 . IDENT_4 IDENT_5 = mock ( org . apache . hadoop . mapreduce . IDENT_2 . IDENT_3 . IDENT_4 . class ) ; java . util . Map < org . apache . hadoop . mapreduce . IDENT_6 , org . apache . hadoop . mapreduce . IDENT_2 . IDENT_3 . IDENT_7 > IDENT_8 = new java . util . IDENT_9 < org . apache . hadoop . mapreduce . IDENT_6 , org . apache . hadoop . mapreduce . IDENT_2 . IDENT_3 . IDENT_7 > ( ) ; org . apache . hadoop . mapreduce . IDENT_6 id = new org . apache . hadoop . mapreduce . METHOD_1 ( STRING_0 , 0 , org . apache . hadoop . mapreduce . IDENT_10 . IDENT_11 , 0 , 0 ) ; org . apache . hadoop . mapreduce . IDENT_2 . IDENT_3 . IDENT_7 info = mock ( org . apache . hadoop . mapreduce . IDENT_2 . IDENT_3 . IDENT_7 . class ) ; when ( info . METHOD_2 ( ) ) . thenReturn ( id ) ; when ( info . METHOD_3 ( ) ) . thenReturn ( INT_0 ) ; IDENT_8 . put ( id , info ) ; id = new org . apache . hadoop . mapreduce . METHOD_1 ( "1" , 0 , org . apache . hadoop . mapreduce . IDENT_10 . IDENT_11 , 1 , 1 ) ; info = mock ( org . apache . hadoop . mapreduce . IDENT_2 . IDENT_3 . IDENT_7 . class ) ; when ( info . METHOD_2 ( ) ) . thenReturn ( id ) ; when ( info . METHOD_3 ( ) ) . thenReturn ( INT_1 ) ; IDENT_8 . put ( id , info ) ; when ( IDENT_5 . METHOD_4 ( ) ) . thenReturn ( IDENT_8 ) ; org . apache . hadoop . mapreduce . v2 . IDENT_12 . IDENT_13 task = new org . apache . hadoop . mapreduce . v2 . IDENT_12 . METHOD_5 ( IDENT_1 , IDENT_5 ) ; org . apache . hadoop . mapreduce . v2 . api . records . IDENT_14 report = task . METHOD_6 ( ) ; "<AssertPlaceHolder>" ; } METHOD_3 ( ) { if ( ( METHOD_7 ( ) ) == false ) { return 0 ; } return IDENT_15 ; }
|
org . junit . Assert . assertTrue ( ( ( report . METHOD_3 ( ) ) == 10 ) )
|
METHOD_0 ( ) { org . apache . hadoop . util . IDENT_0 . METHOD_1 ( ) ; try { IDENT_1 = org . apache . hadoop . mapreduce . v2 . IDENT_2 . IDENT_3 . METHOD_2 ( new java . lang . String [ 0 ] ) ; } catch ( org . apache . hadoop . util . IDENT_0 e ) { "<AssertPlaceHolder>" ; org . apache . hadoop . util . IDENT_0 . METHOD_3 ( ) ; org . junit . Assert . fail ( ) ; } } METHOD_2 ( java . lang . String [ ] ) { java . lang . Thread . METHOD_4 ( new org . apache . hadoop . yarn . METHOD_5 ( ) ) ; org . apache . hadoop . util . StringUtils . METHOD_6 ( org . apache . hadoop . mapreduce . v2 . IDENT_2 . IDENT_3 . class , args , org . apache . hadoop . mapreduce . v2 . IDENT_2 . IDENT_3 . LOG ) ; org . apache . hadoop . mapreduce . v2 . IDENT_2 . IDENT_3 IDENT_4 = null ; try { IDENT_4 = new org . apache . hadoop . mapreduce . v2 . IDENT_2 . METHOD_7 ( ) ; org . apache . hadoop . util . IDENT_5 . get ( ) . METHOD_8 ( new org . apache . hadoop . mapreduce . v2 . IDENT_2 . METHOD_9 ( IDENT_4 ) , org . apache . hadoop . mapreduce . v2 . IDENT_2 . METHOD_7 . IDENT_6 ) ; org . apache . hadoop . yarn . conf . IDENT_7 conf = new org . apache . hadoop . yarn . conf . METHOD_10 ( new org . apache . hadoop . IDENT_8 . METHOD_11 ( ) ) ; new org . apache . hadoop . util . METHOD_12 ( conf , args ) ; IDENT_4 . init ( conf ) ; IDENT_4 . start ( ) ; } catch ( java . lang . Throwable t ) { org . apache . hadoop . mapreduce . v2 . IDENT_2 . METHOD_7 . LOG . METHOD_13 ( STRING_0 , t ) ; org . apache . hadoop . util . IDENT_0 . METHOD_14 ( ( - 1 ) , STRING_0 ) ; } return IDENT_4 ; }
|
org . junit . Assert . assertEquals ( 0 , e . status )
|
METHOD_0 ( ) { { long now = java . lang . System . currentTimeMillis ( ) ; java . lang . String [ ] IDENT_0 = new java . lang . String [ ] { "foo" , "bar" } ; final org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . conf . Configuration ( ) ; org . apache . hadoop . fs . IDENT_1 [ ] fs = new org . apache . hadoop . fs . IDENT_1 [ IDENT_0 . length ] ; for ( int i = 0 ; i < ( IDENT_0 . length ) ; i ++ ) { org . apache . hadoop . security . IDENT_2 IDENT_3 = org . apache . hadoop . security . IDENT_2 . METHOD_1 ( IDENT_0 [ i ] ) ; fs [ i ] = IDENT_3 . METHOD_2 ( new java . security . IDENT_4 < org . apache . hadoop . fs . IDENT_1 > ( ) { public org . apache . hadoop . fs . IDENT_1 run ( ) throws java . io . IOException { return org . apache . hadoop . fs . IDENT_1 . get ( conf ) ; } } ) ; for ( int j = 0 ; j < i ; j ++ ) { "<AssertPlaceHolder>" ; } } org . apache . hadoop . fs . IDENT_1 . METHOD_3 ( ) ; } { try { org . apache . hadoop . fs . IDENT_5 . METHOD_4 ( IDENT_6 . IDENT_7 ) ; } catch ( org . apache . hadoop . fs . java . net IDENT_8 ) { org . apache . hadoop . fs . IDENT_5 . LOG . warn ( ( ( STRING_0 + ( org . apache . hadoop . hdfs . server . IDENT_9 . IDENT_6 . IDENT_7 ) ) + STRING_1 ) , IDENT_8 ) ; } org . apache . hadoop . fs . IDENT_5 . METHOD_4 ( 0 ) ; } } get ( org . apache . hadoop . conf . Configuration ) { return org . apache . hadoop . fs . IDENT_1 . get ( org . apache . hadoop . fs . IDENT_1 . METHOD_5 ( conf ) , conf ) ; }
|
org . junit . Assert . assertFalse ( ( ( fs [ j ] ) == ( fs [ i ] ) ) )
|
METHOD_0 ( ) { org . apache . hadoop . fs . IDENT_0 . IDENT_1 IDENT_2 = METHOD_1 ( false ) ; org . apache . hadoop . fs . IDENT_0 . IDENT_3 IDENT_4 = new org . apache . hadoop . fs . IDENT_0 . METHOD_2 ( IDENT_2 , org . apache . hadoop . fs . IDENT_0 . IDENT_5 . IDENT_6 ) ; int expected = org . apache . hadoop . fs . IDENT_0 . Constants . IDENT_7 . values ( ) . length ; org . apache . hadoop . fs . IDENT_0 . Operation op = null ; java . util . Set < java . lang . String > types = new java . util . HashSet < java . lang . String > ( ) ; org . apache . hadoop . fs . IDENT_8 fs = org . apache . hadoop . fs . IDENT_8 . get ( IDENT_2 . METHOD_3 ( ) ) ; while ( true ) { op = IDENT_4 . select ( 1 , 1 ) ; if ( op == null ) { break ; } op . run ( fs ) ; types . add ( op . getType ( ) ) ; } "<AssertPlaceHolder>" ; } size ( ) { return ( ( java . lang . Integer ) ( new io . IDENT_9 . transaction . handler . METHOD_4 ( io . IDENT_9 . transaction . handler . IDENT_10 . IDENT_11 ) { @ io . IDENT_9 . metadata . IDENT_12 . Override public java . lang . Object METHOD_5 ( ) throws java . io . IOException { io . IDENT_9 . metadata . hdfs . dal . IDENT_13 IDENT_14 = ( ( io . IDENT_9 . metadata . hdfs . dal . IDENT_13 ) ( io . IDENT_9 . metadata . IDENT_15 . METHOD_6 ( io . IDENT_9 . metadata . hdfs . dal . IDENT_13 . class ) ) ) ; return IDENT_14 . METHOD_7 ( ) ; } } . METHOD_8 ( ) ) ) ; }
|
org . junit . Assert . assertEquals ( types . size ( ) , expected )
|
METHOD_0 ( ) { "<AssertPlaceHolder>" ; } METHOD_1 ( ) { return configuration ; }
|
org . junit . Assert . assertNotNull ( binding . METHOD_1 ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . mapreduce . v2 . api . IDENT_0 IDENT_1 = mock ( org . apache . hadoop . mapreduce . v2 . api . IDENT_0 . class ) ; when ( IDENT_1 . METHOD_1 ( METHOD_2 ( ) ) ) . METHOD_3 ( new java . io . IOException ( STRING_0 ) ) ; org . apache . hadoop . IDENT_2 . IDENT_3 IDENT_4 = mock ( org . apache . hadoop . IDENT_2 . IDENT_3 . class ) ; when ( IDENT_4 . METHOD_4 ( org . apache . hadoop . mapreduce . IDENT_5 . METHOD_5 ( IDENT_6 ) . METHOD_6 ( ) ) ) . thenReturn ( null ) ; org . apache . hadoop . IDENT_2 . IDENT_7 IDENT_8 = METHOD_7 ( IDENT_1 , IDENT_4 ) ; try { IDENT_8 . METHOD_8 ( IDENT_6 ) ; org . junit . Assert . fail ( STRING_1 ) ; } catch ( java . io . IOException e ) { "<AssertPlaceHolder>" ; } } getMessage ( ) { return builder . getMessage ( ) ; }
|
org . junit . Assert . assertTrue ( e . getMessage ( ) . contains ( STRING_0 ) )
|
METHOD_0 ( ) { final int IDENT_0 = 100 ; org . apache . hadoop . io . IDENT_1 IDENT_2 = new org . apache . hadoop . io . METHOD_1 ( ( IDENT_0 + 4 ) ) ; org . apache . hadoop . IDENT_3 . IDENT_4 IDENT_5 = new org . apache . hadoop . IDENT_3 . METHOD_2 ( IDENT_2 ) ; for ( int i = 0 ; i < IDENT_0 ; ++ i ) { IDENT_5 . write ( i ) ; } IDENT_5 . close ( ) ; org . apache . hadoop . io . IDENT_6 IDENT_7 = new org . apache . hadoop . io . METHOD_3 ( ) ; IDENT_7 . METHOD_4 ( IDENT_2 . METHOD_5 ( ) , ( IDENT_0 + 4 ) ) ; org . apache . hadoop . IDENT_3 . IDENT_8 IDENT_9 = new org . apache . hadoop . IDENT_3 . METHOD_6 ( IDENT_7 , INT_0 , new org . apache . hadoop . conf . Configuration ( ) ) ; for ( int i = 0 ; i < IDENT_0 ; ++ i ) { "<AssertPlaceHolder>" ; } IDENT_9 . close ( ) ; } read ( ) { if ( ( IDENT_10 ) == null ) { IDENT_10 = new byte [ 1 ] ; } int ret = read ( IDENT_10 , 0 , 1 ) ; return ret <= 0 ? - 1 : ( IDENT_10 [ 0 ] ) & 255 ; }
|
org . junit . Assert . assertEquals ( i , IDENT_9 . read ( ) )
|
METHOD_0 ( ) { java . io . IDENT_0 IDENT_1 = org . mockito . Mockito . mock ( java . io . IDENT_0 . class ) ; org . apache . hadoop . IDENT_2 . IDENT_3 IDENT_4 = new org . apache . hadoop . IDENT_2 . METHOD_1 ( IDENT_1 ) ; org . mockito . Mockito . METHOD_2 ( new java . io . IOException ( STRING_0 ) ) . when ( IDENT_1 ) . METHOD_3 ( ) ; try { IDENT_4 . close ( ) ; org . junit . Assert . fail ( STRING_1 ) ; } catch ( java . io . IOException IDENT_5 ) { "<AssertPlaceHolder>" ; } org . mockito . Mockito . verify ( IDENT_1 ) . close ( ) ; } getMessage ( ) { java . lang . String msg = super . getMessage ( ) ; if ( msg == null ) { return ( ( ( ( STRING_2 + ( ( IDENT_6 ) == null ? "" : STRING_3 + ( IDENT_6 ) ) ) + STRING_4 ) + ( IDENT_7 ) ) + STRING_5 ) + ( count ) ; } else { return msg ; } }
|
org . junit . Assert . assertEquals ( STRING_0 , IDENT_5 . getMessage ( ) )
|
METHOD_0 ( ) { org . apache . camel . component . IDENT_0 . IDENT_1 configuration = new org . apache . camel . component . IDENT_0 . METHOD_1 ( ) ; binding = new org . apache . camel . component . IDENT_0 . METHOD_2 ( configuration ) ; "<AssertPlaceHolder>" ; } METHOD_3 ( ) { return configuration ; }
|
org . junit . Assert . assertSame ( configuration , binding . METHOD_3 ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . IDENT_0 . IDENT_1 . METHOD_1 ( ) ; org . apache . hadoop . fs . Path IDENT_2 = METHOD_2 ( new org . apache . hadoop . fs . Path ( org . apache . hadoop . IDENT_0 . IDENT_1 . IDENT_3 , STRING_0 ) ) ; org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . conf . Configuration ( ) ; conf . set ( IDENT_4 . IDENT_5 , STRING_1 ) ; conf . set ( IDENT_6 . IDENT_7 , STRING_2 3 ) ; conf . METHOD_3 ( IDENT_8 . IDENT_9 , 1 ) ; conf . set ( IDENT_8 . IDENT_10 , STRING_3 ) ; final java . lang . String [ ] args = new java . lang . String [ ] { STRING_4 , STRING_2 3 , STRING_2 2 , IDENT_2 . toString ( ) , STRING_2 1 , "1" , STRING_5 , "1" , STRING_6 , "1" , STRING_7 , "1" } ; int res = - 1 ; try { res = org . apache . hadoop . util . IDENT_11 . run ( conf , new org . apache . hadoop . mapreduce . METHOD_4 ( ) , args ) ; } catch ( java . lang . Exception e ) { System . out . println ( ( STRING_8 + ( e . METHOD_5 ( ) ) ) ) ; e . printStackTrace ( System . out ) ; org . junit . Assert . fail ( STRING_2 0 ) ; } "<AssertPlaceHolder>" ; } fail ( java . lang . String ) { System . err . println ( message ) ; System . err . println ( STRING_9 ) ; throw new java . lang . IllegalArgumentException ( message ) ; }
|
org . junit . Assert . assertEquals ( STRING_2 , 0 , res )
|
METHOD_0 ( ) { org . apache . hadoop . IDENT_0 . IDENT_1 job = new org . apache . hadoop . IDENT_0 . METHOD_1 ( ) ; METHOD_2 ( job ) ; org . apache . hadoop . IDENT_0 . IDENT_2 IDENT_3 = new org . apache . hadoop . IDENT_0 . METHOD_3 ( job , org . apache . hadoop . IDENT_0 . IDENT_4 . IDENT_5 . METHOD_4 ( ) ) ; org . apache . hadoop . IDENT_0 . IDENT_6 IDENT_7 = new org . apache . hadoop . IDENT_0 . METHOD_5 ( job , org . apache . hadoop . IDENT_0 . IDENT_4 . IDENT_5 ) ; org . apache . hadoop . IDENT_0 . IDENT_8 IDENT_9 = new org . apache . hadoop . IDENT_0 . METHOD_6 ( ) ; org . apache . hadoop . IDENT_0 . IDENT_10 . METHOD_7 ( job , IDENT_9 . METHOD_8 ( IDENT_7 ) ) ; IDENT_9 . METHOD_9 ( IDENT_3 ) ; IDENT_9 . METHOD_10 ( IDENT_7 ) ; java . lang . String file = STRING_0 ; org . apache . hadoop . IDENT_0 . IDENT_11 IDENT_12 = IDENT_11 . IDENT_13 ; org . apache . hadoop . fs . IDENT_14 IDENT_15 = org . apache . hadoop . fs . IDENT_14 . METHOD_11 ( job ) ; org . apache . hadoop . IDENT_0 . IDENT_16 IDENT_17 = new org . apache . hadoop . IDENT_0 . METHOD_12 ( ) ; org . apache . hadoop . IDENT_0 . IDENT_18 IDENT_19 = IDENT_17 . METHOD_13 ( IDENT_15 , job , file , IDENT_12 ) ; METHOD_14 ( IDENT_19 , IDENT_12 ) ; IDENT_9 . METHOD_15 ( IDENT_7 ) ; IDENT_9 . METHOD_16 ( IDENT_3 ) ; java . io . File IDENT_20 = new java . io . File ( new org . apache . hadoop . fs . Path ( org . apache . hadoop . IDENT_0 . IDENT_4 . IDENT_21 , file ) . toString ( ) ) ; java . lang . StringBuffer IDENT_22 = new java . lang . StringBuffer ( ) ; IDENT_22 . append ( IDENT_23 ) . append ( STRING_1 ) . append ( IDENT_24 ) . append ( "\n" ) ; IDENT_22 . append ( IDENT_24 ) . append ( "\n" ) ; IDENT_22 . append ( IDENT_25 ) . append ( "\n" ) ; IDENT_22 . append ( IDENT_26 ) . append ( "\n" ) ; IDENT_22 . append ( IDENT_23 ) . append ( "\n" ) ; IDENT_22 . append ( IDENT_26 ) . append ( STRING_1 ) . append ( IDENT_25 ) . append ( "\n" ) ; java . lang . String output = org . apache . hadoop . IDENT_0 . IDENT_27 . METHOD_17 ( IDENT_20 ) ; "<AssertPlaceHolder>" ; } toString ( ) { if ( ( json ) == null ) { return STRING_2 + ( id ) ; } else { return json . toString ( ) ; } }
|
org . junit . Assert . assertEquals ( output , IDENT_22 . toString ( ) )
|
METHOD_0 ( ) { org . IDENT_0 . bean . IDENT_1 IDENT_2 = new org . IDENT_0 . bean . METHOD_1 ( ) ; IDENT_2 . METHOD_2 ( ) ; IDENT_2 . METHOD_3 ( new org . IDENT_0 . bean . IDENT_3 . METHOD_4 ( ( ( short ) ( INT_0 ) ) , ( ( short ) ( 0 ) ) ) ) ; try { org . apache . camel . component . IDENT_4 . IDENT_5 IDENT_6 = binding . METHOD_5 ( IDENT_7 , IDENT_2 ) ; java . util . Map < java . lang . METHOD_4 , java . lang . Object > IDENT_8 = IDENT_6 . METHOD_6 ( IDENT_9 . IDENT_10 , java . util . Map . class ) ; "<AssertPlaceHolder>" ; } catch ( java . lang . Exception e ) { org . junit . Assert . fail ( STRING_0 ) ; } } valueOf ( java . lang . String ) { java . lang . String type = org . apache . camel . util . IDENT_11 . METHOD_7 ( IDENT_12 , STRING_1 ) ; if ( ( type == null ) || ( STRING_2 . equals ( type ) ) ) { return org . apache . camel . util . IDENT_11 . METHOD_8 ( IDENT_12 , STRING_1 , org . apache . camel . component . IDENT_13 . IDENT_14 . IDENT_15 :: new ) ; } else if ( STRING_3 . equals ( type ) ) { return org . apache . camel . util . IDENT_11 . METHOD_8 ( IDENT_12 , STRING_1 , org . apache . camel . component . IDENT_13 . IDENT_14 . IDENT_16 :: new ) ; } return java . util . Optional . empty ( ) ; }
|
org . junit . Assert . assertEquals ( java . lang . METHOD_4 . valueOf ( ( ( short ) ( 0 ) ) ) , IDENT_8 . get ( java . lang . METHOD_4 . valueOf ( ( ( short ) ( INT_0 ) ) ) ) )
|
METHOD_0 ( ) { org . apache . hadoop . IDENT_0 . IDENT_1 . LOG . info ( STRING_0 ) ; org . apache . hadoop . IDENT_0 . IDENT_2 . METHOD_1 ( org . apache . hadoop . IDENT_0 . IDENT_1 . job , IDENT_2 . IDENT_3 . class ) ; IDENT_2 . IDENT_3 . METHOD_2 ( org . apache . hadoop . IDENT_0 . IDENT_1 . job , 1000 ) ; org . apache . hadoop . IDENT_0 . IDENT_1 . fs . delete ( org . apache . hadoop . IDENT_0 . IDENT_1 . IDENT_4 , true ) ; for ( int length = 0 ; length < ( org . apache . hadoop . IDENT_0 . IDENT_1 . IDENT_5 ) ; length += ( org . apache . hadoop . IDENT_0 . IDENT_1 . random . METHOD_3 ( ( ( org . apache . hadoop . IDENT_0 . IDENT_1 . IDENT_5 ) / 10 ) ) ) + 1 ) { org . apache . hadoop . IDENT_0 . IDENT_1 . LOG . info ( ( STRING_1 + length ) ) ; org . apache . hadoop . IDENT_0 . IDENT_1 . METHOD_4 ( length ) ; int count = METHOD_5 ( 1 ) ; org . apache . hadoop . IDENT_0 . IDENT_1 . LOG . info ( ( ( STRING_2 + count ) + STRING_3 ) ) ; int IDENT_6 = length / 1000 ; if ( ( IDENT_6 * 1000 ) != length ) IDENT_6 ++ ; "<AssertPlaceHolder>" ; } org . apache . hadoop . IDENT_0 . IDENT_1 . fs . delete ( org . apache . hadoop . IDENT_0 . IDENT_1 . IDENT_4 , true ) ; } info ( java . lang . String ) { if ( ( LOG ) != null ) { LOG . info ( msg ) ; } else if ( ( LOGGER ) != null ) { LOGGER . info ( msg ) ; } }
|
org . junit . Assert . assertEquals ( count , IDENT_6 )
|
METHOD_0 ( ) { java . lang . String IDENT_0 = null ; org . apache . hadoop . hdfs . IDENT_1 IDENT_2 = null ; org . apache . hadoop . IDENT_3 . IDENT_4 IDENT_5 = null ; org . apache . hadoop . fs . IDENT_6 IDENT_7 = null ; try { final int IDENT_8 = 4 ; final int IDENT_9 = INT_0 ; org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . conf . Configuration ( ) ; IDENT_2 = new org . apache . hadoop . hdfs . IDENT_1 . Builder ( conf ) . build ( ) ; IDENT_7 = IDENT_2 . METHOD_1 ( ) ; IDENT_0 = IDENT_7 . METHOD_2 ( ) . toString ( ) ; IDENT_5 = new org . apache . hadoop . IDENT_3 . METHOD_3 ( IDENT_8 , IDENT_0 , 2 , false ) ; org . apache . hadoop . IDENT_3 . IDENT_10 IDENT_11 = new org . apache . hadoop . IDENT_3 . METHOD_4 ( ) ; boolean result ; result = org . apache . hadoop . IDENT_3 . IDENT_12 . METHOD_5 ( IDENT_7 . METHOD_2 ( ) , IDENT_11 , 3 , 1 ) ; "<AssertPlaceHolder>" ; } finally { if ( IDENT_2 != null ) { IDENT_2 . METHOD_6 ( ) ; } if ( IDENT_5 != null ) { IDENT_5 . METHOD_6 ( ) ; } } } METHOD_2 ( ) { return file . METHOD_7 ( ) ; }
|
org . junit . Assert . assertTrue ( result )
|
METHOD_0 ( ) { org . apache . hadoop . IDENT_0 . IDENT_1 job = new org . apache . hadoop . IDENT_0 . METHOD_1 ( ) ; job . set ( IDENT_2 . IDENT_3 , org . apache . hadoop . IDENT_0 . IDENT_4 . IDENT_5 ) ; org . apache . hadoop . IDENT_0 . IDENT_6 . METHOD_2 ( job , org . apache . hadoop . IDENT_0 . IDENT_4 . IDENT_7 . METHOD_3 ( ) . METHOD_3 ( ) ) ; org . apache . hadoop . IDENT_0 . IDENT_6 . METHOD_4 ( job , org . apache . hadoop . IDENT_0 . IDENT_4 . IDENT_7 ) ; org . apache . hadoop . fs . IDENT_8 fs = org . apache . hadoop . IDENT_0 . IDENT_4 . IDENT_7 . METHOD_5 ( job ) ; if ( ! ( fs . METHOD_6 ( org . apache . hadoop . IDENT_0 . IDENT_4 . IDENT_7 ) ) ) { org . junit . Assert . fail ( STRING_0 ) ; } java . lang . String file = STRING_1 ; org . apache . hadoop . IDENT_0 . IDENT_9 IDENT_10 = IDENT_9 . IDENT_11 ; org . apache . hadoop . IDENT_0 . IDENT_12 < java . lang . Object , java . lang . Object > IDENT_13 = new org . apache . hadoop . IDENT_0 . IDENT_12 < java . lang . Object , java . lang . Object > ( ) ; org . apache . hadoop . IDENT_0 . IDENT_14 < java . lang . Object , java . lang . Object > IDENT_15 = IDENT_13 . METHOD_7 ( org . apache . hadoop . IDENT_0 . IDENT_4 . IDENT_16 , job , file , IDENT_10 ) ; org . apache . hadoop . io . IDENT_17 IDENT_18 = new org . apache . hadoop . io . METHOD_8 ( STRING_2 ) ; org . apache . hadoop . io . METHOD_8 IDENT_19 = new org . apache . hadoop . io . METHOD_8 ( STRING_3 ) ; org . apache . hadoop . io . METHOD_8 IDENT_20 = new org . apache . hadoop . io . METHOD_8 ( STRING_4 ) ; org . apache . hadoop . io . METHOD_8 IDENT_21 = new org . apache . hadoop . io . METHOD_8 ( STRING_5 ) ; org . apache . hadoop . io . IDENT_22 IDENT_23 = org . apache . hadoop . io . IDENT_22 . get ( ) ; try { IDENT_15 . write ( IDENT_18 , IDENT_20 ) ; IDENT_15 . write ( null , IDENT_23 ) ; IDENT_15 . write ( null , IDENT_20 ) ; IDENT_15 . write ( IDENT_23 , IDENT_21 ) ; IDENT_15 . write ( IDENT_19 , IDENT_23 ) ; IDENT_15 . write ( IDENT_18 , null ) ; IDENT_15 . write ( null , null ) ; IDENT_15 . write ( IDENT_19 , IDENT_21 ) ; } finally { IDENT_15 . close ( IDENT_10 ) ; } java . io . File IDENT_24 = new java . io . File ( new org . apache . hadoop . fs . Path ( org . apache . hadoop . IDENT_0 . IDENT_4 . IDENT_7 , file ) . toString ( ) ) ; java . lang . StringBuffer IDENT_25 = new java . lang . StringBuffer ( ) ; IDENT_25 . append ( IDENT_18 ) . append ( STRING_6 ) . append ( IDENT_20 ) . append ( "\n" ) ; IDENT_25 . append ( IDENT_20 ) . append ( "\n" ) ; IDENT_25 . append ( IDENT_21 ) . append ( "\n" ) ; IDENT_25 . append ( IDENT_19 ) . append ( "\n" ) ; IDENT_25 . append ( IDENT_18 ) . append ( "\n" ) ; IDENT_25 . append ( IDENT_19 ) . append ( STRING_6 ) . append ( IDENT_21 ) . append ( "\n" ) ; java . lang . String output = org . apache . hadoop . IDENT_0 . IDENT_26 . METHOD_9 ( IDENT_24 ) ; "<AssertPlaceHolder>" ; } toString ( ) { if ( ( json ) == null ) { return STRING_7 + ( id ) ; } else { return json . toString ( ) ; } }
|
org . junit . Assert . assertEquals ( IDENT_25 . toString ( ) , output )
|
METHOD_0 ( ) { org . apache . hadoop . IDENT_0 . IDENT_1 job = new org . apache . hadoop . IDENT_0 . METHOD_1 ( ) ; java . lang . String IDENT_2 = STRING_0 ; job . set ( STRING_1 , IDENT_2 ) ; job . set ( IDENT_3 . IDENT_4 , org . apache . hadoop . IDENT_0 . IDENT_5 . IDENT_6 ) ; org . apache . hadoop . IDENT_0 . IDENT_7 . METHOD_2 ( job , org . apache . hadoop . IDENT_0 . IDENT_5 . IDENT_8 . METHOD_3 ( ) . METHOD_3 ( ) ) ; org . apache . hadoop . IDENT_0 . IDENT_7 . METHOD_4 ( job , org . apache . hadoop . IDENT_0 . IDENT_5 . IDENT_8 ) ; org . apache . hadoop . fs . IDENT_9 fs = org . apache . hadoop . IDENT_0 . IDENT_5 . IDENT_8 . METHOD_5 ( job ) ; if ( ! ( fs . METHOD_6 ( org . apache . hadoop . IDENT_0 . IDENT_5 . IDENT_8 ) ) ) { org . junit . Assert . fail ( STRING_2 ) ; } java . lang . String file = STRING_3 ; org . apache . hadoop . IDENT_0 . IDENT_10 IDENT_11 = IDENT_10 . IDENT_12 ; org . apache . hadoop . IDENT_0 . IDENT_13 < java . lang . Object , java . lang . Object > IDENT_14 = new org . apache . hadoop . IDENT_0 . IDENT_13 < java . lang . Object , java . lang . Object > ( ) ; org . apache . hadoop . IDENT_0 . IDENT_15 < java . lang . Object , java . lang . Object > IDENT_16 = IDENT_14 . METHOD_7 ( org . apache . hadoop . IDENT_0 . IDENT_5 . IDENT_17 , job , file , IDENT_11 ) ; org . apache . hadoop . io . IDENT_18 IDENT_19 = new org . apache . hadoop . io . METHOD_8 ( STRING_4 ) ; org . apache . hadoop . io . METHOD_8 IDENT_20 = new org . apache . hadoop . io . METHOD_8 ( STRING_5 ) ; org . apache . hadoop . io . METHOD_8 IDENT_21 = new org . apache . hadoop . io . METHOD_8 ( STRING_6 ) ; org . apache . hadoop . io . METHOD_8 IDENT_22 = new org . apache . hadoop . io . METHOD_8 ( STRING_7 ) ; org . apache . hadoop . io . IDENT_23 IDENT_24 = org . apache . hadoop . io . IDENT_23 . get ( ) ; try { IDENT_16 . write ( IDENT_19 , IDENT_21 ) ; IDENT_16 . write ( null , IDENT_24 ) ; IDENT_16 . write ( null , IDENT_21 ) ; IDENT_16 . write ( IDENT_24 , IDENT_22 ) ; IDENT_16 . write ( IDENT_20 , IDENT_24 ) ; IDENT_16 . write ( IDENT_19 , null ) ; IDENT_16 . write ( null , null ) ; IDENT_16 . write ( IDENT_20 , IDENT_22 ) ; } finally { IDENT_16 . close ( IDENT_11 ) ; } java . io . File IDENT_25 = new java . io . File ( new org . apache . hadoop . fs . Path ( org . apache . hadoop . IDENT_0 . IDENT_5 . IDENT_8 , file ) . toString ( ) ) ; java . lang . StringBuffer IDENT_26 = new java . lang . StringBuffer ( ) ; IDENT_26 . append ( IDENT_19 ) . append ( IDENT_2 ) . append ( IDENT_21 ) . append ( "\n" ) ; IDENT_26 . append ( IDENT_21 ) . append ( "\n" ) ; IDENT_26 . append ( IDENT_22 ) . append ( "\n" ) ; IDENT_26 . append ( IDENT_20 ) . append ( "\n" ) ; IDENT_26 . append ( IDENT_19 ) . append ( "\n" ) ; IDENT_26 . append ( IDENT_20 ) . append ( IDENT_2 ) . append ( IDENT_22 ) . append ( "\n" ) ; java . lang . String output = org . apache . hadoop . IDENT_0 . IDENT_27 . METHOD_9 ( IDENT_25 ) ; "<AssertPlaceHolder>" ; } toString ( ) { if ( ( json ) == null ) { return STRING_8 + ( id ) ; } else { return json . toString ( ) ; } }
|
org . junit . Assert . assertEquals ( IDENT_26 . toString ( ) , output )
|
METHOD_0 ( ) { org . apache . camel . component . IDENT_0 . IDENT_1 configuration = new org . apache . camel . component . IDENT_0 . METHOD_1 ( ) ; binding . METHOD_2 ( configuration ) ; "<AssertPlaceHolder>" ; } METHOD_3 ( ) { return configuration ; }
|
org . junit . Assert . assertSame ( configuration , binding . METHOD_3 ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . mapreduce . lib . input . IDENT_0 . LOG . info ( STRING_0 ) ; org . apache . hadoop . mapreduce . lib . input . IDENT_1 . METHOD_1 ( org . apache . hadoop . mapreduce . lib . input . IDENT_0 . job , IDENT_1 . IDENT_2 . class ) ; IDENT_1 . IDENT_2 . METHOD_2 ( org . apache . hadoop . mapreduce . lib . input . IDENT_0 . job . METHOD_3 ( ) , 1000 ) ; org . apache . hadoop . mapreduce . lib . input . IDENT_0 . fs . delete ( org . apache . hadoop . mapreduce . lib . input . IDENT_0 . IDENT_3 , true ) ; for ( int length = 0 ; length < ( org . apache . hadoop . mapreduce . lib . input . IDENT_0 . IDENT_4 ) ; length += ( org . apache . hadoop . mapreduce . lib . input . IDENT_0 . random . METHOD_4 ( ( ( org . apache . hadoop . mapreduce . lib . input . IDENT_0 . IDENT_4 ) / 10 ) ) ) + 1 ) { org . apache . hadoop . mapreduce . lib . input . IDENT_0 . LOG . info ( ( STRING_1 + length ) ) ; org . apache . hadoop . mapreduce . lib . input . IDENT_0 . METHOD_5 ( length ) ; int count = METHOD_6 ( 1 ) ; org . apache . hadoop . mapreduce . lib . input . IDENT_0 . LOG . info ( ( ( STRING_2 + count ) + STRING_3 ) ) ; int IDENT_5 = length / 1000 ; if ( ( IDENT_5 * 1000 ) != length ) IDENT_5 ++ ; "<AssertPlaceHolder>" ; } org . apache . hadoop . mapreduce . lib . input . IDENT_0 . fs . delete ( org . apache . hadoop . mapreduce . lib . input . IDENT_0 . IDENT_3 , true ) ; } info ( java . lang . String ) { if ( ( LOG ) != null ) { LOG . info ( msg ) ; } else if ( ( LOGGER ) != null ) { LOGGER . info ( msg ) ; } }
|
org . junit . Assert . assertEquals ( count , IDENT_5 )
|
METHOD_0 ( ) { org . apache . camel . component . IDENT_0 . IDENT_1 configuration = new org . apache . camel . component . IDENT_0 . METHOD_1 ( ) ; component = new org . apache . camel . component . IDENT_0 . METHOD_2 ( configuration ) ; "<AssertPlaceHolder>" ; } METHOD_3 ( ) { return configuration ; }
|
org . junit . Assert . assertSame ( configuration , component . METHOD_3 ( ) )
|
METHOD_0 ( ) { java . lang . String [ ] args = new java . lang . String [ 2 ] ; args [ 0 ] = org . apache . hadoop . examples . IDENT_0 . IDENT_1 ; args [ 1 ] = org . apache . hadoop . examples . IDENT_0 . IDENT_2 ; org . apache . hadoop . examples . IDENT_3 IDENT_4 = new org . apache . hadoop . examples . METHOD_1 ( ) ; org . apache . hadoop . util . IDENT_5 . run ( new org . apache . hadoop . conf . Configuration ( ) , IDENT_4 , args ) ; double IDENT_6 = IDENT_4 . METHOD_2 ( ) ; org . apache . hadoop . examples . IDENT_0 . IDENT_7 IDENT_8 = new org . apache . hadoop . examples . IDENT_0 . METHOD_3 ( ) ; "<AssertPlaceHolder>" ; } read ( java . nio . ByteBuffer ) { int IDENT_9 = 0 ; if ( IDENT_10 ) { if ( IDENT_11 . METHOD_4 ( ) ) { int IDENT_12 = java . lang . Math . min ( buf . METHOD_5 ( ) , IDENT_11 . METHOD_5 ( ) ) ; METHOD_6 ( IDENT_11 , buf , IDENT_12 ) ; IDENT_9 += IDENT_12 ; } if ( ( ( buf . METHOD_5 ( ) ) >= ( IDENT_13 ) ) && ( ( IDENT_14 ) == 0 ) ) { int len = ( buf . METHOD_5 ( ) ) - ( ( buf . METHOD_5 ( ) ) OP_0 ( IDENT_13 ) ) ; len = java . lang . Math . min ( len , IDENT_11 . METHOD_7 ( ) ) ; int IDENT_15 = buf . METHOD_8 ( ) ; buf . METHOD_8 ( ( ( buf . position ( ) ) + len ) ) ; int IDENT_16 = 0 ; try { IDENT_16 = METHOD_9 ( buf ) ; } finally { buf . METHOD_8 ( IDENT_15 ) ; } if ( IDENT_16 == ( - 1 ) ) { return IDENT_9 ; } else { IDENT_9 += IDENT_16 ; buf . position ( ( ( buf . position ( ) ) + IDENT_16 ) ) ; } } if ( ( ( ( buf . METHOD_5 ( ) ) > 0 ) && ( ( buf . METHOD_5 ( ) ) < ( IDENT_13 ) ) ) || ( ( IDENT_14 ) > 0 ) ) { int IDENT_17 = java . lang . Math . min ( buf . METHOD_5 ( ) , ( ( IDENT_13 ) - ( IDENT_14 ) ) ) ; int IDENT_16 = METHOD_10 ( IDENT_17 ) ; if ( IDENT_16 == ( - 1 ) ) { return IDENT_9 ; } else { int IDENT_12 = java . lang . Math . min ( IDENT_16 , buf . METHOD_5 ( ) ) ; METHOD_6 ( IDENT_11 , buf , IDENT_12 ) ; IDENT_9 += IDENT_12 ; } } } else { IDENT_9 = METHOD_9 ( buf ) ; if ( IDENT_9 > 0 ) { buf . position ( ( ( buf . position ( ) ) + IDENT_9 ) ) ; } } return IDENT_9 ; }
|
org . junit . Assert . assertEquals ( IDENT_6 , IDENT_8 . read ( org . apache . hadoop . examples . IDENT_0 . IDENT_1 ) , 0.0 )
|
METHOD_0 ( ) { org . apache . camel . IDENT_0 context = new org . apache . camel . impl . METHOD_1 ( ) ; component = new org . apache . camel . component . IDENT_1 . METHOD_2 ( context ) ; "<AssertPlaceHolder>" ; } METHOD_3 ( ) { return IDENT_2 ; }
|
org . junit . Assert . assertSame ( context , component . METHOD_3 ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . fs . IDENT_0 . IDENT_1 ex = ( ( org . apache . hadoop . fs . IDENT_0 . IDENT_1 ) ( org . apache . hadoop . fs . IDENT_0 . IDENT_2 . METHOD_1 ( org . apache . hadoop . fs . IDENT_0 . IDENT_1 . class , org . apache . hadoop . fs . IDENT_0 . IDENT_2 . METHOD_2 ( INT_0 ) ) ) ) ; "<AssertPlaceHolder>" ; } getStatusCode ( ) { return IDENT_3 ; }
|
org . junit . Assert . assertEquals ( INT_0 , ex . getStatusCode ( ) )
|
METHOD_0 ( ) { com . amazonaws . IDENT_0 IDENT_1 = new com . amazonaws . METHOD_1 ( STRING_0 ) ; IDENT_1 . METHOD_2 ( INT_0 ) ; org . apache . hadoop . fs . IDENT_2 . IDENT_3 ex = ( ( org . apache . hadoop . fs . IDENT_2 . IDENT_3 ) ( org . apache . hadoop . fs . IDENT_2 . IDENT_4 . METHOD_3 ( org . apache . hadoop . fs . IDENT_2 . IDENT_3 . class , IDENT_1 ) ) ) ; "<AssertPlaceHolder>" ; } getStatusCode ( ) { return IDENT_5 ; }
|
org . junit . Assert . assertEquals ( INT_0 , ex . getStatusCode ( ) )
|
METHOD_0 ( ) { final org . apache . hadoop . tools . IDENT_0 options = new org . apache . hadoop . tools . METHOD_1 ( new org . apache . hadoop . fs . Path ( STRING_0 ) , new org . apache . hadoop . fs . Path ( STRING_1 ) ) ; "<AssertPlaceHolder>" ; } METHOD_2 ( ) { return IDENT_1 ; }
|
org . junit . Assert . assertEquals ( new org . apache . hadoop . fs . Path ( STRING_1 ) , options . METHOD_2 ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = org . apache . hadoop . tools . IDENT_0 . METHOD_1 ( ) ; org . apache . hadoop . fs . Path IDENT_1 = org . apache . hadoop . mapreduce . IDENT_2 . METHOD_2 ( new org . apache . hadoop . mapreduce . METHOD_3 ( conf ) , conf ) ; IDENT_1 . METHOD_4 ( conf ) . METHOD_5 ( IDENT_1 ) ; org . apache . hadoop . fs . Path IDENT_3 = METHOD_6 ( STRING_0 ) ; org . apache . hadoop . fs . Path target = METHOD_6 ( STRING_1 ) ; org . apache . hadoop . tools . IDENT_4 IDENT_5 = new org . apache . hadoop . tools . METHOD_7 ( conf , null ) ; java . lang . String [ ] IDENT_6 = new java . lang . String [ ] { IDENT_3 . toString ( ) , target . toString ( ) } ; IDENT_5 . run ( IDENT_6 ) ; "<AssertPlaceHolder>" ; } exists ( org . apache . hadoop . fs . Path ) { METHOD_8 ( org . apache . hadoop . fs . IDENT_7 . IDENT_8 ) ; return super . exists ( f ) ; }
|
org . junit . Assert . assertTrue ( org . apache . hadoop . tools . IDENT_0 . fs . exists ( target ) )
|
METHOD_0 ( ) { org . apache . camel . component . IDENT_0 . IDENT_1 configuration = new org . apache . camel . component . IDENT_0 . METHOD_1 ( ) ; component . METHOD_2 ( configuration ) ; "<AssertPlaceHolder>" ; } METHOD_3 ( ) { return configuration ; }
|
org . junit . Assert . assertSame ( configuration , component . METHOD_3 ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . conf . Configuration conf = new org . apache . hadoop . conf . Configuration ( ) ; conf . METHOD_1 ( IDENT_0 . IDENT_1 , 2 ) ; org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_4 . METHOD_2 ( conf , true ) ; conf . METHOD_3 ( IDENT_5 . IDENT_6 , true ) ; org . apache . hadoop . mapreduce . IDENT_7 IDENT_8 = new org . apache . hadoop . mapreduce . METHOD_4 ( ) ; org . apache . hadoop . mapreduce . IDENT_9 < org . apache . hadoop . io . IDENT_10 , org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_11 > reader = new org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_12 . METHOD_5 ( ) ; org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_12 . IDENT_13 writer = new org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_12 . METHOD_6 ( ) ; org . apache . hadoop . mapreduce . IDENT_14 IDENT_15 = new org . apache . hadoop . METHOD_7 ( ) ; org . apache . hadoop . mapreduce . IDENT_16 IDENT_17 = new org . apache . hadoop . mapreduce . task . IDENT_18 . METHOD_8 ( ) ; org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_19 split = METHOD_9 ( ) ; org . apache . hadoop . mapreduce . IDENT_20 < org . apache . hadoop . io . IDENT_10 , org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_11 , org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_21 , org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_11 > IDENT_22 = new org . apache . hadoop . mapreduce . task . IDENT_23 < org . apache . hadoop . io . IDENT_10 , org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_11 , org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_21 , org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_11 > ( conf , IDENT_8 , reader , writer , IDENT_15 , IDENT_17 , split ) ; org . apache . hadoop . mapreduce . IDENT_24 . Context ctx = new org . apache . hadoop . mapreduce . lib . map . IDENT_25 < org . apache . hadoop . io . IDENT_10 , org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_11 , org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_21 , org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_11 > ( ) . METHOD_10 ( IDENT_22 ) ; reader . METHOD_11 ( split , ctx ) ; ctx . METHOD_12 ( ) . METHOD_3 ( IDENT_5 . IDENT_6 , true ) ; org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_4 . METHOD_2 ( ctx . METHOD_12 ( ) , true ) ; org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_26 . IDENT_27 mapper = new org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_26 . METHOD_13 ( ) ; mapper . run ( ctx ) ; java . util . Map < org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_21 , org . apache . hadoop . IDENT_2 . IDENT_3 . IDENT_11 > data = writer . METHOD_14 ( ) ; "<AssertPlaceHolder>" ; } size ( ) { return 1 ; }
|
org . junit . Assert . assertEquals ( 2 , data . size ( ) )
|
METHOD_0 ( ) { java . lang . String expected = STRING_0 + ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( ( STRING_1 3 + STRING_2 ) + STRING_3 ) + STRING_4 ) + STRING_4 2 ) + STRING_1 ) + STRING_1 6 ) + STRING_1 9 ) + STRING_5 0 ) + STRING_1 4 ) + STRING_6 ) + STRING_1 2 ) + STRING_4 4 ) + STRING_1 5 ) + STRING_5 1 ) + STRING_7 ) + STRING_1 1 ) + STRING_5 2 ) + STRING_8 ) + STRING_5 4 ) + STRING_4 7 ) + STRING_5 ) + STRING_5 5 ) + STRING_5 3 ) + STRING_4 5 ) + STRING_4 1 ) + STRING_1 8 ) + STRING_1 7 ) + STRING_1 0 ) + STRING_4 0 ) + STRING_4 9 ) + STRING_9 ) + STRING_4 3 ) + STRING_4 8 ) + STRING_4 6 ) ; "<AssertPlaceHolder>" ; } toString ( ) { return ( ( ( STRING_10 + ( id ) ) + STRING_11 ) + ( name ) ) + "]" ; }
|
org . junit . Assert . assertEquals ( expected , configuration . toString ( ) )
|
METHOD_0 ( ) { runner . start ( ) ; runner . METHOD_1 ( new org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_2 . METHOD_2 ( 0 ) ) ; org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_2 . METHOD_2 . IDENT_3 . METHOD_3 ( INT_0 , TimeUnit . IDENT_4 ) ; "<AssertPlaceHolder>" ; } METHOD_1 ( org . apache . hadoop . yarn . server . IDENT_5 . IDENT_1 . IDENT_6 . IDENT_7 ) { int IDENT_8 = 0 ; java . util . Collection < org . apache . hadoop . yarn . server . IDENT_5 . IDENT_1 . common . IDENT_9 . IDENT_10 > nodes = cs . METHOD_4 ( ) . values ( ) ; int start = org . apache . hadoop . yarn . server . IDENT_5 . IDENT_1 . IDENT_6 . IDENT_7 . random . METHOD_5 ( nodes . size ( ) ) ; for ( org . apache . hadoop . yarn . server . IDENT_5 . IDENT_1 . common . IDENT_9 . IDENT_10 node : nodes ) { if ( ( IDENT_8 ++ ) >= start ) { cs . METHOD_6 ( node ) ; } } for ( org . apache . hadoop . yarn . server . IDENT_5 . IDENT_1 . common . IDENT_9 . IDENT_10 node : nodes ) { cs . METHOD_6 ( node ) ; } try { java . lang . Thread . sleep ( cs . METHOD_7 ( ) ) ; } catch ( java . lang . IDENT_11 e ) { } }
|
org . junit . Assert . assertTrue ( org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_2 . METHOD_2 . first )
|
METHOD_0 ( ) { try { try { org . apache . hadoop . fs . IDENT_0 . METHOD_1 ( IDENT_1 . METHOD_2 ( ) ) ; } catch ( java . lang . Exception e ) { } METHOD_3 ( ) ; IDENT_1 . delete ( ) ; org . apache . hadoop . streaming . IDENT_2 job = new org . apache . hadoop . streaming . METHOD_4 ( ) ; job . METHOD_5 ( new org . apache . hadoop . conf . Configuration ( ) ) ; job . run ( METHOD_6 ( ) ) ; java . io . File IDENT_3 = new java . io . File ( IDENT_1 , STRING_0 ) . METHOD_2 ( ) ; java . lang . String output = org . apache . hadoop . streaming . IDENT_4 . METHOD_7 ( IDENT_3 ) ; IDENT_3 . delete ( ) ; System . out . println ( ( STRING_1 + ( map ) ) ) ; System . out . println ( ( STRING_2 + ( IDENT_5 ) ) ) ; System . err . println ( ( STRING_3 + ( IDENT_6 ) ) ) ; System . err . println ( ( STRING_4 + output ) ) ; "<AssertPlaceHolder>" ; } finally { IDENT_7 . delete ( ) ; org . apache . hadoop . fs . IDENT_0 . METHOD_1 ( IDENT_1 . METHOD_2 ( ) ) ; } } println ( java . lang . Object ) { System . out . println ( s ) ; }
|
org . junit . Assert . assertEquals ( IDENT_6 , output )
|
METHOD_0 ( ) { try { try { org . apache . hadoop . fs . IDENT_0 . METHOD_1 ( IDENT_1 . METHOD_2 ( ) ) ; } catch ( java . lang . Exception e ) { } METHOD_3 ( ) ; boolean IDENT_2 = false ; job = new org . apache . hadoop . streaming . METHOD_4 ( METHOD_5 ( ) , IDENT_2 ) ; job . METHOD_6 ( ) ; org . apache . hadoop . streaming . File IDENT_3 = new org . apache . hadoop . streaming . File ( IDENT_1 , STRING_0 ) . METHOD_2 ( ) ; java . lang . String output = org . apache . hadoop . streaming . IDENT_4 . METHOD_7 ( IDENT_3 ) ; IDENT_3 . delete ( ) ; System . err . println ( ( STRING_1 + ( IDENT_5 ) ) ) ; System . err . println ( ( STRING_2 + output ) ) ; "<AssertPlaceHolder>" ; } finally { IDENT_6 . delete ( ) ; org . apache . hadoop . fs . IDENT_0 . METHOD_1 ( IDENT_1 . METHOD_2 ( ) ) ; } } println ( java . lang . Object ) { System . out . println ( s ) ; }
|
org . junit . Assert . assertEquals ( IDENT_5 , output )
|
METHOD_0 ( ) { try { org . apache . hadoop . fs . IDENT_0 . METHOD_1 ( IDENT_1 . METHOD_2 ( ) ) ; } catch ( java . lang . Exception e ) { } try { METHOD_3 ( ) ; boolean IDENT_2 = false ; job = new org . apache . hadoop . streaming . METHOD_4 ( METHOD_5 ( ) , IDENT_2 ) ; job . METHOD_6 ( ) ; org . apache . hadoop . streaming . File IDENT_3 = new org . apache . hadoop . streaming . File ( IDENT_1 , STRING_0 ) . METHOD_2 ( ) ; java . lang . String output = org . apache . hadoop . streaming . IDENT_4 . METHOD_7 ( IDENT_3 ) ; IDENT_3 . delete ( ) ; System . err . println ( ( STRING_1 + ( IDENT_5 ) ) ) ; System . err . println ( ( STRING_2 + output ) ) ; System . err . println ( ( STRING_3 + ( IDENT_5 . compareTo ( output ) ) ) ) ; "<AssertPlaceHolder>" ; } finally { IDENT_6 . delete ( ) ; org . apache . hadoop . fs . IDENT_0 . METHOD_1 ( IDENT_1 . METHOD_2 ( ) ) ; } } compareTo ( org . apache . hadoop . fs . IDENT_7 ) { if ( IDENT_8 == null ) { return 1 ; } return ( hashCode ( ) ) - ( IDENT_8 . hashCode ( ) ) ; }
|
org . junit . Assert . assertEquals ( IDENT_5 , output )
|
METHOD_0 ( ) { java . lang . String IDENT_0 = STRING_0 ; org . apache . hadoop . streaming . File IDENT_1 = null ; try { try { org . apache . hadoop . fs . IDENT_2 . METHOD_1 ( IDENT_3 . METHOD_2 ( ) ) ; } catch ( java . lang . Exception e ) { } METHOD_3 ( ) ; boolean IDENT_4 = false ; job = new org . apache . hadoop . streaming . METHOD_4 ( METHOD_5 ( ) , IDENT_4 ) ; job . METHOD_6 ( ) ; IDENT_1 = new org . apache . hadoop . streaming . File ( IDENT_3 , IDENT_0 ) . METHOD_2 ( ) ; java . lang . String output = org . apache . hadoop . streaming . IDENT_5 . METHOD_7 ( IDENT_1 ) ; System . err . println ( ( STRING_1 + ( IDENT_6 ) ) ) ; System . err . println ( ( STRING_2 + output ) ) ; "<AssertPlaceHolder>" ; } finally { IDENT_7 . delete ( ) ; org . apache . hadoop . fs . IDENT_2 . METHOD_1 ( IDENT_3 . METHOD_2 ( ) ) ; } } println ( java . lang . Object ) { System . out . println ( s ) ; }
|
org . junit . Assert . assertEquals ( IDENT_6 , output )
|
METHOD_0 ( ) { message = new org . apache . camel . component . IDENT_0 . METHOD_1 ( IDENT_1 , null , new org . apache . camel . component . IDENT_0 . METHOD_2 ( ) ) ; "<AssertPlaceHolder>" ; } METHOD_3 ( ) { if ( ( IDENT_2 ) != null ) { return IDENT_2 . METHOD_4 ( ) ; } return null ; }
|
org . junit . Assert . assertNull ( message . METHOD_3 ( ) )
|
METHOD_0 ( ) { org . apache . hadoop . streaming . IDENT_0 job = new org . apache . hadoop . streaming . METHOD_1 ( ) ; job . METHOD_2 ( new org . apache . hadoop . conf . Configuration ( ) ) ; job . run ( METHOD_3 ( ) ) ; java . io . File IDENT_1 = new java . io . File ( IDENT_2 , STRING_0 ) . METHOD_4 ( ) ; java . lang . String output = org . apache . hadoop . streaming . IDENT_3 . METHOD_5 ( IDENT_1 ) ; IDENT_1 . delete ( ) ; System . out . println ( ( STRING_1 + ( map ) ) ) ; System . out . println ( ( STRING_2 + ( IDENT_4 ) ) ) ; System . err . println ( ( STRING_3 + ( IDENT_5 ) ) ) ; System . err . println ( ( STRING_4 + output ) ) ; "<AssertPlaceHolder>" ; } println ( java . lang . Object ) { System . out . println ( s ) ; }
|
org . junit . Assert . assertEquals ( IDENT_5 , output )
|
METHOD_0 ( ) { java . lang . String [ ] args = new java . lang . String [ ] { STRING_0 , org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_2 . IDENT_3 , STRING_1 , "1" , STRING_2 0 , STRING_3 , STRING_2 , STRING_4 , STRING_2 1 , STRING_5 , STRING_6 } ; org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_2 . LOG . info ( STRING_2 2 ) ; org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_4 client = new org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . METHOD_1 ( org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_5 . class . getName ( ) , new org . apache . hadoop . conf . Configuration ( IDENT_6 . METHOD_2 ( ) ) ) ; client . init ( args ) ; org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_2 . LOG . info ( STRING_7 ) ; boolean result = client . run ( ) ; org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_2 . LOG . info ( ( STRING_8 + result ) ) ; "<AssertPlaceHolder>" ; } info ( java . lang . String ) { return getInstance ( org . apache . hadoop . yarn . IDENT_7 . IDENT_8 . class ) . METHOD_3 ( METHOD_3 ) ; }
|
org . junit . Assert . assertTrue ( result )
|
METHOD_0 ( ) { java . lang . String [ ] args = new java . lang . String [ ] { STRING_0 0 , org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_2 . IDENT_3 , STRING_1 , "1" , STRING_0 1 , STRING_2 , STRING_0 , STRING_3 , STRING_0 2 , STRING_4 , STRING_5 , STRING_6 } ; org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_2 . LOG . info ( STRING_0 3 ) ; org . apache . hadoop . conf . Configuration conf = IDENT_4 . METHOD_1 ( ) ; conf . METHOD_2 ( IDENT_5 . IDENT_6 , 2 ) ; org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_7 client = new org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . METHOD_3 ( org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_8 . class . getName ( ) , new org . apache . hadoop . conf . Configuration ( conf ) ) ; client . init ( args ) ; org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_2 . LOG . info ( STRING_7 ) ; boolean result = client . run ( ) ; org . apache . hadoop . yarn . IDENT_0 . IDENT_1 . IDENT_2 . LOG . info ( ( STRING_8 + result ) ) ; "<AssertPlaceHolder>" ; } info ( java . lang . String ) { return getInstance ( org . apache . hadoop . yarn . IDENT_9 . IDENT_10 . class ) . METHOD_4 ( METHOD_4 ) ; }
|
org . junit . Assert . assertTrue ( result )
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.