Print this page
6717509 Need to use bswap/bswapq for byte swap of 64-bit integer on x32/x64 (fix lint)

Split Close
Expand all
Collapse all
          --- old/usr/src/common/crypto/sha1/sha1.c
          +++ new/usr/src/common/crypto/sha1/sha1.c
↓ open down ↓ 622 lines elided ↑ open up ↑
 623  623          sha1word d = ctx->state[3];
 624  624          sha1word e = ctx->state[4];
 625  625  
 626  626  #if     defined(W_ARRAY)
 627  627          sha1word        w[16];
 628  628  #else   /* !defined(W_ARRAY) */
 629  629          sha1word        w_0, w_1, w_2,  w_3,  w_4,  w_5,  w_6,  w_7;
 630  630          sha1word        w_8, w_9, w_10, w_11, w_12, w_13, w_14, w_15;
 631  631  #endif  /* !defined(W_ARRAY) */
 632  632  
 633      -        W(0)  = LOAD_BIG_32(blk +  0);
 634      -        W(1)  = LOAD_BIG_32(blk +  4);
 635      -        W(2)  = LOAD_BIG_32(blk +  8);
 636      -        W(3)  = LOAD_BIG_32(blk + 12);
 637      -        W(4)  = LOAD_BIG_32(blk + 16);
 638      -        W(5)  = LOAD_BIG_32(blk + 20);
 639      -        W(6)  = LOAD_BIG_32(blk + 24);
 640      -        W(7)  = LOAD_BIG_32(blk + 28);
 641      -        W(8)  = LOAD_BIG_32(blk + 32);
 642      -        W(9)  = LOAD_BIG_32(blk + 36);
 643      -        W(10) = LOAD_BIG_32(blk + 40);
 644      -        W(11) = LOAD_BIG_32(blk + 44);
 645      -        W(12) = LOAD_BIG_32(blk + 48);
 646      -        W(13) = LOAD_BIG_32(blk + 52);
 647      -        W(14) = LOAD_BIG_32(blk + 56);
 648      -        W(15) = LOAD_BIG_32(blk + 60);
      633 +        W(0)  = LOAD_BIG_32((void *)(blk +  0));
      634 +        W(1)  = LOAD_BIG_32((void *)(blk +  4));
      635 +        W(2)  = LOAD_BIG_32((void *)(blk +  8));
      636 +        W(3)  = LOAD_BIG_32((void *)(blk + 12));
      637 +        W(4)  = LOAD_BIG_32((void *)(blk + 16));
      638 +        W(5)  = LOAD_BIG_32((void *)(blk + 20));
      639 +        W(6)  = LOAD_BIG_32((void *)(blk + 24));
      640 +        W(7)  = LOAD_BIG_32((void *)(blk + 28));
      641 +        W(8)  = LOAD_BIG_32((void *)(blk + 32));
      642 +        W(9)  = LOAD_BIG_32((void *)(blk + 36));
      643 +        W(10) = LOAD_BIG_32((void *)(blk + 40));
      644 +        W(11) = LOAD_BIG_32((void *)(blk + 44));
      645 +        W(12) = LOAD_BIG_32((void *)(blk + 48));
      646 +        W(13) = LOAD_BIG_32((void *)(blk + 52));
      647 +        W(14) = LOAD_BIG_32((void *)(blk + 56));
      648 +        W(15) = LOAD_BIG_32((void *)(blk + 60));
 649  649  
 650  650  #endif  /* !defined(__sparc) */
 651  651  
 652  652          /*
 653  653           * general optimization:
 654  654           *
 655  655           * even though this approach is described in the standard as
 656  656           * being slower algorithmically, it is 30-40% faster than the
 657  657           * "faster" version under SPARC, because this version has more
 658  658           * of the constraints specified at compile-time and uses fewer
↓ open down ↓ 371 lines elided ↑ open up ↑
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX