]> WPIA git - cassiopeia.git/blobdiff - lib/openssl/crypto/bn/asm/parisc-mont.pl
add: execute openssl fetcher to fetch openssl 1.0.1j
[cassiopeia.git] / lib / openssl / crypto / bn / asm / parisc-mont.pl
diff --git a/lib/openssl/crypto/bn/asm/parisc-mont.pl b/lib/openssl/crypto/bn/asm/parisc-mont.pl
new file mode 100644 (file)
index 0000000..c02ef6f
--- /dev/null
@@ -0,0 +1,995 @@
+#!/usr/bin/env perl
+
+# ====================================================================
+# Written by Andy Polyakov <appro@fy.chalmers.se> for the OpenSSL
+# project. The module is, however, dual licensed under OpenSSL and
+# CRYPTOGAMS licenses depending on where you obtain it. For further
+# details see http://www.openssl.org/~appro/cryptogams/.
+# ====================================================================
+
+# On PA-7100LC this module performs ~90-50% better, less for longer
+# keys, than code generated by gcc 3.2 for PA-RISC 1.1. Latter means
+# that compiler utilized xmpyu instruction to perform 32x32=64-bit
+# multiplication, which in turn means that "baseline" performance was
+# optimal in respect to instruction set capabilities. Fair comparison
+# with vendor compiler is problematic, because OpenSSL doesn't define
+# BN_LLONG [presumably] for historical reasons, which drives compiler
+# toward 4 times 16x16=32-bit multiplicatons [plus complementary
+# shifts and additions] instead. This means that you should observe
+# several times improvement over code generated by vendor compiler
+# for PA-RISC 1.1, but the "baseline" is far from optimal. The actual
+# improvement coefficient was never collected on PA-7100LC, or any
+# other 1.1 CPU, because I don't have access to such machine with
+# vendor compiler. But to give you a taste, PA-RISC 1.1 code path
+# reportedly outperformed code generated by cc +DA1.1 +O3 by factor
+# of ~5x on PA-8600.
+#
+# On PA-RISC 2.0 it has to compete with pa-risc2[W].s, which is
+# reportedly ~2x faster than vendor compiler generated code [according
+# to comment in pa-risc2[W].s]. Here comes a catch. Execution core of
+# this implementation is actually 32-bit one, in the sense that it
+# operates on 32-bit values. But pa-risc2[W].s operates on arrays of
+# 64-bit BN_LONGs... How do they interoperate then? No problem. This
+# module picks halves of 64-bit values in reverse order and pretends
+# they were 32-bit BN_LONGs. But can 32-bit core compete with "pure"
+# 64-bit code such as pa-risc2[W].s then? Well, the thing is that
+# 32x32=64-bit multiplication is the best even PA-RISC 2.0 can do,
+# i.e. there is no "wider" multiplication like on most other 64-bit
+# platforms. This means that even being effectively 32-bit, this
+# implementation performs "64-bit" computational task in same amount
+# of arithmetic operations, most notably multiplications. It requires
+# more memory references, most notably to tp[num], but this doesn't
+# seem to exhaust memory port capacity. And indeed, dedicated PA-RISC
+# 2.0 code path provides virtually same performance as pa-risc2[W].s:
+# it's ~10% better for shortest key length and ~10% worse for longest
+# one.
+#
+# In case it wasn't clear. The module has two distinct code paths:
+# PA-RISC 1.1 and PA-RISC 2.0 ones. Latter features carry-free 64-bit
+# additions and 64-bit integer loads, not to mention specific
+# instruction scheduling. In 64-bit build naturally only 2.0 code path
+# is assembled. In 32-bit application context both code paths are
+# assembled, PA-RISC 2.0 CPU is detected at run-time and proper path
+# is taken automatically. Also, in 32-bit build the module imposes
+# couple of limitations: vector lengths has to be even and vector
+# addresses has to be 64-bit aligned. Normally neither is a problem:
+# most common key lengths are even and vectors are commonly malloc-ed,
+# which ensures alignment.
+#
+# Special thanks to polarhome.com for providing HP-UX account on
+# PA-RISC 1.1 machine, and to correspondent who chose to remain
+# anonymous for testing the code on PA-RISC 2.0 machine.
+\f
+$0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1;
+
+$flavour = shift;
+$output = shift;
+
+open STDOUT,">$output";
+
+if ($flavour =~ /64/) {
+       $LEVEL          ="2.0W";
+       $SIZE_T         =8;
+       $FRAME_MARKER   =80;
+       $SAVED_RP       =16;
+       $PUSH           ="std";
+       $PUSHMA         ="std,ma";
+       $POP            ="ldd";
+       $POPMB          ="ldd,mb";
+       $BN_SZ          =$SIZE_T;
+} else {
+       $LEVEL          ="1.1"; #$LEVEL.="\n\t.ALLOW\t2.0";
+       $SIZE_T         =4;
+       $FRAME_MARKER   =48;
+       $SAVED_RP       =20;
+       $PUSH           ="stw";
+       $PUSHMA         ="stwm";
+       $POP            ="ldw";
+       $POPMB          ="ldwm";
+       $BN_SZ          =$SIZE_T;
+       if (open CONF,"<${dir}../../opensslconf.h") {
+           while(<CONF>) {
+               if (m/#\s*define\s+SIXTY_FOUR_BIT/) {
+                   $BN_SZ=8;
+                   $LEVEL="2.0";
+                   last;
+               }
+           }
+           close CONF;
+       }
+}
+
+$FRAME=8*$SIZE_T+$FRAME_MARKER;        # 8 saved regs + frame marker
+                               #                [+ argument transfer]
+$LOCALS=$FRAME-$FRAME_MARKER;
+$FRAME+=32;                    # local variables
+
+$tp="%r31";
+$ti1="%r29";
+$ti0="%r28";
+
+$rp="%r26";
+$ap="%r25";
+$bp="%r24";
+$np="%r23";
+$n0="%r22";    # passed through stack in 32-bit
+$num="%r21";   # passed through stack in 32-bit
+$idx="%r20";
+$arrsz="%r19";
+
+$nm1="%r7";
+$nm0="%r6";
+$ab1="%r5";
+$ab0="%r4";
+
+$fp="%r3";
+$hi1="%r2";
+$hi0="%r1";
+
+$xfer=$n0;     # accomodates [-16..15] offset in fld[dw]s
+
+$fm0="%fr4";   $fti=$fm0;
+$fbi="%fr5L";
+$fn0="%fr5R";
+$fai="%fr6";   $fab0="%fr7";   $fab1="%fr8";
+$fni="%fr9";   $fnm0="%fr10";  $fnm1="%fr11";
+
+$code=<<___;
+       .LEVEL  $LEVEL
+       .SPACE  \$TEXT\$
+       .SUBSPA \$CODE\$,QUAD=0,ALIGN=8,ACCESS=0x2C,CODE_ONLY
+
+       .EXPORT bn_mul_mont,ENTRY,ARGW0=GR,ARGW1=GR,ARGW2=GR,ARGW3=GR
+       .ALIGN  64
+bn_mul_mont
+       .PROC
+       .CALLINFO       FRAME=`$FRAME-8*$SIZE_T`,NO_CALLS,SAVE_RP,SAVE_SP,ENTRY_GR=6
+       .ENTRY
+       $PUSH   %r2,-$SAVED_RP(%sp)             ; standard prologue
+       $PUSHMA %r3,$FRAME(%sp)
+       $PUSH   %r4,`-$FRAME+1*$SIZE_T`(%sp)
+       $PUSH   %r5,`-$FRAME+2*$SIZE_T`(%sp)
+       $PUSH   %r6,`-$FRAME+3*$SIZE_T`(%sp)
+       $PUSH   %r7,`-$FRAME+4*$SIZE_T`(%sp)
+       $PUSH   %r8,`-$FRAME+5*$SIZE_T`(%sp)
+       $PUSH   %r9,`-$FRAME+6*$SIZE_T`(%sp)
+       $PUSH   %r10,`-$FRAME+7*$SIZE_T`(%sp)
+       ldo     -$FRAME(%sp),$fp
+___
+$code.=<<___ if ($SIZE_T==4);
+       ldw     `-$FRAME_MARKER-4`($fp),$n0
+       ldw     `-$FRAME_MARKER-8`($fp),$num
+       nop
+       nop                                     ; alignment
+___
+$code.=<<___ if ($BN_SZ==4);
+       comiclr,<=      6,$num,%r0              ; are vectors long enough?
+       b               L\$abort
+       ldi             0,%r28                  ; signal "unhandled"
+       add,ev          %r0,$num,$num           ; is $num even?
+       b               L\$abort
+       nop
+       or              $ap,$np,$ti1
+       extru,=         $ti1,31,3,%r0           ; are ap and np 64-bit aligned?
+       b               L\$abort
+       nop
+       nop                                     ; alignment
+       nop
+
+       fldws           0($n0),${fn0}
+       fldws,ma        4($bp),${fbi}           ; bp[0]
+___
+$code.=<<___ if ($BN_SZ==8);
+       comib,>         3,$num,L\$abort         ; are vectors long enough?
+       ldi             0,%r28                  ; signal "unhandled"
+       addl            $num,$num,$num          ; I operate on 32-bit values
+
+       fldws           4($n0),${fn0}           ; only low part of n0
+       fldws           4($bp),${fbi}           ; bp[0] in flipped word order
+___
+$code.=<<___;
+       fldds           0($ap),${fai}           ; ap[0,1]
+       fldds           0($np),${fni}           ; np[0,1]
+
+       sh2addl         $num,%r0,$arrsz
+       ldi             31,$hi0
+       ldo             36($arrsz),$hi1         ; space for tp[num+1]
+       andcm           $hi1,$hi0,$hi1          ; align
+       addl            $hi1,%sp,%sp
+       $PUSH           $fp,-$SIZE_T(%sp)
+
+       ldo             `$LOCALS+16`($fp),$xfer
+       ldo             `$LOCALS+32+4`($fp),$tp
+
+       xmpyu           ${fai}L,${fbi},${fab0}  ; ap[0]*bp[0]
+       xmpyu           ${fai}R,${fbi},${fab1}  ; ap[1]*bp[0]
+       xmpyu           ${fn0},${fab0}R,${fm0}
+
+       addl            $arrsz,$ap,$ap          ; point at the end
+       addl            $arrsz,$np,$np
+       subi            0,$arrsz,$idx           ; j=0
+       ldo             8($idx),$idx            ; j++++
+
+       xmpyu           ${fni}L,${fm0}R,${fnm0} ; np[0]*m
+       xmpyu           ${fni}R,${fm0}R,${fnm1} ; np[1]*m
+       fstds           ${fab0},-16($xfer)
+       fstds           ${fnm0},-8($xfer)
+       fstds           ${fab1},0($xfer)
+       fstds           ${fnm1},8($xfer)
+        flddx          $idx($ap),${fai}        ; ap[2,3]
+        flddx          $idx($np),${fni}        ; np[2,3]
+___
+$code.=<<___ if ($BN_SZ==4);
+       mtctl           $hi0,%cr11              ; $hi0 still holds 31
+       extrd,u,*=      $hi0,%sar,1,$hi0        ; executes on PA-RISC 1.0
+       b               L\$parisc11
+       nop
+___
+$code.=<<___;                                  # PA-RISC 2.0 code-path
+       xmpyu           ${fai}L,${fbi},${fab0}  ; ap[j]*bp[0]
+       xmpyu           ${fni}L,${fm0}R,${fnm0} ; np[j]*m
+       ldd             -16($xfer),$ab0
+       fstds           ${fab0},-16($xfer)
+
+       extrd,u         $ab0,31,32,$hi0
+       extrd,u         $ab0,63,32,$ab0
+       ldd             -8($xfer),$nm0
+       fstds           ${fnm0},-8($xfer)
+        ldo            8($idx),$idx            ; j++++
+        addl           $ab0,$nm0,$nm0          ; low part is discarded
+        extrd,u        $nm0,31,32,$hi1
+\f
+L\$1st
+       xmpyu           ${fai}R,${fbi},${fab1}  ; ap[j+1]*bp[0]
+       xmpyu           ${fni}R,${fm0}R,${fnm1} ; np[j+1]*m
+       ldd             0($xfer),$ab1
+       fstds           ${fab1},0($xfer)
+        addl           $hi0,$ab1,$ab1
+        extrd,u        $ab1,31,32,$hi0
+       ldd             8($xfer),$nm1
+       fstds           ${fnm1},8($xfer)
+        extrd,u        $ab1,63,32,$ab1
+        addl           $hi1,$nm1,$nm1
+       flddx           $idx($ap),${fai}        ; ap[j,j+1]
+       flddx           $idx($np),${fni}        ; np[j,j+1]
+        addl           $ab1,$nm1,$nm1
+        extrd,u        $nm1,31,32,$hi1
+
+       xmpyu           ${fai}L,${fbi},${fab0}  ; ap[j]*bp[0]
+       xmpyu           ${fni}L,${fm0}R,${fnm0} ; np[j]*m
+       ldd             -16($xfer),$ab0
+       fstds           ${fab0},-16($xfer)
+        addl           $hi0,$ab0,$ab0
+        extrd,u        $ab0,31,32,$hi0
+       ldd             -8($xfer),$nm0
+       fstds           ${fnm0},-8($xfer)
+        extrd,u        $ab0,63,32,$ab0
+        addl           $hi1,$nm0,$nm0
+       stw             $nm1,-4($tp)            ; tp[j-1]
+        addl           $ab0,$nm0,$nm0
+        stw,ma         $nm0,8($tp)             ; tp[j-1]
+       addib,<>        8,$idx,L\$1st           ; j++++
+        extrd,u        $nm0,31,32,$hi1
+
+       xmpyu           ${fai}R,${fbi},${fab1}  ; ap[j]*bp[0]
+       xmpyu           ${fni}R,${fm0}R,${fnm1} ; np[j]*m
+       ldd             0($xfer),$ab1
+       fstds           ${fab1},0($xfer)
+        addl           $hi0,$ab1,$ab1
+        extrd,u        $ab1,31,32,$hi0
+       ldd             8($xfer),$nm1
+       fstds           ${fnm1},8($xfer)
+        extrd,u        $ab1,63,32,$ab1
+        addl           $hi1,$nm1,$nm1
+       ldd             -16($xfer),$ab0
+        addl           $ab1,$nm1,$nm1
+       ldd             -8($xfer),$nm0
+        extrd,u        $nm1,31,32,$hi1
+
+        addl           $hi0,$ab0,$ab0
+        extrd,u        $ab0,31,32,$hi0
+       stw             $nm1,-4($tp)            ; tp[j-1]
+        extrd,u        $ab0,63,32,$ab0
+        addl           $hi1,$nm0,$nm0
+       ldd             0($xfer),$ab1
+        addl           $ab0,$nm0,$nm0
+       ldd,mb          8($xfer),$nm1
+        extrd,u        $nm0,31,32,$hi1
+       stw,ma          $nm0,8($tp)             ; tp[j-1]
+
+       ldo             -1($num),$num           ; i--
+       subi            0,$arrsz,$idx           ; j=0
+___
+$code.=<<___ if ($BN_SZ==4);
+       fldws,ma        4($bp),${fbi}           ; bp[1]
+___
+$code.=<<___ if ($BN_SZ==8);
+       fldws           0($bp),${fbi}           ; bp[1] in flipped word order
+___
+$code.=<<___;
+        flddx          $idx($ap),${fai}        ; ap[0,1]
+        flddx          $idx($np),${fni}        ; np[0,1]
+        fldws          8($xfer),${fti}R        ; tp[0]
+       addl            $hi0,$ab1,$ab1
+        extrd,u        $ab1,31,32,$hi0
+        extrd,u        $ab1,63,32,$ab1
+        ldo            8($idx),$idx            ; j++++
+        xmpyu          ${fai}L,${fbi},${fab0}  ; ap[0]*bp[1]
+        xmpyu          ${fai}R,${fbi},${fab1}  ; ap[1]*bp[1]
+       addl            $hi1,$nm1,$nm1
+       addl            $ab1,$nm1,$nm1
+       extrd,u         $nm1,31,32,$hi1
+        fstws,mb       ${fab0}L,-8($xfer)      ; save high part
+       stw             $nm1,-4($tp)            ; tp[j-1]
+
+        fcpy,sgl       %fr0,${fti}L            ; zero high part
+        fcpy,sgl       %fr0,${fab0}L
+       addl            $hi1,$hi0,$hi0
+       extrd,u         $hi0,31,32,$hi1
+        fcnvxf,dbl,dbl ${fti},${fti}           ; 32-bit unsigned int -> double
+        fcnvxf,dbl,dbl ${fab0},${fab0}
+       stw             $hi0,0($tp)
+       stw             $hi1,4($tp)
+
+       fadd,dbl        ${fti},${fab0},${fab0}  ; add tp[0]
+       fcnvfx,dbl,dbl  ${fab0},${fab0}         ; double -> 33-bit unsigned int
+       xmpyu           ${fn0},${fab0}R,${fm0}
+       ldo             `$LOCALS+32+4`($fp),$tp
+L\$outer
+       xmpyu           ${fni}L,${fm0}R,${fnm0} ; np[0]*m
+       xmpyu           ${fni}R,${fm0}R,${fnm1} ; np[1]*m
+       fstds           ${fab0},-16($xfer)      ; 33-bit value
+       fstds           ${fnm0},-8($xfer)
+        flddx          $idx($ap),${fai}        ; ap[2]
+        flddx          $idx($np),${fni}        ; np[2]
+        ldo            8($idx),$idx            ; j++++
+       ldd             -16($xfer),$ab0         ; 33-bit value
+       ldd             -8($xfer),$nm0
+       ldw             0($xfer),$hi0           ; high part
+
+       xmpyu           ${fai}L,${fbi},${fab0}  ; ap[j]*bp[i]
+       xmpyu           ${fni}L,${fm0}R,${fnm0} ; np[j]*m
+        extrd,u        $ab0,31,32,$ti0         ; carry bit
+        extrd,u        $ab0,63,32,$ab0
+       fstds           ${fab1},0($xfer)
+        addl           $ti0,$hi0,$hi0          ; account carry bit
+       fstds           ${fnm1},8($xfer)
+        addl           $ab0,$nm0,$nm0          ; low part is discarded
+       ldw             0($tp),$ti1             ; tp[1]
+        extrd,u        $nm0,31,32,$hi1
+       fstds           ${fab0},-16($xfer)
+       fstds           ${fnm0},-8($xfer)
+\f
+L\$inner
+       xmpyu           ${fai}R,${fbi},${fab1}  ; ap[j+1]*bp[i]
+       xmpyu           ${fni}R,${fm0}R,${fnm1} ; np[j+1]*m
+       ldd             0($xfer),$ab1
+       fstds           ${fab1},0($xfer)
+        addl           $hi0,$ti1,$ti1
+        addl           $ti1,$ab1,$ab1
+       ldd             8($xfer),$nm1
+       fstds           ${fnm1},8($xfer)
+        extrd,u        $ab1,31,32,$hi0
+        extrd,u        $ab1,63,32,$ab1
+       flddx           $idx($ap),${fai}        ; ap[j,j+1]
+       flddx           $idx($np),${fni}        ; np[j,j+1]
+        addl           $hi1,$nm1,$nm1
+        addl           $ab1,$nm1,$nm1
+       ldw             4($tp),$ti0             ; tp[j]
+       stw             $nm1,-4($tp)            ; tp[j-1]
+
+       xmpyu           ${fai}L,${fbi},${fab0}  ; ap[j]*bp[i]
+       xmpyu           ${fni}L,${fm0}R,${fnm0} ; np[j]*m
+       ldd             -16($xfer),$ab0
+       fstds           ${fab0},-16($xfer)
+        addl           $hi0,$ti0,$ti0
+        addl           $ti0,$ab0,$ab0
+       ldd             -8($xfer),$nm0
+       fstds           ${fnm0},-8($xfer)
+        extrd,u        $ab0,31,32,$hi0
+        extrd,u        $nm1,31,32,$hi1
+       ldw             8($tp),$ti1             ; tp[j]
+        extrd,u        $ab0,63,32,$ab0
+        addl           $hi1,$nm0,$nm0
+        addl           $ab0,$nm0,$nm0
+        stw,ma         $nm0,8($tp)             ; tp[j-1]
+       addib,<>        8,$idx,L\$inner         ; j++++
+        extrd,u        $nm0,31,32,$hi1
+
+       xmpyu           ${fai}R,${fbi},${fab1}  ; ap[j]*bp[i]
+       xmpyu           ${fni}R,${fm0}R,${fnm1} ; np[j]*m
+       ldd             0($xfer),$ab1
+       fstds           ${fab1},0($xfer)
+        addl           $hi0,$ti1,$ti1
+        addl           $ti1,$ab1,$ab1
+       ldd             8($xfer),$nm1
+       fstds           ${fnm1},8($xfer)
+        extrd,u        $ab1,31,32,$hi0
+        extrd,u        $ab1,63,32,$ab1
+       ldw             4($tp),$ti0             ; tp[j]
+        addl           $hi1,$nm1,$nm1
+        addl           $ab1,$nm1,$nm1
+       ldd             -16($xfer),$ab0
+       ldd             -8($xfer),$nm0
+        extrd,u        $nm1,31,32,$hi1
+
+       addl            $hi0,$ab0,$ab0
+        addl           $ti0,$ab0,$ab0
+        stw            $nm1,-4($tp)            ; tp[j-1]
+        extrd,u        $ab0,31,32,$hi0
+       ldw             8($tp),$ti1             ; tp[j]
+        extrd,u        $ab0,63,32,$ab0
+        addl           $hi1,$nm0,$nm0
+       ldd             0($xfer),$ab1
+        addl           $ab0,$nm0,$nm0
+       ldd,mb          8($xfer),$nm1
+        extrd,u        $nm0,31,32,$hi1
+        stw,ma         $nm0,8($tp)             ; tp[j-1]
+
+       addib,=         -1,$num,L\$outerdone    ; i--
+       subi            0,$arrsz,$idx           ; j=0
+___
+$code.=<<___ if ($BN_SZ==4);
+       fldws,ma        4($bp),${fbi}           ; bp[i]
+___
+$code.=<<___ if ($BN_SZ==8);
+       ldi             12,$ti0                 ; bp[i] in flipped word order
+       addl,ev         %r0,$num,$num
+       ldi             -4,$ti0
+       addl            $ti0,$bp,$bp
+       fldws           0($bp),${fbi}
+___
+$code.=<<___;
+        flddx          $idx($ap),${fai}        ; ap[0]
+       addl            $hi0,$ab1,$ab1
+        flddx          $idx($np),${fni}        ; np[0]
+        fldws          8($xfer),${fti}R        ; tp[0]
+       addl            $ti1,$ab1,$ab1
+       extrd,u         $ab1,31,32,$hi0
+       extrd,u         $ab1,63,32,$ab1
+
+        ldo            8($idx),$idx            ; j++++
+        xmpyu          ${fai}L,${fbi},${fab0}  ; ap[0]*bp[i]
+        xmpyu          ${fai}R,${fbi},${fab1}  ; ap[1]*bp[i]
+       ldw             4($tp),$ti0             ; tp[j]
+
+       addl            $hi1,$nm1,$nm1
+        fstws,mb       ${fab0}L,-8($xfer)      ; save high part
+       addl            $ab1,$nm1,$nm1
+       extrd,u         $nm1,31,32,$hi1
+        fcpy,sgl       %fr0,${fti}L            ; zero high part
+        fcpy,sgl       %fr0,${fab0}L
+       stw             $nm1,-4($tp)            ; tp[j-1]
+
+        fcnvxf,dbl,dbl ${fti},${fti}           ; 32-bit unsigned int -> double
+        fcnvxf,dbl,dbl ${fab0},${fab0}
+       addl            $hi1,$hi0,$hi0
+        fadd,dbl       ${fti},${fab0},${fab0}  ; add tp[0]
+       addl            $ti0,$hi0,$hi0
+       extrd,u         $hi0,31,32,$hi1
+        fcnvfx,dbl,dbl ${fab0},${fab0}         ; double -> 33-bit unsigned int
+       stw             $hi0,0($tp)
+       stw             $hi1,4($tp)
+        xmpyu          ${fn0},${fab0}R,${fm0}
+
+       b               L\$outer
+       ldo             `$LOCALS+32+4`($fp),$tp
+\f
+L\$outerdone
+       addl            $hi0,$ab1,$ab1
+       addl            $ti1,$ab1,$ab1
+       extrd,u         $ab1,31,32,$hi0
+       extrd,u         $ab1,63,32,$ab1
+
+       ldw             4($tp),$ti0             ; tp[j]
+
+       addl            $hi1,$nm1,$nm1
+       addl            $ab1,$nm1,$nm1
+       extrd,u         $nm1,31,32,$hi1
+       stw             $nm1,-4($tp)            ; tp[j-1]
+
+       addl            $hi1,$hi0,$hi0
+       addl            $ti0,$hi0,$hi0
+       extrd,u         $hi0,31,32,$hi1
+       stw             $hi0,0($tp)
+       stw             $hi1,4($tp)
+
+       ldo             `$LOCALS+32`($fp),$tp
+       sub             %r0,%r0,%r0             ; clear borrow
+___
+$code.=<<___ if ($BN_SZ==4);
+       ldws,ma         4($tp),$ti0
+       extru,=         $rp,31,3,%r0            ; is rp 64-bit aligned?
+       b               L\$sub_pa11
+       addl            $tp,$arrsz,$tp
+L\$sub
+       ldwx            $idx($np),$hi0
+       subb            $ti0,$hi0,$hi1
+       ldwx            $idx($tp),$ti0
+       addib,<>        4,$idx,L\$sub
+       stws,ma         $hi1,4($rp)
+
+       subb            $ti0,%r0,$hi1
+       ldo             -4($tp),$tp
+___
+$code.=<<___ if ($BN_SZ==8);
+       ldd,ma          8($tp),$ti0
+L\$sub
+       ldd             $idx($np),$hi0
+       shrpd           $ti0,$ti0,32,$ti0       ; flip word order
+       std             $ti0,-8($tp)            ; save flipped value
+       sub,db          $ti0,$hi0,$hi1
+       ldd,ma          8($tp),$ti0
+       addib,<>        8,$idx,L\$sub
+       std,ma          $hi1,8($rp)
+
+       extrd,u         $ti0,31,32,$ti0         ; carry in flipped word order
+       sub,db          $ti0,%r0,$hi1
+       ldo             -8($tp),$tp
+___
+$code.=<<___;
+       and             $tp,$hi1,$ap
+       andcm           $rp,$hi1,$bp
+       or              $ap,$bp,$np
+
+       sub             $rp,$arrsz,$rp          ; rewind rp
+       subi            0,$arrsz,$idx
+       ldo             `$LOCALS+32`($fp),$tp
+L\$copy
+       ldd             $idx($np),$hi0
+       std,ma          %r0,8($tp)
+       addib,<>        8,$idx,.-8              ; L\$copy
+       std,ma          $hi0,8($rp)     
+___
+
+if ($BN_SZ==4) {                               # PA-RISC 1.1 code-path
+$ablo=$ab0;
+$abhi=$ab1;
+$nmlo0=$nm0;
+$nmhi0=$nm1;
+$nmlo1="%r9";
+$nmhi1="%r8";
+
+$code.=<<___;
+       b               L\$done
+       nop
+
+       .ALIGN          8
+L\$parisc11
+       xmpyu           ${fai}L,${fbi},${fab0}  ; ap[j]*bp[0]
+       xmpyu           ${fni}L,${fm0}R,${fnm0} ; np[j]*m
+       ldw             -12($xfer),$ablo
+       ldw             -16($xfer),$hi0
+       ldw             -4($xfer),$nmlo0
+       ldw             -8($xfer),$nmhi0
+       fstds           ${fab0},-16($xfer)
+       fstds           ${fnm0},-8($xfer)
+
+        ldo            8($idx),$idx            ; j++++
+        add            $ablo,$nmlo0,$nmlo0     ; discarded
+        addc           %r0,$nmhi0,$hi1
+       ldw             4($xfer),$ablo
+       ldw             0($xfer),$abhi
+       nop
+\f
+L\$1st_pa11
+       xmpyu           ${fai}R,${fbi},${fab1}  ; ap[j+1]*bp[0]
+       flddx           $idx($ap),${fai}        ; ap[j,j+1]
+       xmpyu           ${fni}R,${fm0}R,${fnm1} ; np[j+1]*m
+       flddx           $idx($np),${fni}        ; np[j,j+1]
+        add            $hi0,$ablo,$ablo
+       ldw             12($xfer),$nmlo1
+        addc           %r0,$abhi,$hi0
+       ldw             8($xfer),$nmhi1
+        add            $ablo,$nmlo1,$nmlo1
+       fstds           ${fab1},0($xfer)
+        addc           %r0,$nmhi1,$nmhi1
+       fstds           ${fnm1},8($xfer)
+        add            $hi1,$nmlo1,$nmlo1
+       ldw             -12($xfer),$ablo
+        addc           %r0,$nmhi1,$hi1
+       ldw             -16($xfer),$abhi
+
+       xmpyu           ${fai}L,${fbi},${fab0}  ; ap[j]*bp[0]
+       ldw             -4($xfer),$nmlo0
+       xmpyu           ${fni}L,${fm0}R,${fnm0} ; np[j]*m
+       ldw             -8($xfer),$nmhi0
+        add            $hi0,$ablo,$ablo
+       stw             $nmlo1,-4($tp)          ; tp[j-1]
+        addc           %r0,$abhi,$hi0
+       fstds           ${fab0},-16($xfer)
+        add            $ablo,$nmlo0,$nmlo0
+       fstds           ${fnm0},-8($xfer)
+        addc           %r0,$nmhi0,$nmhi0
+       ldw             0($xfer),$abhi
+        add            $hi1,$nmlo0,$nmlo0
+       ldw             4($xfer),$ablo
+        stws,ma        $nmlo0,8($tp)           ; tp[j-1]
+       addib,<>        8,$idx,L\$1st_pa11      ; j++++
+        addc           %r0,$nmhi0,$hi1
+
+        ldw            8($xfer),$nmhi1
+        ldw            12($xfer),$nmlo1
+       xmpyu           ${fai}R,${fbi},${fab1}  ; ap[j]*bp[0]
+       xmpyu           ${fni}R,${fm0}R,${fnm1} ; np[j]*m
+        add            $hi0,$ablo,$ablo
+       fstds           ${fab1},0($xfer)
+        addc           %r0,$abhi,$hi0
+       fstds           ${fnm1},8($xfer)
+        add            $ablo,$nmlo1,$nmlo1
+       ldw             -16($xfer),$abhi
+        addc           %r0,$nmhi1,$nmhi1
+       ldw             -12($xfer),$ablo
+        add            $hi1,$nmlo1,$nmlo1
+       ldw             -8($xfer),$nmhi0
+        addc           %r0,$nmhi1,$hi1
+       ldw             -4($xfer),$nmlo0
+
+        add            $hi0,$ablo,$ablo
+       stw             $nmlo1,-4($tp)          ; tp[j-1]
+        addc           %r0,$abhi,$hi0
+       ldw             0($xfer),$abhi
+        add            $ablo,$nmlo0,$nmlo0
+       ldw             4($xfer),$ablo
+        addc           %r0,$nmhi0,$nmhi0
+       ldws,mb         8($xfer),$nmhi1
+        add            $hi1,$nmlo0,$nmlo0
+       ldw             4($xfer),$nmlo1
+        addc           %r0,$nmhi0,$hi1
+       stws,ma         $nmlo0,8($tp)           ; tp[j-1]
+
+       ldo             -1($num),$num           ; i--
+       subi            0,$arrsz,$idx           ; j=0
+
+        fldws,ma       4($bp),${fbi}           ; bp[1]
+        flddx          $idx($ap),${fai}        ; ap[0,1]
+        flddx          $idx($np),${fni}        ; np[0,1]
+        fldws          8($xfer),${fti}R        ; tp[0]
+       add             $hi0,$ablo,$ablo
+       addc            %r0,$abhi,$hi0
+        ldo            8($idx),$idx            ; j++++
+        xmpyu          ${fai}L,${fbi},${fab0}  ; ap[0]*bp[1]
+        xmpyu          ${fai}R,${fbi},${fab1}  ; ap[1]*bp[1]
+       add             $hi1,$nmlo1,$nmlo1
+       addc            %r0,$nmhi1,$nmhi1
+       add             $ablo,$nmlo1,$nmlo1
+       addc            %r0,$nmhi1,$hi1
+        fstws,mb       ${fab0}L,-8($xfer)      ; save high part
+       stw             $nmlo1,-4($tp)          ; tp[j-1]
+
+        fcpy,sgl       %fr0,${fti}L            ; zero high part
+        fcpy,sgl       %fr0,${fab0}L
+       add             $hi1,$hi0,$hi0
+       addc            %r0,%r0,$hi1
+        fcnvxf,dbl,dbl ${fti},${fti}           ; 32-bit unsigned int -> double
+        fcnvxf,dbl,dbl ${fab0},${fab0}
+       stw             $hi0,0($tp)
+       stw             $hi1,4($tp)
+
+       fadd,dbl        ${fti},${fab0},${fab0}  ; add tp[0]
+       fcnvfx,dbl,dbl  ${fab0},${fab0}         ; double -> 33-bit unsigned int
+       xmpyu           ${fn0},${fab0}R,${fm0}
+       ldo             `$LOCALS+32+4`($fp),$tp
+L\$outer_pa11
+       xmpyu           ${fni}L,${fm0}R,${fnm0} ; np[0]*m
+       xmpyu           ${fni}R,${fm0}R,${fnm1} ; np[1]*m
+       fstds           ${fab0},-16($xfer)      ; 33-bit value
+       fstds           ${fnm0},-8($xfer)
+        flddx          $idx($ap),${fai}        ; ap[2,3]
+        flddx          $idx($np),${fni}        ; np[2,3]
+       ldw             -16($xfer),$abhi        ; carry bit actually
+        ldo            8($idx),$idx            ; j++++
+       ldw             -12($xfer),$ablo
+       ldw             -8($xfer),$nmhi0
+       ldw             -4($xfer),$nmlo0
+       ldw             0($xfer),$hi0           ; high part
+
+       xmpyu           ${fai}L,${fbi},${fab0}  ; ap[j]*bp[i]
+       xmpyu           ${fni}L,${fm0}R,${fnm0} ; np[j]*m
+       fstds           ${fab1},0($xfer)
+        addl           $abhi,$hi0,$hi0         ; account carry bit
+       fstds           ${fnm1},8($xfer)
+        add            $ablo,$nmlo0,$nmlo0     ; discarded
+       ldw             0($tp),$ti1             ; tp[1]
+        addc           %r0,$nmhi0,$hi1
+       fstds           ${fab0},-16($xfer)
+       fstds           ${fnm0},-8($xfer)
+       ldw             4($xfer),$ablo
+       ldw             0($xfer),$abhi
+\f
+L\$inner_pa11
+       xmpyu           ${fai}R,${fbi},${fab1}  ; ap[j+1]*bp[i]
+       flddx           $idx($ap),${fai}        ; ap[j,j+1]
+       xmpyu           ${fni}R,${fm0}R,${fnm1} ; np[j+1]*m
+       flddx           $idx($np),${fni}        ; np[j,j+1]
+        add            $hi0,$ablo,$ablo
+       ldw             4($tp),$ti0             ; tp[j]
+        addc           %r0,$abhi,$abhi
+       ldw             12($xfer),$nmlo1
+        add            $ti1,$ablo,$ablo
+       ldw             8($xfer),$nmhi1
+        addc           %r0,$abhi,$hi0
+       fstds           ${fab1},0($xfer)
+        add            $ablo,$nmlo1,$nmlo1
+       fstds           ${fnm1},8($xfer)
+        addc           %r0,$nmhi1,$nmhi1
+       ldw             -12($xfer),$ablo
+        add            $hi1,$nmlo1,$nmlo1
+       ldw             -16($xfer),$abhi
+        addc           %r0,$nmhi1,$hi1
+
+       xmpyu           ${fai}L,${fbi},${fab0}  ; ap[j]*bp[i]
+       ldw             8($tp),$ti1             ; tp[j]
+       xmpyu           ${fni}L,${fm0}R,${fnm0} ; np[j]*m
+       ldw             -4($xfer),$nmlo0
+        add            $hi0,$ablo,$ablo
+       ldw             -8($xfer),$nmhi0
+        addc           %r0,$abhi,$abhi
+       stw             $nmlo1,-4($tp)          ; tp[j-1]
+        add            $ti0,$ablo,$ablo
+       fstds           ${fab0},-16($xfer)
+        addc           %r0,$abhi,$hi0
+       fstds           ${fnm0},-8($xfer)
+        add            $ablo,$nmlo0,$nmlo0
+       ldw             4($xfer),$ablo
+        addc           %r0,$nmhi0,$nmhi0
+       ldw             0($xfer),$abhi
+        add            $hi1,$nmlo0,$nmlo0
+        stws,ma        $nmlo0,8($tp)           ; tp[j-1]
+       addib,<>        8,$idx,L\$inner_pa11    ; j++++
+        addc           %r0,$nmhi0,$hi1
+
+       xmpyu           ${fai}R,${fbi},${fab1}  ; ap[j]*bp[i]
+       ldw             12($xfer),$nmlo1
+       xmpyu           ${fni}R,${fm0}R,${fnm1} ; np[j]*m
+       ldw             8($xfer),$nmhi1
+        add            $hi0,$ablo,$ablo
+       ldw             4($tp),$ti0             ; tp[j]
+        addc           %r0,$abhi,$abhi
+       fstds           ${fab1},0($xfer)
+        add            $ti1,$ablo,$ablo
+       fstds           ${fnm1},8($xfer)
+        addc           %r0,$abhi,$hi0
+       ldw             -16($xfer),$abhi
+        add            $ablo,$nmlo1,$nmlo1
+       ldw             -12($xfer),$ablo
+        addc           %r0,$nmhi1,$nmhi1
+       ldw             -8($xfer),$nmhi0
+        add            $hi1,$nmlo1,$nmlo1
+       ldw             -4($xfer),$nmlo0
+        addc           %r0,$nmhi1,$hi1
+
+       add             $hi0,$ablo,$ablo
+        stw            $nmlo1,-4($tp)          ; tp[j-1]
+       addc            %r0,$abhi,$abhi
+        add            $ti0,$ablo,$ablo
+       ldw             8($tp),$ti1             ; tp[j]
+        addc           %r0,$abhi,$hi0
+       ldw             0($xfer),$abhi
+        add            $ablo,$nmlo0,$nmlo0
+       ldw             4($xfer),$ablo
+        addc           %r0,$nmhi0,$nmhi0
+       ldws,mb         8($xfer),$nmhi1
+        add            $hi1,$nmlo0,$nmlo0
+       ldw             4($xfer),$nmlo1
+        addc           %r0,$nmhi0,$hi1
+        stws,ma        $nmlo0,8($tp)           ; tp[j-1]
+
+       addib,=         -1,$num,L\$outerdone_pa11; i--
+       subi            0,$arrsz,$idx           ; j=0
+
+        fldws,ma       4($bp),${fbi}           ; bp[i]
+        flddx          $idx($ap),${fai}        ; ap[0]
+       add             $hi0,$ablo,$ablo
+       addc            %r0,$abhi,$abhi
+        flddx          $idx($np),${fni}        ; np[0]
+        fldws          8($xfer),${fti}R        ; tp[0]
+       add             $ti1,$ablo,$ablo
+       addc            %r0,$abhi,$hi0
+
+        ldo            8($idx),$idx            ; j++++
+        xmpyu          ${fai}L,${fbi},${fab0}  ; ap[0]*bp[i]
+        xmpyu          ${fai}R,${fbi},${fab1}  ; ap[1]*bp[i]
+       ldw             4($tp),$ti0             ; tp[j]
+
+       add             $hi1,$nmlo1,$nmlo1
+       addc            %r0,$nmhi1,$nmhi1
+        fstws,mb       ${fab0}L,-8($xfer)      ; save high part
+       add             $ablo,$nmlo1,$nmlo1
+       addc            %r0,$nmhi1,$hi1
+        fcpy,sgl       %fr0,${fti}L            ; zero high part
+        fcpy,sgl       %fr0,${fab0}L
+       stw             $nmlo1,-4($tp)          ; tp[j-1]
+
+        fcnvxf,dbl,dbl ${fti},${fti}           ; 32-bit unsigned int -> double
+        fcnvxf,dbl,dbl ${fab0},${fab0}
+       add             $hi1,$hi0,$hi0
+       addc            %r0,%r0,$hi1
+        fadd,dbl       ${fti},${fab0},${fab0}  ; add tp[0]
+       add             $ti0,$hi0,$hi0
+       addc            %r0,$hi1,$hi1
+        fcnvfx,dbl,dbl ${fab0},${fab0}         ; double -> 33-bit unsigned int
+       stw             $hi0,0($tp)
+       stw             $hi1,4($tp)
+        xmpyu          ${fn0},${fab0}R,${fm0}
+
+       b               L\$outer_pa11
+       ldo             `$LOCALS+32+4`($fp),$tp
+\f
+L\$outerdone_pa11
+       add             $hi0,$ablo,$ablo
+       addc            %r0,$abhi,$abhi
+       add             $ti1,$ablo,$ablo
+       addc            %r0,$abhi,$hi0
+
+       ldw             4($tp),$ti0             ; tp[j]
+
+       add             $hi1,$nmlo1,$nmlo1
+       addc            %r0,$nmhi1,$nmhi1
+       add             $ablo,$nmlo1,$nmlo1
+       addc            %r0,$nmhi1,$hi1
+       stw             $nmlo1,-4($tp)          ; tp[j-1]
+
+       add             $hi1,$hi0,$hi0
+       addc            %r0,%r0,$hi1
+       add             $ti0,$hi0,$hi0
+       addc            %r0,$hi1,$hi1
+       stw             $hi0,0($tp)
+       stw             $hi1,4($tp)
+
+       ldo             `$LOCALS+32+4`($fp),$tp
+       sub             %r0,%r0,%r0             ; clear borrow
+       ldw             -4($tp),$ti0
+       addl            $tp,$arrsz,$tp
+L\$sub_pa11
+       ldwx            $idx($np),$hi0
+       subb            $ti0,$hi0,$hi1
+       ldwx            $idx($tp),$ti0
+       addib,<>        4,$idx,L\$sub_pa11
+       stws,ma         $hi1,4($rp)
+
+       subb            $ti0,%r0,$hi1
+       ldo             -4($tp),$tp
+       and             $tp,$hi1,$ap
+       andcm           $rp,$hi1,$bp
+       or              $ap,$bp,$np
+
+       sub             $rp,$arrsz,$rp          ; rewind rp
+       subi            0,$arrsz,$idx
+       ldo             `$LOCALS+32`($fp),$tp
+L\$copy_pa11
+       ldwx            $idx($np),$hi0
+       stws,ma         %r0,4($tp)
+       addib,<>        4,$idx,L\$copy_pa11
+       stws,ma         $hi0,4($rp)     
+
+       nop                                     ; alignment
+L\$done
+___
+}
+\f
+$code.=<<___;
+       ldi             1,%r28                  ; signal "handled"
+       ldo             $FRAME($fp),%sp         ; destroy tp[num+1]
+
+       $POP    `-$FRAME-$SAVED_RP`(%sp),%r2    ; standard epilogue
+       $POP    `-$FRAME+1*$SIZE_T`(%sp),%r4
+       $POP    `-$FRAME+2*$SIZE_T`(%sp),%r5
+       $POP    `-$FRAME+3*$SIZE_T`(%sp),%r6
+       $POP    `-$FRAME+4*$SIZE_T`(%sp),%r7
+       $POP    `-$FRAME+5*$SIZE_T`(%sp),%r8
+       $POP    `-$FRAME+6*$SIZE_T`(%sp),%r9
+       $POP    `-$FRAME+7*$SIZE_T`(%sp),%r10
+L\$abort
+       bv      (%r2)
+       .EXIT
+       $POPMB  -$FRAME(%sp),%r3
+       .PROCEND
+       .STRINGZ "Montgomery Multiplication for PA-RISC, CRYPTOGAMS by <appro\@openssl.org>"
+___
+\f
+# Explicitly encode PA-RISC 2.0 instructions used in this module, so
+# that it can be compiled with .LEVEL 1.0. It should be noted that I
+# wouldn't have to do this, if GNU assembler understood .ALLOW 2.0
+# directive...
+
+my $ldd = sub {
+  my ($mod,$args) = @_;
+  my $orig = "ldd$mod\t$args";
+
+    if ($args =~ /%r([0-9]+)\(%r([0-9]+)\),%r([0-9]+)/)                # format 4
+    {  my $opcode=(0x03<<26)|($2<<21)|($1<<16)|(3<<6)|$3;
+       sprintf "\t.WORD\t0x%08x\t; %s",$opcode,$orig;
+    }
+    elsif ($args =~ /(\-?[0-9]+)\(%r([0-9]+)\),%r([0-9]+)/)    # format 5
+    {  my $opcode=(0x03<<26)|($2<<21)|(1<<12)|(3<<6)|$3;
+       $opcode|=(($1&0xF)<<17)|(($1&0x10)<<12);                # encode offset
+       $opcode|=(1<<5)  if ($mod =~ /^,m/);
+       $opcode|=(1<<13) if ($mod =~ /^,mb/);
+       sprintf "\t.WORD\t0x%08x\t; %s",$opcode,$orig;
+    }
+    else { "\t".$orig; }
+};
+
+my $std = sub {
+  my ($mod,$args) = @_;
+  my $orig = "std$mod\t$args";
+
+    if ($args =~ /%r([0-9]+),(\-?[0-9]+)\(%r([0-9]+)\)/)       # format 6
+    {  my $opcode=(0x03<<26)|($3<<21)|($1<<16)|(1<<12)|(0xB<<6);
+       $opcode|=(($2&0xF)<<1)|(($2&0x10)>>4);                  # encode offset
+       $opcode|=(1<<5)  if ($mod =~ /^,m/);
+       $opcode|=(1<<13) if ($mod =~ /^,mb/);
+       sprintf "\t.WORD\t0x%08x\t; %s",$opcode,$orig;
+    }
+    else { "\t".$orig; }
+};
+
+my $extrd = sub {
+  my ($mod,$args) = @_;
+  my $orig = "extrd$mod\t$args";
+
+    # I only have ",u" completer, it's implicitly encoded...
+    if ($args =~ /%r([0-9]+),([0-9]+),([0-9]+),%r([0-9]+)/)    # format 15
+    {  my $opcode=(0x36<<26)|($1<<21)|($4<<16);
+       my $len=32-$3;
+       $opcode |= (($2&0x20)<<6)|(($2&0x1f)<<5);               # encode pos
+       $opcode |= (($len&0x20)<<7)|($len&0x1f);                # encode len
+       sprintf "\t.WORD\t0x%08x\t; %s",$opcode,$orig;
+    }
+    elsif ($args =~ /%r([0-9]+),%sar,([0-9]+),%r([0-9]+)/)     # format 12
+    {  my $opcode=(0x34<<26)|($1<<21)|($3<<16)|(2<<11)|(1<<9);
+       my $len=32-$2;
+       $opcode |= (($len&0x20)<<3)|($len&0x1f);                # encode len
+       $opcode |= (1<<13) if ($mod =~ /,\**=/);
+       sprintf "\t.WORD\t0x%08x\t; %s",$opcode,$orig;
+    }
+    else { "\t".$orig; }
+};
+
+my $shrpd = sub {
+  my ($mod,$args) = @_;
+  my $orig = "shrpd$mod\t$args";
+
+    if ($args =~ /%r([0-9]+),%r([0-9]+),([0-9]+),%r([0-9]+)/)  # format 14
+    {  my $opcode=(0x34<<26)|($2<<21)|($1<<16)|(1<<10)|$4;
+       my $cpos=63-$3;
+       $opcode |= (($cpos&0x20)<<6)|(($cpos&0x1f)<<5);         # encode sa
+       sprintf "\t.WORD\t0x%08x\t; %s",$opcode,$orig;
+    }
+    else { "\t".$orig; }
+};
+
+my $sub = sub {
+  my ($mod,$args) = @_;
+  my $orig = "sub$mod\t$args";
+
+    if ($mod eq ",db" && $args =~ /%r([0-9]+),%r([0-9]+),%r([0-9]+)/) {
+       my $opcode=(0x02<<26)|($2<<21)|($1<<16)|$3;
+       $opcode|=(1<<10);       # e1
+       $opcode|=(1<<8);        # e2
+       $opcode|=(1<<5);        # d
+       sprintf "\t.WORD\t0x%08x\t; %s",$opcode,$orig
+    }
+    else { "\t".$orig; }
+};
+
+sub assemble {
+  my ($mnemonic,$mod,$args)=@_;
+  my $opcode = eval("\$$mnemonic");
+
+    ref($opcode) eq 'CODE' ? &$opcode($mod,$args) : "\t$mnemonic$mod\t$args";
+}
+
+foreach (split("\n",$code)) {
+       s/\`([^\`]*)\`/eval $1/ge;
+       # flip word order in 64-bit mode...
+       s/(xmpyu\s+)($fai|$fni)([LR])/$1.$2.($3 eq "L"?"R":"L")/e if ($BN_SZ==8);
+       # assemble 2.0 instructions in 32-bit mode...
+       s/^\s+([a-z]+)([\S]*)\s+([\S]*)/&assemble($1,$2,$3)/e if ($BN_SZ==4);
+
+       s/\bbv\b/bve/gm if ($SIZE_T==8);
+
+       print $_,"\n";
+}
+close STDOUT;