pkgsrc-Changes-HG archive

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]

[pkgsrc/trunk]: pkgsrc/lang/drscheme Make this compile with gcc-3.3 by replac...



details:   https://anonhg.NetBSD.org/pkgsrc/rev/bc3628451f90
branches:  trunk
changeset: 466244:bc3628451f90
user:      jschauma <jschauma%pkgsrc.org@localhost>
date:      Tue Jan 13 23:40:24 2004 +0000

description:
Make this compile with gcc-3.3 by replacing multi-line strings

"foo
string"

in inline asm with

"foo\n"
"string"

(Noted in agc's latest bulk build.)

diffstat:

 lang/drscheme/distinfo         |    3 +-
 lang/drscheme/patches/patch-aq |  346 +++++++++++++++++++++++++++++++++++++++++
 2 files changed, 348 insertions(+), 1 deletions(-)

diffs (truncated from 364 to 300 lines):

diff -r acb17e5db915 -r bc3628451f90 lang/drscheme/distinfo
--- a/lang/drscheme/distinfo    Tue Jan 13 23:37:27 2004 +0000
+++ b/lang/drscheme/distinfo    Tue Jan 13 23:40:24 2004 +0000
@@ -1,4 +1,4 @@
-$NetBSD: distinfo,v 1.8 2003/05/03 20:57:03 jmc Exp $
+$NetBSD: distinfo,v 1.9 2004/01/13 23:40:24 jschauma Exp $
 
 SHA1 (drscheme/202/plt.src.x.tar.gz) = bb06b6e3f8214db71ebb12e95e744f996622ee06
 Size (drscheme/202/plt.src.x.tar.gz) = 6133563 bytes
@@ -18,3 +18,4 @@
 SHA1 (patch-an) = bca5a79edd48a2670d02496b6da3799c13c0aed6
 SHA1 (patch-ao) = 667593d28a695fd7caee210deb8ee06e72019c40
 SHA1 (patch-ap) = 2395a864a40af34f295e0ed782979d90d75f1ec4
+SHA1 (patch-aq) = 058affb4c3f840f27ce88aba0baaeacba2eb8282
diff -r acb17e5db915 -r bc3628451f90 lang/drscheme/patches/patch-aq
--- /dev/null   Thu Jan 01 00:00:00 1970 +0000
+++ b/lang/drscheme/patches/patch-aq    Tue Jan 13 23:40:24 2004 +0000
@@ -0,0 +1,346 @@
+$NetBSD: patch-aq,v 1.1 2004/01/13 23:40:25 jschauma Exp $
+
+--- ./mzscheme/src/gmp/gmplonglong.h.orig      2001-06-26 12:46:25.000000000 -0400
++++ ./mzscheme/src/gmp/gmplonglong.h   2004-01-13 18:20:47.000000000 -0500
+@@ -231,19 +231,19 @@
+ #define UMUL_TIME 5
+ #else
+ #define umul_ppmm(xh, xl, a, b) \
+-  __asm__ ("%@ Inlined umul_ppmm
+-      mov     %|r0, %2, lsr #16
+-      mov     %|r2, %3, lsr #16
+-      bic     %|r1, %2, %|r0, lsl #16
+-      bic     %|r2, %3, %|r2, lsl #16
+-      mul     %1, %|r1, %|r2
+-      mul     %|r2, %|r0, %|r2
+-      mul     %|r1, %0, %|r1
+-      mul     %0, %|r0, %0
+-      adds    %|r1, %|r2, %|r1
+-      addcs   %0, %0, #65536
+-      adds    %1, %1, %|r1, lsl #16
+-      adc     %0, %0, %|r1, lsr #16"                                  \
++  __asm__ ("%@ Inlined umul_ppmm\n"
++"     mov     %|r0, %2, lsr #16\n"
++"     mov     %|r2, %3, lsr #16\n"
++"     bic     %|r1, %2, %|r0, lsl #16\n"
++"     bic     %|r2, %3, %|r2, lsl #16\n"
++"     mul     %1, %|r1, %|r2\n"
++"     mul     %|r2, %|r0, %|r2\n"
++"     mul     %|r1, %0, %|r1\n"
++"     mul     %0, %|r0, %0\n"
++"     adds    %|r1, %|r2, %|r1\n"
++"     addcs   %0, %0, #65536\n"
++"     adds    %1, %1, %|r1, lsl #16\n"
++"     adc     %0, %0, %|r1, lsr #16"                                  \
+          : "=&r" (xh), "=r" (xl)                                      \
+          : "r" (a), "r" (b)                                           \
+          : "r0", "r1", "r2")
+@@ -360,22 +360,22 @@
+   do {                                                                        \
+     USItype __tmp;                                                    \
+     __asm__ (                                                         \
+-       "ldi           1,%0
+-      extru,=         %1,15,16,%%r0           ; Bits 31..16 zero?
+-      extru,tr        %1,15,16,%1             ; No.  Shift down, skip add.
+-      ldo             16(%0),%0               ; Yes.  Perform add.
+-      extru,=         %1,23,8,%%r0            ; Bits 15..8 zero?
+-      extru,tr        %1,23,8,%1              ; No.  Shift down, skip add.
+-      ldo             8(%0),%0                ; Yes.  Perform add.
+-      extru,=         %1,27,4,%%r0            ; Bits 7..4 zero?
+-      extru,tr        %1,27,4,%1              ; No.  Shift down, skip add.
+-      ldo             4(%0),%0                ; Yes.  Perform add.
+-      extru,=         %1,29,2,%%r0            ; Bits 3..2 zero?
+-      extru,tr        %1,29,2,%1              ; No.  Shift down, skip add.
+-      ldo             2(%0),%0                ; Yes.  Perform add.
+-      extru           %1,30,1,%1              ; Extract bit 1.
+-      sub             %0,%1,%0                ; Subtract it.
+-      " : "=r" (count), "=r" (__tmp) : "1" (x));                      \
++       "ldi           1,%0\n"
++"     extru,=         %1,15,16,%%r0           ; Bits 31..16 zero?\n"
++"     extru,tr        %1,15,16,%1             ; No.  Shift down, skip add.\n"
++"     ldo             16(%0),%0               ; Yes.  Perform add.\n"
++"     extru,=         %1,23,8,%%r0            ; Bits 15..8 zero?\n"
++"     extru,tr        %1,23,8,%1              ; No.  Shift down, skip add.\n"
++"     ldo             8(%0),%0                ; Yes.  Perform add.\n"
++"     extru,=         %1,27,4,%%r0            ; Bits 7..4 zero?\n"
++"     extru,tr        %1,27,4,%1              ; No.  Shift down, skip add.\n"
++"     ldo             4(%0),%0                ; Yes.  Perform add.\n"
++"     extru,=         %1,29,2,%%r0            ; Bits 3..2 zero?\n"
++"     extru,tr        %1,29,2,%1              ; No.  Shift down, skip add.\n"
++"     ldo             2(%0),%0                ; Yes.  Perform add.\n"
++"     extru           %1,30,1,%1              ; Extract bit 1.\n"
++"     sub             %0,%1,%0                ; Subtract it.\n"
++      : "=r" (count), "=r" (__tmp) : "1" (x));                        \
+   } while (0)
+ #endif /* hppa */
+ 
+@@ -529,28 +529,28 @@
+ #else /* for other 68k family members use 16x16->32 multiplication */
+ #define umul_ppmm(xh, xl, a, b) \
+   do { USItype __umul_tmp1, __umul_tmp2;                              \
+-      __asm__ ("| Inlined umul_ppmm
+-      move%.l %5,%3
+-      move%.l %2,%0
+-      move%.w %3,%1
+-      swap    %3
+-      swap    %0
+-      mulu%.w %2,%1
+-      mulu%.w %3,%0
+-      mulu%.w %2,%3
+-      swap    %2
+-      mulu%.w %5,%2
+-      add%.l  %3,%2
+-      jcc     1f
+-      add%.l  %#0x10000,%0
+-1:    move%.l %2,%3
+-      clr%.w  %2
+-      swap    %2
+-      swap    %3
+-      clr%.w  %3
+-      add%.l  %3,%1
+-      addx%.l %2,%0
+-      | End inlined umul_ppmm"                                        \
++      __asm__ ("| Inlined umul_ppmm\n"
++"     move%.l %5,%3\n"
++"     move%.l %2,%0\n"
++"     move%.w %3,%1\n"
++"     swap    %3\n"
++"     swap    %0\n"
++"     mulu%.w %2,%1\n"
++"     mulu%.w %3,%0\n"
++"     mulu%.w %2,%3\n"
++"     swap    %2\n"
++"     mulu%.w %5,%2\n"
++"     add%.l  %3,%2\n"
++"     jcc     1f\n"
++"     add%.l  %#0x10000,%0\n"
++"1:   move%.l %2,%3\n"
++"     clr%.w  %2\n"
++"     swap    %2\n"
++"     swap    %3\n"
++"     clr%.w  %3\n"
++"     add%.l  %3,%1\n"
++"     addx%.l %2,%0\n"
++"     | End inlined umul_ppmm"                                        \
+             : "=&d" ((USItype)(xh)), "=&d" ((USItype)(xl)),           \
+               "=d" (__umul_tmp1), "=&d" (__umul_tmp2)                 \
+             : "%2" ((USItype)(a)), "d" ((USItype)(b)));               \
+@@ -827,26 +827,26 @@
+            "1" ((USItype)(al)), "r" ((USItype)(bl)))
+ #define smul_ppmm(ph, pl, m0, m1) \
+   __asm__ (                                                           \
+-       "s     r2,r2
+-      mts r10,%2
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      m       r2,%3
+-      cas     %0,r2,r0
+-      mfs     r10,%1"                                                 \
++       "s     r2,r2\n"
++"     mts r10,%2\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     m       r2,%3\n"
++"     cas     %0,r2,r0\n"
++"     mfs     r10,%1"                                                 \
+          : "=r" ((USItype)(ph)), "=r" ((USItype)(pl))                 \
+          : "%r" ((USItype)(m0)), "r" ((USItype)(m1))                  \
+          : "r2");                                                     \
+@@ -936,45 +936,45 @@
+   __asm__ ("umul %2,%3,%1;rd %%y,%0" : "=r" (w1), "=r" (w0) : "r" (u), "r" (v))
+ #define UMUL_TIME 5
+ #define udiv_qrnnd(q, r, n1, n0, d) \
+-  __asm__ ("! Inlined udiv_qrnnd
+-      wr      %%g0,%2,%%y     ! Not a delayed write for sparclite
+-      tst     %%g0
+-      divscc  %3,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%%g1
+-      divscc  %%g1,%4,%0
+-      rd      %%y,%1
+-      bl,a 1f
+-      add     %1,%4,%1
+-1:    ! End of inline udiv_qrnnd"                                     \
++  __asm__ ("! Inlined udiv_qrnnd\n"
++"     wr      %%g0,%2,%%y     ! Not a delayed write for sparclite\n"
++"     tst     %%g0\n"
++"     divscc  %3,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%%g1\n"
++"     divscc  %%g1,%4,%0\n"
++"     rd      %%y,%1\n"
++"     bl,a 1f\n"
++"     add     %1,%4,%1\n"
++"1:   ! End of inline udiv_qrnnd"                                     \
+          : "=r" (q), "=r" (r) : "r" (n1), "r" (n0), "rI" (d)
+          : "%g1" __AND_CLOBBER_CC)
+ #define UDIV_TIME 37
+@@ -989,46 +989,46 @@
+ /* Default to sparc v7 versions of umul_ppmm and udiv_qrnnd.  */
+ #ifndef umul_ppmm
+ #define umul_ppmm(w1, w0, u, v) \
+-  __asm__ ("! Inlined umul_ppmm
+-      wr      %%g0,%2,%%y     ! SPARC has 0-3 delay insn after a wr
+-      sra     %3,31,%%g2      ! Don't move this insn
+-      and     %2,%%g2,%%g2    ! Don't move this insn
+-      andcc   %%g0,0,%%g1     ! Don't move this insn
+-      mulscc  %%g1,%3,%%g1
+-      mulscc  %%g1,%3,%%g1
+-      mulscc  %%g1,%3,%%g1
+-      mulscc  %%g1,%3,%%g1
+-      mulscc  %%g1,%3,%%g1
+-      mulscc  %%g1,%3,%%g1
+-      mulscc  %%g1,%3,%%g1
+-      mulscc  %%g1,%3,%%g1
+-      mulscc  %%g1,%3,%%g1
+-      mulscc  %%g1,%3,%%g1
+-      mulscc  %%g1,%3,%%g1
+-      mulscc  %%g1,%3,%%g1
+-      mulscc  %%g1,%3,%%g1
+-      mulscc  %%g1,%3,%%g1



Home | Main Index | Thread Index | Old Index