/* * This file is subject to the terms and conditions of the GNU General Public * License. See the file "COPYING" in the main directory of this archive * for more details. * * Copyright (C) 1994, 1995, 1996, 1998, 1999 by Ralf Baechle * Copyright (C) 1996 David S. Miller (dm@engr.sgi.com) * Copyright (C) 1994, 1995, 1996, by Andreas Busse * Copyright (C) 1999 Silicon Graphics, Inc. */ #include #include #include #include #include #include #include #include #include #include #include .set mips3 #define PF_USEDFPU 0x00100000 /* task used FPU this quantum (SMP) */ #define ST_OFF (KERNEL_STACK_SIZE - 32 - PT_SIZE + PT_STATUS) /* * [jsun] FPU context is saved if and only if the process has used FPU in * the current run (PF_USEDFPU). In any case, the CU1 bit for user space * STATUS register should be 0, so that a process *always* starts its * userland with FPU disabled after each context switch. * * FPU will be enabled as soon as the process accesses FPU again, through * do_cpu() trap. */ /* * task_struct *resume(task_struct *prev, task_struct *next) */ .set noreorder .align 5 LEAF(resume) mfc0 t1, CP0_STATUS sd t1, THREAD_STATUS(a0) cpu_save_nonscratch a0 sd ra, THREAD_REG31(a0) /* * check if we need to save FPU registers */ ld t0, TASK_FLAGS(a0) li t1, PF_USEDFPU and t2, t0, t1 beqz t2, 1f nor t1, zero, t1 /* * clear PF_USEDFPU bit in task flags */ and t0, t0, t1 sd t0, TASK_FLAGS(a0) /* * clear saved user stack CU1 bit */ ld t0, ST_OFF(a0) li t1, ~ST0_CU1 and t0, t0, t1 sd t0, ST_OFF(a0) sll t2, t0, 5 bgez t2, 2f sdc1 $f0, (THREAD_FPU + 0x00)(a0) fpu_save_16odd a0 2: fpu_save_16even a0 t1 # clobbers t1 1: /* * The order of restoring the registers takes care of the race * updating $28, $29 and kernelsp without disabling ints. */ move $28, a1 cpu_restore_nonscratch $28 daddiu a1, $28, KERNEL_STACK_SIZE-32 set_saved_sp a1 t0 mfc0 t1, CP0_STATUS /* Do we really need this? */ li a3, 0xff00 and t1, a3 ld a2, THREAD_STATUS($28) nor a3, $0, a3 and a2, a3 or a2, t1 mtc0 a2, CP0_STATUS jr ra move v0, a0 END(resume) /* * Save a thread's fp context. */ .set noreorder LEAF(_save_fp) mfc0 t0, CP0_STATUS sll t1, t0, 5 bgez t1, 1f # 16 register mode? nop fpu_save_16odd a0 1: fpu_save_16even a0 t1 # clobbers t1 jr ra sdc1 $f0, (THREAD_FPU + 0x00)(a0) END(_save_fp) /* * Restore a thread's fp context. */ LEAF(_restore_fp) mfc0 t0, CP0_STATUS sll t1, t0, 5 bgez t1, 1f # 16 register mode? nop fpu_restore_16odd a0 1: fpu_restore_16even a0, t0 # clobbers t0 jr ra ldc1 $f0, (THREAD_FPU + 0x00)(a0) END(_restore_fp) /* * Load the FPU with signalling NANS. This bit pattern we're using has * the property that no matter whether considered as single or as double * precision represents signaling NANS. * * We initialize fcr31 to rounding to nearest, no exceptions. */ #define FPU_DEFAULT 0x00000000 LEAF(_init_fpu) mfc0 t0, CP0_STATUS li t1, ST0_CU1 or t0, t1 mtc0 t0, CP0_STATUS FPU_ENABLE_HAZARD sll t0, t0, 5 li t1, FPU_DEFAULT ctc1 t1, fcr31 bgez t0, 1f # 16 / 32 register mode? li t0, -1 dmtc1 t0, $f1 dmtc1 t0, $f3 dmtc1 t0, $f5 dmtc1 t0, $f7 dmtc1 t0, $f9 dmtc1 t0, $f11 dmtc1 t0, $f13 dmtc1 t0, $f15 dmtc1 t0, $f17 dmtc1 t0, $f19 dmtc1 t0, $f21 dmtc1 t0, $f23 dmtc1 t0, $f25 dmtc1 t0, $f27 dmtc1 t0, $f29 dmtc1 t0, $f31 1: dmtc1 t0, $f0 dmtc1 t0, $f2 dmtc1 t0, $f4 dmtc1 t0, $f6 dmtc1 t0, $f8 dmtc1 t0, $f10 dmtc1 t0, $f12 dmtc1 t0, $f14 dmtc1 t0, $f16 dmtc1 t0, $f18 dmtc1 t0, $f20 dmtc1 t0, $f22 dmtc1 t0, $f24 dmtc1 t0, $f26 dmtc1 t0, $f28 jr ra dmtc1 t0, $f30 END(_init_fpu)