projects
/
bootcensus
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
fixed int86 to return correctly from BIOS calls that internally change
[bootcensus]
/
src
/
boot
/
boot2.s
diff --git
a/src/boot/boot2.s
b/src/boot/boot2.s
index
2a163e1
..
7067a8f
100644
(file)
--- a/
src/boot/boot2.s
+++ b/
src/boot/boot2.s
@@
-615,16
+615,20
@@
e820_looptop:
cmp $0x534d4150, %eax
jnz e820_fail
cmp $0x534d4150, %eax
jnz e820_fail
+ # skip areas starting above 4GB as we won't be able to use them
+ cmpl $0, 4(%edi)
+ jnz e820_skip
+
+ # only care for type 1 (usable ram), otherwise ignore
+ cmpl $1, 16(%edi)
+ jnz e820_skip
+
mov buffer, %eax
mov $boot_mem_map, %esi
mov boot_mem_map_size, %ebp
# again, that's [ebp * 8 + esi]
mov %eax, (%esi,%ebp,8)
mov buffer, %eax
mov $boot_mem_map, %esi
mov boot_mem_map_size, %ebp
# again, that's [ebp * 8 + esi]
mov %eax, (%esi,%ebp,8)
- # only care for type 1 (usable ram), otherwise ignore
- cmpl $1, 16(%edi)
- jnz e820_skip
-
# skip areas with 0 size (also clamp size to 4gb)
# test high 32bits
cmpl $0, 12(%edi)
# skip areas with 0 size (also clamp size to 4gb)
# test high 32bits
cmpl $0, 12(%edi)
@@
-759,6
+763,9
@@
rmidt: .short 0x3ff
saved_esp: .long 0
saved_ebp: .long 0
saved_esp: .long 0
saved_ebp: .long 0
+saved_eax: .long 0
+saved_es: .word 0
+saved_ds: .word 0
# drop back to unreal mode to call 16bit interrupt
.global int86
# drop back to unreal mode to call 16bit interrupt
.global int86
@@
-799,17
+806,24
@@
int86:
mov %ebp, saved_ebp
mov 12(%ebp), %esp
popal
mov %ebp, saved_ebp
mov 12(%ebp), %esp
popal
+ popfw
+ pop %es
+ pop %ds
+ # ignore fs and gs for now, don't think I'm going to need them
mov saved_esp, %esp
mov saved_esp, %esp
+ # move to the real-mode stack, accessible from ss=0
+ # just in case the BIOS call screws up our unreal mode
+ mov $0x7be0, %esp
+
# call 16bit interrupt
int_op: int $0
# call 16bit interrupt
int_op: int $0
- mov saved_ebp, %ebp
- mov 12(%ebp), %esp
- add $34, %esp
- pushfw
- pushal
- mov saved_esp, %esp
+ # save all registers that we'll clobber before having the
+ # chance to populate the int86regs structure
+ mov %eax, saved_eax
+ mov %ds, saved_ds
+ mov %es, saved_es
# re-enable protection
mov %cr0, %eax
# re-enable protection
mov %cr0, %eax
@@
-826,6
+840,18
@@
int_op: int $0
mov %ax, %ss
nop
mov %ax, %ss
nop
+ mov saved_ebp, %ebp
+ mov 12(%ebp), %esp
+ add $38, %esp
+ mov saved_ds, %ax
+ pushw %ax
+ mov saved_es, %ax
+ pushw %ax
+ pushfw
+ mov saved_eax, %eax
+ pushal
+ mov saved_esp, %esp
+
# restore 32bit interrupt descriptor table
lidt (saved_idtr)
sti
# restore 32bit interrupt descriptor table
lidt (saved_idtr)
sti
@@
-837,3
+863,5
@@
int_op: int $0
# buffer used by the track loader ... to load tracks.
.align 16
buffer:
# buffer used by the track loader ... to load tracks.
.align 16
buffer:
+ .global low_mem_buffer
+low_mem_buffer: