blob: e146c9731bd9efa69df12cfb09bb8a0c59e8342f [file] [log] [blame]
// Misc utility functions.
//
// Copyright (C) 2008,2009 Kevin O'Connor <kevin@koconnor.net>
//
// This file may be distributed under the terms of the GNU LGPLv3 license.
#include "util.h" // call16
#include "bregs.h" // struct bregs
#include "config.h" // BUILD_STACK_ADDR
/****************************************************************
* 16bit calls
****************************************************************/
// Call a function with a specified register state. Note that on
// return, the interrupt enable/disable flag may be altered.
inline void
call16(struct bregs *callregs)
{
if (!MODESEGMENT && getesp() > BUILD_STACK_ADDR)
panic("call16 with invalid stack\n");
asm volatile(
#if MODE16 == 1
"calll __call16\n"
"cli\n"
"cld"
#else
"calll __call16_from32"
#endif
: "+a" (callregs), "+m" (*callregs)
:
: "ebx", "ecx", "edx", "esi", "edi", "cc", "memory");
}
inline void
call16big(struct bregs *callregs)
{
ASSERT32FLAT();
if (getesp() > BUILD_STACK_ADDR)
panic("call16 with invalid stack\n");
asm volatile(
"calll __call16big_from32"
: "+a" (callregs), "+m" (*callregs)
:
: "ebx", "ecx", "edx", "esi", "edi", "cc", "memory");
}
inline void
__call16_int(struct bregs *callregs, u16 offset)
{
if (MODESEGMENT)
callregs->code.seg = GET_SEG(CS);
else
callregs->code.seg = SEG_BIOS;
callregs->code.offset = offset;
call16(callregs);
}
// 16bit trampoline for enabling irqs from 32bit mode.
ASM16(
" .global trampoline_checkirqs\n"
"trampoline_checkirqs:\n"
" rep ; nop\n"
" lretw"
);
void
check_irqs(void)
{
if (MODE16) {
asm volatile(
"sti\n"
"nop\n"
"rep ; nop\n"
"cli\n"
"cld\n"
: : :"memory");
} else {
extern void trampoline_checkirqs();
struct bregs br;
br.flags = F_IF;
br.code.seg = SEG_BIOS;
br.code.offset = (u32)&trampoline_checkirqs;
call16big(&br);
}
}
/****************************************************************
* String ops
****************************************************************/
// Sum the bytes in the specified area.
u8
checksum_far(u16 buf_seg, void *buf_far, u32 len)
{
SET_SEG(ES, buf_seg);
u32 i;
u8 sum = 0;
for (i=0; i<len; i++)
sum += GET_VAR(ES, ((u8*)buf_far)[i]);
return sum;
}
u8
checksum(void *buf, u32 len)
{
return checksum_far(GET_SEG(SS), buf, len);
}
size_t
strlen(const char *s)
{
if (__builtin_constant_p(s))
return __builtin_strlen(s);
const char *p = s;
while (*p)
p++;
return p-s;
}
// Compare two areas of memory.
int
memcmp(const void *s1, const void *s2, size_t n)
{
while (n) {
if (*(u8*)s1 != *(u8*)s2)
return *(u8*)s1 < *(u8*)s2 ? -1 : 1;
s1++;
s2++;
n--;
}
return 0;
}
// Compare two strings.
int
strcmp(const char *s1, const char *s2)
{
for (;;) {
if (*s1 != *s2)
return *s1 < *s2 ? -1 : 1;
if (! *s1)
return 0;
s1++;
s2++;
}
}
inline void
memset_far(u16 d_seg, void *d_far, u8 c, size_t len)
{
SET_SEG(ES, d_seg);
asm volatile(
"rep stosb %%es:(%%di)"
: "+c"(len), "+D"(d_far)
: "a"(c)
: "cc", "memory");
}
inline void
memset16_far(u16 d_seg, void *d_far, u16 c, size_t len)
{
len /= 2;
SET_SEG(ES, d_seg);
asm volatile(
"rep stosw %%es:(%%di)"
: "+c"(len), "+D"(d_far)
: "a"(c)
: "cc", "memory");
}
void *
memset(void *s, int c, size_t n)
{
while (n)
((char *)s)[--n] = c;
return s;
}
inline void
memcpy_far(u16 d_seg, void *d_far, u16 s_seg, const void *s_far, size_t len)
{
SET_SEG(ES, d_seg);
u16 bkup_ds;
asm volatile(
"movw %%ds, %w0\n"
"movw %w4, %%ds\n"
"rep movsb (%%si),%%es:(%%di)\n"
"movw %w0, %%ds"
: "=&r"(bkup_ds), "+c"(len), "+S"(s_far), "+D"(d_far)
: "r"(s_seg)
: "cc", "memory");
}
void *
#undef memcpy
memcpy(void *d1, const void *s1, size_t len)
#if MODESEGMENT == 0
#define memcpy __builtin_memcpy
#endif
{
SET_SEG(ES, GET_SEG(SS));
void *d = d1;
if (((u32)d1 | (u32)s1 | len) & 3) {
// non-aligned memcpy
asm volatile(
"rep movsb (%%esi),%%es:(%%edi)"
: "+c"(len), "+S"(s1), "+D"(d)
: : "cc", "memory");
return d1;
}
// Common case - use 4-byte copy
len /= 4;
asm volatile(
"rep movsl (%%esi),%%es:(%%edi)"
: "+c"(len), "+S"(s1), "+D"(d)
: : "cc", "memory");
return d1;
}
// Copy from memory mapped IO. IO mem is very slow, so yield
// periodically. 'len' must be 4 byte aligned.
void
iomemcpy(void *d, const void *s, u32 len)
{
yield();
while (len) {
u32 copylen = len;
if (copylen > 2048)
copylen = 2048;
len -= copylen;
copylen /= 4;
asm volatile(
"rep movsl (%%esi),%%es:(%%edi)"
: "+c"(copylen), "+S"(s), "+D"(d)
: : "cc", "memory");
yield();
}
}
void *
memmove(void *d, const void *s, size_t len)
{
if (s >= d)
return memcpy(d, s, len);
d += len-1;
s += len-1;
while (len--) {
*(char*)d = *(char*)s;
d--;
s--;
}
return d;
}
// Copy a string - truncating it if necessary.
char *
strtcpy(char *dest, const char *src, size_t len)
{
char *d = dest;
while (--len && *src != '\0')
*d++ = *src++;
*d = '\0';
return dest;
}
/****************************************************************
* Keyboard calls
****************************************************************/
// Wait for 'usec' microseconds using (with irqs enabled) using int 1586.
void
biosusleep(u32 usec)
{
struct bregs br;
memset(&br, 0, sizeof(br));
br.flags = F_IF;
br.ah = 0x86;
br.cx = usec >> 16;
br.dx = usec;
call16_int(0x15, &br);
}
// See if a keystroke is pending in the keyboard buffer.
static int
check_for_keystroke(void)
{
struct bregs br;
memset(&br, 0, sizeof(br));
br.flags = F_IF;
br.ah = 1;
start_preempt();
call16_int(0x16, &br);
finish_preempt();
return !(br.flags & F_ZF);
}
// Return a keystroke - waiting forever if necessary.
static int
get_raw_keystroke(void)
{
struct bregs br;
memset(&br, 0, sizeof(br));
br.flags = F_IF;
start_preempt();
call16_int(0x16, &br);
finish_preempt();
return br.ah;
}
// Read a keystroke - waiting up to 'msec' milliseconds.
int
get_keystroke(int msec)
{
for (;;) {
if (check_for_keystroke())
return get_raw_keystroke();
if (msec <= 0)
return -1;
start_preempt();
biosusleep(50*1000);
finish_preempt();
msec -= 50;
}
}