diff --git a/contrib/toolchain/gcc/5x/libgcc/Makefile b/contrib/toolchain/gcc/5x/libgcc/Makefile
index 6986d7ee20..590d85e977 100644
--- a/contrib/toolchain/gcc/5x/libgcc/Makefile
+++ b/contrib/toolchain/gcc/5x/libgcc/Makefile
@@ -29,7 +29,7 @@ CFLAGS+= -Wold-style-definition $(CFLAGS_OPT)
DECNUMINC = -Iconfig/libbid -DENABLE_DECIMAL_BID_FORMAT
-INCLUDES = -I. -I../gcc -I../include $(DECNUMINC)
+INCLUDES = -I. -I../gcc -I../include -I$(SDK_DIR)/sources/newlib/libc/include $(DECNUMINC)
gcc_compile = $(CC) $(INCLUDES) $(CFLAGS)
@@ -66,7 +66,10 @@ TPBIT_FUNCS = _pack_tf _unpack_tf _addsub_tf _mul_tf _div_tf \
LIB2ADDEH = unwind-dw2.c unwind-dw2-fde.c unwind-sjlj.c unwind-c.c
LIB2ADDEHSTATIC = $(LIB2ADDEH)
-LIB2ADD = config/i386/gthr-kos32.c
+LIB2ADD = config/i386/gthr-kos32.c \
+ config/i386/kos32-app.c \
+ config/i386/libc-loader.c
+
EH_MODEL = dw2
CUSTOM_CRTSTUFF = yes
@@ -97,12 +100,14 @@ crtfastmath.o: config/i386/crtfastmath.c
$(gcc_compile) -mfxsr -msse -c $<
-LIB1ASMSRC = i386/chkstk.S
-LIB1ASMFUNCS = _chkstk _chkstk_ms
+LIB1ASMSRC = i386/start.S
+LIB1ASMFUNCS = _chkstk _chkstk_ms _start
DFP_ENABLE = true
LIB2ADD += config/i386/cpuinfo.c
+LIB2ADD += config/i386/sfp-exceptions.c
+
softfp_float_modes := tf
softfp_int_modes := si di ti
diff --git a/contrib/toolchain/gcc/5x/libgcc/config/i386/chkstk.S b/contrib/toolchain/gcc/5x/libgcc/config/i386/chkstk.S
deleted file mode 100644
index f2029cce7d..0000000000
--- a/contrib/toolchain/gcc/5x/libgcc/config/i386/chkstk.S
+++ /dev/null
@@ -1,60 +0,0 @@
-
-.section .text
-
-#ifdef L_chkstk
- .global ___chkstk
- .global __alloca
-
-___chkstk:
-__alloca:
- pushl %ecx /* save temp */
- leal 8(%esp), %ecx /* point past return addr */
- subl %eax, %ecx
- cmpl %fs:8, %ecx # check low stack limit
- jb 1f
-
- movl %esp, %eax /* save old stack pointer */
- movl %ecx, %esp /* decrement stack */
- movl (%eax), %ecx /* recover saved temp */
- movl 4(%eax), %eax /* recover return address */
-
- /* Push the return value back. Doing this instead of just
- jumping to %eax preserves the cached call-return stack
- used by most modern processors. */
- pushl %eax
- ret
-1:
- int3 #trap to debugger
- .ascii "Stack overflow"
-#endif
-
-#ifdef L_chkstk_ms
- .global ___chkstk_ms
-
-___chkstk_ms:
- pushl %ecx /* save temp */
- pushl %eax
- cmpl $0x1000, %eax /* > 4k ?*/
- leal 12(%esp), %ecx /* point past return addr */
- jb 2f
-1:
- subl $0x1000, %ecx /* yes, move pointer down 4k*/
- cmpl %fs:8, %ecx /* check low stack limit */
- jb 3f
-
- orl $0x0, (%ecx) /* probe there */
- subl $0x1000, %eax /* decrement count */
- cmpl $0x1000, %eax
- ja 1b /* and do it again */
-
-2:
- subl %eax, %ecx
- orl $0x0, (%ecx) /* less than 4k, just peek here */
-
- popl %eax
- popl %ecx
- ret
-3:
- int3 #trap to debugger
- .ascii "Stack overflow"
-#endif
diff --git a/contrib/toolchain/gcc/5x/libgcc/config/i386/gthr-kos32.c b/contrib/toolchain/gcc/5x/libgcc/config/i386/gthr-kos32.c
index 2de5c4674f..f7d2bf4090 100644
--- a/contrib/toolchain/gcc/5x/libgcc/config/i386/gthr-kos32.c
+++ b/contrib/toolchain/gcc/5x/libgcc/config/i386/gthr-kos32.c
@@ -20,6 +20,7 @@ a copy of the GCC Runtime Library Exception along with this program;
see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
. */
+#include
#include "gthr-kos32.h"
#define FUTEX_INIT 0
@@ -27,27 +28,12 @@ see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
#define FUTEX_WAIT 2
#define FUTEX_WAKE 3
-unsigned int tls_alloc(void);
-int tls_free(unsigned int key);
-void *tls_get(unsigned int key);
-void *tls_set(unsigned int key, void *val);
-
#define exchange_acquire(ptr, new) \
__atomic_exchange_4((ptr), (new), __ATOMIC_ACQUIRE)
#define exchange_release(ptr, new) \
__atomic_exchange_4((ptr), (new), __ATOMIC_RELEASE)
-
-static inline void yield(void)
-{
- __asm__ __volatile__(
- "int $0x40"
- ::"a"(68), "b"(1));
-};
-
-
-
int __gthr_kos32_once (__gthread_once_t *once, void (*func) (void))
{
if (once == NULL || func == NULL)
@@ -95,15 +81,12 @@ int __gthr_kos32_key_delete (__gthread_key_t key)
void* __gthr_kos32_getspecific (__gthread_key_t key)
{
- void *ptr;
- ptr = tls_get(key);
- return ptr;
+ return tls_get(key);
}
int __gthr_kos32_setspecific (__gthread_key_t key, const void *ptr)
{
- tls_set(key, CONST_CAST2(void *, const void *, ptr));
- return 0;
+ return tls_set(key, CONST_CAST2(void *, const void *, ptr));
}
void __gthr_kos32_mutex_init_function (__gthread_mutex_t *mutex)
@@ -151,7 +134,7 @@ int __gthr_kos32_mutex_trylock (__gthread_mutex_t *mutex)
{
int zero = 0;
- return __atomic_compare_exchange_4(&mutex->lock, &zero, 1,0,__ATOMIC_ACQUIRE,__ATOMIC_RELAXED);
+ return !__atomic_compare_exchange_4(&mutex->lock, &zero, 1,0,__ATOMIC_ACQUIRE,__ATOMIC_RELAXED);
}
int __gthr_kos32_mutex_unlock (__gthread_mutex_t *mutex)
@@ -173,79 +156,102 @@ int __gthr_kos32_mutex_unlock (__gthread_mutex_t *mutex)
void __gthr_kos32_recursive_mutex_init_function (__gthread_recursive_mutex_t *mutex)
{
-// mutex->counter = -1;
- mutex->depth = 0;
- mutex->owner = 0;
-// mutex->sema = CreateSemaphoreW (NULL, 0, 65535, NULL);
+ int handle;
+
+ mutex->lock = 0;
+
+ __asm__ volatile(
+ "int $0x40\t"
+ :"=a"(handle)
+ :"a"(77),"b"(FUTEX_INIT),"c"(mutex));
+ mutex->handle = handle;
+
+ mutex->depth = 0;
+ mutex->owner = 0;
}
-#if 0
-int
-__gthr_win32_recursive_mutex_lock (__gthread_recursive_mutex_t *mutex)
+int __gthr_kos32_recursive_mutex_lock (__gthread_recursive_mutex_t *mutex)
{
- DWORD me = GetCurrentThreadId();
- if (InterlockedIncrement (&mutex->counter) == 0)
+ int tmp;
+
+ unsigned long me = (unsigned long)tls_get(TLS_KEY_LOW_STACK);
+
+ if( __sync_fetch_and_add(&mutex->lock, 1) == 0)
{
- mutex->depth = 1;
- mutex->owner = me;
+ mutex->depth = 1;
+ mutex->owner = me;
+ return 0;
}
- else if (mutex->owner == me)
+ else if (mutex->owner == me)
{
- InterlockedDecrement (&mutex->counter);
- ++(mutex->depth);
+ __sync_fetch_and_sub(&mutex->lock, 1);
+ ++(mutex->depth);
}
- else if (WaitForSingleObject (mutex->sema, INFINITE) == WAIT_OBJECT_0)
+ else while (exchange_acquire (&mutex->lock, 2) != 0)
{
- mutex->depth = 1;
- mutex->owner = me;
- }
- else
- {
- /* WaitForSingleObject returns WAIT_FAILED, and we can only do
- some best-effort cleanup here. */
- InterlockedDecrement (&mutex->counter);
- return 1;
- }
- return 0;
+ __asm__ volatile(
+ "int $0x40\t\n"
+ :"=a"(tmp)
+ :"a"(77),"b"(FUTEX_WAIT),
+ "c"(mutex->handle),"d"(2),"S"(0));
+ mutex->depth = 1;
+ mutex->owner = me;
+ };
+
+ return 0;
}
-int
-__gthr_win32_recursive_mutex_trylock (__gthread_recursive_mutex_t *mutex)
+int __gthr_kos32_recursive_mutex_trylock (__gthread_recursive_mutex_t *mutex)
{
- DWORD me = GetCurrentThreadId();
- if (__GTHR_W32_InterlockedCompareExchange (&mutex->counter, 0, -1) < 0)
+ unsigned long me = (unsigned long)tls_get(TLS_KEY_LOW_STACK);
+ int zero = 0;
+
+ if(__atomic_compare_exchange_4(&mutex->lock, &zero, 1,0,__ATOMIC_ACQUIRE,__ATOMIC_RELAXED))
{
- mutex->depth = 1;
- mutex->owner = me;
+ mutex->depth = 1;
+ mutex->owner = me;
}
- else if (mutex->owner == me)
- ++(mutex->depth);
- else
- return 1;
+ else if (mutex->owner == me)
+ ++(mutex->depth);
+ else
+ return 1;
- return 0;
+ return 0;
}
-int
-__gthr_win32_recursive_mutex_unlock (__gthread_recursive_mutex_t *mutex)
+int __gthr_kos32_recursive_mutex_unlock (__gthread_recursive_mutex_t *mutex)
{
- --(mutex->depth);
- if (mutex->depth == 0)
+ --(mutex->depth);
+
+ if (mutex->depth == 0)
{
- mutex->owner = 0;
+ int prev;
- if (InterlockedDecrement (&mutex->counter) >= 0)
- return ReleaseSemaphore (mutex->sema, 1, NULL) ? 0 : 1;
- }
+ prev = exchange_release (&mutex->lock, 0);
- return 0;
+ if (prev != 1)
+ {
+ __asm__ volatile(
+ "int $0x40\t"
+ :"=a"(prev)
+ :"a"(77),"b"(FUTEX_WAKE),
+ "c"(mutex->handle),"d"(1));
+ };
+ mutex->owner = 0;
+ };
+
+ return 0;
}
-int
-__gthr_win32_recursive_mutex_destroy (__gthread_recursive_mutex_t *mutex)
+int __gthr_kos32_recursive_mutex_destroy (__gthread_recursive_mutex_t *mutex)
{
- CloseHandle ((HANDLE) mutex->sema);
- return 0;
+ int retval;
+
+ __asm__ volatile(
+ "int $0x40\t"
+ :"=a"(retval)
+ :"a"(77),"b"(FUTEX_DESTROY),"c"(mutex->handle));
+
+ return 0;
}
-#endif
diff --git a/contrib/toolchain/gcc/5x/libgcc/config/i386/kos32-app.c b/contrib/toolchain/gcc/5x/libgcc/config/i386/kos32-app.c
new file mode 100644
index 0000000000..94b5c0dde5
--- /dev/null
+++ b/contrib/toolchain/gcc/5x/libgcc/config/i386/kos32-app.c
@@ -0,0 +1,31 @@
+/*
+ * This file has no copyright assigned and is placed in the Public Domain.
+ * This file is a part of the kos32-runtime package.
+ * No warranty is given; refer to the file DISCLAIMER within the package.
+ *
+ * Source code for the startup proceedures used by all programs. This code
+ * is compiled to make crt1.o, which should be located in the library path.
+ *
+ */
+
+void *load_libc(void);
+void* get_entry_point(void *raw);
+
+void __attribute__((noreturn))
+__crt_startup (void)
+{
+ void __attribute__((noreturn)) (*entry)(void *img);
+ void *img;
+
+ img = load_libc();
+
+ if(!img)
+ {
+ asm ("int $0x40" ::"a"(-1));
+ };
+
+ entry = get_entry_point(img);
+ entry(img);
+}
+
+
diff --git a/contrib/toolchain/gcc/5x/libgcc/config/i386/libc-loader.c b/contrib/toolchain/gcc/5x/libgcc/config/i386/libc-loader.c
new file mode 100644
index 0000000000..cd393bc8b9
--- /dev/null
+++ b/contrib/toolchain/gcc/5x/libgcc/config/i386/libc-loader.c
@@ -0,0 +1,202 @@
+#include
+#include
+
+#include "pe.h"
+
+#define unlikely(x) __builtin_expect(!!(x), 0)
+
+void* load_libc(void);
+
+static inline void sec_copy(void *dst, void *src, size_t len)
+{
+ __asm__ __volatile__ (
+ "shrl $2, %%ecx \n\t"
+ "rep movsl"
+ :
+ :"c"(len),"S"(src),"D"(dst)
+ :"cc");
+ __asm__ __volatile__ (
+ ""
+ :::"ecx","esi","edi");
+};
+
+static inline int IsPowerOf2(uint32_t val)
+{
+ if(val == 0)
+ return 0;
+ return (val & (val - 1)) == 0;
+}
+
+static int validate_pe(void *raw, size_t raw_size)
+{
+ PIMAGE_DOS_HEADER dos;
+ PIMAGE_NT_HEADERS32 nt;
+
+ dos = (PIMAGE_DOS_HEADER)raw;
+
+ if( !raw || raw_size < sizeof(IMAGE_DOS_HEADER) )
+ return 0;
+
+ if( dos->e_magic != IMAGE_DOS_SIGNATURE || dos->e_lfanew <= 0)
+ return 0;
+
+ nt = MakePtr( PIMAGE_NT_HEADERS32, dos, dos->e_lfanew);
+
+ if( (uint32_t)nt < (uint32_t)raw)
+ return 0;
+
+ if(nt->Signature != IMAGE_NT_SIGNATURE)
+ return 0;
+
+ if(nt->FileHeader.Machine != IMAGE_FILE_MACHINE_I386)
+ return 0;
+
+ if(nt->OptionalHeader.Magic != IMAGE_NT_OPTIONAL_HDR32_MAGIC)
+ return 0;
+
+ if(nt->OptionalHeader.SectionAlignment < 4096)
+ {
+ if(nt->OptionalHeader.FileAlignment != nt->OptionalHeader.SectionAlignment)
+ return 0;
+ }
+ else if(nt->OptionalHeader.SectionAlignment < nt->OptionalHeader.FileAlignment)
+ return 0;
+
+ if(!IsPowerOf2(nt->OptionalHeader.SectionAlignment) ||
+ !IsPowerOf2(nt->OptionalHeader.FileAlignment))
+ return 0;
+
+ if(nt->FileHeader.NumberOfSections > 96)
+ return 0;
+
+ return 1;
+}
+
+static void* create_image(void *raw)
+{
+ PIMAGE_DOS_HEADER dos;
+ PIMAGE_NT_HEADERS32 nt;
+ PIMAGE_SECTION_HEADER img_sec;
+
+ void *img_base;
+ unsigned int i;
+
+ dos = (PIMAGE_DOS_HEADER)raw;
+ nt = MakePtr( PIMAGE_NT_HEADERS32, dos, dos->e_lfanew);
+
+ img_base = user_alloc(nt->OptionalHeader.SizeOfImage);
+
+ if(unlikely(img_base == NULL))
+ return 0;
+
+ sec_copy(img_base, raw, nt->OptionalHeader.SizeOfHeaders);
+
+ img_sec = MakePtr(PIMAGE_SECTION_HEADER, nt, sizeof(IMAGE_NT_HEADERS32));
+
+ for(i=0; i< nt->FileHeader.NumberOfSections; i++)
+ {
+ void *src_ptr;
+ void *dest_ptr;
+
+ if ( img_sec->SizeOfRawData && img_sec->PointerToRawData )
+ {
+ src_ptr = MakePtr(void*, raw, img_sec->PointerToRawData);
+ dest_ptr = MakePtr(void*, img_base, img_sec->VirtualAddress);
+ sec_copy(dest_ptr, src_ptr, img_sec->SizeOfRawData);
+ };
+
+ img_sec++;
+ };
+
+ if(nt->OptionalHeader.DataDirectory[5].Size)
+ {
+ PIMAGE_BASE_RELOCATION reloc;
+
+ uint32_t delta = (uint32_t)img_base - nt->OptionalHeader.ImageBase;
+
+ reloc = MakePtr(PIMAGE_BASE_RELOCATION, img_base,
+ nt->OptionalHeader.DataDirectory[5].VirtualAddress);
+
+ while ( reloc->SizeOfBlock != 0 )
+ {
+ uint32_t cnt;
+ uint16_t *entry;
+ uint16_t reltype;
+ uint32_t offs;
+
+ cnt = (reloc->SizeOfBlock - sizeof(*reloc))/sizeof(uint16_t);
+ entry = MakePtr( uint16_t*, reloc, sizeof(*reloc) );
+
+ for ( i=0; i < cnt; i++ )
+ {
+ uint16_t *p16;
+ uint32_t *p32;
+
+ reltype = (*entry & 0xF000) >> 12;
+ offs = (*entry & 0x0FFF) + reloc->VirtualAddress;
+ switch(reltype)
+ {
+ case 1:
+ p16 = MakePtr(uint16_t*, img_base, offs);
+ *p16+= (uint16_t)(delta>>16);
+ break;
+ case 2:
+ p16 = MakePtr(uint16_t*, img_base, offs);
+ *p16+= (uint16_t)delta;
+ break;
+ case 3:
+ p32 = MakePtr(uint32_t*, img_base, offs);
+ *p32+= delta;
+ }
+ entry++;
+ }
+ reloc = MakePtr(PIMAGE_BASE_RELOCATION, reloc,reloc->SizeOfBlock);
+ };
+// printf("unmap base %p offset %x %d page(s)\n",
+// img_base,
+// nt->OptionalHeader.DataDirectory[5].VirtualAddress,
+// (nt->OptionalHeader.DataDirectory[5].Size+4095)>>12);
+
+ user_unmap(img_base,nt->OptionalHeader.DataDirectory[5].VirtualAddress,
+ nt->OptionalHeader.DataDirectory[5].Size);
+ };
+ return img_base;
+};
+
+void* get_entry_point(void *raw)
+{
+ PIMAGE_DOS_HEADER dos;
+ PIMAGE_NT_HEADERS32 nt;
+
+ dos = (PIMAGE_DOS_HEADER)raw;
+ nt = MakePtr( PIMAGE_NT_HEADERS32, dos, dos->e_lfanew);
+
+ return MakePtr(void*, raw, nt->OptionalHeader.AddressOfEntryPoint);
+};
+
+
+void* load_libc(void)
+{
+ void *raw_img;
+ size_t raw_size;
+ void *img_base = NULL;
+ ufile_t uf;
+
+ uf = load_file("/kolibrios/lib/libc.dll");
+
+ raw_img = uf.data;
+ raw_size = uf.size;
+
+ if(raw_img == NULL)
+ return NULL;
+
+ if(validate_pe(raw_img, raw_size) != 0)
+ img_base = create_image(raw_img);
+
+ user_free(raw_img);
+
+ return img_base;
+
+}
+
+
diff --git a/contrib/toolchain/gcc/5x/libgcc/config/i386/list.h b/contrib/toolchain/gcc/5x/libgcc/config/i386/list.h
new file mode 100644
index 0000000000..3a43161306
--- /dev/null
+++ b/contrib/toolchain/gcc/5x/libgcc/config/i386/list.h
@@ -0,0 +1,707 @@
+#ifndef _LINUX_LIST_H
+#define _LINUX_LIST_H
+
+
+/*
+ * Simple doubly linked list implementation.
+ *
+ * Some of the internal functions ("__xxx") are useful when
+ * manipulating whole lists rather than single entries, as
+ * sometimes we already know the next/prev entries and we can
+ * generate better code by using them directly rather than
+ * using the generic single-entry routines.
+ */
+
+/**
+ * container_of - cast a member of a structure out to the containing structure
+ * @ptr: the pointer to the member.
+ * @type: the type of the container struct this is embedded in.
+ * @member: the name of the member within the struct.
+ *
+ */
+#define container_of(ptr, type, member) ({ \
+ const typeof( ((type *)0)->member ) *__mptr = (ptr); \
+ (type *)( (char *)__mptr - offsetof(type,member) );})
+
+#define LIST_POISON1 ((struct list_head*)0xFFFF0100)
+#define LIST_POISON2 ((struct list_head*)0xFFFF0200)
+
+#define prefetch(x) __builtin_prefetch(x)
+
+struct list_head {
+ struct list_head *next, *prev;
+};
+
+#define LIST_HEAD_INIT(name) { &(name), &(name) }
+
+#define LIST_HEAD(name) \
+ struct list_head name = LIST_HEAD_INIT(name)
+
+static inline void INIT_LIST_HEAD(struct list_head *list)
+{
+ list->next = list;
+ list->prev = list;
+}
+
+/*
+ * Insert a new entry between two known consecutive entries.
+ *
+ * This is only for internal list manipulation where we know
+ * the prev/next entries already!
+ */
+#ifndef CONFIG_DEBUG_LIST
+static inline void __list_add(struct list_head *new,
+ struct list_head *prev,
+ struct list_head *next)
+{
+ next->prev = new;
+ new->next = next;
+ new->prev = prev;
+ prev->next = new;
+}
+#else
+extern void __list_add(struct list_head *new,
+ struct list_head *prev,
+ struct list_head *next);
+#endif
+
+/**
+ * list_add - add a new entry
+ * @new: new entry to be added
+ * @head: list head to add it after
+ *
+ * Insert a new entry after the specified head.
+ * This is good for implementing stacks.
+ */
+static inline void list_add(struct list_head *new, struct list_head *head)
+{
+ __list_add(new, head, head->next);
+}
+
+
+/**
+ * list_add_tail - add a new entry
+ * @new: new entry to be added
+ * @head: list head to add it before
+ *
+ * Insert a new entry before the specified head.
+ * This is useful for implementing queues.
+ */
+static inline void list_add_tail(struct list_head *new, struct list_head *head)
+{
+ __list_add(new, head->prev, head);
+}
+
+/*
+ * Delete a list entry by making the prev/next entries
+ * point to each other.
+ *
+ * This is only for internal list manipulation where we know
+ * the prev/next entries already!
+ */
+static inline void __list_del(struct list_head * prev, struct list_head * next)
+{
+ next->prev = prev;
+ prev->next = next;
+}
+
+/**
+ * list_del - deletes entry from list.
+ * @entry: the element to delete from the list.
+ * Note: list_empty() on entry does not return true after this, the entry is
+ * in an undefined state.
+ */
+#ifndef CONFIG_DEBUG_LIST
+static inline void list_del(struct list_head *entry)
+{
+ __list_del(entry->prev, entry->next);
+ entry->next = LIST_POISON1;
+ entry->prev = LIST_POISON2;
+}
+#else
+extern void list_del(struct list_head *entry);
+#endif
+
+/**
+ * list_replace - replace old entry by new one
+ * @old : the element to be replaced
+ * @new : the new element to insert
+ *
+ * If @old was empty, it will be overwritten.
+ */
+static inline void list_replace(struct list_head *old,
+ struct list_head *new)
+{
+ new->next = old->next;
+ new->next->prev = new;
+ new->prev = old->prev;
+ new->prev->next = new;
+}
+
+static inline void list_replace_init(struct list_head *old,
+ struct list_head *new)
+{
+ list_replace(old, new);
+ INIT_LIST_HEAD(old);
+}
+
+/**
+ * list_del_init - deletes entry from list and reinitialize it.
+ * @entry: the element to delete from the list.
+ */
+static inline void list_del_init(struct list_head *entry)
+{
+ __list_del(entry->prev, entry->next);
+ INIT_LIST_HEAD(entry);
+}
+
+/**
+ * list_move - delete from one list and add as another's head
+ * @list: the entry to move
+ * @head: the head that will precede our entry
+ */
+static inline void list_move(struct list_head *list, struct list_head *head)
+{
+ __list_del(list->prev, list->next);
+ list_add(list, head);
+}
+
+/**
+ * list_move_tail - delete from one list and add as another's tail
+ * @list: the entry to move
+ * @head: the head that will follow our entry
+ */
+static inline void list_move_tail(struct list_head *list,
+ struct list_head *head)
+{
+ __list_del(list->prev, list->next);
+ list_add_tail(list, head);
+}
+
+/**
+ * list_is_last - tests whether @list is the last entry in list @head
+ * @list: the entry to test
+ * @head: the head of the list
+ */
+static inline int list_is_last(const struct list_head *list,
+ const struct list_head *head)
+{
+ return list->next == head;
+}
+
+/**
+ * list_empty - tests whether a list is empty
+ * @head: the list to test.
+ */
+static inline int list_empty(const struct list_head *head)
+{
+ return head->next == head;
+}
+
+/**
+ * list_empty_careful - tests whether a list is empty and not being modified
+ * @head: the list to test
+ *
+ * Description:
+ * tests whether a list is empty _and_ checks that no other CPU might be
+ * in the process of modifying either member (next or prev)
+ *
+ * NOTE: using list_empty_careful() without synchronization
+ * can only be safe if the only activity that can happen
+ * to the list entry is list_del_init(). Eg. it cannot be used
+ * if another CPU could re-list_add() it.
+ */
+static inline int list_empty_careful(const struct list_head *head)
+{
+ struct list_head *next = head->next;
+ return (next == head) && (next == head->prev);
+}
+
+/**
+ * list_is_singular - tests whether a list has just one entry.
+ * @head: the list to test.
+ */
+static inline int list_is_singular(const struct list_head *head)
+{
+ return !list_empty(head) && (head->next == head->prev);
+}
+
+static inline void __list_cut_position(struct list_head *list,
+ struct list_head *head, struct list_head *entry)
+{
+ struct list_head *new_first = entry->next;
+ list->next = head->next;
+ list->next->prev = list;
+ list->prev = entry;
+ entry->next = list;
+ head->next = new_first;
+ new_first->prev = head;
+}
+
+/**
+ * list_cut_position - cut a list into two
+ * @list: a new list to add all removed entries
+ * @head: a list with entries
+ * @entry: an entry within head, could be the head itself
+ * and if so we won't cut the list
+ *
+ * This helper moves the initial part of @head, up to and
+ * including @entry, from @head to @list. You should
+ * pass on @entry an element you know is on @head. @list
+ * should be an empty list or a list you do not care about
+ * losing its data.
+ *
+ */
+static inline void list_cut_position(struct list_head *list,
+ struct list_head *head, struct list_head *entry)
+{
+ if (list_empty(head))
+ return;
+ if (list_is_singular(head) &&
+ (head->next != entry && head != entry))
+ return;
+ if (entry == head)
+ INIT_LIST_HEAD(list);
+ else
+ __list_cut_position(list, head, entry);
+}
+
+static inline void __list_splice(const struct list_head *list,
+ struct list_head *prev,
+ struct list_head *next)
+{
+ struct list_head *first = list->next;
+ struct list_head *last = list->prev;
+
+ first->prev = prev;
+ prev->next = first;
+
+ last->next = next;
+ next->prev = last;
+}
+
+/**
+ * list_splice - join two lists, this is designed for stacks
+ * @list: the new list to add.
+ * @head: the place to add it in the first list.
+ */
+static inline void list_splice(const struct list_head *list,
+ struct list_head *head)
+{
+ if (!list_empty(list))
+ __list_splice(list, head, head->next);
+}
+
+/**
+ * list_splice_tail - join two lists, each list being a queue
+ * @list: the new list to add.
+ * @head: the place to add it in the first list.
+ */
+static inline void list_splice_tail(struct list_head *list,
+ struct list_head *head)
+{
+ if (!list_empty(list))
+ __list_splice(list, head->prev, head);
+}
+
+/**
+ * list_splice_init - join two lists and reinitialise the emptied list.
+ * @list: the new list to add.
+ * @head: the place to add it in the first list.
+ *
+ * The list at @list is reinitialised
+ */
+static inline void list_splice_init(struct list_head *list,
+ struct list_head *head)
+{
+ if (!list_empty(list)) {
+ __list_splice(list, head, head->next);
+ INIT_LIST_HEAD(list);
+ }
+}
+
+/**
+ * list_splice_tail_init - join two lists and reinitialise the emptied list
+ * @list: the new list to add.
+ * @head: the place to add it in the first list.
+ *
+ * Each of the lists is a queue.
+ * The list at @list is reinitialised
+ */
+static inline void list_splice_tail_init(struct list_head *list,
+ struct list_head *head)
+{
+ if (!list_empty(list)) {
+ __list_splice(list, head->prev, head);
+ INIT_LIST_HEAD(list);
+ }
+}
+
+/**
+ * list_entry - get the struct for this entry
+ * @ptr: the &struct list_head pointer.
+ * @type: the type of the struct this is embedded in.
+ * @member: the name of the list_struct within the struct.
+ */
+#define list_entry(ptr, type, member) \
+ container_of(ptr, type, member)
+
+/**
+ * list_first_entry - get the first element from a list
+ * @ptr: the list head to take the element from.
+ * @type: the type of the struct this is embedded in.
+ * @member: the name of the list_struct within the struct.
+ *
+ * Note, that list is expected to be not empty.
+ */
+#define list_first_entry(ptr, type, member) \
+ list_entry((ptr)->next, type, member)
+
+/**
+ * list_for_each - iterate over a list
+ * @pos: the &struct list_head to use as a loop cursor.
+ * @head: the head for your list.
+ */
+#define list_for_each(pos, head) \
+ for (pos = (head)->next; prefetch(pos->next), pos != (head); \
+ pos = pos->next)
+
+/**
+ * __list_for_each - iterate over a list
+ * @pos: the &struct list_head to use as a loop cursor.
+ * @head: the head for your list.
+ *
+ * This variant differs from list_for_each() in that it's the
+ * simplest possible list iteration code, no prefetching is done.
+ * Use this for code that knows the list to be very short (empty
+ * or 1 entry) most of the time.
+ */
+#define __list_for_each(pos, head) \
+ for (pos = (head)->next; pos != (head); pos = pos->next)
+
+/**
+ * list_for_each_prev - iterate over a list backwards
+ * @pos: the &struct list_head to use as a loop cursor.
+ * @head: the head for your list.
+ */
+#define list_for_each_prev(pos, head) \
+ for (pos = (head)->prev; prefetch(pos->prev), pos != (head); \
+ pos = pos->prev)
+
+/**
+ * list_for_each_safe - iterate over a list safe against removal of list entry
+ * @pos: the &struct list_head to use as a loop cursor.
+ * @n: another &struct list_head to use as temporary storage
+ * @head: the head for your list.
+ */
+#define list_for_each_safe(pos, n, head) \
+ for (pos = (head)->next, n = pos->next; pos != (head); \
+ pos = n, n = pos->next)
+
+/**
+ * list_for_each_prev_safe - iterate over a list backwards safe against removal of list entry
+ * @pos: the &struct list_head to use as a loop cursor.
+ * @n: another &struct list_head to use as temporary storage
+ * @head: the head for your list.
+ */
+#define list_for_each_prev_safe(pos, n, head) \
+ for (pos = (head)->prev, n = pos->prev; \
+ prefetch(pos->prev), pos != (head); \
+ pos = n, n = pos->prev)
+
+/**
+ * list_for_each_entry - iterate over list of given type
+ * @pos: the type * to use as a loop cursor.
+ * @head: the head for your list.
+ * @member: the name of the list_struct within the struct.
+ */
+#define list_for_each_entry(pos, head, member) \
+ for (pos = list_entry((head)->next, typeof(*pos), member); \
+ prefetch(pos->member.next), &pos->member != (head); \
+ pos = list_entry(pos->member.next, typeof(*pos), member))
+
+/**
+ * list_for_each_entry_reverse - iterate backwards over list of given type.
+ * @pos: the type * to use as a loop cursor.
+ * @head: the head for your list.
+ * @member: the name of the list_struct within the struct.
+ */
+#define list_for_each_entry_reverse(pos, head, member) \
+ for (pos = list_entry((head)->prev, typeof(*pos), member); \
+ prefetch(pos->member.prev), &pos->member != (head); \
+ pos = list_entry(pos->member.prev, typeof(*pos), member))
+
+/**
+ * list_prepare_entry - prepare a pos entry for use in list_for_each_entry_continue()
+ * @pos: the type * to use as a start point
+ * @head: the head of the list
+ * @member: the name of the list_struct within the struct.
+ *
+ * Prepares a pos entry for use as a start point in list_for_each_entry_continue().
+ */
+#define list_prepare_entry(pos, head, member) \
+ ((pos) ? : list_entry(head, typeof(*pos), member))
+
+/**
+ * list_for_each_entry_continue - continue iteration over list of given type
+ * @pos: the type * to use as a loop cursor.
+ * @head: the head for your list.
+ * @member: the name of the list_struct within the struct.
+ *
+ * Continue to iterate over list of given type, continuing after
+ * the current position.
+ */
+#define list_for_each_entry_continue(pos, head, member) \
+ for (pos = list_entry(pos->member.next, typeof(*pos), member); \
+ prefetch(pos->member.next), &pos->member != (head); \
+ pos = list_entry(pos->member.next, typeof(*pos), member))
+
+/**
+ * list_for_each_entry_continue_reverse - iterate backwards from the given point
+ * @pos: the type * to use as a loop cursor.
+ * @head: the head for your list.
+ * @member: the name of the list_struct within the struct.
+ *
+ * Start to iterate over list of given type backwards, continuing after
+ * the current position.
+ */
+#define list_for_each_entry_continue_reverse(pos, head, member) \
+ for (pos = list_entry(pos->member.prev, typeof(*pos), member); \
+ prefetch(pos->member.prev), &pos->member != (head); \
+ pos = list_entry(pos->member.prev, typeof(*pos), member))
+
+/**
+ * list_for_each_entry_from - iterate over list of given type from the current point
+ * @pos: the type * to use as a loop cursor.
+ * @head: the head for your list.
+ * @member: the name of the list_struct within the struct.
+ *
+ * Iterate over list of given type, continuing from current position.
+ */
+#define list_for_each_entry_from(pos, head, member) \
+ for (; prefetch(pos->member.next), &pos->member != (head); \
+ pos = list_entry(pos->member.next, typeof(*pos), member))
+
+/**
+ * list_for_each_entry_safe - iterate over list of given type safe against removal of list entry
+ * @pos: the type * to use as a loop cursor.
+ * @n: another type * to use as temporary storage
+ * @head: the head for your list.
+ * @member: the name of the list_struct within the struct.
+ */
+#define list_for_each_entry_safe(pos, n, head, member) \
+ for (pos = list_entry((head)->next, typeof(*pos), member), \
+ n = list_entry(pos->member.next, typeof(*pos), member); \
+ &pos->member != (head); \
+ pos = n, n = list_entry(n->member.next, typeof(*n), member))
+
+/**
+ * list_for_each_entry_safe_continue
+ * @pos: the type * to use as a loop cursor.
+ * @n: another type * to use as temporary storage
+ * @head: the head for your list.
+ * @member: the name of the list_struct within the struct.
+ *
+ * Iterate over list of given type, continuing after current point,
+ * safe against removal of list entry.
+ */
+#define list_for_each_entry_safe_continue(pos, n, head, member) \
+ for (pos = list_entry(pos->member.next, typeof(*pos), member), \
+ n = list_entry(pos->member.next, typeof(*pos), member); \
+ &pos->member != (head); \
+ pos = n, n = list_entry(n->member.next, typeof(*n), member))
+
+/**
+ * list_for_each_entry_safe_from
+ * @pos: the type * to use as a loop cursor.
+ * @n: another type * to use as temporary storage
+ * @head: the head for your list.
+ * @member: the name of the list_struct within the struct.
+ *
+ * Iterate over list of given type from current point, safe against
+ * removal of list entry.
+ */
+#define list_for_each_entry_safe_from(pos, n, head, member) \
+ for (n = list_entry(pos->member.next, typeof(*pos), member); \
+ &pos->member != (head); \
+ pos = n, n = list_entry(n->member.next, typeof(*n), member))
+
+/**
+ * list_for_each_entry_safe_reverse
+ * @pos: the type * to use as a loop cursor.
+ * @n: another type * to use as temporary storage
+ * @head: the head for your list.
+ * @member: the name of the list_struct within the struct.
+ *
+ * Iterate backwards over list of given type, safe against removal
+ * of list entry.
+ */
+#define list_for_each_entry_safe_reverse(pos, n, head, member) \
+ for (pos = list_entry((head)->prev, typeof(*pos), member), \
+ n = list_entry(pos->member.prev, typeof(*pos), member); \
+ &pos->member != (head); \
+ pos = n, n = list_entry(n->member.prev, typeof(*n), member))
+
+/*
+ * Double linked lists with a single pointer list head.
+ * Mostly useful for hash tables where the two pointer list head is
+ * too wasteful.
+ * You lose the ability to access the tail in O(1).
+ */
+
+struct hlist_head {
+ struct hlist_node *first;
+};
+
+struct hlist_node {
+ struct hlist_node *next, **pprev;
+};
+
+#define HLIST_HEAD_INIT { .first = NULL }
+#define HLIST_HEAD(name) struct hlist_head name = { .first = NULL }
+#define INIT_HLIST_HEAD(ptr) ((ptr)->first = NULL)
+static inline void INIT_HLIST_NODE(struct hlist_node *h)
+{
+ h->next = NULL;
+ h->pprev = NULL;
+}
+
+static inline int hlist_unhashed(const struct hlist_node *h)
+{
+ return !h->pprev;
+}
+
+static inline int hlist_empty(const struct hlist_head *h)
+{
+ return !h->first;
+}
+
+static inline void __hlist_del(struct hlist_node *n)
+{
+ struct hlist_node *next = n->next;
+ struct hlist_node **pprev = n->pprev;
+ *pprev = next;
+ if (next)
+ next->pprev = pprev;
+}
+
+static inline void hlist_del(struct hlist_node *n)
+{
+ __hlist_del(n);
+ n->next = (struct hlist_node*)LIST_POISON1;
+ n->pprev = (struct hlist_node**)LIST_POISON2;
+}
+
+static inline void hlist_del_init(struct hlist_node *n)
+{
+ if (!hlist_unhashed(n)) {
+ __hlist_del(n);
+ INIT_HLIST_NODE(n);
+ }
+}
+
+static inline void hlist_add_head(struct hlist_node *n, struct hlist_head *h)
+{
+ struct hlist_node *first = h->first;
+ n->next = first;
+ if (first)
+ first->pprev = &n->next;
+ h->first = n;
+ n->pprev = &h->first;
+}
+
+/* next must be != NULL */
+static inline void hlist_add_before(struct hlist_node *n,
+ struct hlist_node *next)
+{
+ n->pprev = next->pprev;
+ n->next = next;
+ next->pprev = &n->next;
+ *(n->pprev) = n;
+}
+
+static inline void hlist_add_after(struct hlist_node *n,
+ struct hlist_node *next)
+{
+ next->next = n->next;
+ n->next = next;
+ next->pprev = &n->next;
+
+ if(next->next)
+ next->next->pprev = &next->next;
+}
+
+/*
+ * Move a list from one list head to another. Fixup the pprev
+ * reference of the first entry if it exists.
+ */
+static inline void hlist_move_list(struct hlist_head *old,
+ struct hlist_head *new)
+{
+ new->first = old->first;
+ if (new->first)
+ new->first->pprev = &new->first;
+ old->first = NULL;
+}
+
+#define hlist_entry(ptr, type, member) container_of(ptr,type,member)
+
+#define hlist_for_each(pos, head) \
+ for (pos = (head)->first; pos && ({ prefetch(pos->next); 1; }); \
+ pos = pos->next)
+
+#define hlist_for_each_safe(pos, n, head) \
+ for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \
+ pos = n)
+
+/**
+ * hlist_for_each_entry - iterate over list of given type
+ * @tpos: the type * to use as a loop cursor.
+ * @pos: the &struct hlist_node to use as a loop cursor.
+ * @head: the head for your list.
+ * @member: the name of the hlist_node within the struct.
+ */
+#define hlist_for_each_entry(tpos, pos, head, member) \
+ for (pos = (head)->first; \
+ pos && ({ prefetch(pos->next); 1;}) && \
+ ({ tpos = hlist_entry(pos, typeof(*tpos), member); 1;}); \
+ pos = pos->next)
+
+/**
+ * hlist_for_each_entry_continue - iterate over a hlist continuing after current point
+ * @tpos: the type * to use as a loop cursor.
+ * @pos: the &struct hlist_node to use as a loop cursor.
+ * @member: the name of the hlist_node within the struct.
+ */
+#define hlist_for_each_entry_continue(tpos, pos, member) \
+ for (pos = (pos)->next; \
+ pos && ({ prefetch(pos->next); 1;}) && \
+ ({ tpos = hlist_entry(pos, typeof(*tpos), member); 1;}); \
+ pos = pos->next)
+
+/**
+ * hlist_for_each_entry_from - iterate over a hlist continuing from current point
+ * @tpos: the type * to use as a loop cursor.
+ * @pos: the &struct hlist_node to use as a loop cursor.
+ * @member: the name of the hlist_node within the struct.
+ */
+#define hlist_for_each_entry_from(tpos, pos, member) \
+ for (; pos && ({ prefetch(pos->next); 1;}) && \
+ ({ tpos = hlist_entry(pos, typeof(*tpos), member); 1;}); \
+ pos = pos->next)
+
+/**
+ * hlist_for_each_entry_safe - iterate over list of given type safe against removal of list entry
+ * @tpos: the type * to use as a loop cursor.
+ * @pos: the &struct hlist_node to use as a loop cursor.
+ * @n: another &struct hlist_node to use as temporary storage
+ * @head: the head for your list.
+ * @member: the name of the hlist_node within the struct.
+ */
+#define hlist_for_each_entry_safe(tpos, pos, n, head, member) \
+ for (pos = (head)->first; \
+ pos && ({ n = pos->next; 1; }) && \
+ ({ tpos = hlist_entry(pos, typeof(*tpos), member); 1;}); \
+ pos = n)
+
+#endif
diff --git a/contrib/toolchain/gcc/5x/libgcc/config/i386/pe.h b/contrib/toolchain/gcc/5x/libgcc/config/i386/pe.h
new file mode 100644
index 0000000000..edd886538b
--- /dev/null
+++ b/contrib/toolchain/gcc/5x/libgcc/config/i386/pe.h
@@ -0,0 +1,188 @@
+
+typedef unsigned short WORD;
+typedef unsigned int DWORD;
+typedef unsigned int LONG;
+typedef unsigned char BYTE;
+
+#define IMAGE_DOS_SIGNATURE 0x5A4D
+#define IMAGE_NT_SIGNATURE 0x00004550
+#define IMAGE_NT_OPTIONAL_HDR32_MAGIC 0x10b
+
+#pragma pack(push,2)
+typedef struct _IMAGE_DOS_HEADER
+{
+ WORD e_magic;
+ WORD e_cblp;
+ WORD e_cp;
+ WORD e_crlc;
+ WORD e_cparhdr;
+ WORD e_minalloc;
+ WORD e_maxalloc;
+ WORD e_ss;
+ WORD e_sp;
+ WORD e_csum;
+ WORD e_ip;
+ WORD e_cs;
+ WORD e_lfarlc;
+ WORD e_ovno;
+ WORD e_res[4];
+ WORD e_oemid;
+ WORD e_oeminfo;
+ WORD e_res2[10];
+ LONG e_lfanew;
+} IMAGE_DOS_HEADER,*PIMAGE_DOS_HEADER;
+#pragma pack(pop)
+
+
+#pragma pack(push,4)
+typedef struct _IMAGE_FILE_HEADER
+{
+ WORD Machine;
+ WORD NumberOfSections;
+ DWORD TimeDateStamp;
+ DWORD PointerToSymbolTable;
+ DWORD NumberOfSymbols;
+ WORD SizeOfOptionalHeader;
+ WORD Characteristics;
+} IMAGE_FILE_HEADER, *PIMAGE_FILE_HEADER;
+
+#define IMAGE_FILE_DLL 0x2000
+
+#define IMAGE_FILE_MACHINE_I386 0x014c /* Intel 386 or later processors
+ and compatible processors */
+typedef struct _IMAGE_DATA_DIRECTORY {
+ DWORD VirtualAddress;
+ DWORD Size;
+} IMAGE_DATA_DIRECTORY,*PIMAGE_DATA_DIRECTORY;
+
+#define IMAGE_NUMBEROF_DIRECTORY_ENTRIES 16
+
+typedef struct _IMAGE_OPTIONAL_HEADER {
+ WORD Magic;
+ BYTE MajorLinkerVersion;
+ BYTE MinorLinkerVersion;
+ DWORD SizeOfCode;
+ DWORD SizeOfInitializedData;
+ DWORD SizeOfUninitializedData;
+ DWORD AddressOfEntryPoint;
+ DWORD BaseOfCode;
+ DWORD BaseOfData;
+ DWORD ImageBase;
+ DWORD SectionAlignment;
+ DWORD FileAlignment;
+ WORD MajorOperatingSystemVersion;
+ WORD MinorOperatingSystemVersion;
+ WORD MajorImageVersion;
+ WORD MinorImageVersion;
+ WORD MajorSubsystemVersion;
+ WORD MinorSubsystemVersion;
+ DWORD Win32VersionValue;
+ DWORD SizeOfImage;
+ DWORD SizeOfHeaders;
+ DWORD CheckSum;
+ WORD Subsystem;
+ WORD DllCharacteristics;
+ DWORD SizeOfStackReserve;
+ DWORD SizeOfStackCommit;
+ DWORD SizeOfHeapReserve;
+ DWORD SizeOfHeapCommit;
+ DWORD LoaderFlags;
+ DWORD NumberOfRvaAndSizes;
+ IMAGE_DATA_DIRECTORY DataDirectory[IMAGE_NUMBEROF_DIRECTORY_ENTRIES];
+} IMAGE_OPTIONAL_HEADER,*PIMAGE_OPTIONAL_HEADER;
+
+#pragma pack(pop)
+
+
+#pragma pack(push,4)
+typedef struct _IMAGE_NT_HEADERS
+{
+ DWORD Signature;
+ IMAGE_FILE_HEADER FileHeader;
+ IMAGE_OPTIONAL_HEADER OptionalHeader;
+} IMAGE_NT_HEADERS32,*PIMAGE_NT_HEADERS32;
+
+#define IMAGE_SIZEOF_SHORT_NAME 8
+
+typedef struct _IMAGE_SECTION_HEADER
+{
+ BYTE Name[IMAGE_SIZEOF_SHORT_NAME];
+ union
+ {
+ DWORD PhysicalAddress;
+ DWORD VirtualSize;
+ } Misc;
+ DWORD VirtualAddress;
+ DWORD SizeOfRawData;
+ DWORD PointerToRawData;
+ DWORD PointerToRelocations;
+ DWORD PointerToLinenumbers;
+ WORD NumberOfRelocations;
+ WORD NumberOfLinenumbers;
+ DWORD Characteristics;
+} IMAGE_SECTION_HEADER,*PIMAGE_SECTION_HEADER;
+#pragma pack(pop)
+
+#define IMAGE_SCN_CNT_INITIALIZED_DATA 0x00000040
+#define IMAGE_SCN_CNT_UNINITIALIZED_DATA 0x00000080
+#define IMAGE_SCN_MEM_SHARED 0x10000000
+#define IMAGE_SCN_MEM_EXECUTE 0x20000000
+#define IMAGE_SCN_MEM_WRITE 0x80000000
+
+#pragma pack(push,4)
+typedef struct _IMAGE_BASE_RELOCATION {
+ DWORD VirtualAddress;
+ DWORD SizeOfBlock;
+} IMAGE_BASE_RELOCATION,*PIMAGE_BASE_RELOCATION;
+#pragma pack(pop)
+
+typedef struct _IMAGE_IMPORT_DESCRIPTOR
+{
+ union
+ {
+ DWORD Characteristics;
+ DWORD OriginalFirstThunk;
+ };
+ DWORD TimeDateStamp;
+ DWORD ForwarderChain;
+ DWORD Name;
+ DWORD FirstThunk;
+} IMAGE_IMPORT_DESCRIPTOR,*PIMAGE_IMPORT_DESCRIPTOR;
+
+typedef struct _IMAGE_THUNK_DATA32
+{
+ union
+ {
+ DWORD ForwarderString;
+ DWORD Function;
+ DWORD Ordinal;
+ DWORD AddressOfData;
+ } u1;
+} IMAGE_THUNK_DATA32,*PIMAGE_THUNK_DATA32;
+
+typedef struct _IMAGE_IMPORT_BY_NAME
+{
+ WORD Hint;
+ BYTE Name[1];
+} IMAGE_IMPORT_BY_NAME,*PIMAGE_IMPORT_BY_NAME;
+
+#define IMAGE_ORDINAL_FLAG 0x80000000
+
+typedef struct _IMAGE_EXPORT_DIRECTORY {
+ DWORD Characteristics;
+ DWORD TimeDateStamp;
+ WORD MajorVersion;
+ WORD MinorVersion;
+ DWORD Name;
+ DWORD Base;
+ DWORD NumberOfFunctions;
+ DWORD NumberOfNames;
+ DWORD AddressOfFunctions;
+ DWORD AddressOfNames;
+ DWORD AddressOfNameOrdinals;
+} IMAGE_EXPORT_DIRECTORY,*PIMAGE_EXPORT_DIRECTORY;
+
+
+#define MakePtr( cast, ptr, addValue ) (cast)( (uint32_t)(ptr) + (uint32_t)(addValue) )
+
+
diff --git a/contrib/toolchain/gcc/5x/libgcc/config/i386/start.S b/contrib/toolchain/gcc/5x/libgcc/config/i386/start.S
new file mode 100644
index 0000000000..8596934b82
--- /dev/null
+++ b/contrib/toolchain/gcc/5x/libgcc/config/i386/start.S
@@ -0,0 +1,112 @@
+
+#tls:0 pid process id
+#tls:4 tid reserved for thread slot
+#tls:8 thread's stack low limit
+#tls:12 thread's stack high limit
+#tls:16 reseved for libc
+
+#ifdef L_start
+
+ .section .init
+ .global __start
+
+.align 4
+__start:
+ movl $68, %eax
+ movl $12, %ebx
+ lea __size_of_stack_reserve__, %ecx
+ addl $4095, %ecx #load stack size
+ andl $-4096, %ecx #align to page granularity
+ int $0x40 #and allocate
+ testl %eax, %eax
+ jz 1f
+
+ addl %eax, %ecx
+ movl %eax, %fs:8
+ movl %ecx, %fs:12 #save stack base - low limit
+ #save stack top - high limit
+ movl %ecx, %esp #reload stack
+
+ subl $1024, %esp
+
+ movl $9, %eax
+ movl %esp, %ebx
+ movl $-1, %ecx
+ int $0x40
+
+ movl 30(%ebx), %eax
+ movl %eax, %fs:0 #save pid
+
+ addl $1024, %esp
+
+ jmp ___crt_startup
+1:
+ int3 #trap to debugger
+
+ .ascii "No enough memory for stack allocation"
+
+#endif
+
+
+#ifdef L_chkstk
+
+ .section .text
+ .global ___chkstk
+ .global __alloca
+
+___chkstk:
+__alloca:
+ pushl %ecx /* save temp */
+ leal 8(%esp), %ecx /* point past return addr */
+ subl %eax, %ecx
+ cmpl %fs:8, %ecx # check low stack limit
+ jb 1f
+
+ movl %esp, %eax /* save old stack pointer */
+ movl %ecx, %esp /* decrement stack */
+ movl (%eax), %ecx /* recover saved temp */
+ movl 4(%eax), %eax /* recover return address */
+
+ /* Push the return value back. Doing this instead of just
+ jumping to %eax preserves the cached call-return stack
+ used by most modern processors. */
+ pushl %eax
+ ret
+1:
+ int3 #trap to debugger
+ .ascii "Stack overflow"
+#endif
+
+#ifdef L_chkstk_ms
+
+ .section .text
+ .global ___chkstk_ms
+
+___chkstk_ms:
+ pushl %ecx /* save temp */
+ pushl %eax
+ cmpl $0x1000, %eax /* > 4k ?*/
+ leal 12(%esp), %ecx /* point past return addr */
+ jb 2f
+1:
+ subl $0x1000, %ecx /* yes, move pointer down 4k*/
+ cmpl %fs:8, %ecx /* check low stack limit */
+ jb 3f
+
+ orl $0x0, (%ecx) /* probe there */
+ subl $0x1000, %eax /* decrement count */
+ cmpl $0x1000, %eax
+ ja 1b /* and do it again */
+
+2:
+ subl %eax, %ecx
+ orl $0x0, (%ecx) /* less than 4k, just peek here */
+
+ popl %eax
+ popl %ecx
+ ret
+3:
+ int3 #trap to debugger
+ .ascii "Stack overflow"
+#endif
+