os/kernelhwsrv/kernel/eka/include/x86hlp_gcc.inl
author sl
Tue, 10 Jun 2014 14:32:02 +0200 (2014-06-10)
changeset 1 260cb5ec6c19
permissions -rw-r--r--
Update contrib.
     1 // Copyright (c) 2007-2009 Nokia Corporation and/or its subsidiary(-ies).
     2 // All rights reserved.
     3 // This component and the accompanying materials are made available
     4 // under the terms of the License "Eclipse Public License v1.0"
     5 // which accompanies this distribution, and is available
     6 // at the URL "http://www.eclipse.org/legal/epl-v10.html".
     7 //
     8 // Initial Contributors:
     9 // Nokia Corporation - initial contribution.
    10 //
    11 // Contributors:
    12 //
    13 // Description:
    14 // e32\common\x86\x86hlp_gcc.inl
    15 // If there are no exports then GCC 3.4.x does not generate a .reloc 
    16 // section, without which rombuild can't relocate the .code section
    17 // to its ROM address. Your ROM then goes boom early in the boot sequence.
    18 // This unused export forces the PE to be generated with a .reloc section.
    19 // 
    20 //
    21 
    22 EXPORT_C void __ignore_this_export()
    23 	{
    24 	}
    25 
    26 static void DivisionByZero()
    27 	{
    28 	asm("int 0");
    29 	}
    30 
    31 extern "C" {
    32 
    33 void __NAKED__ _alloca()
    34 {
    35 	// GCC passes the param in eax and expects no return value
    36 	asm("pop ecx");
    37 	asm("sub esp, eax");
    38 	asm("push ecx");
    39 	asm("ret");
    40 }
    41 
    42 void __NAKED__ _allmul()
    43 //
    44 // Multiply two 64 bit integers returning a 64 bit result
    45 // On entry:
    46 //		[esp+4], [esp+8] = arg 1
    47 //		[esp+12], [esp+16] = arg 1
    48 // Return result in edx:eax
    49 // Remove arguments from stack
    50 //
    51 	{
    52 	asm("mov eax, [esp+4]");		// eax = low1
    53 	asm("mul dword ptr [esp+16]");	// edx:eax = low1*high2
    54 	asm("mov ecx, eax");			// keep low 32 bits of product
    55 	asm("mov eax, [esp+8]");		// eax = high1
    56 	asm("mul dword ptr [esp+12]");	// edx:eax = high1*low2
    57 	asm("add ecx, eax");			// accumulate low 32 bits of product
    58 	asm("mov eax, [esp+4]");		// eax = low1
    59 	asm("mul dword ptr [esp+12]");	// edx:eax = low1*low2
    60 	asm("add edx, ecx");			// add cross terms to high 32 bits
    61 	asm("ret");
    62 	}
    63 
    64 void __NAKED__ udiv64_divby0()
    65 	{
    66 	asm("int 0");					// division by zero exception
    67 	asm("ret");
    68 	}
    69 
    70 __NAKED__ /*LOCAL_C*/ void UDiv64()
    71 	{
    72 	// unsigned divide edx:eax by edi:esi
    73 	// quotient in ebx:eax, remainder in edi:edx
    74 	// ecx, ebp, esi also modified
    75 	asm("test edi, edi");
    76 	asm("jnz short UDiv64a");			// branch if divisor >= 2^32
    77 	asm("test esi, esi");
    78 	asm("jz %a0": : "i"(&DivisionByZero)); // if divisor=0, branch to error routine
    79 	asm("mov ebx, eax");				// ebx=dividend low
    80 	asm("mov eax, edx");				// eax=dividend high
    81 	asm("xor edx, edx");				// edx=0
    82 	asm("div esi");						// quotient high now in eax
    83 	asm("xchg eax, ebx");				// quotient high in ebx, dividend low in eax
    84 	asm("div esi");						// quotient now in ebx:eax, remainder in edi:edx
    85 	asm("ret");
    86 	asm("UDiv64e:");
    87 	asm("xor eax, eax");				// set result to 0xFFFFFFFF
    88 	asm("dec eax");
    89 	asm("jmp short UDiv64f");
    90 	asm("UDiv64a:");
    91 	asm("js short UDiv64b");			// skip if divisor msb set
    92 	asm("bsr ecx, edi");				// ecx=bit number of divisor msb - 32
    93 	asm("inc cl");
    94 	asm("push edi");					// save divisor high
    95 	asm("push esi");					// save divisor low
    96 	asm("shrd esi, edi, cl");			// shift divisor right so that msb is bit 31
    97 	asm("mov ebx, edx");				// dividend into ebx:ebp
    98 	asm("mov ebp, eax");
    99 	asm("shrd eax, edx, cl");			// shift dividend right same number of bits
   100 	asm("shr edx, cl");
   101 	asm("cmp edx, esi");				// check if approx quotient will be 2^32
   102 	asm("jae short UDiv64e");			// if so, true result must be 0xFFFFFFFF
   103 	asm("div esi");						// approximate quotient now in eax
   104 	asm("UDiv64f:");
   105 	asm("mov ecx, eax");				// into ecx
   106 	asm("mul edi");						// multiply approx. quotient by divisor high
   107 	asm("mov esi, eax");				// ls dword into esi, ms into edi
   108 	asm("mov edi, edx");
   109 	asm("mov eax, ecx");				// approx. quotient into eax
   110 	asm("mul dword ptr [esp]");			// multiply approx. quotient by divisor low
   111 	asm("add edx, esi");				// edi:edx:eax now equals approx. quotient * divisor
   112 	asm("adc edi, 0");
   113 	asm("xor esi, esi");
   114 	asm("sub ebp, eax");				// subtract dividend - approx. quotient *divisor
   115 	asm("sbb ebx, edx");
   116 	asm("sbb esi, edi");
   117 	asm("jnc short UDiv64c");			// if no borrow, result OK
   118 	asm("dec ecx");						// else result is one too big
   119 	asm("add ebp, [esp]");				// and add divisor to get correct remainder
   120 	asm("adc ebx, [esp+4]");
   121 	asm("UDiv64c:");
   122 	asm("mov eax, ecx");				// result into ebx:eax, remainder into edi:edx
   123 	asm("mov edi, ebx");
   124 	asm("mov edx, ebp");
   125 	asm("xor ebx, ebx");
   126 	asm("add esp, 8");					// remove temporary values from stack
   127 	asm("ret");
   128 	asm("UDiv64b:");
   129 	asm("mov ebx, 1");
   130 	asm("sub eax, esi");				// subtract divisor from dividend
   131 	asm("sbb edx, edi");
   132 	asm("jnc short UDiv64d");			// if no borrow, result=1, remainder in edx:eax
   133 	asm("add eax, esi");				// else add back
   134 	asm("adc edx, edi");
   135 	asm("dec ebx");						// and decrement quotient
   136 	asm("UDiv64d:");
   137 	asm("mov edi, edx");				// remainder into edi:edx
   138 	asm("mov edx, eax");
   139 	asm("mov eax, ebx");				// result in ebx:eax
   140 	asm("xor ebx, ebx");
   141 	asm("ret");
   142 	}
   143 
   144 __NAKED__ void _aulldvrm()
   145 //
   146 // Divide two 64 bit unsigned integers, returning a 64 bit result
   147 // and a 64 bit remainder
   148 //
   149 // On entry:
   150 //		[esp+4], [esp+8] = dividend
   151 //		[esp+12], [esp+16] = divisor
   152 //
   153 // Return (dividend / divisor) in edx:eax
   154 // Return (dividend % divisor) in ebx:ecx
   155 //
   156 // Remove arguments from stack
   157 //
   158 	{
   159 	asm("push ebp");
   160 	asm("push edi");
   161 	asm("push esi");
   162 	asm("mov eax, [esp+16]");
   163 	asm("mov edx, [esp+20]");
   164 	asm("mov esi, [esp+24]");
   165 	asm("mov edi, [esp+28]");
   166 	asm("call %a0": : "i"(&UDiv64));
   167 	asm("mov ecx, edx");
   168 	asm("mov edx, ebx");
   169 	asm("mov ebx, edi");
   170 	asm("pop esi");
   171 	asm("pop edi");
   172 	asm("pop ebp");
   173 	asm("ret");
   174 	}
   175 
   176 __NAKED__ void _alldvrm()
   177 //
   178 // Divide two 64 bit signed integers, returning a 64 bit result
   179 // and a 64 bit remainder
   180 //
   181 // On entry:
   182 //		[esp+4], [esp+8] = dividend
   183 //		[esp+12], [esp+16] = divisor
   184 //
   185 // Return (dividend / divisor) in edx:eax
   186 // Return (dividend % divisor) in ebx:ecx
   187 //
   188 // Remove arguments from stack
   189 //
   190 	{
   191 	asm("push ebp");
   192 	asm("push edi");
   193 	asm("push esi");
   194 	asm("mov eax, [esp+16]");
   195 	asm("mov edx, [esp+20]");
   196 	asm("mov esi, [esp+24]");
   197 	asm("mov edi, [esp+28]");
   198 	asm("test edx, edx");
   199 	asm("jns alldrvm_dividend_nonnegative");
   200 	asm("neg edx");
   201 	asm("neg eax");
   202 	asm("sbb edx, 0");
   203 	asm("alldrvm_dividend_nonnegative:");
   204 	asm("test edi, edi");
   205 	asm("jns alldrvm_divisor_nonnegative");
   206 	asm("neg edi");
   207 	asm("neg esi");
   208 	asm("sbb edi, 0");
   209 	asm("alldrvm_divisor_nonnegative:");
   210 	asm("call %a0": : "i"(&UDiv64));
   211 	asm("mov ebp, [esp+20]");
   212 	asm("mov ecx, edx");
   213 	asm("xor ebp, [esp+28]");
   214 	asm("mov edx, ebx");
   215 	asm("mov ebx, edi");
   216 	asm("jns alldrvm_quotient_nonnegative");
   217 	asm("neg edx");
   218 	asm("neg eax");
   219 	asm("sbb edx, 0");
   220 	asm("alldrvm_quotient_nonnegative:");
   221 	asm("cmp dword ptr [esp+20], 0");
   222 	asm("jns alldrvm_rem_nonnegative");
   223 	asm("neg ebx");
   224 	asm("neg ecx");
   225 	asm("sbb ebx, 0");
   226 	asm("alldrvm_rem_nonnegative:");
   227 	asm("pop esi");
   228 	asm("pop edi");
   229 	asm("pop ebp");
   230 	asm("ret");
   231 	}
   232 
   233 //__NAKED__ void _aulldiv()
   234 __NAKED__ void __udivdi3 ()
   235 //
   236 // Divide two 64 bit unsigned integers returning a 64 bit result
   237 // On entry:
   238 //		[esp+4], [esp+8] = dividend
   239 //		[esp+12], [esp+16] = divisor
   240 // Return result in edx:eax
   241 // Remove arguments from stack
   242 //
   243 	{
   244 	asm("push ebp");
   245 	asm("push edi");
   246 	asm("push esi");
   247 	asm("push ebx");
   248 	asm("mov eax, [esp+20]");
   249 	asm("mov edx, [esp+24]");
   250 	asm("mov esi, [esp+28]");
   251 	asm("mov edi, [esp+32]");
   252 	asm("call %a0": : "i"(&UDiv64));
   253 	asm("mov edx, ebx");
   254 	asm("pop ebx");
   255 	asm("pop esi");
   256 	asm("pop edi");
   257 	asm("pop ebp");
   258 	asm("ret");
   259 	}
   260 
   261 
   262 __NAKED__ void __divdi3()
   263 
   264 //
   265 // Divide two 64 bit signed integers returning a 64 bit result
   266 // On entry:
   267 //		[esp+4], [esp+8] = dividend
   268 //		[esp+12], [esp+16] = divisor
   269 // Return result in edx:eax
   270 // Remove arguments from stack
   271 //
   272 	{
   273 	asm("push ebp");
   274 	asm("push edi");
   275 	asm("push esi");
   276 	asm("push ebx");
   277 	asm("mov eax, [esp+20]");
   278 	asm("mov edx, [esp+24]");
   279 	asm("mov esi, [esp+28]");
   280 	asm("mov edi, [esp+32]");
   281 	asm("test edx, edx");
   282 	asm("jns divdi_dividend_nonnegative");
   283 	asm("neg edx");
   284 	asm("neg eax");
   285 	asm("sbb edx, 0");
   286 	asm("divdi_dividend_nonnegative:");
   287 	asm("test edi, edi");
   288 	asm("jns divdi_divisor_nonnegative");
   289 	asm("neg edi");
   290 	asm("neg esi");
   291 	asm("sbb edi, 0");
   292 	asm("divdi_divisor_nonnegative:");
   293 	asm("call %a0": : "i"(&UDiv64));
   294 	asm("mov ecx, [esp+24]");
   295 	asm("mov edx, ebx");
   296 	asm("xor ecx, [esp+32]");
   297 	asm("jns divdi_quotient_nonnegative");
   298 	asm("neg edx");
   299 	asm("neg eax");
   300 	asm("sbb edx, 0");
   301 	asm("divdi_quotient_nonnegative:");
   302 	asm("pop ebx");
   303 	asm("pop esi");
   304 	asm("pop edi");
   305 	asm("pop ebp");
   306 	asm("ret");
   307 	}
   308 
   309 __NAKED__ void __umoddi3()
   310 //
   311 // Divide two 64 bit unsigned integers and return 64 bit remainder
   312 // On entry:
   313 //		[esp+4], [esp+8] = dividend
   314 //		[esp+12], [esp+16] = divisor
   315 // Return result in edx:eax
   316 // Remove arguments from stack
   317 //
   318 	{
   319 	asm("push ebp");
   320 	asm("push edi");
   321 	asm("push esi");
   322 	asm("push ebx");
   323 	asm("mov eax, [esp+20]");
   324 	asm("mov edx, [esp+24]");
   325 	asm("mov esi, [esp+28]");
   326 	asm("mov edi, [esp+32]");
   327 	asm("call %a0": : "i"(&UDiv64));
   328 	asm("mov eax, edx");
   329 	asm("mov edx, edi");
   330 	asm("pop ebx");
   331 	asm("pop esi");
   332 	asm("pop edi");
   333 	asm("pop ebp");
   334 	asm("ret");
   335 	}
   336 
   337 __NAKED__ void __moddi3()
   338 //
   339 // Divide two 64 bit signed integers and return 64 bit remainder
   340 // On entry:
   341 //		[esp+4], [esp+8] = dividend
   342 //		[esp+12], [esp+16] = divisor
   343 // Return result in edx:eax
   344 // Remove arguments from stack
   345 //
   346 	{
   347 	asm("push ebp");
   348 	asm("push edi");
   349 	asm("push esi");
   350 	asm("push ebx");
   351 	asm("mov eax, [esp+20]");
   352 	asm("mov edx, [esp+24]");
   353 	asm("mov esi, [esp+28]");
   354 	asm("mov edi, [esp+32]");
   355 	asm("test edx, edx");
   356 	asm("jns dividend_nonnegative");
   357 	asm("neg edx");
   358 	asm("neg eax");
   359 	asm("sbb edx, 0");
   360 	asm("dividend_nonnegative:");
   361 	asm("test edi, edi");
   362 	asm("jns divisor_nonnegative");
   363 	asm("neg edi");
   364 	asm("neg esi");
   365 	asm("sbb edi, 0");
   366 	asm("divisor_nonnegative:");
   367 	asm("call %a0": : "i"(&UDiv64));
   368 	asm("mov eax, edx");
   369 	asm("mov edx, edi");
   370 	asm("cmp dword ptr [esp+24], 0");
   371 	asm("jns rem_nonnegative");
   372 	asm("neg edx");
   373 	asm("neg eax");
   374 	asm("sbb edx, 0");
   375 	asm("rem_nonnegative:");
   376 	asm("pop ebx");
   377 	asm("pop esi");
   378 	asm("pop edi");
   379 	asm("pop ebp");
   380 	asm("ret");
   381 	}
   382 
   383 __NAKED__ void _allshr()
   384 //
   385 // Arithmetic shift right EDX:EAX by CL
   386 //
   387 	{
   388 	asm("cmp cl, 64");
   389 	asm("jae asr_count_ge_64");
   390 	asm("cmp cl, 32");
   391 	asm("jae asr_count_ge_32");
   392 	asm("shrd eax, edx, cl");
   393 	asm("sar edx, cl");
   394 	asm("ret");
   395 	asm("asr_count_ge_32:");
   396 	asm("sub cl, 32");
   397 	asm("mov eax, edx");
   398 	asm("cdq");
   399 	asm("sar eax, cl");
   400 	asm("ret");
   401 	asm("asr_count_ge_64:");
   402 	asm("sar edx, 32");
   403 	asm("mov eax, edx");
   404 	asm("ret");
   405 	}
   406 
   407 __NAKED__ void _allshl()
   408 //
   409 // shift left EDX:EAX by CL
   410 //
   411 	{
   412 	asm("cmp cl, 64");
   413 	asm("jae lsl_count_ge_64");
   414 	asm("cmp cl, 32");
   415 	asm("jae lsl_count_ge_32");
   416 	asm("shld edx, eax, cl");
   417 	asm("shl eax, cl");
   418 	asm("ret");
   419 	asm("lsl_count_ge_32:");
   420 	asm("sub cl, 32");
   421 	asm("mov edx, eax");
   422 	asm("xor eax, eax");
   423 	asm("shl edx, cl");
   424 	asm("ret");
   425 	asm("lsl_count_ge_64:");
   426 	asm("xor edx, edx");
   427 	asm("xor eax, eax");
   428 	asm("ret");
   429 	}
   430 
   431 __NAKED__ void _aullshr()
   432 //
   433 // Logical shift right EDX:EAX by CL
   434 //
   435 	{
   436 	asm("cmp cl, 64");
   437 	asm("jae lsr_count_ge_64");
   438 	asm("cmp cl, 32");
   439 	asm("jae lsr_count_ge_32");
   440 	asm("shrd eax, edx, cl");
   441 	asm("shr edx, cl");
   442 	asm("ret");
   443 	asm("lsr_count_ge_32:");
   444 	asm("sub cl, 32");
   445 	asm("mov eax, edx");
   446 	asm("xor edx, edx");
   447 	asm("shr eax, cl");
   448 	asm("ret");
   449 	asm("lsr_count_ge_64:");
   450 	asm("xor edx, edx");
   451 	asm("xor eax, eax");
   452 	asm("ret");
   453 	}
   454 
   455 }