os/kernelhwsrv/kernel/eka/nkernsmp/nk_irq.cpp
author sl@SLION-WIN7.fritz.box
Fri, 15 Jun 2012 03:10:57 +0200
changeset 0 bde4ae8d615e
permissions -rw-r--r--
First public contribution.
     1 // Copyright (c) 2007-2009 Nokia Corporation and/or its subsidiary(-ies).
     2 // All rights reserved.
     3 // This component and the accompanying materials are made available
     4 // under the terms of the License "Eclipse Public License v1.0"
     5 // which accompanies this distribution, and is available
     6 // at the URL "http://www.eclipse.org/legal/epl-v10.html".
     7 //
     8 // Initial Contributors:
     9 // Nokia Corporation - initial contribution.
    10 //
    11 // Contributors:
    12 //
    13 // Description:
    14 // e32\nkernsmp\nk_irq.cpp
    15 // 
    16 //
    17 
    18 /**
    19  @file
    20  @internalTechnology
    21 */
    22 
    23 #include <e32cmn.h>
    24 #include <e32cmn_private.h>
    25 #include "nk_priv.h"
    26 #include <nk_irq.h>
    27 
    28 NIrq		Irq[NK_MAX_IRQS];
    29 NIrqHandler	Handlers[NK_MAX_IRQ_HANDLERS];
    30 NIrqHandler* NIrqHandler::FirstFree;
    31 
    32 extern "C" void send_irq_ipi(TSubScheduler*);
    33 
    34 void StepCookie(volatile TUint16& p, TInt n)
    35 	{
    36 	TUint32 x = p<<17;
    37 	while(n--)
    38 		{
    39 		TUint32 y = x;
    40 		x<<=1;
    41 		y^=x;
    42 		x |= ((y>>31)<<17);
    43 		}
    44 	p = (TUint16)(x>>17);
    45 	}
    46 
    47 NIrq::NIrq()
    48 	:	iNIrqLock(TSpinLock::EOrderNIrq)
    49 	{
    50 	iIState = EWait;
    51 	iEventsPending = 0;
    52 	iEnabledEvents = 0;
    53 	iHwId = 0;
    54 	iX = 0;
    55 	}
    56 
    57 TInt NIrq::BindRaw(NIsr aIsr, TAny* aPtr)
    58 	{
    59 	// Call only from thread context
    60 	TInt r = KErrNone;
    61 	Wait();
    62 	iNIrqLock.LockOnly();
    63 	if (iStaticFlags & EShared)
    64 		{
    65 		r = KErrAccessDenied;
    66 		goto error;
    67 		}
    68 	if ( (iIState & ERaw) || !iHandlers.IsEmpty())
    69 		{
    70 		r = KErrInUse;
    71 		goto error;
    72 		}
    73 	iHandlers.iA.iNext = (SDblQueLink*)aIsr;
    74 	iHandlers.iA.iPrev = (SDblQueLink*)aPtr;
    75 	__e32_atomic_ior_rel32(&iIState, ERaw);
    76 error:
    77 	iNIrqLock.UnlockOnly();
    78 	Done();
    79 	return r;
    80 	}
    81 
    82 TInt NIrq::UnbindRaw()
    83 	{
    84 	// Call only from thread context
    85 	TInt r = DisableRaw(TRUE);
    86 	if (r != KErrNone)
    87 		return r;
    88 	Wait();
    89 	iNIrqLock.LockOnly();
    90 	if (iIState & ERaw)
    91 		{
    92 		iHandlers.iA.iNext = 0;
    93 		iHandlers.iA.iPrev = 0;
    94 		++iGeneration;	// release anyone still waiting in Disable()
    95 		__e32_atomic_and_rel32(&iIState, ~(ERaw|EUnbind));
    96 		}
    97 	iNIrqLock.UnlockOnly();
    98 	Done();
    99 	return r;
   100 	}
   101 
   102 TInt NIrq::DisableRaw(TBool aUnbind)
   103 	{
   104 	TBool wait = FALSE;
   105 	TInt r = KErrNone;
   106 	TInt irq = __SPIN_LOCK_IRQSAVE(iNIrqLock);
   107 	if (!(iIState & ERaw))
   108 		r = KErrGeneral;
   109 	else
   110 		{
   111 		wait = TRUE;
   112 		if (aUnbind)
   113 			__e32_atomic_ior_acq32(&iIState, EUnbind);
   114 		if (!(iEnabledEvents & 1))
   115 			{
   116 			iEnabledEvents |= 1;
   117 			HwDisable();
   118 //			wait = TRUE;
   119 			}
   120 		}
   121 	__SPIN_UNLOCK_IRQRESTORE(iNIrqLock,irq);
   122 	TInt c = NKern::CurrentContext();
   123 	if (wait && c!=NKern::EInterrupt)
   124 		{
   125 		// wait for currently running handler to finish or interrupt to be reenabled
   126 		if (c==NKern::EThread)
   127 			NKern::ThreadEnterCS();
   128 		HwWaitCpus();	// ensure other CPUs have had a chance to accept any outstanding interrupts
   129 		TUint32 g = iGeneration;
   130 		while ( ((iIState >> 16) || HwPending()) && (iGeneration == g))
   131 			{
   132 			__chill();
   133 			}
   134 		if (c==NKern::EThread)
   135 			NKern::ThreadLeaveCS();
   136 		}
   137 	return r;
   138 	}
   139 
   140 TInt NIrq::EnableRaw()
   141 	{
   142 	TInt r = KErrNone;
   143 	TInt irq = __SPIN_LOCK_IRQSAVE(iNIrqLock);
   144 	if (!(iIState & ERaw))
   145 		r = KErrGeneral;
   146 	else if (iIState & EUnbind)
   147 		r = KErrNotReady;
   148 	else if (iEnabledEvents & 1)
   149 		{
   150 		iEnabledEvents = 0;
   151 		HwEnable();
   152 		++iGeneration;
   153 		}
   154 	__SPIN_UNLOCK_IRQRESTORE(iNIrqLock,irq);
   155 	return r;
   156 	}
   157 
   158 TInt NIrq::Bind(NIrqHandler* aH)
   159 	{
   160 	// Call only from thread context
   161 	TInt r = KErrInUse;
   162 	Wait();
   163 	if (!(iIState & ERaw))
   164 		{
   165 		r = KErrNone;
   166 		TBool empty = iHandlers.IsEmpty();
   167 		TBool shared = iStaticFlags & EShared;
   168 		TBool exclusive = iIState & NIrqHandler::EExclusive;
   169 		if (!empty)
   170 			{
   171 			if (!shared || exclusive)
   172 				{
   173 				r = KErrAccessDenied;
   174 				goto error;
   175 				}
   176 			NIrqHandler* h = _LOFF(iHandlers.First(), NIrqHandler, iIrqLink);
   177 			if (h->iHState & NIrqHandler::EExclusive)
   178 				{
   179 				r = KErrAccessDenied;
   180 				goto error;
   181 				}
   182 			}
   183 		aH->iIrq = this;
   184 		iHandlers.Add(&aH->iIrqLink);
   185 		}
   186 error:
   187 	Done();
   188 	return r;
   189 	}
   190 
   191 void NIrq::HwIsr()
   192 	{
   193 	TRACE_IRQ12(16, this, iVector, iIState);
   194 	TBool eoi_done = FALSE;
   195 	TUint32 rcf0 = EnterIsr();		// for initial run count
   196 	TUint32 rcf1 = iIState;			// might have changed while we were waiting in EnterIsr()
   197 	if (rcf1 & ERaw)
   198 		{
   199 		if (!(rcf1 & EUnbind))
   200 			{
   201 			NIsr f = (NIsr)iHandlers.iA.iNext;
   202 			TAny* p = iHandlers.iA.iPrev;
   203 			(*f)(p);
   204 			}
   205 		HwEoi();
   206 		IsrDone();
   207 		return;
   208 		}
   209 	if (rcf0 >> 16)
   210 		{
   211 		HwEoi();
   212 		return;
   213 		}
   214 	if (!(iStaticFlags & ELevel))
   215 		{
   216 		eoi_done = TRUE;
   217 		HwEoi();
   218 		}
   219 	do	{
   220 		// Handler list can't be touched now
   221 		SDblQueLink* anchor = &iHandlers.iA;
   222 		SDblQueLink* p = anchor->iNext;
   223 		while (p != anchor)
   224 			{
   225 			NIrqHandler* h = _LOFF(p, NIrqHandler, iIrqLink);
   226 			h->Activate(1);
   227 			p = p->iNext;
   228 			}
   229 		if (!eoi_done)
   230 			{
   231 			eoi_done = TRUE;
   232 			HwEoi();
   233 			}
   234 		if ((iStaticFlags & ELevel) && iEventsPending)
   235 			{
   236 			// For a level triggered interrupt make sure interrupt is disabled until
   237 			// all pending event handlers have run, to avoid a continuous interrupt.
   238 			TInt irq = __SPIN_LOCK_IRQSAVE(iNIrqLock);
   239 			if (iEventsPending)
   240 				{
   241 				iEnabledEvents |= 1;
   242 				HwDisable();
   243 				}
   244 			__SPIN_UNLOCK_IRQRESTORE(iNIrqLock,irq);
   245 			}
   246 		} while (IsrDone());
   247 	}
   248 
   249 void NIrqHandler::Activate(TInt aCount)
   250 	{
   251 	TUint32 orig = DoActivate(aCount);
   252 	TRACE_IRQ12(17, this, orig, aCount);
   253 	if (orig & (EDisable|EUnbind|EActive))
   254 		return;	// disabled or already active
   255 	if (iTied)
   256 		{
   257 		// we need to enforce mutual exclusion between the event handler
   258 		// and the tied thread or thread group, so the event handler must
   259 		// run on the CPU to which the thread or group is currently attached
   260 		// once the event has been attached to that CPU, the thread/group
   261 		// can't be migrated until the event handler completes.
   262 		// need a pending event count for the tied thread/group
   263 		// so we know when the thread/group can be migrated
   264 		TInt tied_cpu = iTied->BeginTiedEvent();
   265 		TInt this_cpu = NKern::CurrentCpu();
   266 		if (tied_cpu != this_cpu)
   267 			{
   268 			__e32_atomic_add_acq32(&iIrq->iEventsPending, 1);
   269 			TheSubSchedulers[tied_cpu].QueueEventAndKick(this);
   270 			// FIXME: move IRQ over to tied CPU if this is the only handler for that IRQ
   271 			//			what to do about shared IRQs?
   272 			return;
   273 			}
   274 		}
   275 	// event can run on this CPU so run it now
   276 	if (aCount)
   277 		{
   278 		orig = EventBegin();
   279 		TRACE_IRQ8(18, this, orig);
   280 		(*iFn)(iPtr);
   281 		orig = EventDone();
   282 		TRACE_IRQ8(19, this, orig);
   283 		if (!(orig & EActive))
   284 			{
   285 			if (iTied)
   286 				iTied->EndTiedEvent();
   287 			return;	// that was last occurrence or event now disabled
   288 			}
   289 		}
   290 	__e32_atomic_add_ord32(&iIrq->iEventsPending, 1);
   291 //	add event to this cpu
   292 	SubScheduler().QueueEventAndKick(this);
   293 	}
   294 
   295 
   296 NIrqHandler::NIrqHandler()
   297 	{
   298 	iIrqLink.iNext = 0;
   299 	iIrq = 0;
   300 	iTied = 0;
   301 	iHState = EDisable|EBind|ENotReady|EEventHandlerIrq;
   302 	iFn = 0;
   303 	iPtr = 0;
   304 	memclr(iNIrqHandlerSpare, sizeof(iNIrqHandlerSpare));
   305 	}
   306 
   307 void NIrqHandler::Free()
   308 	{
   309 	NKern::Lock();
   310 	NEventHandler::TiedLock.LockOnly();
   311 	if (!iTied)	// Only free if iTied has been cleared
   312 		{
   313 		iIrqLink.iNext = FirstFree;
   314 		FirstFree = this;
   315 		}
   316 	NEventHandler::TiedLock.UnlockOnly();
   317 	NKern::Unlock();
   318 	}
   319 
   320 NIrqHandler* NIrqHandler::Alloc()
   321 	{
   322 	NKern::Lock();
   323 	NEventHandler::TiedLock.LockOnly();
   324 	NIrqHandler* p = FirstFree;
   325 	if (p)
   326 		FirstFree = (NIrqHandler*)p->iIrqLink.iNext;
   327 	NEventHandler::TiedLock.UnlockOnly();
   328 	NKern::Unlock();
   329 	if (p)
   330 		new (p) NIrqHandler();
   331 	return p;
   332 	}
   333 
   334 TInt NIrqHandler::Enable(TInt aHandle)
   335 	{
   336 	// call from any context
   337 	TBool reactivate = FALSE;
   338 	TInt r = KErrNotReady;
   339 	NIrq* pI = iIrq;
   340 	if (!pI)
   341 		return KErrNotReady;
   342 	TInt irq = __SPIN_LOCK_IRQSAVE(pI->iNIrqLock);	// OK since NIrq's are never deleted
   343 	if (iIrq==pI && TUint(aHandle)==iHandle)	// check handler not unbound
   344 		{
   345 		TUint32 orig = DoSetEnabled();	// clear EDisable and EBind provided neither EUnbind nor ENotReady set
   346 		if (!(orig & (EUnbind|ENotReady)))
   347 			{
   348 			r = KErrNone;
   349 			if (orig & EDisable)	// check not already enabled
   350 				{
   351 				++iGeneration;
   352 				TUint32 n = pI->iEnabledEvents;
   353 				pI->iEnabledEvents += 2;
   354 				if (n==0)
   355 					pI->HwEnable();	// enable HW interrupt if this is first handler to be enabled
   356 				if ((orig >> 16) && !(orig & EActive))
   357 					// replay remembered interrupt(s)
   358 					reactivate = TRUE;
   359 				}
   360 			}
   361 		}
   362 	if (reactivate)
   363 		{
   364 		pI->iNIrqLock.UnlockOnly();
   365 		Activate(0);
   366 		pI->iNIrqLock.LockOnly();
   367 		}
   368 	__SPIN_UNLOCK_IRQRESTORE(pI->iNIrqLock,irq);
   369 	return r;
   370 	}
   371 
   372 TInt NIrqHandler::Disable(TBool aUnbind, TInt aHandle)
   373 	{
   374 	// call from any context
   375 	NIrq* pI = iIrq;
   376 	if (!pI)
   377 		return KErrGeneral;
   378 	TInt irq = __SPIN_LOCK_IRQSAVE(pI->iNIrqLock);	// OK since NIrq's are never deleted
   379 	if (iIrq != pI || TUint(aHandle)!=iHandle)	// check handler not unbound
   380 		{
   381 		__SPIN_UNLOCK_IRQRESTORE(pI->iNIrqLock,irq);
   382 		return KErrGeneral;
   383 		}
   384 	TInt r = aUnbind ? KErrGeneral : KErrNone;
   385 	TUint32 f = aUnbind ? EUnbind|EDisable : EDisable;
   386 	TUint32 orig = __e32_atomic_ior_acq32(&iHState, f);
   387 	TUint32 g = iGeneration;
   388 	if (!(orig & EDisable))	// check not already disabled
   389 		{
   390 		pI->iEnabledEvents -= 2;
   391 		if (!pI->iEnabledEvents)
   392 			pI->HwDisable();	// disable HW interrupt if no more enabled handlers
   393 		}
   394 	if (aUnbind && !(orig & EUnbind))
   395 		{
   396 		volatile TUint16& cookie = *(volatile TUint16*)(((TUint8*)&iHandle)+2);
   397 		StepCookie(cookie, 1);
   398 		r = KErrNone;
   399 		}
   400 	__SPIN_UNLOCK_IRQRESTORE(pI->iNIrqLock,irq);
   401 	if (NKern::CurrentContext() != NKern::EInterrupt)
   402 		{
   403 		// wait for currently running handler to finish or interrupt to be reenabled
   404  		while ((iHState & EActive) && (iGeneration == g))
   405 			{
   406 			__chill();
   407 			}
   408 		}
   409 	return r;
   410 	}
   411 
   412 TInt NIrqHandler::Unbind(TInt aId, NSchedulable* aTied)
   413 	{
   414 	TInt r = Disable(TRUE, aId);	// waits for any current activation of ISR to finish
   415 	if (r==KErrNone || aTied)	// returns KErrGeneral if someone else already unbound this interrupt handler
   416 		{
   417 		// Possible race condition here between tied thread termination and interrupt unbind.
   418 		// We need to be sure that the iTied field must be NULL before the tied thread/group
   419 		// is destroyed.
   420 		NKern::Lock();
   421 		NEventHandler::TiedLock.LockOnly();	// this guarantees pH->iTied cannot change
   422 		NSchedulable* t = iTied;
   423 		if (t)
   424 			{
   425 			// We need to guarantee the object pointed to by t cannot be deleted until we
   426 			// have finished with it.
   427 			t->AcqSLock();
   428 			if (iTiedLink.iNext)
   429 				{
   430 				iTiedLink.Deque();
   431 				iTiedLink.iNext = 0;
   432 				iTied = 0;
   433 				}
   434 			if (aTied && aTied==t)
   435 				iTied = 0;
   436 			t->RelSLock();
   437 			}
   438 		NEventHandler::TiedLock.UnlockOnly();
   439 		NKern::Unlock();
   440 		}
   441 	if (r==KErrNone)
   442 		{
   443 		DoUnbind();
   444 		Free();
   445 		}
   446 	return r;
   447 	}
   448 
   449 void NIrqHandler::DoUnbind()
   450 	{
   451 	// Call only from thread context
   452 	NIrq* pI = iIrq;
   453 	pI->Wait();
   454 	iIrqLink.Deque();
   455 	iIrq = 0;
   456 	pI->Done();
   457 	}
   458 
   459 TBool TSubScheduler::QueueEvent(NEventHandler* aEvent)
   460 	{
   461 	TInt irq = __SPIN_LOCK_IRQSAVE(iEventHandlerLock);
   462 	TBool pending = iEventHandlersPending;
   463 	iEventHandlersPending = TRUE;
   464 	iEventHandlers.Add(aEvent);
   465 	__SPIN_UNLOCK_IRQRESTORE(iEventHandlerLock,irq);
   466 	return !pending;
   467 	}
   468 
   469 void TSubScheduler::QueueEventAndKick(NEventHandler* aEvent)
   470 	{
   471 	if (QueueEvent(aEvent))
   472 		{
   473 		// extra barrier ?
   474 		send_irq_ipi(this);
   475 		}
   476 	}
   477 
   478 extern "C" void run_event_handlers(TSubScheduler* aS)
   479 	{
   480 	while (aS->iEventHandlersPending)
   481 		{
   482 		TInt irq = __SPIN_LOCK_IRQSAVE(aS->iEventHandlerLock);
   483 		if (aS->iEventHandlers.IsEmpty())
   484 			{
   485 			aS->iEventHandlersPending = FALSE;
   486 			__SPIN_UNLOCK_IRQRESTORE(aS->iEventHandlerLock, irq);
   487 			break;
   488 			}
   489 		NIrqHandler* h = (NIrqHandler*)aS->iEventHandlers.First()->Deque();
   490 		if (aS->iEventHandlers.IsEmpty())
   491 			aS->iEventHandlersPending = FALSE;
   492 		TInt type = h->iHType;
   493 		NSchedulable* tied = h->iTied;
   494 		if (type == NEventHandler::EEventHandlerNTimer)
   495 			{
   496 			NEventFn f = h->iFn;
   497 			TAny* p = h->iPtr;
   498 			mb();	// make sure dequeue observed and iFn,iPtr,iTied sampled before state change observed
   499 			h->i8888.iHState1 = NTimer::EIdle; // can't touch timer again after this
   500 			__SPIN_UNLOCK_IRQRESTORE(aS->iEventHandlerLock, irq);
   501 			(*f)(p);
   502 			if (tied)
   503 				tied->EndTiedEvent();
   504 			continue;
   505 			}
   506 		__SPIN_UNLOCK_IRQRESTORE(aS->iEventHandlerLock, irq);
   507 		TBool requeue = TRUE;
   508 		switch (h->iHType)
   509 			{
   510 			case NEventHandler::EEventHandlerIrq:
   511 				{
   512 				TUint32 orig;
   513 				// event can run on this CPU so run it now
   514 				// if event tied, migration of tied thread/group will have been blocked
   515 				orig = h->EventBegin();
   516 				TRACE_IRQ8(20, h, orig);
   517 				(*h->iFn)(h->iPtr);
   518 				TRACE_IRQ4(21, h);
   519 				if (!(h->iHState & NIrqHandler::ERunCountMask))	// if run count still nonzero, definitely still active
   520 					{
   521 					NIrq* pI = h->iIrq;
   522 					irq = __SPIN_LOCK_IRQSAVE(pI->iNIrqLock);
   523 					orig = h->EventDone();
   524 					TRACE_IRQ8(22, h, orig);
   525 					if (!(orig & NIrqHandler::EActive))
   526 						{
   527 						// handler is no longer active - can't touch it again
   528 						// pI is OK since NIrq's are never deleted/reused
   529 						requeue = FALSE;
   530 						if (__e32_atomic_add_rel32(&pI->iEventsPending, TUint32(-1)) == 1)
   531 							{
   532 							if (pI->iEnabledEvents & 1)
   533 								{
   534 								pI->iEnabledEvents &= ~1;
   535 								if (pI->iEnabledEvents)
   536 									pI->HwEnable();
   537 								}
   538 							}
   539 						}
   540 					__SPIN_UNLOCK_IRQRESTORE(pI->iNIrqLock,irq);
   541 					}
   542 				break;
   543 				}
   544 			default:
   545 				__KTRACE_OPT(KPANIC,DEBUGPRINT("h=%08x",h));
   546 				__NK_ASSERT_ALWAYS(0);
   547 			}
   548 		if (tied && !requeue)
   549 			{
   550 			// If the tied thread/group has no more tied events outstanding
   551 			// and has a migration pending, trigger the migration now.
   552 			// Atomically change the tied_cpu to the target CPU here. An IDFC
   553 			// can then effect the migration.
   554 			// Note that the tied code can't run in parallel with us until
   555 			// the tied_cpu is changed. However it could run as soon as the
   556 			// tied_cpu is changed (e.g. if added to ready list after change)
   557 			tied->EndTiedEvent();
   558 			}
   559 		if (requeue)
   560 			{
   561 			// still pending so put it back on the queue
   562 			// leave interrupt disabled (if so) and migration of tied thread/group blocked
   563 			aS->QueueEvent(h);
   564 			}
   565 		}
   566 	}
   567 
   568 /******************************************************************************
   569  * Public interrupt management functions
   570  ******************************************************************************/
   571 
   572 void NKern::InterruptInit0()
   573 	 {
   574 	 TInt i;
   575 	 TUint16 cookie = 1;
   576 	 NIrqHandler::FirstFree = 0;
   577 	 for (i=NK_MAX_IRQ_HANDLERS-1; i>=0; --i)
   578 		 {
   579 		 StepCookie(cookie, 61);
   580 		 NIrqHandler* h = &::Handlers[i];
   581 		__KTRACE_OPT(KBOOT,DEBUGPRINT("NIrqHandler[%d] at %08x", i, h));
   582 		 h->iGeneration = 0;
   583 		 h->iHandle = (cookie << 16) | i;
   584 		 h->iIrqLink.iNext = NIrqHandler::FirstFree;
   585 		 NIrqHandler::FirstFree = h;
   586 		 }
   587 	 NIrq::HwInit0();
   588 	 }
   589 
   590 EXPORT_C TInt NKern::InterruptInit(TInt aId, TUint32 aFlags, TInt aVector, TUint32 aHwId, TAny* aExt)
   591 	{
   592 	__KTRACE_OPT(KBOOT,DEBUGPRINT("NKII: ID=%02x F=%08x V=%03x HWID=%08x X=%08x", aId, aFlags, aVector, aHwId, aExt));
   593 	TRACE_IRQ12(0, (aId|(aVector<<16)), aFlags, aHwId);
   594 	if (TUint(aId) >= TUint(NK_MAX_IRQS))
   595   		return KErrArgument;
   596 	NIrq* pI = &Irq[aId];
   597 	__KTRACE_OPT(KBOOT,DEBUGPRINT("NIrq[%02x] at %08x", aId, pI));
   598 	TRACE_IRQ8(1, aId, pI);
   599 	new (pI) NIrq;
   600 	pI->iX = (NIrqX*)aExt;
   601 	pI->iIndex = (TUint16)aId;
   602 	pI->iHwId = aHwId;
   603 	pI->iVector = aVector;
   604 	pI->iStaticFlags = (TUint16)(aFlags & 0x13);
   605 	if (aFlags & NKern::EIrqInit_Count)
   606 		pI->iIState |= NIrq::ECount;
   607 	pI->HwInit();
   608 	__e32_atomic_and_rel32(&pI->iIState, ~NIrq::EWait);
   609 	return KErrNone;
   610 	}
   611 
   612 EXPORT_C TInt NKern::InterruptBind(TInt aId, NIsr aIsr, TAny* aPtr, TUint32 aFlags, NSchedulable* aTied)
   613 	{
   614 	__KTRACE_OPT(KNKERN,DEBUGPRINT(">NKIB: ID=%02x ISR=%08x(%08x) F=%08x T=%T", aId, aIsr, aPtr, aFlags, aTied));
   615 	TRACE_IRQ12(2, aId, aIsr, aPtr);
   616 	TRACE_IRQ12(3, aId, aFlags, aTied);
   617 	CHECK_PRECONDITIONS(MASK_THREAD_STANDARD,"NKern::InterruptBind");
   618 	if (TUint(aId) >= TUint(NK_MAX_IRQS))
   619 		{
   620 		TRACE_IRQ8(4, aId, KErrArgument);
   621 		return KErrArgument;
   622 		}
   623 	NIrq* pI = &Irq[aId];
   624 	NIrqHandler* pH = 0;
   625 	NSchedulable* pT = 0;
   626 	if (aFlags & NKern::EIrqBind_Tied)
   627 		{
   628 		if (!aTied)
   629 			aTied = NKern::CurrentThread();
   630 		pT = aTied;
   631 		}
   632 	TInt r = KErrNoMemory;
   633 	TInt handle = 0;
   634 	NKern::ThreadEnterCS();
   635 	if (!(aFlags & NKern::EIrqBind_Raw))
   636 		{
   637 		pH = NIrqHandler::Alloc();
   638 		if (!pH)
   639 			goto out;
   640 		pH->iFn = aIsr;
   641 		pH->iPtr = aPtr;
   642 		__e32_atomic_add_ord32(&pH->iGeneration, 1);
   643 		if (aFlags & EIrqBind_Exclusive)
   644 			pH->iHState |= NIrqHandler::EExclusive;
   645 		if (aFlags & EIrqBind_Count)
   646 			pH->iHState |= NIrqHandler::ECount;
   647 		r = pI->Bind(pH);
   648 		if (r==KErrNone)
   649 			{
   650 			handle = pH->iHandle;
   651 			// We assume that aTied cannot disappear entirely before we return
   652 			if (pT)
   653 				{
   654 				NKern::Lock();
   655 				r = pT->AddTiedEvent(pH);
   656 				NKern::Unlock();
   657 				}
   658 			if (r!=KErrNone)
   659 				{
   660 				// unbind
   661 				pH->DoUnbind();
   662 				}
   663 			}
   664 		if (r!=KErrNone)
   665 			pH->Free();
   666 		}
   667 	else
   668 		{
   669 		if (aFlags & NKern::EIrqBind_Tied)
   670 			r = KErrNotSupported;
   671 		else
   672 			r = pI->BindRaw(aIsr, aPtr);
   673 		}
   674 out:
   675 	if (r==KErrNone)
   676 		{
   677 		// clear ENotReady so handler can be enabled
   678 		__e32_atomic_and_rel32(&pH->iHState, ~NIrqHandler::ENotReady);
   679 		r = handle;
   680 		}
   681 	NKern::ThreadLeaveCS();
   682 	__KTRACE_OPT(KNKERN,DEBUGPRINT("<NKIB: %08x", r));
   683 	TRACE_IRQ8(4, aId, r);
   684 	return r;
   685 	}
   686 
   687 TInt NIrq::FromHandle(TInt& aHandle, NIrq*& aIrq, NIrqHandler*& aHandler)
   688 	{
   689 	TRACE_IRQ4(5, aHandle);
   690 	aIrq = 0;
   691 	aHandler = 0;
   692 	NIrqHandler* pH = 0;
   693 	NIrqHandler* pH2 = 0;
   694 	NIrq* pI = 0;
   695 	SDblQueLink* anchor = 0;
   696 	TUint32 i;
   697 	TInt r = KErrArgument;
   698 	if (aHandle & NKern::EIrqCookieMask)
   699 		{
   700 		i = aHandle & NKern::EIrqIndexMask;
   701 		if (i>=NK_MAX_IRQ_HANDLERS)
   702 			goto out;
   703 		pH = &::Handlers[i];
   704 		if (pH->iHandle != TUint(aHandle))
   705 			goto out;
   706 		aHandler = pH;
   707 		aIrq = pH->iIrq;
   708 		r = KErrNone;
   709 		goto out;
   710 		}
   711 	if (TUint32(aHandle)>=NK_MAX_IRQS)
   712 		goto out;
   713 	pI = &::Irq[aHandle];
   714 	if (pI->iIState & NIrq::ERaw)
   715 		{
   716 		aIrq = pI;
   717 		r = KErrNone;
   718 		goto out;
   719 		}
   720 	if (pI->iStaticFlags & NIrq::EShared)
   721 		goto out;
   722 	anchor = &pI->iHandlers.iA;
   723 	pH = _LOFF(anchor->iNext, NIrqHandler, iIrqLink);
   724 	i = pH - ::Handlers;
   725 	if (i>=NK_MAX_IRQ_HANDLERS)
   726 		goto out;
   727 	pH2 = &::Handlers[i];
   728 	if (pH2 != pH)
   729 		goto out;
   730 	if (pH->iIrq != pI || anchor->iPrev != anchor->iNext)
   731 		goto out;
   732 	aHandle = pH->iHandle;
   733 	aHandler = pH;
   734 	aIrq = pI;
   735 	r = KErrNone;
   736 out:
   737 	TRACE_IRQ4(6, r);
   738 	TRACE_IRQ12(7, aHandle, aIrq, aHandler);
   739 	return r;
   740 	}
   741 
   742 EXPORT_C TInt NKern::InterruptUnbind(TInt aId)
   743 	{
   744 	TRACE_IRQ4(8, aId);
   745 	__KTRACE_OPT(KNKERN,DEBUGPRINT(">NKIU: ID=%08x", aId));
   746 	CHECK_PRECONDITIONS(MASK_THREAD_STANDARD,"NKern::InterruptUnbind");
   747 	NIrq* pI;
   748 	NIrqHandler* pH;
   749 	TInt r = NIrq::FromHandle(aId, pI, pH);
   750 	if (r!=KErrNone)
   751 		return r;
   752 	NKern::ThreadEnterCS();
   753 	if (!pH)
   754 		{
   755 		// raw ISR
   756 		r = pI->UnbindRaw();
   757 		}
   758 	else
   759 		{
   760 		r = pH->Unbind(aId, 0);
   761 		}
   762 	NKern::ThreadLeaveCS();
   763 	TRACE_IRQ4(9, r);
   764 	return r;
   765 	}
   766 
   767 EXPORT_C TInt NKern::InterruptEnable(TInt aId)
   768 	{
   769 	__KTRACE_OPT(KNKERN,DEBUGPRINT(">NKIE: ID=%08x", aId));
   770 	TRACE_IRQ4(10, aId);
   771 	NIrq* pI;
   772 	NIrqHandler* pH;
   773 	TInt r = NIrq::FromHandle(aId, pI, pH);
   774 	if (r==KErrNone)
   775 		r = pH ? pH->Enable(aId) : pI->EnableRaw();
   776 	TRACE_IRQ4(11, r);
   777 	return r;
   778 	}
   779 
   780 EXPORT_C TInt NKern::InterruptDisable(TInt aId)
   781 	{
   782 	__KTRACE_OPT(KNKERN,DEBUGPRINT(">NKID: ID=%08x", aId));
   783 	TRACE_IRQ4(12, aId);
   784 	NIrq* pI;
   785 	NIrqHandler* pH;
   786 	TInt r = NIrq::FromHandle(aId, pI, pH);
   787 	if (r==KErrNone)
   788 		r = pH ? pH->Disable(FALSE, aId) : pI->DisableRaw(FALSE);
   789 	TRACE_IRQ4(13, r);
   790 	return r;
   791 	}
   792 
   793 EXPORT_C TInt NKern::InterruptClear(TInt aId)
   794 	{
   795 	__KTRACE_OPT(KNKERN,DEBUGPRINT(">NKIC: ID=%08x", aId));
   796 	return KErrNotSupported;
   797 	}
   798 
   799 EXPORT_C TInt NKern::InterruptSetPriority(TInt aId, TInt aPri)
   800 	{
   801 	__KTRACE_OPT(KNKERN,DEBUGPRINT(">NKIS: ID=%08x PRI=%08x", aId, aPri));
   802 	return KErrNotSupported;
   803 	}
   804 
   805 EXPORT_C TInt NKern::InterruptSetCpuMask(TInt aId, TUint32 aMask)
   806 	{
   807 	__KTRACE_OPT(KNKERN,DEBUGPRINT(">NKIM: ID=%08x M=%08x", aId, aMask));
   808 	return KErrNotSupported;
   809 	}
   810 
   811 EXPORT_C void NKern::Interrupt(TInt aId)
   812 	{
   813 	__NK_ASSERT_ALWAYS(TUint(aId) < TUint(NK_MAX_IRQS));
   814 	NIrq* pI = &Irq[aId];
   815 	pI->HwIsr();
   816 	}
   817