posix: pthread: replace irq_lock with spinlock

We shouldn't use swapping with an interrupt lock held
as it works incorrectly on SMP platforms.

Fix that by replacing irq_lock with spinlock for pthread
subsystem.

NOTE: we fix that in a simple way with single spinlock
for mutex / cond_var / barrier. That could be improved
later (i.e. split it for several spinlocks).

Signed-off-by: Eugeniy Paltsev <Eugeniy.Paltsev@synopsys.com>
Signed-off-by: Evgeniy Paltsev <PaltsevEvgeniy@gmail.com>
diff --git a/lib/posix/pthread_barrier.c b/lib/posix/pthread_barrier.c
index 2126256..dcf9287 100644
--- a/lib/posix/pthread_barrier.c
+++ b/lib/posix/pthread_barrier.c
@@ -9,9 +9,11 @@
 #include <ksched.h>
 #include <wait_q.h>
 
+extern struct k_spinlock z_pthread_spinlock;
+
 int pthread_barrier_wait(pthread_barrier_t *b)
 {
-	unsigned int key = irq_lock();
+	k_spinlock_key_t key = k_spin_lock(&z_pthread_spinlock);
 	int ret = 0;
 
 	b->count++;
@@ -22,10 +24,10 @@
 		while (z_waitq_head(&b->wait_q)) {
 			_ready_one_thread(&b->wait_q);
 		}
-		z_reschedule_irqlock(key);
+		z_reschedule(&z_pthread_spinlock, key);
 		ret = PTHREAD_BARRIER_SERIAL_THREAD;
 	} else {
-		(void) z_pend_curr_irqlock(key, &b->wait_q, K_FOREVER);
+		(void) z_pend_curr(&z_pthread_spinlock, key, &b->wait_q, K_FOREVER);
 	}
 
 	return ret;
diff --git a/lib/posix/pthread_cond.c b/lib/posix/pthread_cond.c
index e84f270..b397c4a 100644
--- a/lib/posix/pthread_cond.c
+++ b/lib/posix/pthread_cond.c
@@ -9,6 +9,8 @@
 #include <wait_q.h>
 #include <posix/pthread.h>
 
+extern struct k_spinlock z_pthread_spinlock;
+
 int64_t timespec_to_timeoutms(const struct timespec *abstime);
 
 static int cond_wait(pthread_cond_t *cv, pthread_mutex_t *mut,
@@ -16,12 +18,13 @@
 {
 	__ASSERT(mut->lock_count == 1U, "");
 
-	int ret, key = irq_lock();
+	int ret;
+	k_spinlock_key_t key = k_spin_lock(&z_pthread_spinlock);
 
 	mut->lock_count = 0U;
 	mut->owner = NULL;
 	_ready_one_thread(&mut->wait_q);
-	ret = z_pend_curr_irqlock(key, &cv->wait_q, timeout);
+	ret = z_pend_curr(&z_pthread_spinlock, key, &cv->wait_q, timeout);
 
 	/* FIXME: this extra lock (and the potential context switch it
 	 * can cause) could be optimized out.  At the point of the
@@ -49,23 +52,23 @@
 
 int pthread_cond_signal(pthread_cond_t *cv)
 {
-	int key = irq_lock();
+	k_spinlock_key_t key = k_spin_lock(&z_pthread_spinlock);
 
 	_ready_one_thread(&cv->wait_q);
-	z_reschedule_irqlock(key);
+	z_reschedule(&z_pthread_spinlock, key);
 
 	return 0;
 }
 
 int pthread_cond_broadcast(pthread_cond_t *cv)
 {
-	int key = irq_lock();
+	k_spinlock_key_t key = k_spin_lock(&z_pthread_spinlock);
 
 	while (z_waitq_head(&cv->wait_q)) {
 		_ready_one_thread(&cv->wait_q);
 	}
 
-	z_reschedule_irqlock(key);
+	z_reschedule(&z_pthread_spinlock, key);
 
 	return 0;
 }
diff --git a/lib/posix/pthread_mutex.c b/lib/posix/pthread_mutex.c
index 3e541b2..e56fca3 100644
--- a/lib/posix/pthread_mutex.c
+++ b/lib/posix/pthread_mutex.c
@@ -9,6 +9,8 @@
 #include <wait_q.h>
 #include <posix/pthread.h>
 
+struct k_spinlock z_pthread_spinlock;
+
 int64_t timespec_to_timeoutms(const struct timespec *abstime);
 
 #define MUTEX_MAX_REC_LOCK 32767
@@ -22,13 +24,14 @@
 
 static int acquire_mutex(pthread_mutex_t *m, k_timeout_t timeout)
 {
-	int rc = 0, key = irq_lock();
+	int rc = 0;
+	k_spinlock_key_t key = k_spin_lock(&z_pthread_spinlock);
 
 	if (m->lock_count == 0U && m->owner == NULL) {
 		m->lock_count++;
 		m->owner = pthread_self();
 
-		irq_unlock(key);
+		k_spin_unlock(&z_pthread_spinlock, key);
 		return 0;
 	} else if (m->owner == pthread_self()) {
 		if (m->type == PTHREAD_MUTEX_RECURSIVE &&
@@ -41,16 +44,16 @@
 			rc = EINVAL;
 		}
 
-		irq_unlock(key);
+		k_spin_unlock(&z_pthread_spinlock, key);
 		return rc;
 	}
 
 	if (K_TIMEOUT_EQ(timeout, K_NO_WAIT)) {
-		irq_unlock(key);
+		k_spin_unlock(&z_pthread_spinlock, key);
 		return EINVAL;
 	}
 
-	rc = z_pend_curr_irqlock(key, &m->wait_q, timeout);
+	rc = z_pend_curr(&z_pthread_spinlock, key, &m->wait_q, timeout);
 	if (rc != 0) {
 		rc = ETIMEDOUT;
 	}
@@ -121,17 +124,17 @@
  */
 int pthread_mutex_unlock(pthread_mutex_t *m)
 {
-	unsigned int key = irq_lock();
+	k_spinlock_key_t key = k_spin_lock(&z_pthread_spinlock);
 
 	k_tid_t thread;
 
 	if (m->owner != pthread_self()) {
-		irq_unlock(key);
+		k_spin_unlock(&z_pthread_spinlock, key);
 		return EPERM;
 	}
 
 	if (m->lock_count == 0U) {
-		irq_unlock(key);
+		k_spin_unlock(&z_pthread_spinlock, key);
 		return EINVAL;
 	}
 
@@ -144,13 +147,13 @@
 			m->lock_count++;
 			arch_thread_return_value_set(thread, 0);
 			z_ready_thread(thread);
-			z_reschedule_irqlock(key);
+			z_reschedule(&z_pthread_spinlock, key);
 			return 0;
 		}
 		m->owner = NULL;
 
 	}
-	irq_unlock(key);
+	k_spin_unlock(&z_pthread_spinlock, key);
 	return 0;
 }