#error "fair rwlock needs more bits per long to deal with that many CPUs"
#endif
+/* Test with no contention duration, in seconds */
+#define SINGLE_WRITER_TEST_DURATION 10
+#define SINGLE_READER_TEST_DURATION 10
+#define MULTIPLE_READERS_TEST_DURATION 10
+
/* Test duration, in seconds */
-#define TEST_DURATION 600
+#define TEST_DURATION 60
#define THREAD_ROFFSET 1UL
#define THREAD_RMASK ((NR_CPUS - 1) * THREAD_ROFFSET)
#define HARDIRQ_WOFFSET HARDIRQ_WMASK
#define NR_VARS 100
-#define NR_WRITERS 3
-#define NR_READERS 6
-#define NR_INTERRUPT_READERS 2
-
-/* Writer iteration delay, in ms. 0 for busy loop. */
-#define WRITER_DELAY 1
+#define NR_WRITERS 2
+#define NR_TRYLOCK_WRITERS 2
+#define NR_READERS 4
+#define NR_TRYLOCK_READERS 2
+#define NR_INTERRUPT_READERS 1
+#define NR_TRYLOCK_INTERRUPT_READERS 1
+
+/*
+ * Writer iteration delay, in us. 0 for busy loop. Caution : writers can
+ * starve readers.
+ */
+#define WRITER_DELAY 10
static int var[NR_VARS];
static struct task_struct *reader_threads[NR_READERS];
+static struct task_struct *trylock_reader_threads[NR_TRYLOCK_READERS];
static struct task_struct *writer_threads[NR_WRITERS];
+static struct task_struct *trylock_writer_threads[NR_TRYLOCK_WRITERS];
static struct task_struct *interrupt_reader;
+static struct task_struct *trylock_interrupt_reader;
static struct fair_rwlock frwlock = {
.value = ATOMIC_LONG_INIT(0),
};
+static cycles_t cycles_calibration_min,
+ cycles_calibration_avg,
+ cycles_calibration_max;
+
+static inline cycles_t calibrate_cycles(cycles_t cycles)
+{
+ return cycles - cycles_calibration_avg;
+}
+
struct proc_dir_entry *pentry = NULL;
static int reader_thread(void *data)
int i;
int prev, cur;
unsigned long iter = 0;
- cycles_t time1, time2, delaymax = 0;
+ cycles_t time1, time2, delay, delaymax = 0, delaymin = ULLONG_MAX,
+ delayavg = 0;
printk("reader_thread/%lu runnning\n", (unsigned long)data);
do {
iter++;
preempt_disable(); /* for get_cycles accuracy */
+ rdtsc_barrier();
time1 = get_cycles();
+ rdtsc_barrier();
+
fair_read_lock(&frwlock);
+
+ rdtsc_barrier();
time2 = get_cycles();
- delaymax = max(delaymax, time2 - time1);
+ rdtsc_barrier();
+ delay = time2 - time1;
+ delaymax = max(delaymax, delay);
+ delaymin = min(delaymin, delay);
+ delayavg += delay;
prev = var[0];
for (i = 1; i < NR_VARS; i++) {
cur = var[i];
preempt_enable(); /* for get_cycles accuracy */
//msleep(100);
} while (!kthread_should_stop());
+ delayavg /= iter;
printk("reader_thread/%lu iterations : %lu, "
- "max contention %llu cycles\n",
- (unsigned long)data, iter, delaymax);
+ "lock delay [min,avg,max] %llu,%llu,%llu cycles\n",
+ (unsigned long)data, iter,
+ calibrate_cycles(delaymin),
+ calibrate_cycles(delayavg),
+ calibrate_cycles(delaymax));
return 0;
}
+static int trylock_reader_thread(void *data)
+{
+ int i;
+ int prev, cur;
+ unsigned long iter = 0, success_iter = 0;
+
+ printk("trylock_reader_thread/%lu runnning\n", (unsigned long)data);
+ do {
+ while (!fair_read_trylock(&frwlock))
+ iter++;
+ success_iter++;
+ prev = var[0];
+ for (i = 1; i < NR_VARS; i++) {
+ cur = var[i];
+ if (cur != prev)
+ printk(KERN_ALERT
+ "Unequal cur %d/prev %d at i %d, iter %lu "
+ "in thread\n", cur, prev, i, iter);
+ }
+ fair_read_unlock(&frwlock);
+ //msleep(100);
+ } while (!kthread_should_stop());
+ printk("trylock_reader_thread/%lu iterations : %lu, "
+ "successful iterations : %lu\n",
+ (unsigned long)data, iter, success_iter);
+ return 0;
+}
+
+DEFINE_PER_CPU(cycles_t, int_delaymin);
+DEFINE_PER_CPU(cycles_t, int_delayavg);
DEFINE_PER_CPU(cycles_t, int_delaymax);
+DEFINE_PER_CPU(cycles_t, int_ipi_nr);
static void interrupt_reader_ipi(void *data)
{
int i;
int prev, cur;
cycles_t time1, time2;
- cycles_t *delaymax;
+ cycles_t *delaymax, *delaymin, *delayavg, *ipi_nr, delay;
/*
* Skip the ipi caller, not in irq context.
return;
delaymax = &per_cpu(int_delaymax, smp_processor_id());
+ delaymin = &per_cpu(int_delaymin, smp_processor_id());
+ delayavg = &per_cpu(int_delayavg, smp_processor_id());
+ ipi_nr = &per_cpu(int_ipi_nr, smp_processor_id());
+
+ rdtsc_barrier();
time1 = get_cycles();
+ rdtsc_barrier();
+
fair_read_lock(&frwlock);
+
+ rdtsc_barrier();
time2 = get_cycles();
- *delaymax = max(*delaymax, time2 - time1);
+ rdtsc_barrier();
+ delay = time2 - time1;
+ *delaymax = max(*delaymax, delay);
+ *delaymin = min(*delaymin, delay);
+ *delayavg += delay;
+ (*ipi_nr)++;
+ prev = var[0];
+ for (i = 1; i < NR_VARS; i++) {
+ cur = var[i];
+ if (cur != prev)
+ printk(KERN_ALERT
+ "Unequal cur %d/prev %d at i %d in interrupt\n",
+ cur, prev, i);
+ }
+ fair_read_unlock(&frwlock);
+}
+
+DEFINE_PER_CPU(unsigned long, trylock_int_iter);
+DEFINE_PER_CPU(unsigned long, trylock_int_success);
+
+static void trylock_interrupt_reader_ipi(void *data)
+{
+ int i;
+ int prev, cur;
+
+ /*
+ * Skip the ipi caller, not in irq context.
+ */
+ if (!in_irq())
+ return;
+
+ per_cpu(trylock_int_iter, smp_processor_id())++;
+ while (!fair_read_trylock(&frwlock))
+ per_cpu(trylock_int_iter, smp_processor_id())++;
+ per_cpu(trylock_int_success, smp_processor_id())++;
prev = var[0];
for (i = 1; i < NR_VARS; i++) {
cur = var[i];
fair_read_unlock(&frwlock);
}
+
static int interrupt_reader_thread(void *data)
{
unsigned long iter = 0;
int i;
+ for_each_online_cpu(i) {
+ per_cpu(int_delaymax, i) = 0;
+ per_cpu(int_delaymin, i) = ULLONG_MAX;
+ per_cpu(int_delayavg, i) = 0;
+ per_cpu(int_ipi_nr, i) = 0;
+ }
do {
iter++;
on_each_cpu(interrupt_reader_ipi, NULL, 0);
printk("interrupt_reader_thread/%lu iterations : %lu\n",
(unsigned long)data, iter);
for_each_online_cpu(i) {
+ per_cpu(int_delayavg, i) /= per_cpu(int_ipi_nr, i);
printk("interrupt readers on CPU %i, "
- "max contention : %llu cycles\n",
- i, per_cpu(int_delaymax, i));
+ "lock delay [min,avg,max] %llu,%llu,%llu cycles\n",
+ i,
+ calibrate_cycles(per_cpu(int_delaymin, i)),
+ calibrate_cycles(per_cpu(int_delayavg, i)),
+ calibrate_cycles(per_cpu(int_delaymax, i)));
+ }
+ return 0;
+}
+
+static int trylock_interrupt_reader_thread(void *data)
+{
+ unsigned long iter = 0;
+ int i;
+
+ do {
+ iter++;
+ on_each_cpu(trylock_interrupt_reader_ipi, NULL, 0);
+ msleep(100);
+ } while (!kthread_should_stop());
+ printk("trylock_interrupt_reader_thread/%lu iterations : %lu\n",
+ (unsigned long)data, iter);
+ for_each_online_cpu(i) {
+ printk("trylock interrupt readers on CPU %i, "
+ "iterations %lu, "
+ "successful iterations : %lu\n",
+ i, per_cpu(trylock_int_iter, i),
+ per_cpu(trylock_int_success, i));
+ per_cpu(trylock_int_iter, i) = 0;
+ per_cpu(trylock_int_success, i) = 0;
}
return 0;
}
int i;
int new;
unsigned long iter = 0;
- cycles_t time1, time2, delaymax = 0;
+ cycles_t time1, time2, delay, delaymax = 0, delaymin = ULLONG_MAX,
+ delayavg = 0;
printk("writer_thread/%lu runnning\n", (unsigned long)data);
do {
iter++;
preempt_disable(); /* for get_cycles accuracy */
+ rdtsc_barrier();
time1 = get_cycles();
+ rdtsc_barrier();
+
fair_write_lock_irq(&frwlock);
//fair_write_lock(&frwlock);
+
+ rdtsc_barrier();
time2 = get_cycles();
- delaymax = max(delaymax, time2 - time1);
+ rdtsc_barrier();
+ delay = time2 - time1;
+ delaymax = max(delaymax, delay);
+ delaymin = min(delaymin, delay);
+ delayavg += delay;
new = (int)get_cycles();
for (i = 0; i < NR_VARS; i++) {
var[i] = new;
fair_write_unlock_irq(&frwlock);
preempt_enable(); /* for get_cycles accuracy */
if (WRITER_DELAY > 0)
- msleep(WRITER_DELAY);
+ udelay(WRITER_DELAY);
} while (!kthread_should_stop());
+ delayavg /= iter;
printk("writer_thread/%lu iterations : %lu, "
- "max contention %llu cycles\n",
- (unsigned long)data, iter, delaymax);
+ "lock delay [min,avg,max] %llu,%llu,%llu cycles\n",
+ (unsigned long)data, iter,
+ calibrate_cycles(delaymin),
+ calibrate_cycles(delayavg),
+ calibrate_cycles(delaymax));
+ return 0;
+}
+
+static int trylock_writer_thread(void *data)
+{
+ int i;
+ int new;
+ unsigned long iter = 0, success = 0;
+
+ printk("trylock_writer_thread/%lu runnning\n", (unsigned long)data);
+ do {
+ fair_write_subscribe(&frwlock);
+ while (!fair_write_trylock_subscribed_irq(&frwlock)) {
+ iter++;
+ continue;
+ }
+ success++;
+ //fair_write_lock(&frwlock);
+ new = (int)get_cycles();
+ for (i = 0; i < NR_VARS; i++) {
+ var[i] = new;
+ }
+ //fair_write_unlock(&frwlock);
+ fair_write_unlock_irq(&frwlock);
+ if (WRITER_DELAY > 0)
+ udelay(WRITER_DELAY);
+ } while (!kthread_should_stop());
+ printk("trylock_writer_thread/%lu iterations : %lu, "
+ "successful iterations : %lu\n",
+ (unsigned long)data, iter, success);
return 0;
}
BUG_ON(!reader_threads[i]);
}
+ for (i = 0; i < NR_TRYLOCK_READERS; i++) {
+ printk("starting trylock reader thread %lu\n", i);
+ trylock_reader_threads[i] = kthread_run(trylock_reader_thread,
+ (void *)i, "frwlock_trylock_reader");
+ BUG_ON(!trylock_reader_threads[i]);
+ }
+
+
printk("starting interrupt reader %lu\n", i);
interrupt_reader = kthread_run(interrupt_reader_thread, NULL,
"frwlock_interrupt_reader");
+ printk("starting trylock interrupt reader %lu\n", i);
+ trylock_interrupt_reader = kthread_run(trylock_interrupt_reader_thread,
+ NULL, "frwlock_trylock_interrupt_reader");
for (i = 0; i < NR_WRITERS; i++) {
printk("starting writer thread %lu\n", i);
"frwlock_writer");
BUG_ON(!writer_threads[i]);
}
+ for (i = 0; i < NR_TRYLOCK_WRITERS; i++) {
+ printk("starting trylock writer thread %lu\n", i);
+ trylock_writer_threads[i] = kthread_run(trylock_writer_thread,
+ (void *)i, "frwlock_trylock_writer");
+ BUG_ON(!trylock_writer_threads[i]);
+ }
}
static void fair_rwlock_stop(void)
{
unsigned long i;
- for (i = 0; i < NR_WRITERS; i++) {
+ for (i = 0; i < NR_WRITERS; i++)
kthread_stop(writer_threads[i]);
- }
-
- for (i = 0; i < NR_READERS; i++) {
+ for (i = 0; i < NR_TRYLOCK_WRITERS; i++)
+ kthread_stop(trylock_writer_threads[i]);
+ for (i = 0; i < NR_READERS; i++)
kthread_stop(reader_threads[i]);
- }
-
+ for (i = 0; i < NR_TRYLOCK_READERS; i++)
+ kthread_stop(trylock_reader_threads[i]);
kthread_stop(interrupt_reader);
+ kthread_stop(trylock_interrupt_reader);
}
static int my_open(struct inode *inode, struct file *file)
{
+ unsigned long i;
+ cycles_t time1, time2, delay;
+
+ printk("** get_cycles calibration **\n");
+ cycles_calibration_min = ULLONG_MAX;
+ cycles_calibration_avg = 0;
+ cycles_calibration_max = 0;
+
+ local_irq_disable();
+ for (i = 0; i < 10; i++) {
+ rdtsc_barrier();
+ time1 = get_cycles();
+ rdtsc_barrier();
+ rdtsc_barrier();
+ time2 = get_cycles();
+ rdtsc_barrier();
+ delay = time2 - time1;
+ cycles_calibration_min = min(cycles_calibration_min, delay);
+ cycles_calibration_avg += delay;
+ cycles_calibration_max = max(cycles_calibration_max, delay);
+ }
+ cycles_calibration_avg /= 10;
+ local_irq_enable();
+
+ printk("get_cycles takes [min,avg,max] %llu,%llu,%llu cycles, "
+ "results calibrated on avg\n",
+ cycles_calibration_min,
+ cycles_calibration_avg,
+ cycles_calibration_max);
+
+ printk("** Single writer test, no contention **\n");
+ writer_threads[0] = kthread_run(writer_thread, (void *)0,
+ "frwlock_writer");
+ BUG_ON(!writer_threads[0]);
+ ssleep(SINGLE_WRITER_TEST_DURATION);
+ kthread_stop(writer_threads[0]);
+
+ printk("** Single reader test, no contention **\n");
+ reader_threads[0] = kthread_run(reader_thread, (void *)0,
+ "frwlock_reader");
+ BUG_ON(!reader_threads[0]);
+ ssleep(SINGLE_READER_TEST_DURATION);
+ kthread_stop(reader_threads[0]);
+
+ printk("** Multiple readers test, no contention **\n");
+ for (i = 0; i < NR_READERS; i++) {
+ printk("starting reader thread %lu\n", i);
+ reader_threads[i] = kthread_run(reader_thread, (void *)i,
+ "frwlock_reader");
+ BUG_ON(!reader_threads[i]);
+ }
+ ssleep(SINGLE_READER_TEST_DURATION);
+ for (i = 0; i < NR_READERS; i++)
+ kthread_stop(reader_threads[i]);
+
+ printk("** High contention test **\n");
perform_test("fair-rwlock-create", fair_rwlock_create);
- ssleep(600);
+ ssleep(TEST_DURATION);
perform_test("fair-rwlock-stop", fair_rwlock_stop);
return -EPERM;