RustMP/src/bin/test_ad_hoc.rs

143 lines
4.5 KiB
Rust

use std::sync::*;
//use std::sync::RwLock;
use std::sync::atomic::*;
use std::thread;
//use rustmp::rmp_parallel_for;
/*
* A basic sequential function that we want to modify.
* This is written in the most human readable way possible. Not compatable
* with `#[rmp_parallel_for]`.
*/
fn seq_main() {
let mut counter = 0;
for i in 0..4 {
println!("Index {}: Hello from loop {}!", counter, i);
counter += 1;
}
}
/*
* What the user should write for an RustMP parallel program
* The parallel section needs to be in its own function, due to current
* Rust limitations with custom attributes on expressions.
* See <https://github.com/rust-lang/rust/issues/54727> for details
*
* Function should be valid Rust regardless of whether macro is applied.
* Whether Arc gets applied automatically or manually is up for debate.
*/
fn aug_main() {
let mut counter = 0;
//#[rmp_parallel_for(shared(counter) schedule(static, 1))]
fn _loop(counter: &mut i32) {
for i in 0..4 {
println!("Index {}: Hello from loop {}!", counter, i);
*counter += 1;
}} _loop(&mut counter);
}
/*
* What `#[rmp_parallel_for]` should convert the function into
* Current implementation is still ad-hoc, but hopefully the
* macro would expand the function as designed.
*
* Number of threads __rmp_internal_max_threads = 4
*/
fn rmp_main() {
let mut counter = 0;
fn _loop(counter: &mut i32) {
// Startup - Populate environment variables using env::var
let __rmp_internal_max_threads = 4;
// Startup - Populate macro parameters
let __rmp_internal_block_size = 1;
// Startup - Initialize required arrays
let mut __rmp_internal_threads_arr = vec![];
let mut __rmp_internal_iter_arr = vec![];
for _ in 0..__rmp_internal_max_threads {
__rmp_internal_iter_arr.push(vec![]);
}
let mut __rmp_internal_curr_block_size = 0;
let mut __rmp_internal_curr_block_thread = 0;
// Startup - Promote shared mutables into Arc references
// Idea - Possible optimization based on type? RwLock is expensive.
let __rmp_var_counter = Arc::new(AtomicI32::new(*counter));
// Execution - Precompute the iterations for each loop
// The 0..4 here should be parsed from the original tokens
for __rmp_internal_i in 0..4 {
__rmp_internal_iter_arr[__rmp_internal_curr_block_thread].push(__rmp_internal_i);
__rmp_internal_curr_block_size += 1;
if __rmp_internal_curr_block_size >= __rmp_internal_block_size {
__rmp_internal_curr_block_thread = (__rmp_internal_curr_block_thread + 1) % __rmp_internal_max_threads;
}
}
// Execution - Spawn threads with loop contents
for __rmp_internal_iter in __rmp_internal_iter_arr {
// Clone used Arcs here
let __rmp_var_counter = Arc::clone(&__rmp_var_counter);
// Spawn threads
__rmp_internal_threads_arr.push(thread::spawn(move || {
for i in __rmp_internal_iter {
// Having separate load and fetch_add should be a data race,
// However, I believe OpenMP also treats it as a data race,
// so its fine to have this issue
// Need to implement #[rmp_critical] to update it correctly
println!("Index {}: Hello from loop {}!", __rmp_var_counter.load(Ordering::SeqCst), i);
__rmp_var_counter.fetch_add(1, Ordering::SeqCst);
}
}));
}
// Cleanup - Wait for threads
for __rmp_internal_thread in __rmp_internal_threads_arr {
let _ = __rmp_internal_thread.join();
}
// Cleanup - Restore variables from Arc references
*counter = __rmp_var_counter.load(Ordering::SeqCst);
} _loop(&mut counter);
}
/*
* A basic parallel function written by hand in the most human readable way
* possible.
*/
fn par_main() {
let counter = Arc::new(AtomicIsize::new(0));
let mut children = vec![];
for i in 0..4 {
let counter = Arc::clone(&counter);
children.push(thread::spawn(move || {
let index = counter.fetch_add(1, Ordering::SeqCst);
println!("Index {}: Hello from loop {}!", index, i);
}));
}
for child in children {
let _ = child.join();
}
}
fn main() {
println!("Running Sequential Version:");
seq_main();
println!("\nRunning Augmented Sequential Version:");
aug_main();
println!("\nRunning Ad-hoc Parallel Version:");
par_main();
println!("\nRunning Augmented Parallel Version:");
rmp_main();
}