Skip to content

Commit

Permalink
Formatting
Browse files Browse the repository at this point in the history
  • Loading branch information
vilukissa68 committed May 6, 2024
1 parent 0d3f47b commit d310d7c
Show file tree
Hide file tree
Showing 2 changed files with 47 additions and 43 deletions.
55 changes: 30 additions & 25 deletions examples/hpc/dla-driver/examples/mac_benchmark.rs
Original file line number Diff line number Diff line change
Expand Up @@ -3,38 +3,38 @@

extern crate alloc;

use headsail_bsp::{rt::entry, sprint, sprintln, init_alloc};
use dla_driver::*;
use headsail_bsp::{init_alloc, rt::entry, sprint, sprintln};
use panic_halt as _;

use rand::{Rng, SeedableRng};
use rand::rngs::SmallRng;
use rand::RngCore;
use rand::{Rng, SeedableRng};

use alloc::vec::*;

macro_rules! conv2d_out_parameters_height {
(($input_h:expr, $kernel_h:expr, $padding_h:expr, $dilation_h:expr, $stride_h:expr)) => {
{
($input_h + 2 * $padding_h - $dilation_h * ($kernel_h - 1) - 1) / $stride_h + 1
}
};
(($input_h:expr, $kernel_h:expr, $padding_h:expr, $dilation_h:expr, $stride_h:expr)) => {{
($input_h + 2 * $padding_h - $dilation_h * ($kernel_h - 1) - 1) / $stride_h + 1
}};
}

macro_rules! conv2d_out_parameters_width {
(($input_w:expr, $kernel_w:expr, $padding_w:expr, $dilation_w:expr, $stride_w:expr)) => {
{
($input_w + 2 * $padding_w - $dilation_w * ($kernel_w - 1) - 1) / $stride_w + 1
}
};
(($input_w:expr, $kernel_w:expr, $padding_w:expr, $dilation_w:expr, $stride_w:expr)) => {{
($input_w + 2 * $padding_w - $dilation_w * ($kernel_w - 1) - 1) / $stride_w + 1
}};
}

fn conv2d_output_parameters(input: (usize, usize), kernel: (usize, usize), padding: (usize, usize),
dilation: (usize, usize), stride: (usize, usize)) -> (usize, usize) {
fn conv2d_output_parameters(
input: (usize, usize),
kernel: (usize, usize),
padding: (usize, usize),
dilation: (usize, usize),
stride: (usize, usize),
) -> (usize, usize) {
let w_out = (input.0 + 2 * padding.0 - dilation.0 * (kernel.0 - 1) - 1) / stride.0 + 1;
let h_out = (input.1 + 2 * padding.1 - dilation.1 * (kernel.1 - 1) - 1) / stride.1 + 1;
(w_out, h_out)

}

fn generate_random_array(buffer: &mut [u8], size: usize) {
Expand All @@ -47,7 +47,7 @@ fn generate_random_array(buffer: &mut [u8], size: usize) {
fn generate_random_matrix(height: usize, width: usize, seed: u64) -> Vec<u8> {
let mut res: Vec<u8> = Vec::new();
let mut rng = SmallRng::seed_from_u64(seed);
for i in 0..(height*width) {
for i in 0..(height * width) {
res.push((rng.next_u64() & 0xFF) as u8);
}
res
Expand All @@ -56,18 +56,25 @@ fn generate_random_matrix(height: usize, width: usize, seed: u64) -> Vec<u8> {
fn generate_random_matrix_small(height: usize, width: usize, seed: u64) -> Vec<u8> {
let mut res: Vec<u8> = Vec::new();
let mut rng = SmallRng::seed_from_u64(seed);
for i in 0..(height*width) {
for i in 0..(height * width) {
res.push((rng.next_u64() & 0x1) as u8);
}
res
}

fn run_random_layer(dla: &mut Dla, in_w: usize, in_h: usize, k_w: usize, k_h: usize, seed: u64) -> Vec<u8> {
fn run_random_layer(
dla: &mut Dla,
in_w: usize,
in_h: usize,
k_w: usize,
k_h: usize,
seed: u64,
) -> Vec<u8> {
// Generate input and kernel
dla.init_layer();

let mut input = generate_random_matrix(in_w, in_h, seed);
let mut kernel = generate_random_matrix_small(k_w, k_h, seed*2);
let mut kernel = generate_random_matrix_small(k_w, k_h, seed * 2);

dla.set_kernel_size(1, k_w, k_h);
dla.set_input_size(1, in_w, in_h);
Expand All @@ -76,19 +83,17 @@ fn run_random_layer(dla: &mut Dla, in_w: usize, in_h: usize, k_w: usize, k_h: us
dla.write_kernel(&mut kernel);

// Calculate output size
let (w_out, h_out) = conv2d_output_parameters((in_w, in_h), (k_w, k_h), (0,0), (1,1), (1,1));
let (w_out, h_out) = conv2d_output_parameters((in_w, in_h), (k_w, k_h), (0, 0), (1, 1), (1, 1));

dla.kernel_data_ready(true);
dla.input_data_ready(true);

// Print the matrix
sprintln!("Waiting for calculation");
while !dla.handle_handshake() {
}
while !dla.handle_handshake() {}
sprintln!("Calculation ready");
let output: Vec<u8> = dla.read_output(w_out * h_out);
let output: Vec<u8> = dla.read_output(w_out * h_out);
output

}

#[entry]
Expand All @@ -102,7 +107,7 @@ fn main() -> ! {
dla.set_pp_clip(8);

for x in 0..2 {
let res = run_random_layer(&mut dla, 8,8,2,2, x*x);
let res = run_random_layer(&mut dla, 8, 8, 2, 2, x * x);
for x in res {
sprint!("{:?} ", x);
}
Expand Down
35 changes: 17 additions & 18 deletions examples/hpc/dla-driver/src/lib.rs
Original file line number Diff line number Diff line change
Expand Up @@ -45,23 +45,23 @@ pub enum MemoryBank {
impl MemoryBank {
fn addr(&self) -> usize {
match self {
MemoryBank::BANK0 => MEMORY_BANK_0_OFFSET,
MemoryBank::BANK1 => MEMORY_BANK_1_OFFSET,
MemoryBank::BANK2 => MEMORY_BANK_2_OFFSET,
MemoryBank::BANK3 => MEMORY_BANK_3_OFFSET,
MemoryBank::BANK4 => MEMORY_BANK_4_OFFSET,
MemoryBank::BANK5 => MEMORY_BANK_5_OFFSET,
MemoryBank::BANK6 => MEMORY_BANK_6_OFFSET,
MemoryBank::BANK7 => MEMORY_BANK_7_OFFSET,
MemoryBank::BANK8 => MEMORY_BANK_8_OFFSET,
MemoryBank::BANK9 => MEMORY_BANK_9_OFFSET,
MemoryBank::BANK10 => MEMORY_BANK_10_OFFSET,
MemoryBank::BANK11 => MEMORY_BANK_11_OFFSET,
MemoryBank::BANK12 => MEMORY_BANK_12_OFFSET,
MemoryBank::BANK13 => MEMORY_BANK_13_OFFSET,
MemoryBank::BANK14 => MEMORY_BANK_14_OFFSET,
MemoryBank::BANK15 => MEMORY_BANK_15_OFFSET,
_ => 0,
MemoryBank::BANK0 => MEMORY_BANK_0_OFFSET,
MemoryBank::BANK1 => MEMORY_BANK_1_OFFSET,
MemoryBank::BANK2 => MEMORY_BANK_2_OFFSET,
MemoryBank::BANK3 => MEMORY_BANK_3_OFFSET,
MemoryBank::BANK4 => MEMORY_BANK_4_OFFSET,
MemoryBank::BANK5 => MEMORY_BANK_5_OFFSET,
MemoryBank::BANK6 => MEMORY_BANK_6_OFFSET,
MemoryBank::BANK7 => MEMORY_BANK_7_OFFSET,
MemoryBank::BANK8 => MEMORY_BANK_8_OFFSET,
MemoryBank::BANK9 => MEMORY_BANK_9_OFFSET,
MemoryBank::BANK10 => MEMORY_BANK_10_OFFSET,
MemoryBank::BANK11 => MEMORY_BANK_11_OFFSET,
MemoryBank::BANK12 => MEMORY_BANK_12_OFFSET,
MemoryBank::BANK13 => MEMORY_BANK_13_OFFSET,
MemoryBank::BANK14 => MEMORY_BANK_14_OFFSET,
MemoryBank::BANK15 => MEMORY_BANK_15_OFFSET,
_ => 0,
}
}
fn value(&self) -> usize {
Expand All @@ -85,7 +85,6 @@ impl MemoryBank {
_ => 0,
}
}

}

#[derive(Copy, Clone)]
Expand Down

0 comments on commit d310d7c

Please sign in to comment.