From ea5a1a47c051cbe22e8c522142553e5126d739d9 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Wed, 16 Oct 2024 23:38:56 +0530
Subject: [PATCH 01/31] Update README.md
---
src/README.md | 20 ++++++++++----------
1 file changed, 10 insertions(+), 10 deletions(-)
diff --git a/src/README.md b/src/README.md
index 5c4074a..5318030 100644
--- a/src/README.md
+++ b/src/README.md
@@ -5,43 +5,43 @@ A list of functions that have been implemented can be found here :-
>This C++ code implements LU factorization using OpenMP for parallel execution of matrix updates. It optimizes the decomposition by distributing computations for the lower (L) and upper (U) triangular matrices across multiple threads.
### 2) Maximum element search
->The code for this function can be found in [max.cpp](max.cpp), and input for the following can be found in input.cpp
+>The code for this function can be found in [max.cpp](src/max.cpp), and input for the following can be found in input.cpp
The code uses OpenMP for parallel programming to find the maximum element in an array. The search is distributed across multiple threads, improving performance by dividing the workload.
### 3) Matrix Matrix Multiplication
->The code for the following function can be found in [mm.cpp](mm.cpp)
+>The code for the following function can be found in [mm.cpp](src/mm.cpp)
This code performs matrix-matrix multiplication using OpenMP to parallelize the computation across multiple threads. It optimizes the multiplication process for large matrices, reducing execution time by distributing the workload across available CPU cores.
### 4) Montecarlo Method
->The code for the following function can be found in [montecarlo.cpp](montecarlo.cpp)
+>The code for the following function can be found in [montecarlo.cpp](src/montecarlo.cpp)
The code estimates the value of Pi using the Monte Carlo method with OpenMP for parallel processing. It simulates random points within a unit square and counts how many fall within the unit circle, then uses multiple threads to improve performance and speed up the estimation process.
### 5) Matrix Vector Multiplication
->The code for the following function can be found in [mv.cpp](mv.cpp)
+>The code for the following function can be found in [mv.cpp](src/mv.cpp)
The code performs matrix-vector multiplication using OpenMP for parallel processing. The dynamic scheduling with a chunk size of 16 distributes the computation of each row of the matrix across multiple threads, optimizing the execution for large-scale data by balancing the load dynamically.
### 6) Product of elements of an array
->The code for the following function can be found in [prod.cpp](prod.cpp)
+>The code for the following function can be found in [prod.cpp](src/prod.cpp)
This C++ code calculates the product of elements in an array using OpenMP to parallelize the computation. It optimizes large product calculations by summing the logarithms of array elements in parallel and exponentiating the result to obtain the final product, reducing potential overflow risks.
### 7) Pi reduction
->The code for the following function can be found in [pi-reduction.cpp](pi-reduction.cpp)
+>The code for the following function can be found in [pi-reduction.cpp](src/pi-reduction.cpp)
This C++ code estimates the value of Pi using numerical integration with the OpenMP library for parallelization. It divides the computation of the integral into multiple threads, summing partial results in parallel using a reduction clause to optimize the performance and accuracy when calculating Pi across a large number of steps.
### 8) Calculation of Standard Deviation
->The code for the following function can be found in [standard_dev.cpp](standard_dev.cpp)
+>The code for the following function can be found in [standard_dev.cpp](src/standard_dev.cpp)
This C++ code calculates the standard deviation of a dataset using OpenMP for parallel processing. It first computes the mean in parallel, then calculates the variance by summing the squared differences from the mean, distributing both tasks across multiple threads to improve performance with large datasets.
### 9) Sum of elements of an array
->The code for the following function can be found in [sum2.cpp](sum2.cpp)
+>The code for the following function can be found in [sum2.cpp](src/sum2.cpp)
This C++ code computes the sum of a large array (with 10 million elements) in parallel using OpenMP. It divides the workload among multiple threads based on the total number of threads, each thread calculates a partial sum, and the results are combined in a critical section to avoid race conditions. The execution time for the sum computation is also measured and displayed.
### 10) Vector-Vector Dot product calculation
->The code for the following function can be found in [vvd.cpp](vvd.cpp)
+>The code for the following function can be found in [vvd.cpp](src/vvd.cpp)
This C++ code calculates the dot product of two arrays using OpenMP for parallelization. It initializes two arrays, A and B, each containing 1000 elements set to 1. The dot product is computed in parallel using a dynamic scheduling strategy, with a chunk size of 100, and the results are combined using a reduction operation. The final result is printed to the console.
### 11) Sum calculation (wrong as pragma barrier is not calculated)
->The code for the following function can be found in [wrong_sum.cpp](wrong.cpp)
+>The code for the following function can be found in [wrong_sum.cpp](src/wrong.cpp)
This C++ code computes the sum of an array using OpenMP with task-based parallelism. It initializes an array of size 600 with all elements set to 1. The code divides the summation task into segments of size 100, allowing multiple threads to process these segments concurrently. The results from each task are accumulated into a shared variable sum using a critical section to prevent data races.
## 0.2) Compilation
From c14016ce650474f269dbe6329127ae470d7c95c1 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Wed, 16 Oct 2024 23:57:03 +0530
Subject: [PATCH 02/31] Update README.md
---
README.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/README.md b/README.md
index 0b3ac9f..89eaab6 100644
--- a/README.md
+++ b/README.md
@@ -87,7 +87,7 @@ Here you can find out more about `MPICH`: [https://www.mpich.org/](https://www.m
| | |-- data.txt
| |-- Contributing.md
| |-- LU_factorisation.cpp
- | |-- README.md
+ | |-- [README.md](src/README.md)
| |-- algoimage.png
| |-- broadcast.cpp
| |-- concatenate.cpp
From f6bee836276cc245fea3ba173a6c1f3bc6e4e4bb Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 00:04:40 +0530
Subject: [PATCH 03/31] Update README.md
---
README.md | 5 +++--
1 file changed, 3 insertions(+), 2 deletions(-)
diff --git a/README.md b/README.md
index 89eaab6..73e0dca 100644
--- a/README.md
+++ b/README.md
@@ -87,7 +87,7 @@ Here you can find out more about `MPICH`: [https://www.mpich.org/](https://www.m
| | |-- data.txt
| |-- Contributing.md
| |-- LU_factorisation.cpp
- | |-- [README.md](src/README.md)
+ | |-- README.md
| |-- algoimage.png
| |-- broadcast.cpp
| |-- concatenate.cpp
@@ -146,7 +146,8 @@ Here you can find out more about `MPICH`: [https://www.mpich.org/](https://www.m
| |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
|-- Application
| |-- page-rank.cpp
-
+### mote
+> Information about Functions in main is provided in [READMA.md](src/README.md)
## Resources
- [Parallelization with MPI and OpenMPI](http://compphysics.github.io/ComputationalPhysics2/doc/LectureNotes/_build/html/parallelization.html#)
- [OpenMP](https://medium.com/swlh/openmp-on-ubuntu-1145355eeb2)
From bab89f4bf0ba5e5940ca11a64d29346c9f666717 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:20:29 +0530
Subject: [PATCH 04/31] Update README.md
---
README.md | 7 +++----
1 file changed, 3 insertions(+), 4 deletions(-)
diff --git a/README.md b/README.md
index 73e0dca..adf481f 100644
--- a/README.md
+++ b/README.md
@@ -146,9 +146,8 @@ Here you can find out more about `MPICH`: [https://www.mpich.org/](https://www.m
| |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
|-- Application
| |-- page-rank.cpp
-### mote
+### Note
> Information about Functions in main is provided in [READMA.md](src/README.md)
-## Resources
-- [Parallelization with MPI and OpenMPI](http://compphysics.github.io/ComputationalPhysics2/doc/LectureNotes/_build/html/parallelization.html#)
-- [OpenMP](https://medium.com/swlh/openmp-on-ubuntu-1145355eeb2)
+> For contributing to this repo kindly go through the guidelines provided in [Contributing.md](src/Contributing.md)
+
From 93dabc0aff5f49a9c63b5949dd4d96a8955ee2c7 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:20:53 +0530
Subject: [PATCH 05/31] Update README.md
---
README.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/README.md b/README.md
index adf481f..e11abad 100644
--- a/README.md
+++ b/README.md
@@ -147,7 +147,7 @@ Here you can find out more about `MPICH`: [https://www.mpich.org/](https://www.m
|-- Application
| |-- page-rank.cpp
### Note
-> Information about Functions in main is provided in [READMA.md](src/README.md)
+> Information about Functions in main is provided in [READMA.md](src/README.md)
> For contributing to this repo kindly go through the guidelines provided in [Contributing.md](src/Contributing.md)
From fd8be1edaab15740ccf8da99014349707062c5a9 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:23:00 +0530
Subject: [PATCH 06/31] Update README.md
---
README.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/README.md b/README.md
index e11abad..4a98bac 100644
--- a/README.md
+++ b/README.md
@@ -147,7 +147,7 @@ Here you can find out more about `MPICH`: [https://www.mpich.org/](https://www.m
|-- Application
| |-- page-rank.cpp
### Note
-> Information about Functions in main is provided in [READMA.md](src/README.md)
+> Information about Functions in main is provided in [README.md](src/README.md)
> For contributing to this repo kindly go through the guidelines provided in [Contributing.md](src/Contributing.md)
From 2b40e4b5780c93615e6cce376dc93c973ccaf231 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:25:46 +0530
Subject: [PATCH 07/31] Update Contributing.md
---
src/Contributing.md | 39 ++++++++++++++++++++++++++-------------
1 file changed, 26 insertions(+), 13 deletions(-)
diff --git a/src/Contributing.md b/src/Contributing.md
index dda2152..914dc68 100644
--- a/src/Contributing.md
+++ b/src/Contributing.md
@@ -3,19 +3,32 @@
Please note that this project is released with a **HELLO-FOSS**.
By participating in this project you agree to abide by its terms.
-Ensure your pull request adheres to the following guidelines:
-
-- Before submitting, please ensure that similar suggestions haven't already been made by searching through previous contributions.
-- Open source applications submitted must include an English-language README.md, a screenshot of the app in the README, and provide binaries for at least one operating system, ideally covering macOS, Linux, and Windows.- Submitted packages should be tested and documented.
-- Make an individual pull request for each suggestion.
-- Any submitted packages must be properly tested and come with clear documentation.
-- New categories, or improvements to the existing categorization are welcome.
-- Keep descriptions short and simple, but descriptive.
-- Start the description with a capital and end with a full stop/period.
-- Check your spelling and grammar.
-- Make sure your text editor is set to remove trailing whitespace.
-- The pull request should have a useful title and include a link to the package and why it should be included.
+If you would like to contribute to the project, please follow these guidelines:
+1. Fork the original WnCC repository to your personal account.
+
+2. Clone the forked repository locally.
+
+3. Create a new branch for your feature or bug fix.
+
+4. Make the necessary changes and commit them.
+
+5. Push your changes to your forked repository.
+
+6. Submit a pull request to the main repository with your branch, explaining the changes you made and any additional information that might be helpful for review.
+
+# Usage
+> Clone the Git repository:
+
+```shell
+ # Clone your fork of the GitHub Repo
+ git clone https://github.com/your_username/SoC-Portal.git
+```
+
+> Follow the installation and compilation steps provided in [README.md](README.md)
By following these guidelines, you help maintain the quality and organization of the project!
-
+
+## Resources
+- [Parallelization with MPI and OpenMPI](http://compphysics.github.io/ComputationalPhysics2/doc/LectureNotes/_build/html/parallelization.html#)
+- [OpenMP](https://medium.com/swlh/openmp-on-ubuntu-1145355eeb2)
***HAPPY LEARNING 😀😀😀***
From 8b170a089a597bc9eba132b5b0a714cdf7999fc9 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:26:12 +0530
Subject: [PATCH 08/31] Update Contributing.md
---
src/Contributing.md | 1 -
1 file changed, 1 deletion(-)
diff --git a/src/Contributing.md b/src/Contributing.md
index 914dc68..f1b56bb 100644
--- a/src/Contributing.md
+++ b/src/Contributing.md
@@ -24,7 +24,6 @@ If you would like to contribute to the project, please follow these guidelines:
# Clone your fork of the GitHub Repo
git clone https://github.com/your_username/SoC-Portal.git
```
-
> Follow the installation and compilation steps provided in [README.md](README.md)
By following these guidelines, you help maintain the quality and organization of the project!
From 60727e757779825610d54d48f0984c137c5767da Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:27:40 +0530
Subject: [PATCH 09/31] Update Contributing.md
---
src/Contributing.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/src/Contributing.md b/src/Contributing.md
index f1b56bb..9d65e3f 100644
--- a/src/Contributing.md
+++ b/src/Contributing.md
@@ -24,7 +24,7 @@ If you would like to contribute to the project, please follow these guidelines:
# Clone your fork of the GitHub Repo
git clone https://github.com/your_username/SoC-Portal.git
```
-> Follow the installation and compilation steps provided in [README.md](README.md)
+> Follow the installation and compilation steps provided in [README.md](https://github.com/LoPA607/Hello-Foss-CPP/README.md)
By following these guidelines, you help maintain the quality and organization of the project!
## Resources
From 6eab7679ba8875a0b71b29fc69604dc7304af3cd Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:29:39 +0530
Subject: [PATCH 10/31] Delete README.md
---
README.md | 153 ------------------------------------------------------
1 file changed, 153 deletions(-)
delete mode 100644 README.md
diff --git a/README.md b/README.md
deleted file mode 100644
index 4a98bac..0000000
--- a/README.md
+++ /dev/null
@@ -1,153 +0,0 @@
-# Parallel_Programming
-Parallel programming is often used interchangeably with parallel processing or in connection with parallel computing, which refers to systems that enhance the efficiency of parallel execution. In parallel programming, tasks are divided and executed simultaneously across multiple computers or multiple CPU cores. This approach is essential for large-scale projects where speed and precision are crucial. Although complex, parallel programming enables developers, researchers, and users to perform research and analysis much faster than programs limited to processing tasks sequentially
-
-## Project Description
-This project demonstrates the use of parallel programming techniques in C++ with (`OpenMP`) and (`MPI`) to efficiently execute computations across multiple threads. It includes examples of parallel loops, task scheduling, and thread synchronization. The goal is to showcase performance improvements in computational tasks using OpenMP and MPI directives.
-
-## **1. OpenMP (Open Multi-Processing)**
-
-`OpenMP` is a multi-platform programming interface that enables multiprocessing programming. `OpenMP` can be used in C++, C and Fortran languages, including different architectures like Windows and Unix. It consists of compilator directives that have an impact on code execution.
-
-The `OpenMP` interface is a component of the GNU Compiler Collection (`GCC`), a set of open-source compilers developed by the GNU Project. `GCC` compiler is therefore highly recommended for use with `OpenMP`, although it is not required (there is an Intel compiler that also support `OpenMP`).
-
-**INSTALLATION AND CONFIGURATION ON LINUX SYSTEMS:**
-
-Start the terminal and update the repository:
-
-```bash
->>> sudo apt-get update
-```
-
-Then install the `build-essential` package, including `gcc`, `g++` and `make`:
-
-```bash
->>> sudo apt-get install build-essential
-```
-
-## **2. MPI**
-
-Message Passing Interface (`MPI`) is a communication protocol standard for transferring messages between parallel program processes on one or more computers. `MPI` is currently the most widely used communication model in clusters of computers and supercomputers.
-
-There are several implementations of `MPI`, including `OpenMPI`, `MPICH` and `MSMPI`. On Linux, we can choose from `OpenMPI` and `MPICH`, while `MSMPI` is a Windows implementation. Before going any further, we should ensure that we have the `GCC` compiler installed.
-
-**INSTALLATION AND CONFIGURATION OF `MPICH` ON LINUX SYSTEMS:**
-
-Start the terminal and update the repository:
-
-```bash
->>> sudo apt-get update
-```
-
-We then install the `mpich` package:
-
-```bash
->>> sudo apt-get install mpich
-```
-
-We can now check the version of the installed `MPI` (this will actually be the `GCC` version):
-
-```bash
->>> mpic++ --version
-```
-
-Here you can find out more about `MPICH`: [https://www.mpich.org/](https://www.mpich.org/).
-
-
-**THE INSTALLATION PROCESS UNDER WINDOWS IS COMPLEX, AND I DO NOT RECOMMEND USING MPI WITH THE WINDOWS PLATFORM**
-
-**NOTE**
-> Linux is often recommended for compiling parallel programs
-> For optimal performance and seamless development, we recommend using Linux for building and running this project.
-
-## Project Tree
-> This project contains two branches: `main` and `Advanced`.
-> The `main` branch includes operations from the NumPy module that have been translated into C++ using OpenMP and MPI for parallel processing.
-
-### **main**
-
- .
- |-- include
- | |-- array_concat.h
- | |-- dot_product.h
- | |-- find_max.h
- | |-- lu_decomposition.h
- | |-- matrix_mult.h
- | |-- matrix_vector.h
- | |-- montecarlo.h
- | |-- parallel_sum.h
- | |-- pi_calculator.h
- | |-- product_log.h
- | |-- statistics.h
- | |-- sum2.h
- | |-- sum_task.h
- |
- |-- src
- | |-- Histogram/
- | | |-- histo.cpp
- | | |-- data.txt
- | |-- Contributing.md
- | |-- LU_factorisation.cpp
- | |-- README.md
- | |-- algoimage.png
- | |-- broadcast.cpp
- | |-- concatenate.cpp
- | |-- input.txt(for max.cpp)
- | |-- max.cpp
- | |-- mm.cpp (Matrix Multiplication implementation)
- | |-- montecarlo.cpp
- | |-- mv.cpp (Matrix Vector multiplication)
- | |-- pi-reduction.cpp (Parallel pi computation using OpenMP reduction)
- | |-- prod.cpp (Parallel product computation)
- | |-- standard_dev.cpp
- | |-- sum2.cpp
- | |-- sum_task.cpp
- | |-- vvd.cpp (Vector-Vector Dot product)
- | |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
-
-### **Advanced**
-
- .
- |-- include
- | |-- array_concat.h
- | |-- dot_product.h
- | |-- find_max.h
- | |-- lu_decomposition.h
- | |-- matrix_mult.h
- | |-- matrix_vector.h
- | |-- montecarlo.h
- | |-- parallel_sum.h
- | |-- pi_calculator.h
- | |-- product_log.h
- | |-- statistics.h
- | |-- sum2.h
- | |-- sum_task.h
- |
- |-- src
- | |-- Histogram/
- | | |-- histo.cpp
- | | |-- data.txt
- | |-- Contributing.md
- | |-- LU_factorisation.cpp
- | |-- README.md
- | |-- algoimage.png
- | |-- broadcast.cpp
- | |-- concatenate.cpp
- | |-- input.txt(for max.cpp)
- | |-- max.cpp
- | |-- mm.cpp (Matrix Multiplication implementation)
- | |-- montecarlo.cpp
- | |-- mv.cpp (Matrix Vector multiplication)
- | |-- pi-reduction.cpp (Parallel pi computation using OpenMP reduction)
- | |-- prod.cpp (Parallel product computation)
- | |-- standard_dev.cpp
- | |-- sum2.cpp
- | |-- sum_task.cpp
- | |-- vvd.cpp (Vector-Vector Dot product)
- | |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
- |-- Application
- | |-- page-rank.cpp
-### Note
-> Information about Functions in main is provided in [README.md](src/README.md)
-> For contributing to this repo kindly go through the guidelines provided in [Contributing.md](src/Contributing.md)
-
-
From 9f1a195b207e4b1b98e93f0d7ed71014a1322131 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:30:11 +0530
Subject: [PATCH 11/31] Create Intro.md
---
Intro.md | 153 +++++++++++++++++++++++++++++++++++++++++++++++++++++++
1 file changed, 153 insertions(+)
create mode 100644 Intro.md
diff --git a/Intro.md b/Intro.md
new file mode 100644
index 0000000..4a98bac
--- /dev/null
+++ b/Intro.md
@@ -0,0 +1,153 @@
+# Parallel_Programming
+Parallel programming is often used interchangeably with parallel processing or in connection with parallel computing, which refers to systems that enhance the efficiency of parallel execution. In parallel programming, tasks are divided and executed simultaneously across multiple computers or multiple CPU cores. This approach is essential for large-scale projects where speed and precision are crucial. Although complex, parallel programming enables developers, researchers, and users to perform research and analysis much faster than programs limited to processing tasks sequentially
+
+## Project Description
+This project demonstrates the use of parallel programming techniques in C++ with (`OpenMP`) and (`MPI`) to efficiently execute computations across multiple threads. It includes examples of parallel loops, task scheduling, and thread synchronization. The goal is to showcase performance improvements in computational tasks using OpenMP and MPI directives.
+
+## **1. OpenMP (Open Multi-Processing)**
+
+`OpenMP` is a multi-platform programming interface that enables multiprocessing programming. `OpenMP` can be used in C++, C and Fortran languages, including different architectures like Windows and Unix. It consists of compilator directives that have an impact on code execution.
+
+The `OpenMP` interface is a component of the GNU Compiler Collection (`GCC`), a set of open-source compilers developed by the GNU Project. `GCC` compiler is therefore highly recommended for use with `OpenMP`, although it is not required (there is an Intel compiler that also support `OpenMP`).
+
+**INSTALLATION AND CONFIGURATION ON LINUX SYSTEMS:**
+
+Start the terminal and update the repository:
+
+```bash
+>>> sudo apt-get update
+```
+
+Then install the `build-essential` package, including `gcc`, `g++` and `make`:
+
+```bash
+>>> sudo apt-get install build-essential
+```
+
+## **2. MPI**
+
+Message Passing Interface (`MPI`) is a communication protocol standard for transferring messages between parallel program processes on one or more computers. `MPI` is currently the most widely used communication model in clusters of computers and supercomputers.
+
+There are several implementations of `MPI`, including `OpenMPI`, `MPICH` and `MSMPI`. On Linux, we can choose from `OpenMPI` and `MPICH`, while `MSMPI` is a Windows implementation. Before going any further, we should ensure that we have the `GCC` compiler installed.
+
+**INSTALLATION AND CONFIGURATION OF `MPICH` ON LINUX SYSTEMS:**
+
+Start the terminal and update the repository:
+
+```bash
+>>> sudo apt-get update
+```
+
+We then install the `mpich` package:
+
+```bash
+>>> sudo apt-get install mpich
+```
+
+We can now check the version of the installed `MPI` (this will actually be the `GCC` version):
+
+```bash
+>>> mpic++ --version
+```
+
+Here you can find out more about `MPICH`: [https://www.mpich.org/](https://www.mpich.org/).
+
+
+**THE INSTALLATION PROCESS UNDER WINDOWS IS COMPLEX, AND I DO NOT RECOMMEND USING MPI WITH THE WINDOWS PLATFORM**
+
+**NOTE**
+> Linux is often recommended for compiling parallel programs
+> For optimal performance and seamless development, we recommend using Linux for building and running this project.
+
+## Project Tree
+> This project contains two branches: `main` and `Advanced`.
+> The `main` branch includes operations from the NumPy module that have been translated into C++ using OpenMP and MPI for parallel processing.
+
+### **main**
+
+ .
+ |-- include
+ | |-- array_concat.h
+ | |-- dot_product.h
+ | |-- find_max.h
+ | |-- lu_decomposition.h
+ | |-- matrix_mult.h
+ | |-- matrix_vector.h
+ | |-- montecarlo.h
+ | |-- parallel_sum.h
+ | |-- pi_calculator.h
+ | |-- product_log.h
+ | |-- statistics.h
+ | |-- sum2.h
+ | |-- sum_task.h
+ |
+ |-- src
+ | |-- Histogram/
+ | | |-- histo.cpp
+ | | |-- data.txt
+ | |-- Contributing.md
+ | |-- LU_factorisation.cpp
+ | |-- README.md
+ | |-- algoimage.png
+ | |-- broadcast.cpp
+ | |-- concatenate.cpp
+ | |-- input.txt(for max.cpp)
+ | |-- max.cpp
+ | |-- mm.cpp (Matrix Multiplication implementation)
+ | |-- montecarlo.cpp
+ | |-- mv.cpp (Matrix Vector multiplication)
+ | |-- pi-reduction.cpp (Parallel pi computation using OpenMP reduction)
+ | |-- prod.cpp (Parallel product computation)
+ | |-- standard_dev.cpp
+ | |-- sum2.cpp
+ | |-- sum_task.cpp
+ | |-- vvd.cpp (Vector-Vector Dot product)
+ | |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
+
+### **Advanced**
+
+ .
+ |-- include
+ | |-- array_concat.h
+ | |-- dot_product.h
+ | |-- find_max.h
+ | |-- lu_decomposition.h
+ | |-- matrix_mult.h
+ | |-- matrix_vector.h
+ | |-- montecarlo.h
+ | |-- parallel_sum.h
+ | |-- pi_calculator.h
+ | |-- product_log.h
+ | |-- statistics.h
+ | |-- sum2.h
+ | |-- sum_task.h
+ |
+ |-- src
+ | |-- Histogram/
+ | | |-- histo.cpp
+ | | |-- data.txt
+ | |-- Contributing.md
+ | |-- LU_factorisation.cpp
+ | |-- README.md
+ | |-- algoimage.png
+ | |-- broadcast.cpp
+ | |-- concatenate.cpp
+ | |-- input.txt(for max.cpp)
+ | |-- max.cpp
+ | |-- mm.cpp (Matrix Multiplication implementation)
+ | |-- montecarlo.cpp
+ | |-- mv.cpp (Matrix Vector multiplication)
+ | |-- pi-reduction.cpp (Parallel pi computation using OpenMP reduction)
+ | |-- prod.cpp (Parallel product computation)
+ | |-- standard_dev.cpp
+ | |-- sum2.cpp
+ | |-- sum_task.cpp
+ | |-- vvd.cpp (Vector-Vector Dot product)
+ | |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
+ |-- Application
+ | |-- page-rank.cpp
+### Note
+> Information about Functions in main is provided in [README.md](src/README.md)
+> For contributing to this repo kindly go through the guidelines provided in [Contributing.md](src/Contributing.md)
+
+
From 1d640aecca0b340ef57da7efee5cdd257614e013 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:31:04 +0530
Subject: [PATCH 12/31] Update Contributing.md
---
src/Contributing.md | 3 ++-
1 file changed, 2 insertions(+), 1 deletion(-)
diff --git a/src/Contributing.md b/src/Contributing.md
index 9d65e3f..ba73414 100644
--- a/src/Contributing.md
+++ b/src/Contributing.md
@@ -24,7 +24,8 @@ If you would like to contribute to the project, please follow these guidelines:
# Clone your fork of the GitHub Repo
git clone https://github.com/your_username/SoC-Portal.git
```
-> Follow the installation and compilation steps provided in [README.md](https://github.com/LoPA607/Hello-Foss-CPP/README.md)
+> Follow the installation and compilation steps provided in [README.md](Intro.md)
+
By following these guidelines, you help maintain the quality and organization of the project!
## Resources
From 2cd8853b18eaf28cf72e4454b6d627c7c1f5c592 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:31:30 +0530
Subject: [PATCH 13/31] Update Contributing.md
---
src/Contributing.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/src/Contributing.md b/src/Contributing.md
index ba73414..2d74441 100644
--- a/src/Contributing.md
+++ b/src/Contributing.md
@@ -24,7 +24,7 @@ If you would like to contribute to the project, please follow these guidelines:
# Clone your fork of the GitHub Repo
git clone https://github.com/your_username/SoC-Portal.git
```
-> Follow the installation and compilation steps provided in [README.md](Intro.md)
+> Follow the installation and compilation steps provided in [Intro.md](Intro.md)
By following these guidelines, you help maintain the quality and organization of the project!
From 77de7a195a33527ab114e9f67cde609d4101ecda Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:32:36 +0530
Subject: [PATCH 14/31] Update Contributing.md
From daae223a3aca00f7c4b024f5b78b3c77e660daf2 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:33:02 +0530
Subject: [PATCH 15/31] Update Contributing.md
---
src/Contributing.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/src/Contributing.md b/src/Contributing.md
index 2d74441..e33276c 100644
--- a/src/Contributing.md
+++ b/src/Contributing.md
@@ -24,7 +24,7 @@ If you would like to contribute to the project, please follow these guidelines:
# Clone your fork of the GitHub Repo
git clone https://github.com/your_username/SoC-Portal.git
```
-> Follow the installation and compilation steps provided in [Intro.md](Intro.md)
+> Follow the installation and compilation steps provided in [Intro.md](Hello-Foss-CPP/Intro.md)
By following these guidelines, you help maintain the quality and organization of the project!
From 58053c2140802d574e47c076ce0ff49748cc911c Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:34:58 +0530
Subject: [PATCH 16/31] Create Introduction.md
---
Introduction.md | 151 ++++++++++++++++++++++++++++++++++++++++++++++++
1 file changed, 151 insertions(+)
create mode 100644 Introduction.md
diff --git a/Introduction.md b/Introduction.md
new file mode 100644
index 0000000..5249d90
--- /dev/null
+++ b/Introduction.md
@@ -0,0 +1,151 @@
+# Parallel_Programming
+Parallel programming is often used interchangeably with parallel processing or in connection with parallel computing, which refers to systems that enhance the efficiency of parallel execution. In parallel programming, tasks are divided and executed simultaneously across multiple computers or multiple CPU cores. This approach is essential for large-scale projects where speed and precision are crucial. Although complex, parallel programming enables developers, researchers, and users to perform research and analysis much faster than programs limited to processing tasks sequentially
+
+## Project Description
+This project demonstrates the use of parallel programming techniques in C++ with (`OpenMP`) and (`MPI`) to efficiently execute computations across multiple threads. It includes examples of parallel loops, task scheduling, and thread synchronization. The goal is to showcase performance improvements in computational tasks using OpenMP and MPI directives.
+
+## **1. OpenMP (Open Multi-Processing)**
+
+`OpenMP` is a multi-platform programming interface that enables multiprocessing programming. `OpenMP` can be used in C++, C and Fortran languages, including different architectures like Windows and Unix. It consists of compilator directives that have an impact on code execution.
+
+The `OpenMP` interface is a component of the GNU Compiler Collection (`GCC`), a set of open-source compilers developed by the GNU Project. `GCC` compiler is therefore highly recommended for use with `OpenMP`, although it is not required (there is an Intel compiler that also support `OpenMP`).
+
+**INSTALLATION AND CONFIGURATION ON LINUX SYSTEMS:**
+
+Start the terminal and update the repository:
+
+```bash
+>>> sudo apt-get update
+```
+
+Then install the `build-essential` package, including `gcc`, `g++` and `make`:
+
+```bash
+>>> sudo apt-get install build-essential
+```
+
+## **2. MPI**
+
+Message Passing Interface (`MPI`) is a communication protocol standard for transferring messages between parallel program processes on one or more computers. `MPI` is currently the most widely used communication model in clusters of computers and supercomputers.
+
+There are several implementations of `MPI`, including `OpenMPI`, `MPICH` and `MSMPI`. On Linux, we can choose from `OpenMPI` and `MPICH`, while `MSMPI` is a Windows implementation. Before going any further, we should ensure that we have the `GCC` compiler installed.
+
+**INSTALLATION AND CONFIGURATION OF `MPICH` ON LINUX SYSTEMS:**
+
+Start the terminal and update the repository:
+
+```bash
+>>> sudo apt-get update
+```
+
+We then install the `mpich` package:
+
+```bash
+>>> sudo apt-get install mpich
+```
+
+We can now check the version of the installed `MPI` (this will actually be the `GCC` version):
+
+```bash
+>>> mpic++ --version
+```
+
+Here you can find out more about `MPICH`: [https://www.mpich.org/](https://www.mpich.org/).
+
+
+**THE INSTALLATION PROCESS UNDER WINDOWS IS COMPLEX, AND I DO NOT RECOMMEND USING MPI WITH THE WINDOWS PLATFORM**
+
+**NOTE**
+> Linux is often recommended for compiling parallel programs
+> For optimal performance and seamless development, we recommend using Linux for building and running this project.
+
+## Project Tree
+> This project contains two branches: `main` and `Advanced`.
+> The `main` branch includes operations from the NumPy module that have been translated into C++ using OpenMP and MPI for parallel processing.
+
+### **main**
+
+ .
+ |-- include
+ | |-- array_concat.h
+ | |-- dot_product.h
+ | |-- find_max.h
+ | |-- lu_decomposition.h
+ | |-- matrix_mult.h
+ | |-- matrix_vector.h
+ | |-- montecarlo.h
+ | |-- parallel_sum.h
+ | |-- pi_calculator.h
+ | |-- product_log.h
+ | |-- statistics.h
+ | |-- sum2.h
+ | |-- sum_task.h
+ |
+ |-- src
+ | |-- Histogram/
+ | | |-- histo.cpp
+ | | |-- data.txt
+ | |-- Contributing.md
+ | |-- LU_factorisation.cpp
+ | |-- README.md
+ | |-- algoimage.png
+ | |-- broadcast.cpp
+ | |-- concatenate.cpp
+ | |-- input.txt(for max.cpp)
+ | |-- max.cpp
+ | |-- mm.cpp (Matrix Multiplication implementation)
+ | |-- montecarlo.cpp
+ | |-- mv.cpp (Matrix Vector multiplication)
+ | |-- pi-reduction.cpp (Parallel pi computation using OpenMP reduction)
+ | |-- prod.cpp (Parallel product computation)
+ | |-- standard_dev.cpp
+ | |-- sum2.cpp
+ | |-- sum_task.cpp
+ | |-- vvd.cpp (Vector-Vector Dot product)
+ | |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
+
+### **Advanced**
+
+ .
+ |-- include
+ | |-- array_concat.h
+ | |-- dot_product.h
+ | |-- find_max.h
+ | |-- lu_decomposition.h
+ | |-- matrix_mult.h
+ | |-- matrix_vector.h
+ | |-- montecarlo.h
+ | |-- parallel_sum.h
+ | |-- pi_calculator.h
+ | |-- product_log.h
+ | |-- statistics.h
+ | |-- sum2.h
+ | |-- sum_task.h
+ |
+ |-- src
+ | |-- Histogram/
+ | | |-- histo.cpp
+ | | |-- data.txt
+ | |-- Contributing.md
+ | |-- LU_factorisation.cpp
+ | |-- README.md
+ | |-- algoimage.png
+ | |-- broadcast.cpp
+ | |-- concatenate.cpp
+ | |-- input.txt(for max.cpp)
+ | |-- max.cpp
+ | |-- mm.cpp (Matrix Multiplication implementation)
+ | |-- montecarlo.cpp
+ | |-- mv.cpp (Matrix Vector multiplication)
+ | |-- pi-reduction.cpp (Parallel pi computation using OpenMP reduction)
+ | |-- prod.cpp (Parallel product computation)
+ | |-- standard_dev.cpp
+ | |-- sum2.cpp
+ | |-- sum_task.cpp
+ | |-- vvd.cpp (Vector-Vector Dot product)
+ | |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
+ |-- Application
+ | |-- page-rank.cpp
+### Note
+> Information about Functions in main is provided in [README.md](src/README.md)
+> For contributing to this repo kindly go through the guidelines provided in [Contributing.md](src/Contributing.md)
From 1c00ed896e329829be954c499ad6c19328ff8bfb Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:35:09 +0530
Subject: [PATCH 17/31] Delete Intro.md
---
Intro.md | 153 -------------------------------------------------------
1 file changed, 153 deletions(-)
delete mode 100644 Intro.md
diff --git a/Intro.md b/Intro.md
deleted file mode 100644
index 4a98bac..0000000
--- a/Intro.md
+++ /dev/null
@@ -1,153 +0,0 @@
-# Parallel_Programming
-Parallel programming is often used interchangeably with parallel processing or in connection with parallel computing, which refers to systems that enhance the efficiency of parallel execution. In parallel programming, tasks are divided and executed simultaneously across multiple computers or multiple CPU cores. This approach is essential for large-scale projects where speed and precision are crucial. Although complex, parallel programming enables developers, researchers, and users to perform research and analysis much faster than programs limited to processing tasks sequentially
-
-## Project Description
-This project demonstrates the use of parallel programming techniques in C++ with (`OpenMP`) and (`MPI`) to efficiently execute computations across multiple threads. It includes examples of parallel loops, task scheduling, and thread synchronization. The goal is to showcase performance improvements in computational tasks using OpenMP and MPI directives.
-
-## **1. OpenMP (Open Multi-Processing)**
-
-`OpenMP` is a multi-platform programming interface that enables multiprocessing programming. `OpenMP` can be used in C++, C and Fortran languages, including different architectures like Windows and Unix. It consists of compilator directives that have an impact on code execution.
-
-The `OpenMP` interface is a component of the GNU Compiler Collection (`GCC`), a set of open-source compilers developed by the GNU Project. `GCC` compiler is therefore highly recommended for use with `OpenMP`, although it is not required (there is an Intel compiler that also support `OpenMP`).
-
-**INSTALLATION AND CONFIGURATION ON LINUX SYSTEMS:**
-
-Start the terminal and update the repository:
-
-```bash
->>> sudo apt-get update
-```
-
-Then install the `build-essential` package, including `gcc`, `g++` and `make`:
-
-```bash
->>> sudo apt-get install build-essential
-```
-
-## **2. MPI**
-
-Message Passing Interface (`MPI`) is a communication protocol standard for transferring messages between parallel program processes on one or more computers. `MPI` is currently the most widely used communication model in clusters of computers and supercomputers.
-
-There are several implementations of `MPI`, including `OpenMPI`, `MPICH` and `MSMPI`. On Linux, we can choose from `OpenMPI` and `MPICH`, while `MSMPI` is a Windows implementation. Before going any further, we should ensure that we have the `GCC` compiler installed.
-
-**INSTALLATION AND CONFIGURATION OF `MPICH` ON LINUX SYSTEMS:**
-
-Start the terminal and update the repository:
-
-```bash
->>> sudo apt-get update
-```
-
-We then install the `mpich` package:
-
-```bash
->>> sudo apt-get install mpich
-```
-
-We can now check the version of the installed `MPI` (this will actually be the `GCC` version):
-
-```bash
->>> mpic++ --version
-```
-
-Here you can find out more about `MPICH`: [https://www.mpich.org/](https://www.mpich.org/).
-
-
-**THE INSTALLATION PROCESS UNDER WINDOWS IS COMPLEX, AND I DO NOT RECOMMEND USING MPI WITH THE WINDOWS PLATFORM**
-
-**NOTE**
-> Linux is often recommended for compiling parallel programs
-> For optimal performance and seamless development, we recommend using Linux for building and running this project.
-
-## Project Tree
-> This project contains two branches: `main` and `Advanced`.
-> The `main` branch includes operations from the NumPy module that have been translated into C++ using OpenMP and MPI for parallel processing.
-
-### **main**
-
- .
- |-- include
- | |-- array_concat.h
- | |-- dot_product.h
- | |-- find_max.h
- | |-- lu_decomposition.h
- | |-- matrix_mult.h
- | |-- matrix_vector.h
- | |-- montecarlo.h
- | |-- parallel_sum.h
- | |-- pi_calculator.h
- | |-- product_log.h
- | |-- statistics.h
- | |-- sum2.h
- | |-- sum_task.h
- |
- |-- src
- | |-- Histogram/
- | | |-- histo.cpp
- | | |-- data.txt
- | |-- Contributing.md
- | |-- LU_factorisation.cpp
- | |-- README.md
- | |-- algoimage.png
- | |-- broadcast.cpp
- | |-- concatenate.cpp
- | |-- input.txt(for max.cpp)
- | |-- max.cpp
- | |-- mm.cpp (Matrix Multiplication implementation)
- | |-- montecarlo.cpp
- | |-- mv.cpp (Matrix Vector multiplication)
- | |-- pi-reduction.cpp (Parallel pi computation using OpenMP reduction)
- | |-- prod.cpp (Parallel product computation)
- | |-- standard_dev.cpp
- | |-- sum2.cpp
- | |-- sum_task.cpp
- | |-- vvd.cpp (Vector-Vector Dot product)
- | |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
-
-### **Advanced**
-
- .
- |-- include
- | |-- array_concat.h
- | |-- dot_product.h
- | |-- find_max.h
- | |-- lu_decomposition.h
- | |-- matrix_mult.h
- | |-- matrix_vector.h
- | |-- montecarlo.h
- | |-- parallel_sum.h
- | |-- pi_calculator.h
- | |-- product_log.h
- | |-- statistics.h
- | |-- sum2.h
- | |-- sum_task.h
- |
- |-- src
- | |-- Histogram/
- | | |-- histo.cpp
- | | |-- data.txt
- | |-- Contributing.md
- | |-- LU_factorisation.cpp
- | |-- README.md
- | |-- algoimage.png
- | |-- broadcast.cpp
- | |-- concatenate.cpp
- | |-- input.txt(for max.cpp)
- | |-- max.cpp
- | |-- mm.cpp (Matrix Multiplication implementation)
- | |-- montecarlo.cpp
- | |-- mv.cpp (Matrix Vector multiplication)
- | |-- pi-reduction.cpp (Parallel pi computation using OpenMP reduction)
- | |-- prod.cpp (Parallel product computation)
- | |-- standard_dev.cpp
- | |-- sum2.cpp
- | |-- sum_task.cpp
- | |-- vvd.cpp (Vector-Vector Dot product)
- | |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
- |-- Application
- | |-- page-rank.cpp
-### Note
-> Information about Functions in main is provided in [README.md](src/README.md)
-> For contributing to this repo kindly go through the guidelines provided in [Contributing.md](src/Contributing.md)
-
-
From aee7bbdca17cca92a4e144917289ace543ede247 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:35:21 +0530
Subject: [PATCH 18/31] Create README.md
---
README.md | 151 ++++++++++++++++++++++++++++++++++++++++++++++++++++++
1 file changed, 151 insertions(+)
create mode 100644 README.md
diff --git a/README.md b/README.md
new file mode 100644
index 0000000..5249d90
--- /dev/null
+++ b/README.md
@@ -0,0 +1,151 @@
+# Parallel_Programming
+Parallel programming is often used interchangeably with parallel processing or in connection with parallel computing, which refers to systems that enhance the efficiency of parallel execution. In parallel programming, tasks are divided and executed simultaneously across multiple computers or multiple CPU cores. This approach is essential for large-scale projects where speed and precision are crucial. Although complex, parallel programming enables developers, researchers, and users to perform research and analysis much faster than programs limited to processing tasks sequentially
+
+## Project Description
+This project demonstrates the use of parallel programming techniques in C++ with (`OpenMP`) and (`MPI`) to efficiently execute computations across multiple threads. It includes examples of parallel loops, task scheduling, and thread synchronization. The goal is to showcase performance improvements in computational tasks using OpenMP and MPI directives.
+
+## **1. OpenMP (Open Multi-Processing)**
+
+`OpenMP` is a multi-platform programming interface that enables multiprocessing programming. `OpenMP` can be used in C++, C and Fortran languages, including different architectures like Windows and Unix. It consists of compilator directives that have an impact on code execution.
+
+The `OpenMP` interface is a component of the GNU Compiler Collection (`GCC`), a set of open-source compilers developed by the GNU Project. `GCC` compiler is therefore highly recommended for use with `OpenMP`, although it is not required (there is an Intel compiler that also support `OpenMP`).
+
+**INSTALLATION AND CONFIGURATION ON LINUX SYSTEMS:**
+
+Start the terminal and update the repository:
+
+```bash
+>>> sudo apt-get update
+```
+
+Then install the `build-essential` package, including `gcc`, `g++` and `make`:
+
+```bash
+>>> sudo apt-get install build-essential
+```
+
+## **2. MPI**
+
+Message Passing Interface (`MPI`) is a communication protocol standard for transferring messages between parallel program processes on one or more computers. `MPI` is currently the most widely used communication model in clusters of computers and supercomputers.
+
+There are several implementations of `MPI`, including `OpenMPI`, `MPICH` and `MSMPI`. On Linux, we can choose from `OpenMPI` and `MPICH`, while `MSMPI` is a Windows implementation. Before going any further, we should ensure that we have the `GCC` compiler installed.
+
+**INSTALLATION AND CONFIGURATION OF `MPICH` ON LINUX SYSTEMS:**
+
+Start the terminal and update the repository:
+
+```bash
+>>> sudo apt-get update
+```
+
+We then install the `mpich` package:
+
+```bash
+>>> sudo apt-get install mpich
+```
+
+We can now check the version of the installed `MPI` (this will actually be the `GCC` version):
+
+```bash
+>>> mpic++ --version
+```
+
+Here you can find out more about `MPICH`: [https://www.mpich.org/](https://www.mpich.org/).
+
+
+**THE INSTALLATION PROCESS UNDER WINDOWS IS COMPLEX, AND I DO NOT RECOMMEND USING MPI WITH THE WINDOWS PLATFORM**
+
+**NOTE**
+> Linux is often recommended for compiling parallel programs
+> For optimal performance and seamless development, we recommend using Linux for building and running this project.
+
+## Project Tree
+> This project contains two branches: `main` and `Advanced`.
+> The `main` branch includes operations from the NumPy module that have been translated into C++ using OpenMP and MPI for parallel processing.
+
+### **main**
+
+ .
+ |-- include
+ | |-- array_concat.h
+ | |-- dot_product.h
+ | |-- find_max.h
+ | |-- lu_decomposition.h
+ | |-- matrix_mult.h
+ | |-- matrix_vector.h
+ | |-- montecarlo.h
+ | |-- parallel_sum.h
+ | |-- pi_calculator.h
+ | |-- product_log.h
+ | |-- statistics.h
+ | |-- sum2.h
+ | |-- sum_task.h
+ |
+ |-- src
+ | |-- Histogram/
+ | | |-- histo.cpp
+ | | |-- data.txt
+ | |-- Contributing.md
+ | |-- LU_factorisation.cpp
+ | |-- README.md
+ | |-- algoimage.png
+ | |-- broadcast.cpp
+ | |-- concatenate.cpp
+ | |-- input.txt(for max.cpp)
+ | |-- max.cpp
+ | |-- mm.cpp (Matrix Multiplication implementation)
+ | |-- montecarlo.cpp
+ | |-- mv.cpp (Matrix Vector multiplication)
+ | |-- pi-reduction.cpp (Parallel pi computation using OpenMP reduction)
+ | |-- prod.cpp (Parallel product computation)
+ | |-- standard_dev.cpp
+ | |-- sum2.cpp
+ | |-- sum_task.cpp
+ | |-- vvd.cpp (Vector-Vector Dot product)
+ | |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
+
+### **Advanced**
+
+ .
+ |-- include
+ | |-- array_concat.h
+ | |-- dot_product.h
+ | |-- find_max.h
+ | |-- lu_decomposition.h
+ | |-- matrix_mult.h
+ | |-- matrix_vector.h
+ | |-- montecarlo.h
+ | |-- parallel_sum.h
+ | |-- pi_calculator.h
+ | |-- product_log.h
+ | |-- statistics.h
+ | |-- sum2.h
+ | |-- sum_task.h
+ |
+ |-- src
+ | |-- Histogram/
+ | | |-- histo.cpp
+ | | |-- data.txt
+ | |-- Contributing.md
+ | |-- LU_factorisation.cpp
+ | |-- README.md
+ | |-- algoimage.png
+ | |-- broadcast.cpp
+ | |-- concatenate.cpp
+ | |-- input.txt(for max.cpp)
+ | |-- max.cpp
+ | |-- mm.cpp (Matrix Multiplication implementation)
+ | |-- montecarlo.cpp
+ | |-- mv.cpp (Matrix Vector multiplication)
+ | |-- pi-reduction.cpp (Parallel pi computation using OpenMP reduction)
+ | |-- prod.cpp (Parallel product computation)
+ | |-- standard_dev.cpp
+ | |-- sum2.cpp
+ | |-- sum_task.cpp
+ | |-- vvd.cpp (Vector-Vector Dot product)
+ | |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
+ |-- Application
+ | |-- page-rank.cpp
+### Note
+> Information about Functions in main is provided in [README.md](src/README.md)
+> For contributing to this repo kindly go through the guidelines provided in [Contributing.md](src/Contributing.md)
From fe8ea6a0f7b49bc29aa7aaa006b448649e5ecc9f Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:35:30 +0530
Subject: [PATCH 19/31] Delete Introduction.md
---
Introduction.md | 151 ------------------------------------------------
1 file changed, 151 deletions(-)
delete mode 100644 Introduction.md
diff --git a/Introduction.md b/Introduction.md
deleted file mode 100644
index 5249d90..0000000
--- a/Introduction.md
+++ /dev/null
@@ -1,151 +0,0 @@
-# Parallel_Programming
-Parallel programming is often used interchangeably with parallel processing or in connection with parallel computing, which refers to systems that enhance the efficiency of parallel execution. In parallel programming, tasks are divided and executed simultaneously across multiple computers or multiple CPU cores. This approach is essential for large-scale projects where speed and precision are crucial. Although complex, parallel programming enables developers, researchers, and users to perform research and analysis much faster than programs limited to processing tasks sequentially
-
-## Project Description
-This project demonstrates the use of parallel programming techniques in C++ with (`OpenMP`) and (`MPI`) to efficiently execute computations across multiple threads. It includes examples of parallel loops, task scheduling, and thread synchronization. The goal is to showcase performance improvements in computational tasks using OpenMP and MPI directives.
-
-## **1. OpenMP (Open Multi-Processing)**
-
-`OpenMP` is a multi-platform programming interface that enables multiprocessing programming. `OpenMP` can be used in C++, C and Fortran languages, including different architectures like Windows and Unix. It consists of compilator directives that have an impact on code execution.
-
-The `OpenMP` interface is a component of the GNU Compiler Collection (`GCC`), a set of open-source compilers developed by the GNU Project. `GCC` compiler is therefore highly recommended for use with `OpenMP`, although it is not required (there is an Intel compiler that also support `OpenMP`).
-
-**INSTALLATION AND CONFIGURATION ON LINUX SYSTEMS:**
-
-Start the terminal and update the repository:
-
-```bash
->>> sudo apt-get update
-```
-
-Then install the `build-essential` package, including `gcc`, `g++` and `make`:
-
-```bash
->>> sudo apt-get install build-essential
-```
-
-## **2. MPI**
-
-Message Passing Interface (`MPI`) is a communication protocol standard for transferring messages between parallel program processes on one or more computers. `MPI` is currently the most widely used communication model in clusters of computers and supercomputers.
-
-There are several implementations of `MPI`, including `OpenMPI`, `MPICH` and `MSMPI`. On Linux, we can choose from `OpenMPI` and `MPICH`, while `MSMPI` is a Windows implementation. Before going any further, we should ensure that we have the `GCC` compiler installed.
-
-**INSTALLATION AND CONFIGURATION OF `MPICH` ON LINUX SYSTEMS:**
-
-Start the terminal and update the repository:
-
-```bash
->>> sudo apt-get update
-```
-
-We then install the `mpich` package:
-
-```bash
->>> sudo apt-get install mpich
-```
-
-We can now check the version of the installed `MPI` (this will actually be the `GCC` version):
-
-```bash
->>> mpic++ --version
-```
-
-Here you can find out more about `MPICH`: [https://www.mpich.org/](https://www.mpich.org/).
-
-
-**THE INSTALLATION PROCESS UNDER WINDOWS IS COMPLEX, AND I DO NOT RECOMMEND USING MPI WITH THE WINDOWS PLATFORM**
-
-**NOTE**
-> Linux is often recommended for compiling parallel programs
-> For optimal performance and seamless development, we recommend using Linux for building and running this project.
-
-## Project Tree
-> This project contains two branches: `main` and `Advanced`.
-> The `main` branch includes operations from the NumPy module that have been translated into C++ using OpenMP and MPI for parallel processing.
-
-### **main**
-
- .
- |-- include
- | |-- array_concat.h
- | |-- dot_product.h
- | |-- find_max.h
- | |-- lu_decomposition.h
- | |-- matrix_mult.h
- | |-- matrix_vector.h
- | |-- montecarlo.h
- | |-- parallel_sum.h
- | |-- pi_calculator.h
- | |-- product_log.h
- | |-- statistics.h
- | |-- sum2.h
- | |-- sum_task.h
- |
- |-- src
- | |-- Histogram/
- | | |-- histo.cpp
- | | |-- data.txt
- | |-- Contributing.md
- | |-- LU_factorisation.cpp
- | |-- README.md
- | |-- algoimage.png
- | |-- broadcast.cpp
- | |-- concatenate.cpp
- | |-- input.txt(for max.cpp)
- | |-- max.cpp
- | |-- mm.cpp (Matrix Multiplication implementation)
- | |-- montecarlo.cpp
- | |-- mv.cpp (Matrix Vector multiplication)
- | |-- pi-reduction.cpp (Parallel pi computation using OpenMP reduction)
- | |-- prod.cpp (Parallel product computation)
- | |-- standard_dev.cpp
- | |-- sum2.cpp
- | |-- sum_task.cpp
- | |-- vvd.cpp (Vector-Vector Dot product)
- | |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
-
-### **Advanced**
-
- .
- |-- include
- | |-- array_concat.h
- | |-- dot_product.h
- | |-- find_max.h
- | |-- lu_decomposition.h
- | |-- matrix_mult.h
- | |-- matrix_vector.h
- | |-- montecarlo.h
- | |-- parallel_sum.h
- | |-- pi_calculator.h
- | |-- product_log.h
- | |-- statistics.h
- | |-- sum2.h
- | |-- sum_task.h
- |
- |-- src
- | |-- Histogram/
- | | |-- histo.cpp
- | | |-- data.txt
- | |-- Contributing.md
- | |-- LU_factorisation.cpp
- | |-- README.md
- | |-- algoimage.png
- | |-- broadcast.cpp
- | |-- concatenate.cpp
- | |-- input.txt(for max.cpp)
- | |-- max.cpp
- | |-- mm.cpp (Matrix Multiplication implementation)
- | |-- montecarlo.cpp
- | |-- mv.cpp (Matrix Vector multiplication)
- | |-- pi-reduction.cpp (Parallel pi computation using OpenMP reduction)
- | |-- prod.cpp (Parallel product computation)
- | |-- standard_dev.cpp
- | |-- sum2.cpp
- | |-- sum_task.cpp
- | |-- vvd.cpp (Vector-Vector Dot product)
- | |-- wrong_sum.cpp (Demo of a wrong summation example for learning purposes)
- |-- Application
- | |-- page-rank.cpp
-### Note
-> Information about Functions in main is provided in [README.md](src/README.md)
-> For contributing to this repo kindly go through the guidelines provided in [Contributing.md](src/Contributing.md)
From f6300762108174638bd79c028b2693a895a99386 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:38:19 +0530
Subject: [PATCH 20/31] Update Contributing.md
---
src/Contributing.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/src/Contributing.md b/src/Contributing.md
index e33276c..0a09455 100644
--- a/src/Contributing.md
+++ b/src/Contributing.md
@@ -24,7 +24,7 @@ If you would like to contribute to the project, please follow these guidelines:
# Clone your fork of the GitHub Repo
git clone https://github.com/your_username/SoC-Portal.git
```
-> Follow the installation and compilation steps provided in [Intro.md](Hello-Foss-CPP/Intro.md)
+> Follow the installation and compilation steps provided in [README.md](README.md) and [Intro](./README.md)
By following these guidelines, you help maintain the quality and organization of the project!
From d666baa21ca2f975223fd2a9c1ae1851e48dba2f Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:41:35 +0530
Subject: [PATCH 21/31] Update Contributing.md
---
src/Contributing.md | 3 ++-
1 file changed, 2 insertions(+), 1 deletion(-)
diff --git a/src/Contributing.md b/src/Contributing.md
index 0a09455..2a30729 100644
--- a/src/Contributing.md
+++ b/src/Contributing.md
@@ -24,11 +24,12 @@ If you would like to contribute to the project, please follow these guidelines:
# Clone your fork of the GitHub Repo
git clone https://github.com/your_username/SoC-Portal.git
```
-> Follow the installation and compilation steps provided in [README.md](README.md) and [Intro](./README.md)
+> Follow the installation and compilation steps provided in Introduction page and [README.md](README.md).
By following these guidelines, you help maintain the quality and organization of the project!
## Resources
- [Parallelization with MPI and OpenMPI](http://compphysics.github.io/ComputationalPhysics2/doc/LectureNotes/_build/html/parallelization.html#)
- [OpenMP](https://medium.com/swlh/openmp-on-ubuntu-1145355eeb2)
+
***HAPPY LEARNING 😀😀😀***
From 6e5759df7699c1b85f53b7d33805777b3cfdb47c Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Thu, 17 Oct 2024 11:42:53 +0530
Subject: [PATCH 22/31] Update Contributing.md
---
src/Contributing.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/src/Contributing.md b/src/Contributing.md
index 2a30729..fec17b2 100644
--- a/src/Contributing.md
+++ b/src/Contributing.md
@@ -22,7 +22,7 @@ If you would like to contribute to the project, please follow these guidelines:
```shell
# Clone your fork of the GitHub Repo
- git clone https://github.com/your_username/SoC-Portal.git
+ git clone https://github.com/your_username/Hello-Foss-CPP.git
```
> Follow the installation and compilation steps provided in Introduction page and [README.md](README.md).
From daaed414df31c8524c5cc3eb69da5c791e6bc412 Mon Sep 17 00:00:00 2001
From: Pratyaksh Bhardwaj
Date: Thu, 17 Oct 2024 20:56:46 +0530
Subject: [PATCH 23/31] updated contribution.md
---
src/Contributing.md => Contributing.md | 0
1 file changed, 0 insertions(+), 0 deletions(-)
rename src/Contributing.md => Contributing.md (100%)
diff --git a/src/Contributing.md b/Contributing.md
similarity index 100%
rename from src/Contributing.md
rename to Contributing.md
From 74b069f61ad2176b6b3a9530271fa4dcfb8cb81b Mon Sep 17 00:00:00 2001
From: Aditya Khambete
Date: Thu, 17 Oct 2024 23:33:57 +0530
Subject: [PATCH 24/31] Update max.cpp
---
src/max.cpp | 12 +++++++++---
1 file changed, 9 insertions(+), 3 deletions(-)
diff --git a/src/max.cpp b/src/max.cpp
index 02e7854..a458971 100644
--- a/src/max.cpp
+++ b/src/max.cpp
@@ -18,14 +18,13 @@ int findMax(int elementsToProcess, const std::vector& data) {
int main(int argc, char** argv) {
MPI_Init(&argc, &argv);
- double time = MPI_Wtime(); // Start timing
-
int processorsNr;
MPI_Comm_size(MPI_COMM_WORLD, &processorsNr);
int processId;
MPI_Comm_rank(MPI_COMM_WORLD, &processId);
int buff;
+ double time;
if (processId == 0) {
int max = 0; // Global maximum
@@ -66,6 +65,10 @@ int main(int argc, char** argv) {
MPI_Send(data.data() + startIdx, buff, MPI_INT, i, 2, MPI_COMM_WORLD);
}
+ // Synchronize all processes before starting the timing
+ MPI_Barrier(MPI_COMM_WORLD);
+ time = MPI_Wtime(); // Start timing
+
// Master process finds its own max
max = findMax(elementsToEachProcess + remainder, data);
@@ -78,8 +81,8 @@ int main(int argc, char** argv) {
}
}
- std::cout << "Global maximum is " << max << std::endl;
time = MPI_Wtime() - time; // Stop timing
+ std::cout << "Global maximum is " << max << std::endl;
std::cout << "Time elapsed: " << time << " seconds" << std::endl;
}
@@ -89,6 +92,9 @@ int main(int argc, char** argv) {
std::vector dataToProcess(buff);
MPI_Recv(dataToProcess.data(), buff, MPI_INT, 0, 2, MPI_COMM_WORLD, MPI_STATUS_IGNORE);
+ // Synchronize before starting the computation
+ MPI_Barrier(MPI_COMM_WORLD);
+
int theMax = findMax(buff, dataToProcess);
// Send local max to master process
From 1afe25f7f9065fe0e2b7f1c323d5013f186d6f81 Mon Sep 17 00:00:00 2001
From: Samanth Martis
Date: Fri, 18 Oct 2024 18:41:00 +0530
Subject: [PATCH 25/31] Fixed redirect in README
The link for the contributing.md file was wrong.
---
README.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/README.md b/README.md
index 5249d90..45bcf9a 100644
--- a/README.md
+++ b/README.md
@@ -148,4 +148,4 @@ Here you can find out more about `MPICH`: [https://www.mpich.org/](https://www.m
| |-- page-rank.cpp
### Note
> Information about Functions in main is provided in [README.md](src/README.md)
-> For contributing to this repo kindly go through the guidelines provided in [Contributing.md](src/Contributing.md)
+> For contributing to this repo kindly go through the guidelines provided in [Contributing.md](Contributing.md)
From 7f8c090a550d3c7d04aaef1103514de58c2586ec Mon Sep 17 00:00:00 2001
From: cs293-ahilkhancse
Date: Fri, 18 Oct 2024 22:40:06 +0530
Subject: [PATCH 26/31] montecarlo same random number issue resolved
---
src/montecarlo.cpp | 9 +++++++++
1 file changed, 9 insertions(+)
diff --git a/src/montecarlo.cpp b/src/montecarlo.cpp
index c7d38da..f315408 100644
--- a/src/montecarlo.cpp
+++ b/src/montecarlo.cpp
@@ -2,16 +2,25 @@
#include
#include
#include
+#include
using namespace std;
+unsigned long long getCurrentTimeInMilliseconds() {
+ return std::chrono::duration_cast(
+ std::chrono::high_resolution_clock::now().time_since_epoch()).count();
+}
+
void montecarlo(int n, int num_threads) {
int pCircle = 0, pSquare = 0;
double x, y, d;
int i;
+ unsigned long long seed = getCurrentTimeInMilliseconds();
+ srand(seed); // Seeding with a modified value
// Parallelize the loop using OpenMP
#pragma omp parallel for private(x, y, d, i) reduction(+:pCircle, pSquare) num_threads(num_threads)
for(i = 0; i < n; i++) {
+
// Generate random points between 0 and 1
x = (double)rand() / RAND_MAX;
y = (double)rand() / RAND_MAX;
From 12f174fb8de387bfd7fa2b1609c5918bddd5ef07 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Sat, 19 Oct 2024 14:26:31 +0530
Subject: [PATCH 27/31] Update wrong_sum.cpp
---
src/wrong_sum.cpp | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/src/wrong_sum.cpp b/src/wrong_sum.cpp
index 81ce347..fd04ade 100644
--- a/src/wrong_sum.cpp
+++ b/src/wrong_sum.cpp
@@ -7,7 +7,7 @@ int parallel_sum(const int arr[], int size, int step_size) {
#pragma omp parallel
{
- #pragma omp for nowait
+
for (int i = 0; i < size; i += step_size) {
int start = i, end = i + step_size - 1;
From eef3f0719f807330985671ab19af49f79251492a Mon Sep 17 00:00:00 2001
From: cs293-ahilkhancse
Date: Sun, 20 Oct 2024 02:40:21 +0530
Subject: [PATCH 28/31] concatenation bug for large arrays resolved
---
concat | Bin 0 -> 17384 bytes
src/concatenate.cpp | 23 ++++++++++++++++++-----
2 files changed, 18 insertions(+), 5 deletions(-)
create mode 100755 concat
diff --git a/concat b/concat
new file mode 100755
index 0000000000000000000000000000000000000000..a6339bd52dfe406c4fc8fc8267503ea6d15722be
GIT binary patch
literal 17384
zcmeHPeQ;dWb-%l|un7K;6FV`6;7tOIz-WrHgFVKj_
zu>GC;?sdUnq}zw>d=J@@W?AG+6H)p~aOeLlfunfM(+Tw$$+gp{E1
zc2xlhiFUCH-3`jfew6n5lwEy5x)E-A
zj|B^++}>TVqr80B!o$2zwdXYs1(i>xYB%fh1~g)A-O0GQt3`WI;^e$<_rCsAzo7w0hDY%o
z4fm0>-%Mn>KU`yx0W`JM==(xMdUdtdC0jEh<6*-bNv94Xj;Ik%r^APgWbB}kh$V;3
z5!ql{B$YKqYIMvPj+sU>J4yzza5N*aj0@XyXK$Y|7EXr~iCDt6vMk6LV?b>V#gpMg
z{2}Doz#JRzW+F{6>qn`s%^M(i*`Bc_Fz
zhy}|i+Jeo3qG1P%V-U@*|jZ8)z;uv(RJWJ=gw~EZV9$qItLCITcD5P$G_!R
zg^1Q+#qo<3q8{H=^Wocve|{+^ZCaQ7;x)`U`du{RGyAz4v8>gL|4>-ydg8ATZ>xFf
zgKdFj&|epiucB4;mkN9C*f+7vEf;lI_{e8=WBf{575Vx{?Y*pYAr|9jeuO)K72;Z@
zZ{G9`TuZZ+77mO20m2$yuNIQ+HT6tlBwU2QKfglwjO}Th}Q|FyX+}EteEL>cBatBsl88G4OVI+<|}0Mum97fva0V
zDSg_3uXN}?>%cMb>~h?JueMPkrX4uV$6Bdnpq7DJ27X&J@b;SP{zspDvtG~F{nM>N
z=#NaB{_?Co`EvbBvPL^?7m(8Y_iUan9HseVHcuCf()`zKo-P!n`3G#CE)b>pu+7tjp)`M&&C>;;
zG~a3SbRj6sx7a*g07~KpH
z{QGzv;EeBG7yp)vf8EKqyw#mMbDy63H+}N_Lf?SaGTm}opX!*vu$9-p2F-Vdf@?N?
z9ZA6inCf^Bh(1+UBC>taT#u>oJk>EZt}GYUY?>gwmlTEmX{o>UMItwSpyw9!ng6^)
zpIKa{`%dU*J}}q8!6V8+eYr3s`@!}!e5K?77>@b-_dhW=ap3BbW_P?6zJ=K%brlWd(`3;TI
zv$ohWt>?9AJ*O4(+N(Kj0UXq|f&!j^n8{wU+m5;QlD-|Pfu#+$mfMnUOQLO-Xgr_g&dIlW-F=7WXE2_5Ov_^O@o#D
z$KrixVNyHi@Vp09i3W=mzNu1A^gqF5%66a~#Q^al}yEoi}l76{sFL5mh7
zB>iJMZ4yltUa)Jqd
z@44bU{C4JEF3y1FPUPmk-T8}4xfw-gx1TuNz2-+3b0y#G)VliZvq!X?Hm62)rDMEL
z<$LG!2-c0ZDeAqR?=9%Qix?Ci4fOW?bBF)-!pCG{CTDyXes~m@kmld*yyruPy$+31
z*i||(S8P-r5MCYFThW2f=d`nV?JP~8$_`}D<$KTiE_(IiEP8SF@GFj9JYCrfd7ZSc
zjFg31sb!#+fm#M?8K`BTmVsIZe)}>&@6Q7W+jq^f9rQO>dUW2ffE!WJMNqn?;A;R*9!Ss^?Fq44-gaoRW27m#p;ILt2cap%~c2MC&V2aZ@YQR
z^`9k}>?!Vc#Oiz>OSp+oGjNKX3$^taK4;M0^8{+E8-CBE8%ww{9#LVQ_C7iJg5pHXPS+Q=lw|Z972Wu6iVgY
z56UT`FyCycegaD!d10PRZ#xkIGj1%M1$sh9b)bMV)(s2vTN3SNmU-nP^x2qvtIu
zyr)NXOEjp47++8UbKC)C&wG2g{p&1ucK`6|U-l2`(iDF!l6;?XgujqTp5w%0pX9Go
z^7j-^e|rA^D7-)0?YmhGU`WwEMekR1OwkENk1G0vqR%Qit>~6Q
z1(cZfrd89YdEgffqQSYT4Ol?oAs4x2
z`iY}*O#1(naQ{yJRN@=#b;Z)gItJV|v{LjU@CI?UXb7>UMNb1?>%Y9+Z)2VXpf6nd
z|3doILcp@;b>U6m^2L$`+utj&;C&n|e^qUwl<$ThT`6%}&93wVU+ceG@b8KR_e=cp
zI5QsjqvYSeR`7L#{rn-}%sA*X9{2(Di{d}6jA&j``7ZK9IG^{q(SpUx9)9YjzEX0K
z8$9rk2YwIXeq0q@>(?M~%A0$9lZ02<*Zy@6|K9>m^F)LkL83Z`JT$=1fzC?&`?B53
z$K|+(|2L%njhDq)^3cB$6Rg^JHF@B=2Yw&$KxOEXd&ooon;!aq?t#DH;YajG(q>CA
zgd0NB6O4>1I=j47*3=H!wDm5rqUTBoE;YtY*|ml
z%vcmV)2nLG{`k1sjvq^!=|f^D9UhGt(d_8xA(%Kg1Da-qYHDaGk;49b6;~vcj9@o?
z5}#mXY)tOy?`b!Tc#15+;^e}TN?*L+h~pD9XvhChk|O*y8ZFRhYmqzkO&oS0uW{Pk
zA2#DSY!OaIamE4%AhHq3H^*^sVo%SVJ3D(|k{vX<4UY!U&)xevdo@GbyNgb17&s4P
zXu1ORUHgQw>&tsPd%L@Ahj;GY-LK)4Md!{QjSgAh>_kXz{a2g$jaW2nhT%%($Qa5d
zBN+o{IB?+OPMkG}#*;=i69ZduOu=%3^8<8Lz&fSykx4*
zB*X_a#z<2WuDk17a9YMN-7l4M3@XzZ7gAmpj@o$|<#O{IL-ah0OE2zK
zt+D~vNh(bFdDcQLU9J=T96t06$@cBao@p~FIf{DhvypA6Iqlu&T(7q^tqLxK6@{K3
zoc8?u!Zc8!P4~u?$YQH+#lw&&;R
zI509{dwws#l-~nDMHcemEDwO8XKc3T=W3?6C}obH^_V^kIX$y8&(G~lpHTJ`KgG%M
zD_GE`L&_=JFDM130TnQ}&-UE^Zz=no$|1j(V9M_?$ewK6_TK|WO|XCazE$17^7|U7
z$l@OV?*Mh$2b3Mtf@NlZ^KCKHr(E{98Mca|tj7+RXZozmzJcu&on=Es*^cQ?UH1IE
zz*OfKPA+?P`-01!pJ$lz`!Lpb=kFC|&*LuwN>fpLTomlK|2yE6f3~kzc1-V4c1+y%
z{|tfCeni#5Leh%mVXTZ6BkPr69aXkZ!y2$qY9#FrPw1qfs5&N6%nO*^7r@aU%)n!*lp~B`}WNY2Qc}CX|17O*n4XlG3wpb^EJ!SQ;B$1};|oFVFIvn*aa+
literal 0
HcmV?d00001
diff --git a/src/concatenate.cpp b/src/concatenate.cpp
index 7bbd123..fab03ad 100644
--- a/src/concatenate.cpp
+++ b/src/concatenate.cpp
@@ -5,7 +5,7 @@
using namespace std;
// Function to concatenate two arrays in parallel
-void concatenate(int arr1[], int arr2[], int n1, int n2, int arr3[]) {
+void concatenate(int* arr1, int* arr2, int n1, int n2, int* arr3) {
#pragma omp parallel for
for (int i = 0; i < n1; i++) {
arr3[i] = arr1[i];
@@ -18,10 +18,19 @@ void concatenate(int arr1[], int arr2[], int n1, int n2, int arr3[]) {
}
int main() {
- int n1 = 5, n2 = 5;
- int arr1[n1] = {1, 2, 3, 4, 5};
- int arr2[n2] = {6, 7, 8, 9, 10};
- int arr3[n1 + n2];
+
+ int n1 = 5000000, n2 = 5000000;
+
+ int* arr1 = new int[n1];
+ int* arr2 = new int[n2];
+ int* arr3 = new int[n1+n2];
+
+ // initialisation
+ for(int i = 0; i < n1; i++)
+ {
+ arr1[i] = 0;
+ arr2[i] = 1;
+ }
// Set the number of threads
omp_set_num_threads(2);
@@ -35,5 +44,9 @@ int main() {
}
cout << endl;
+ delete[] arr1;
+ delete[] arr2;
+ delete[] arr3;
+
return 0;
}
\ No newline at end of file
From cbe2086129c545da408c1f715350ebca62815777 Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Wed, 23 Oct 2024 13:43:22 +0530
Subject: [PATCH 29/31] Update README.md
---
src/README.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/src/README.md b/src/README.md
index 5318030..500bc67 100644
--- a/src/README.md
+++ b/src/README.md
@@ -48,7 +48,7 @@ This C++ code computes the sum of an array using OpenMP with task-based parallel
>
```shell
# compile using g++ for Openmp
-g++ - sum2.cpp -o sum2
+g++ -I../include sum2.cpp -o sum2 -fopenmp
./sum2
# compile using g++ for MPI
From e3df0e4ea2dcdb6bdd7a839f2af5797ab96f40ef Mon Sep 17 00:00:00 2001
From: Lopu <162079114+LoPA607@users.noreply.github.com>
Date: Wed, 23 Oct 2024 13:45:52 +0530
Subject: [PATCH 30/31] Update README.md
---
src/README.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/src/README.md b/src/README.md
index 500bc67..34ea6ea 100644
--- a/src/README.md
+++ b/src/README.md
@@ -48,7 +48,7 @@ This C++ code computes the sum of an array using OpenMP with task-based parallel
>
```shell
# compile using g++ for Openmp
-g++ -I../include sum2.cpp -o sum2 -fopenmp
+g++ sum2.cpp -o sum2 -fopenmp
./sum2
# compile using g++ for MPI
From 4401ebeba210081b9a0e6b48222830cb7e6edc34 Mon Sep 17 00:00:00 2001
From: Geet Sethi
Date: Sat, 2 Nov 2024 18:20:31 +0530
Subject: [PATCH 31/31] handle empty matrices in broadcast
---
src/broadcast.cpp | 88 +++++++++++++++++++++++++----------------------
1 file changed, 46 insertions(+), 42 deletions(-)
diff --git a/src/broadcast.cpp b/src/broadcast.cpp
index c291276..70bb6c8 100644
--- a/src/broadcast.cpp
+++ b/src/broadcast.cpp
@@ -1,56 +1,60 @@
-#include
#include
-#include
+
+#include
#include
+#include
// Function to broadcast two matrices
-void broadcast(const std::vector>& A, std::vector>& B) {
- size_t rowsA = A.size();
- size_t colsA = A[0].size();
- size_t rowsB = B.size();
- size_t colsB = B[0].size();
-
- if (rowsA != rowsB && rowsB != 1) {
- throw std::invalid_argument("Incompatible dimensions for broadcasting");
- }
- if (colsA != colsB && colsB != 1) {
- throw std::invalid_argument("Incompatible dimensions for broadcasting");
- }
+void broadcast(const std::vector>& A,
+ std::vector>& B) {
+ size_t rowsA = A.size();
+ size_t colsA = A[0].size();
+ size_t rowsB = B.size();
+ size_t colsB = B[0].size();
- if (rowsB == 1) {
- B.resize(rowsA, B[0]);
- }
- if (colsB == 1) {
- for (auto& row : B) {
- row.resize(colsA, row[0]);
- }
+ if (rowsA == 0 || colsA == 0 || rowsB == 0 || colsB == 0) {
+ throw std::invalid_argument("Empty matrix cannot be broadcasted");
+ }
+ if (rowsA != rowsB && rowsB != 1) {
+ throw std::invalid_argument("Incompatible dimensions for broadcasting");
+ }
+ if (colsA != colsB && colsB != 1) {
+ throw std::invalid_argument("Incompatible dimensions for broadcasting");
+ }
+
+ if (rowsB == 1) {
+ B.resize(rowsA, B[0]);
+ }
+ if (colsB == 1) {
+ for (auto& row : B) {
+ row.resize(colsA, row[0]);
}
+ }
-
- #pragma omp parallel for
- for (size_t i = 0; i < rowsA; i++) {
- #pragma omp parallel for
- for (size_t j = 0; j < colsA; j++) {
- B[i][j] = A[i][j]+B[i][j];
- }
+#pragma omp parallel for
+ for (size_t i = 0; i < rowsA; i++) {
+#pragma omp parallel for
+ for (size_t j = 0; j < colsA; j++) {
+ B[i][j] = A[i][j] + B[i][j];
}
+ }
}
int main() {
- std::vector> A = {{1, 2, 3}, {4, 5, 6}, {7, 8, 9}};
- std::vector> B = {{1,2,3}}; // B has only one row
-
- try {
- broadcast(A, B);
- for (const auto& row : B) {
- for (const auto& elem : row) {
- std::cout << elem << " ";
- }
- std::cout << std::endl;
- }
- } catch (const std::invalid_argument& e) {
- std::cerr << "Error: " << e.what() << std::endl;
+ std::vector> A = {{1, 2, 3}, {4, 5, 6}, {7, 8, 9}};
+ std::vector> B = {{1, 2, 3}}; // B has only one row
+
+ try {
+ broadcast(A, B);
+ for (const auto& row : B) {
+ for (const auto& elem : row) {
+ std::cout << elem << " ";
+ }
+ std::cout << std::endl;
}
+ } catch (const std::invalid_argument& e) {
+ std::cerr << "Error: " << e.what() << std::endl;
+ }
- return 0;
+ return 0;
}