#ParallelComputing

2025-06-11

๐Ÿ“ธ Full house at the OpenMP BOF at #ISC25 โ€” over 140 attendees joined us in Hamburg! ๐ŸŽ‰

Our session "What to Expect from OpenMP API Version 6.0" covered:

โœ… A dive into key features of OpenMP 6.0
โœ… A preview of 6.1 and 7.0
โœ… Updates from toolchain developers
โœ… Lively Q&A to help shape future OpenMP directions

Thanks to everyone who contributed โ€” your feedback is powering the future of parallel programming! ๐Ÿ’ก

#OpenMP #HPC #ISC2025 #OpenMP6 #ParallelComputing #Supercomputing

2025-06-10

Weโ€™re excited to welcome NextSilicon to the OpenMP Architecture Review Board! ๐ŸŽ‰

Their Intelligent Compute Architecture blends adaptive computing with self-optimizing hardware/software and open frameworks like OpenMP. Together, weโ€™re shaping a future of performant, portable, shared-memory parallelism. ๐Ÿ’ป๐ŸŒ

Read the press release:
tinyurl.com/yksfbrah

#OpenMP #NextSilicon #HPC #OpenStandards #ParallelComputing

2025-06-06

Join us at #ISC25 for the tutorial โ€œAdvanced OpenMP: Performance and 6.0 Featuresโ€ on Friday, June 13, 9:00โ€“13:00 CEST in Hall Y12, 2nd Floor, Hamburg Congress Center.

Learn how to boost OpenMP code performance on NUMA systems and accelerators, and get hands-on insights into vectorization, data locality, and the latest features in OpenMP 6.0.

Ideal for developers who want to go beyond the basics!

#HPC #OpenMP #ISC2025 #ParallelComputing

2025-05-27

Just published the post "Parallel and distributed computing in GNU Health." :gnu: ๐Ÿฅ
meanmicio.org/2025/05/27/paral
#ParallelComputing #GNUHealth #Tryton #OpenScience #GNU

Mille e Una Avventura1001avventura@mastodon.uno
2025-05-26

Vediamo le funzionalitร  del Fortran introdotte nel 2008, con i CoArray, e nel 2018 per scoprire come si possono sfruttare tutti i core delle nostre CPU abbreviando i tempi di calcoli scientifici complessi. #fortran #parallelcomputing #multithreading
youtube.com/watch?v=78_12a89MW

Chapel Programming Languagechapelprogramminglanguage
2025-05-05

๐Ÿš€ Excited to announce that hachyderm.io/@mppf and Shreyas from the project will be at in Chicago this week! Come find us, talk all things parallel computing, open source, and the future of high-performance software.

๐ŸŽ™๏ธ Donโ€™t miss our presentation on Day 1 โ€” check the full schedule at events.linuxfoundation.org/hps

Big thanks to The Linux Foundation and HPSF for hosting!

Assn for Computing MachineryACM@mastodon.acm.org
2025-05-01

๐Ÿ† Hugo Krawczyk โ€“ ACM Paris Kanellakis Theory and Practice Award
For pioneering and lasting contributions to the theoretical foundations of cryptographically secure communications, and to the protocols that form the security foundations of the Internet.
๐Ÿ”— bit.ly/4jBJjHX

๐Ÿ‘ Congratulations to all the awardees shaping the future of computing!

#ACMTechnicalAwards #Cryptography #ParallelComputing #ComputerScience

2025-04-27

I achieved a speedup of 10000 compared to the CPU implementation. I'm quite happy with that. Note that the chart uses a logarithmic scale, otherwise most of the runs would not be visible.

For example, the CPU took a bit more than 30 minutes to solve the biggest input, the parallel (CPU) version took a bit more than 8 minutes. Meanwhile, my best handwritten implementation on the GPU takes less than 200ms for the same problem, and the version using the optimized cuBLAS library takes just 64ms.

Edit: upload chart with white background, the transparent version didn't fare well with my Fediverse clients.

#computerScience #parallelComputing #cuda

A bar chart showing the runtime of different matrix multiplication implementations, from a matrix size of 64 up to 4096, doubling every time. To summarize, here are the runtimes in ms for size 4096:

CPU time: 1952765 
CPU OMP time: 498996
GPU GM time : 434 
GPU CUBLAS time: 64 
GPU SHM COL time: 2456 
GPU SHM TILE time: 191
2025-04-14

Hi R people! Could you suggest some guides and links on how to set up parallel processing with `foreach` that can work both on Windows and Linux/Mac? I'm searching for guides on the net, but most of them seem to have become obsolete. Thank you!

#rstats #ParallelComputing

2025-04-03

๐Ÿš€ Intel Developer Tools v2025.1 is here with new OpenMP 6.0 features!

The Intelยฎ Fortran Compiler enhances #OpenMP 6.0 support with two powerful additions: WORKDISTRIBUTE for efficient thread-level work distribution, and INTERCHANGE to reorder loop nests for improved parallelism and optimization.

A big win for HPC and embedded devs!

intel.com/content/www/us/en/de

#Fortran #HPC #Embedded #ParallelComputing

N-gated Hacker Newsngate
2025-03-21

๐Ÿ‘จโ€๐Ÿ’ป Oh, the eternal quest for a "good" parallel computer, like a quest for a unicorn that can also do your taxes. ๐Ÿฆ„๐Ÿ’ป Apparently, GPUs are only good for "predictable" tasks โ€“ maybe like predicting the inevitable death of your dreams for a universally versatile chip. ๐Ÿ˜‚ Why not just ask for a toaster that can handle your email while you're at it? ๐Ÿž๐Ÿ“ง
raphlinus.github.io/gpu/2025/0

2025-03-19

another example (Newton-Raphson zooming for Mandelbrot set):

with 7 threads : 105.26 Watts * 12.2 seconds = 1284.6 Joules
with 1 thread : 54.17 Watts * 52.1 seconds = 2822.3 Joules
sleeping / idle overhead : 26.01 Watts
7 threads minus overhead (105 - 26.01)W * 12.2s = 964 Joules
1 thread minus overhead (54.17 - 26.01)W * 52.1s = 1467.2 Joules
if the machine would be on/idle anyway: 105.26W * 12.2s + 26.01 W * (52.1 - 12.2)s = 2321.9 Joules

thus using more threads saves energy even when parallelism efficiency is far from perfect: best to get in/out as quickly as possible so you can turn the machine off (ideal case) / leave it fully idle (second best)

power consumption doesn't scale linearly with load (a little load increases a lot vs baseline: 1 thread doubles idle power consumption, but high load doesn't increase much more: 7 threads quadruples idle power consumption)

measured with turbostat on Debian, AMD 2700X CPU with default CPU scaling governor, usual browser/email/etc running too

#ParallelComputing #Energy #permacomputing #turbostat

N-gated Hacker Newsngate
2025-03-12

๐Ÿš€๐Ÿ’ป "I went to an event, so naturally, my entire existence now revolves around repackaging basic sorting algorithms with . Because, really, what else is parallel computing for besides impressing friends at parties? ๐Ÿค“"
ashwanirathee.com/blog/2025/so

2025-02-15

#ParallelComputing: not what it's cracked up to be

Alexandre B A Villares ๐Ÿvillares@ciberlandia.pt
2025-02-10

ยซ#Joblib is a set of tools to provide lightweight pipelining in #Python In particular:

Joblib is optimized to be fast and robust on large data in particular and has specific optimizations for #numpy arrays. It is BSD-licensed.ยป

joblib.readthedocs.io/en/stabl

2025-01-06

๐Ÿš€ Exciting news for HPC ๐Ÿš€

Michael Klemm sat down with Doug Eadline of HPCwire to explore OpenMP 6.0! ๐ŸŽ™๏ธ

This release introduces features that simplify parallel programming and drive co-processor-agnostic acceleration for unmatched portable performance. Whether scaling or optimizing, OpenMP 6.0 delivers.

๐Ÿ”— Watch now: hpcwire.com/livewire-interview

#HPC #OpenMP #ParallelComputing #OpenMP60 @dontknow

2024-11-19

๐Ÿš€ Visit the OpenMP booth #3015 at SC24!

๐ŸŒŸ Meet OpenMP experts like Michael Klemm (API 6.0, Fortran), Ruud van der Pas (NUMA, Performance), Tim Mattson (Python, GPUs), and Tom Deakin (GPUs, Reliability). Each offers deep insights tailored for all skill levels.

๐ŸŽฅ Watch the attached video for their schedules. Enhance your OpenMP skills and connect with the community!

#OpenMP #SC24 #HPC #ParallelComputing

2024-11-17

๐Ÿš€ Parallel Python Made Easy! ๐Ÿ

We're hosting a hands-on tutorial on PyOMP, a system bringing OpenMP parallelism to Python! By combining OpenMP directives (as strings) with Numba's JIT compiler, PyOMP taps into LLVM's OpenMP support, delivering C-like performance in Python's simplicity.

Our participants are mastering this game-changing tool to supercharge their workflows.

Stay tuned for updates!

#Python #OpenMP #PyOMP #ParallelComputing #Numba #HPC

Client Info

Server: https://mastodon.social
Version: 2025.04
Repository: https://github.com/cyevgeniy/lmst