If you need assistance, please send an email to forum at 4hv dot org. To ensure your email is not marked as spam, please include the phrase "4hv help" in the subject line. You can also find assistance via IRC, at irc.shadowworld.net, room #hvcomm.
Support 4hv.org!
Donate:
4hv.org is hosted on a dedicated server. Unfortunately, this server costs and we rely on the help of site members to keep 4hv.org running. Please consider donating. We will place your name on the thanks list and you'll be helping to keep 4hv.org alive and free for everyone. Members whose names appear in red bold have donated recently. Green bold denotes those who have recently donated to keep the server carbon neutral.
Special Thanks To:
Aaron Holmes
Aaron Wheeler
Adam Horden
Alan Scrimgeour
Andre
Andrew Haynes
Anonymous000
asabase
Austin Weil
barney
Barry
Bert Hickman
Bill Kukowski
Blitzorn
Brandon Paradelas
Bruce Bowling
BubeeMike
Byong Park
Cesiumsponge
Chris F.
Chris Hooper
Corey Worthington
Derek Woodroffe
Dalus
Dan Strother
Daniel Davis
Daniel Uhrenholt
datasheetarchive
Dave Billington
Dave Marshall
David F.
Dennis Rogers
drelectrix
Dr. John Gudenas
Dr. Spark
E.TexasTesla
eastvoltresearch
Eirik Taylor
Erik Dyakov
Erlend^SE
Finn Hammer
Firebug24k
GalliumMan
Gary Peterson
George Slade
GhostNull
Gordon Mcknight
Graham Armitage
Grant
GreySoul
Henry H
IamSmooth
In memory of Leo Powning
Jacob Cash
James Howells
James Pawson
Jeff Greenfield
Jeff Thomas
Jesse Frost
Jim Mitchell
jlr134
Joe Mastroianni
John Forcina
John Oberg
John Willcutt
Jon Newcomb
klugesmith
Leslie Wright
Lutz Hoffman
Mads Barnkob
Martin King
Mats Karlsson
Matt Gibson
Matthew Guidry
mbd
Michael D'Angelo
Mikkel
mileswaldron
mister_rf
Neil Foster
Nick de Smith
Nick Soroka
nicklenorp
Nik
Norman Stanley
Patrick Coleman
Paul Brodie
Paul Jordan
Paul Montgomery
Ped
Peter Krogen
Peter Terren
PhilGood
Richard Feldman
Robert Bush
Royce Bailey
Scott Fusare
Scott Newman
smiffy
Stella
Steven Busic
Steve Conner
Steve Jones
Steve Ward
Sulaiman
Thomas Coyle
Thomas A. Wallace
Thomas W
Timo
Torch
Ulf Jonsson
vasil
Vaxian
vladi mazzilli
wastehl
Weston
William Kim
William N.
William Stehl
Wesley Venis
The aforementioned have contributed financially to the continuing triumph of 4hv.org. They are deserving of my most heartfelt thanks.
Registered Member #11591
Joined: Wed Mar 20 2013, 08:20PM
Location: UK
Posts: 556
I was following this on kick-starter a while ago. It is a custom ASIC with 16 or 64 individual processing cores, for parallel processing, like hash cracking and machine vision. I wouldn't mind one. The 16 core processor is the co-processor, the host-processor has an arm CPU combined with an FPGA, for lots of power!
Registered Member #30
Joined: Fri Feb 03 2006, 10:52AM
Location: Glasgow, Scotland
Posts: 6706
Yes, it is a very powerful (for its size, cost and power budget) parallel computer. The host processor is a Xilinx Zynq, which is a dual core ARM (running Linux) combined with a FPGA on the same chip. I think the FPGA is mainly used to handle communications between the host processor, the Epiphany chip and the external memory, but you can add your own IP for things like display controllers. The FPGA is supported by Xilinx's free toolset, so you don't need to buy anything to start messing with it.
Then you have the Epiphany ASIC. Each of the parallel processing cores is a quite capable device with 32K of local memory and hardware floating point. You can program it in C/C++ using GCC. It supports OpenMP and OpenCL, industry standard parallel processing APIs, or you can make your own. Communication between cores is by the shared memory model.
Registered Member #65
Joined: Thu Feb 09 2006, 06:43AM
Location:
Posts: 1155
This architecture physically separates core silicon from memory, and therefore must eventually be limited by its cache. A shared memory model would make this contention and pipeline-miss problem exponentially more problematic.
OpenCL is a buggy abstraction layer with a nonzero transactional cost. Anything that runs with overlapping subproblems essentially thrashes the coordinator algorithms.
The $25 Raspberry Pi has a 16 core GPU (now open source) with DMA access on the SoC. Even with spacial locality, the gains for some types of problems have a nonlinear relationship to actual performance improvements. However, the FFT works great if you don't need some sort of real-time DSP:
For the $120 price (and the memory copy model), anyone can buy an older PCIx16 based 200+ core 1GHz+ nVidia card with CUDA . Note the same restrictions will hold.... and your problem might still be solved first by a i7 CPU from years ago.
There are several FPGA with ASIC CPU hybrid chips that have been around for awhile, but in general are only appropriate for a few set of problems like live video/audio/rf-signal stream processing.
The GCC will usually only guarantee functionality, but it is not even close to a high performance compiler with the machine code it outputs.
Registered Member #30
Joined: Fri Feb 03 2006, 10:52AM
Location: Glasgow, Scotland
Posts: 6706
While this is true, it is really a critique of the concept of parallel computing itself, not the Epiphany architecture as such. Not all problems in computing can be parallelised efficiently, and none of them can be parallelised automatically, you have to figure that part out yourself.
Registered Member #11591
Joined: Wed Mar 20 2013, 08:20PM
Location: UK
Posts: 556
Carbon_Rod wrote ...
... For the $120 price (and the memory copy model), anyone can buy an older PCIx16 based 200+ core 1GHz+ nVidia card with CUDA . Note the same restrictions will hold.... and your problem might still be solved first by a i7 CPU from years ago. ...
The parallella is a lot more efficient than an old GPU
Registered Member #11591
Joined: Wed Mar 20 2013, 08:20PM
Location: UK
Posts: 556
GFLOPS/W (or GFLOP/J) is how you usually define the efficiency of parallel processors. I was just thinking it would be difficult to put a Nvidia GPU in a battery powered machine vision robot/device, for example.
Registered Member #65
Joined: Thu Feb 09 2006, 06:43AM
Location:
Posts: 1155
I have seen teams that had several nVidia GPU based laptops running on robot platforms. More importantly, they actually are able to push the design cycle forward without orphan standards, or board support package problems.
The new Tegra cores are a nice little SoC, but the ASIC codecs are usually where most of the energy efficiency gains are made. New mobile ARM core SoC processors are more common than desktops now...
This site is powered by e107, which is released under the GNU GPL License. All work on this site, except where otherwise noted, is licensed under a Creative Commons Attribution-ShareAlike 2.5 License. By submitting any information to this site, you agree that anything submitted will be so licensed. Please read our Disclaimer and Policies page for information on your rights and responsibilities regarding this site.