Parsytec
Type of business | Public |
---|---|
Available in | German |
Founded | 1985 |
Headquarters | , |
Area served | North America, South America, Europe, Asia Pacific |
Founder(s) | Falk-Dietrich Kübler, Gerhard Peise, Bernd Wolf |
Services | Surface inspection systems |
URL | http://www.parsytec.de |
Isra Vision Parsytec AG, a subsidiary of Isra Vision, was originally founded in 1985 as Parsytec (parallel system technology) in Aachen, Germany.
Parsytec gained recognition in the late 1980s and early 1990s as a manufacturer of transputer-based parallel systems. Its product lineup ranged from single transputer plug-in boards for IBM PCs to large, massively parallel systems with thousands of transputers (or processors), such as the Parsytec GC. Some sources describe the latter as ultracomputer-sized, scalable multicomputers (smC).[1][2]
As part of ISRA VISION AG, the company now focuses on solutions in the machine vision and industrial image processing sectors. ISRA Parsytec products are primarily used for quality and surface inspection, particularly in the metal and paper industries.
History
[edit]Parsytec was founded in 1985 in Aachen, Germany, by Falk-Dietrich Kübler, Gerhard H. Peise, and Bernd Wolff, with an 800,000 DM grant from the Federal Ministry for Research and Technology (BMFT).[3]
Unlike SUPRENUM, Parsytec focused its systems, particularly in pattern recognition, on industrial applications such as surface inspection. As a result, the company not only captured a significant market share in European academia but also attracted numerous industrial customers, including many outside Germany. By 1988, exports accounted for approximately one-third of Parsytec's revenue. The company's turnover figures were as follows: zero in 1985, 1.5 million DM in 1986, 5.2 million DM in 1988, 9 million DM in 1989, 15 million DM in 1990, and 17 million USD in 1991.
To allow Parsytec to focus on research and development, a separate entity, ParaCom, was established to handle sales and marketing operations. While Parsytec/ParaCom maintained its headquarters in Aachen, Germany, it also operated subsidiary sales offices in Chemnitz (Germany), Southampton (United Kingdom), Chicago (USA), St Petersburg (Russia), and Moscow (Russia).[4] In Japan, Parsytec's machines were distributed by Matsushita.[3]
Between 1988 and 1994, Parsytec developed an impressive range of transputer-based computers, culminating in the "Parsytec GC" (GigaCluster). This system was available in configurations ranging from 64 to 16,384 transputers.[5]
Parsytec went public in mid-1999 with an initial public offering (IPO) on the German Stock Exchange in Frankfurt.
On 30 April 2006, founder Falk-D. Kübler left the company.[6]
In July 2007,[7] ISRA VISION AG acquired 52.6% of Parsytec AG.[8] The delisting of Parsytec shares from the stock market began in December of the same year, and since 18 April 2008, Parsytec shares have no longer been listed on the stock exchange.[9]
While Parsytec had a workforce of roughly 130 staff in the early 1990s, the ISRA VISION Group employed more than 500 people in 2012/2013.[10]
Today, the core business of ISRA Parsytec within the ISRA VISION Group is the development and distribution of surface inspection systems for strip products in the metal and paper industries.
Products/Computers
[edit]Parsytec's product range included:
- Megaframe (T414/T800) – One per board, up to ten boards in a rack or as plug-in boards
- MultiCluster (T800) – Up to 64 processors in a single rack
- SuperCluster (T800) – 16 to 1,024 processors in a single frame
- GigaCluster (planned: T9000; realized: T800 or MPC 601) – 64 to 16,384 processors in "cubes"
- x'plorer (T800 or MPC 601)
- Cognitive Computer (MPC 604 and Intel Pentium Pro)
- Powermouse (MPC 604)
In total, approximately 700 stand-alone systems (SC and GC) had been shipped.
Initially, Parsytec participated in the GPMIMD (General Purpose MIMD)[11] project under the umbrella of the ESPRIT program,[12] both of which were funded by the European Commission's Directorate for Science. However, after significant disagreements with other participants—Meiko, Parsys, Inmos, and Telmat—regarding the choice of a common physical architecture, Parsytec left the project and announced its own T9000-based machine, the GC. Due to Inmos' issues with the T9000, Parsytec was forced to switch to a system using a combination of Motorola MPC 601 CPUs and Inmos T805 processors. This led to the development of Parsytec's "hybrid" systems (e.g., GC/PP), where transputers were used as communication processors while the computational tasks were offloaded to the PowerPCs.
Parsytec's cluster systems were operated by an external workstation, typically a SUN workstation (e.g., Sun-4).[13]
There is considerable confusion regarding the names of Parsytec products. This is partly due to the architecture, but also because of the aforementioned unavailability of the Inmos T9000, which forced Parsytec to use the T805 and PowerPC processors instead. Systems equipped with PowerPC processors were given the prefix "Power."
The architecture of GC systems is based on self-contained GigaCubes. The basic architectural element of a Parsytec system was a cluster, which consisted, among other components, of four transputers/processors (i.e., a cluster is a node in the classical sense).
A GigaCube (sometimes referred to as a supernode or meganode)[14] consisted of four clusters (nodes), each with 16 Inmos T805 transputers (30 MHz), RAM (up to 4 MB per T805), and an additional redundant T805 (the 17th processor). It also included local link connections and four Inmos C004 routing chips. Hardware fault tolerance was achieved by linking each T805 to a different C004.[15] The unusual spelling of x'plorer led to variations like xPlorer, and the Gigacluster is sometimes referred to as the GigaCube or Grand Challenge.
Megaframe
[edit]Megaframe[16][17] was the product name for a family of transputer-based parallel processing modules,[18] some of which could be used to upgrade an IBM PC.[19] As a standalone system, a Megaframe could hold up to ten processor modules. Different versions of the modules were available, such as one featuring a 32-bit transputer T414 with floating-point hardware (Motorola 68881), 1 MB of RAM (80 nanosecond access time), and a throughput of 10 MIPS, or one with four 16-bit transputers (T22x) with 64 kB of RAM. Additionally, cards for special features were offered, including a graphics processor with a resolution of 1280 x 1024 pixels and an I/O "cluster" with terminal and SCSI interfaces.[20]
Multicluster
[edit]The MultiCluster-1 series consisted of statically configurable systems that could be tailored to specific user requirements, such as the number of processors, amount of memory, I/O configuration, and system topology. The required processor topology could be configured using UniLink connections, fed through a special backplane. Additionally, four external sockets were provided.
Multicluster-2 used network configuration units (NCUs) that provided flexible, dynamically configurable interconnection networks. The multiuser environment could support up to eight users through Parsytec's multiple virtual architecture software. The NCU design was based on the Inmos crossbar switch, the C004, which offers full crossbar connectivity for up to 16 transputers. Each NCU, made of C004s, connected up to 96 UniLinks, linking internal as well as external transputers and other I/O subsystems. MultiCluster-2 allowed for the configuration of various fixed interconnection topologies, such as tree or mesh structures.[14]
SuperCluster
[edit]SuperCluster[21] had a hierarchical, cluster-based design. A basic unit was a 16-transputer T800, fully connected cluster, and larger systems included additional levels of NCUs to form the necessary connections. The Network Configuration Manager (NCM) software controlled the NCUs and dynamically established the required connections. Each transputer could be equipped with 1 to 32 MB of dynamic RAM, with single-error correction and double-error detection.[14]
GigaCluster
[edit]The GigaCluster (GC) was a parallel computer produced in the early 1990s. A GigaCluster was composed of GigaCubes.[22]
Designed for the Inmos T9000 transputers, the GigaCluster could never be launched as originally planned, as the Inmos T9000 transputers never made it to market on time. This led to the development of the GC/PP (PowerPlus), in which two Motorola MPC 601 (80 MHz) were used as the dedicated CPUs, supported by four Inmos T805 transputers (30 MHz).[23]
While the GC/PP was a hybrid system, the GCel ("entry level") was based solely on the T805.[24][25] The GCel was designed to be upgradeable to the T9000 transputers (had they arrived in time), thus becoming a full GC. Since the T9000 was Inmos' evolutionary successor to the T800, the upgrade was planned to be simple and straightforward. This was because, firstly, both transputers shared the same instruction set, and secondly, they had a similar performance ratio of compute power to communication throughput. A theoretical speed-up factor of 10 was expected,[22] but in the end, it was never achieved.
The network structure of the GC was a two-dimensional lattice, with an inter-communication speed between the nodes (i.e., clusters in Parsytec's terminology) of 20 Mbit/s. For its time, the concept of the GC was exceptionally modular and scalable.
A so-called GigaCube was a module that was already a one gigaflop system and served as the building block for larger systems. A module (or "cube" in Parsytec's terminology) contained:
- Four clusters, each equipped with:
- 16 transputers (plus one additional transputer for redundancy, making it 17 transputers per cluster),
- 4 wormhole routing chips (C104 for the planned T9000 and C004 for the realized T805),
- A dedicated power supply and communication ports.
By combining modules (or cubes, respectively), one could theoretically connect up to 16,384 processors to create a very powerful system.
Typical installations included:
System | Number of CPUs | Number of GigaCubes |
---|---|---|
GC-1 | 64 | 1 |
GC-2 | 256 | 4 |
GC-3 | 1024 | 16 |
GC-4 | 4096 | 48 |
GC-5 | 16384 | 256 |
The two largest installations of the GC that were actually shipped had 1,024 processors (16 modules, with 64 transputers per module) and were operated at the data centers of the Universities of Cologne and Paderborn. In October 2004, the system at Paderborn was transferred to the Heinz Nixdorf Museums Forum,[26] where it is now inoperable.
The power consumption of a system with 1,024 processors was approximately 27 kW, and its weight was nearly a ton. In 1992, the system was priced at around 1.5 million DM. While the smaller versions, up to GC-3, were air-cooled, water cooling was mandatory for the larger systems.
In 1992, a GC with 1,024 processors ranked on the TOP500 list[27] of the world's fastest supercomputer installations. In Germany alone, it was the 22nd fastest computer.
In 1995, there were nine Parsytec computers on the TOP500 list, including two GC/PP 192 installations, which ranked 117th and 188th.[28]
In 1996, they still ranked 230th and 231st on the TOP500 list.[29][30]
x'plorer
[edit]The x'plorer model came in two versions: The initial version featured 16 transputers, each with access to 4 MB of RAM, and was called x'plorer. Later, when Parsytec switched to the PPC architecture, it was renamed POWERx'plorer and featured 8 MPC 601 CPUs. Both models were housed in the same desktop case, designed by Via 4 Design.[31]
In any model, the x'plorer was essentially a single "slice" — which Parsytec referred to as a cluster[32] — of a GigaCube (PPC or Transputer), with the smallest version (GC-1) using 4 of these clusters. As a result, some referred to it as a "GC-0.25."[33]
The POWERx'plorer was based on 8 processing units arranged in a 2D mesh. Each processing unit included:
- One 80 MHz MPC 601 processor,
- 8 MB of local memory, and
- A transputer for establishing and maintaining communication links.[34]
Cognitive Computer
[edit]The Parsytec CC (Cognitive Computer) system[35][36][37] was an autonomous unit at the card rack level.
The CC card rack subsystem provided the system with its infrastructure, including power supply and cooling. The system could be configured as a standard 19-inch rack-mounted unit, which accepted various 6U plug-in modules.
The CC system[38] was a distributed memory, message-passing parallel computer and is globally classified in the MIMD category of parallel computers.
There were two different versions available:
- CCe: Based on the Motorola MPC 604 processor running at 133 MHz with 512 KB L2 cache. The modules were connected together at 1 Gbit/s using high-speed (HS) link technology according to the IEEE 1355 standard, allowing data transfer at up to 75 MB/s. The communication controller was integrated into the processor nodes through the PCI bus. The system board used the MPC 105 chip for memory control, DRAM refresh, and memory decoding for banks of DRAM and/or Flash. The CPU bus speed was limited to 66 MHz, while the PCI bus speed was a maximum of 33 MHz.
- CCi: Based on the Intel Pentium Pro, its core elements were dual Pentium Pro-based motherboards (at 266 MHz), which were interconnected using several high-speed networks. Each dual motherboard had 128 MB of memory. Each node had a peak performance of 200 MFLOPS. The product spectrum included single-processor or SMP boards, up to a 144-node system, a wide variety of PCI cards, and different communication solutions (Gigabit HS-Link, Myrinet, SCI, ATM, or Fast Ethernet). The operating systems were Windows NT 4.0 and ParsyFrame (with an optional UNIX environment).[39]
In all CC systems, the nodes were directly connected to the same router, which implemented an active hardware 8x8 crossbar switch for up to 8 connections using the 40 MB/s high-speed link.
Regarding the CCe, the software was based on IBM's AIX 4.1 UNIX operating system, along with Parsytec's parallel programming environment, Embedded PARIX (EPX).[40] This setup combined a standard UNIX environment (including compilers, tools, and libraries) with an advanced software development environment. The system was integrated into the local area network using standard Ethernet. As a result, a CC node had a peak performance of 266 MFLOPS. The peak performance of the 8-node CC system installed at Geneva University Hospital was therefore 2.1 GFLOPS.[41]
Powermouse
[edit]Powermouse was another scalable system that consisted of modules and individual components. It was a straightforward extension of the x'plorer system.[39] Each module (dimensions: 9 cm x 21 cm x 45 cm) contained four MPC 604 processors (200/300 MHz) and 64 MB of RAM, achieving a peak performance of 2.4 GFLOPS. A separate communication processor (T425) equipped with 4 MB of RAM[42] controlled the data flow in four directions to other modules in the system. The bandwidth of a single node was 9 MB/s.
For about 35,000 DM, a basic system consisting of 16 CPUs (i.e., four modules) could provide a total computing power of 9.6 Gflop/s. As with all Parsytec products, Powermouse required a Sun Sparcstation as the front-end.
All software, including PARIX with C++ and Fortran 77 compilers and debuggers (alternatively providing MPI or PVM as user interfaces), was included.[43]
Operating system
[edit]The operating system used was PARIX (PARallel UnIX extensions)[44] – PARIXT8 for the T80x transputers and PARIXT9 for the T9000 transputers, respectively. Based on UNIX, PARIX[45] supported remote procedure calls and was compliant with the POSIX standard. PARIX provided UNIX functionality at the front-end (e.g., a Sun SPARCstation, which had to be purchased separately) with library extensions for the needs of the parallel system at the back-end, which was the Parsytec product itself (connected to the front-end for operation). The PARIX software package included components for the program development environment (compilers, tools, etc.) and the runtime environment (libraries). PARIX offered various types of synchronous and asynchronous communication.
In addition, Parsytec provided a parallel programming environment called Embedded PARIX (EPX).[40]
To develop parallel applications using EPX, data streams and function tasks were allocated to a network of nodes. The data handling between processors required only a few system calls. Standard routines for synchronous communication, such as send and receive, were available, as well as asynchronous system calls. The full set of EPX calls formed the EPX application programming interface (API). The destination for any message transfer was defined through a virtual channel that ended at any user-defined process. Virtual channels were managed by EPX and could be defined by the user. The actual message delivery system utilized the router.[41] Additionally, COSY (Concurrent Operating SYstem)[46] and Helios could also be run on the machines. Helios supported Parsytec's special reset mechanism out of the box.
See also
[edit]References
[edit]- ^ Massively Parallel Computers: Why Not Parallel Computers for the Masses? G. Bell at microsoft.com
- ^ Alternating-Direction Line-Relaxation Methods on Multicomputers J. Hofhaus et al., SIAM J. ScI. COMPUT. Vol. 17, No. 2, pp. 454-478, March 1996
- ^ a b Duell der Zahlenfresser at zeit.de (German)
- ^ Parsytec GmbH at new-npac.org
- ^ Parsytec Article at GeekDot.com
- ^ Annual Statement of Accounts 2006[permanent dead link ] at parsytec.de
- ^ ISRA Vision übernimmt Parsytec Jul 23, 2007 at finanznachrichten.de (German)
- ^ ISRA VISION AG - Erwerb der Mehrheit an der Parsytec AG Jul 24, 2007 at equinet-ag.de (German)
- ^ Investor Relations at ISRA at parsytec.de
- ^ Annual Report 2012/2013 May 05, 2014 at isravision.com
- ^ General-Purpose MIMD Machines at cordis.europa.eu
- ^ European programme (EEC) for research and development in information technologies (ESPRIT), 1984-1988
- ^ A Framework for Characterising Parallel Systems for Performance Evaluation Efstathios Papaefstathiou (1995)
- ^ a b c ESN Information Bulletin 92-08 at dtic.mil
- ^ Hypecube Solutions for Conjugate Directions, J. E. Hartman (1991) at dtic.mil
- ^ MEGAFRAME TPM-1 - Hardware Documentation Ver. 1.2 (1987) at classiccmp.org
- ^ MEGAFRAME MTM-2 - Hardware Documentation Ver. 1.3 (1987) at classiccmp.org
- ^ MEGAFRAME Familie Archived 2013-02-10 at archive.today May 1987 at computerwoche.de (German)
- ^ Ram Meenakshisundaram's Transputer Home Page at classiccmp.org
- ^ Transputersystem ist frei konfigurierbar Archived 2008-04-02 at the Wayback Machine Mar 1987 at computerwoche.de (German)
- ^ Tobias K. [DE] (2005-11-18), parsytec SuperCluster SC 320, retrieved 2024-11-23
- ^ a b Gigacube Article at GeekDot.com
- ^ The Parsytec Power Plus at netlib.org
- ^ Programmierung und Anwendung der Parallelrechnersysteme Parsytec SC und Parsytec GC/PP B. Heiming, 1996, Technical University Hamburg-Harburg (German)
- ^ Synthesizing massive parallel simulation systems to estimate switching activity in finite state machines[permanent dead link ] W. Bachmann et al., Darmstadt University of Technology
- ^ Homepage of the Heinz Nixdorf Museum Forum
- ^ http://www.top500.org TOP500 List
- ^ Top500 List 1996
- ^ Lecture Notes on Applied Parallel Computing Archived 2010-08-16 at the Wayback Machine at ocw.mit.edu
- ^ Viel hilft viel: Die neuen Supercomputer haben Billigprozessoren wie der PC nebenan - aber zu Tausenden at zeit.de (German)
- ^ iF Online Exhibition - Via 4 Design at ifdesign.de
- ^ "Picture of cluster". www.parallab.uib.no. Archived from the original on 2006-08-19.
- ^ x'plorer Article at GeekDot.com
- ^ Experimental Study on Time and Space Sharing on the PowerXplorer S. Bani-Ahmad, Ubiquitous Computing and Communication Journal Vol. 3 No. 1
- ^ "Современные системы фирмы Parsytec". www.ccas.ru. Retrieved 2024-11-23.
- ^ Parsytec CC Series (Hardware.Documentation), Rev. 1.2 (1996) Parsytec GmbH
- ^ The Parsytec CC series at netlib.org
- ^ Target Machines Archived 2004-08-12 at the Wayback Machine at http://parallel.di.uoa.gr
- ^ a b Parallel Computing Hardware Archived 2012-06-16 at the Wayback Machine at ssd.sscc.ru
- ^ a b Embedded Parix Ver. 1.9.2, Software Documentation (1996)
- ^ a b Implementation of an Environment for Monte Carlo simulation of Fully 3-D Positron Tomography on a High-Performance Parallel Platform H. Zaidi, Parallel Computing, Vol. 24 (1998), pp. 1523-1536
- ^ System Parsytec Power Mouse in CSA Archived 2013-04-16 at archive.today Dec 15, 1998 at csa.ru
- ^ Parsytec liefert Baukasten für Parallelrechner Archived 2013-02-10 at archive.today Sep 19, 1997 at comuterwoche.de (German)
- ^ PARIX Release 1.2 Software Documentation March 1993
- ^ "Seite nicht erreichbar - Universität Osnabrück". www.informatik.uni-osnabrueck.de. Retrieved 2024-12-01.
- ^ COSY – ein Betriebssystem für hochparallele Computer R. Butenuth at uni-paderborn.de (German)
External links
[edit]- Homepage of ISRA VISION PARSYTEC AG
- Ram Meenakshisundaram's Transputer Home Page at classiccmp.org
- 16384 Prozessoren bringen 400 Gflops Transputer-Superrechner von Parsytec als neuer Weltmeister Article at computerwoche.de (German)
- Zur Strategie von Parsytec Kuebler: "In zehn Jahren rechnen die meisten Computer parallel" Oct 1, 1993, at computerwoche.de (German)
- The FTMPS-Project: Design and Implementation of Fault-tolerance Techniques for Massively Parallel Systems[permanent dead link ] J. Vounckx et al.
- Homepage of Via 4 Design