|By JCN Newswire||
|January 23, 2014 11:13 PM EST||
Tokyo, Jan 24, 2014 - (JCN Newswire) - Fujitsu today announced that it has completed construction of a new computer system for the Institute for Cosmic Ray Research (ICRR) of the University of Tokyo. This x86 cluster system, forming the backbone for ICRR's research on cosmic rays(1), began operation on January 1, 2014.
The new system will store and analyze observational data on cosmic rays, which include protons, gamma rays, and neutrinos, and will be used for ICRR's various domestic and international joint research projects in astrophysics and particle physics.
The computation server is an x86 cluster comprised of 136 FUJITSU Server PRIMERGY CX250 S2 servers, acting in tandem with a high-speed distributed file system of 33 FUJITSU Storage ETERNUS DX80 S2 units running the FEFS(2) file system for speed, reliability, and scalability.
Integer-math performance of the computation server is rated at 96,426 using the SPECint_rate2006 benchmark(3), about ten times the performance of the previous system. It also boasts a file server capacity of 4.4 petabytes, which is roughly 3.2 times more capacity than the previous system, and has a data-transfer speed of 18 GB/s, an approximately 30-fold improvement.
This Fujitsu system will support the KAGRA Project(4) and contribute to advances in other cutting-edge research in astrophysics and particle physics.
ICRR installed its first x86 cluster system in January 1991 as a collaborative research environment for multiple projects in the field of particle physics and cosmic-ray research, conducted primarily by the University of Tokyo. At locations both in and outside of Japan, ICRR has conducted research on high-energy cosmic rays, high-energy gamma rays, and neutrinos; for example, the Super-Kamiokande(5) Project has discovered neutrino oscillations and the Tibet AS-gamma Experiment used cosmic rays to observe the sun's cosmic-ray shadow(6).
Together with expanding its research in the fields of astrophysics and particle physics, ICRR plans to move forward on new experimental projects, such as the KAGRA project, which aims to detect the gravitational wave for the first time in history, and the ground-based gamma ray observatory, CTA(7), which is designed to help understand galactic and extragalactic very-high energy gamma-ray sources.
To meet the needs of these new projects, ICRR required a new computer system that could keep pace with dramatic growth expected in the volume of observational data and of analyses. It would be essential for the new system to boost analytical performance, have a greater storage capacity and to enhance data input and output performance.
To satisfy these requirements, the new system started with new hardware that has approximately ten times the integer-math performance of its predecessor, enough storage capacity to accommodate roughly six years' worth of observational data in cosmic-ray research, and the FEFS scalable file system to sidestep the bottlenecks created by centralized data I/O loads, all of which result in an efficient environment that enables more accurate research.
The new system offers two main functions: as a computation server that handles data analysis, and as a file server that stores observational data.
- Computation server
An x86 cluster of 136 PRIMERGY CX250 S2 servers with up-to-date CPUs results in integer-math performance rated at 96,426 using the SPECint_rate2006 benchmark, an approximately tenfold improvement over its predecessor.
- File server
With 10 FUJITSU Server PRIMERGY RX300 S7 servers and 33 ETERNUS DX80 S2 storage units, the system has 4.4 PB of storage capacity and an 18-GB/s data-transfer speed, giving it roughly 3.2 times the storage capacity of its predecessor and about 30 times the data transfer performance. The FEFS scalable file system includes functions that allow storage capacity and bandwidth to be allocated on a per-user and per-group basis, creating a research environment optimized for multiple research projects. The FEFS file system demonstrated its value in Super-Kamiokande's system in 2011.
Additional components of the system include 12 FUJITSU Server PRIMERGY RX200 S8 servers used as a login environment for 450 ICRR researchers, a FUJITSU Storage ETERNUS NR1000 F3220 storage system with 22 TB for networked storage, 15x86 servers that provide e-mail, web, and other network services, a storage unit, and additional network switches.
The new system will be used in ICRR's projects both in and outside of Japan, and will help illuminate the mechanisms underlying ultra-high energy phenomena such as where and how high-energy cosmic rays originate and accelerate, through deeper, broader studies of the cosmos. Furthermore, inquiries into the properties of neutrinos and into dark matter particles will lead to even further advances in the study of cosmic rays and particle physics. The world's first gravitational wave detections promise to reveal new aspects of the universe, bringing us closer to testing the theory of general relativity, and to probe the beginning of the universe.
Composition of the New System
ICRR will be the primary user of the new system for multiple research projects.
Comment from Kimihiro Okumura, Associate Professor, ICRR, University of Tokyo
"At ICRR, we expect to begin many new research projects. Accordingly, it is anticipated that there will be a significant increase in observational data. To handle the storage and high-speed processing of this data, we needed to upgrade the performance of our computer system. In particular, our goal was to dramatically increase the speed of data transmission, and for a high-performance, highly reliable storage system that meets these requirements, we employed the Fujitsu FEFS file system. We expect the research groups using the computer system to announce many new observational results in the future that generate new insights into hitherto unknown phenomena relating to the universe and cosmic ray particles."
(1) Cosmic rays:Nuclei and elementary particles originating both inside and outside the Milky Way galaxy. Where they were made and how they are accelerated to such high energies is still not understood.
(2) FEFS:A high-speed distributed file system that can be shared over as many as 100,000 nodes.
(3) 96,426 using the SPECint_rate2006 benchmark: The estimated value using SPECint_rate2006 as a benchmark of CPU performance.
(4) KAGRA Project:The Large-Scale Cryogenic Gravitational Wave Telescope. Selected in 2012 for the Leading-edge Research Infrastructure Program by the Ministry of Education, Sports, Science, Culture, and Technology, facilities for this ICRR project are currently under construction, and due to go into full operation in 2017-2018.
(5) Super-Kamiokande:A facility consisting of a subterranean pool used as a Cherenkov-ray collector located in Kamioka, Gifu Prefecture, Japan.6 Sun's cosmic-ray shadow:A phenomenon in which cosmic rays (high-energy charged particles that move through space) are deflected by the sun's magnetic field.7 Ground-based gamma ray observatory, CTA:A future ground-based gamma-ray observatory planned to come online in 2020 with 10 times the sensitivity of existing gamma-ray telescopes. The sites will be decided in March 2013, and the construction will start in 2015.
About Fujitsu Limited
Fujitsu is the leading Japanese information and communication technology (ICT) company offering a full range of technology products, solutions and services. Approximately 170,000 Fujitsu people support customers in more than 100 countries. We use our experience and the power of ICT to shape the future of society with our customers. Fujitsu Limited (TSE: 6702) reported consolidated revenues of 4.4 trillion yen (US$47 billion) for the fiscal year ended March 31, 2013 For more information, please see www.fujitsu.com.
Source: Fujitsu Limited
Fujitsu Limited Public and Investor Relations www.fujitsu.com/global/news/contacts/ +81-3-3215-5259
Copyright 2014 JCN Newswire. All rights reserved. www.japancorp.net
Many private cloud projects were built to deliver self-service access to development and test resources. While those clouds delivered faster access to resources, they lacked visibility, control and security needed for production deployments. In their session at 18th Cloud Expo, Steve Anderson, Product Manager at BMC Software, and Rick Lefort, Principal Technical Marketing Consultant at BMC Software, will discuss how a cloud designed for production operations not only helps accelerate developer...
May. 6, 2016 07:00 PM EDT Reads: 1,405
In his session at 18th Cloud Expo, Sagi Brody, Chief Technology Officer at Webair Internet Development Inc., will focus on real world deployments of DDoS mitigation strategies in every layer of the network. He will give an overview of methods to prevent these attacks and best practices on how to provide protection in complex cloud platforms. He will also outline what we have found in our experience managing and running thousands of Linux and Unix managed service platforms and what specifically c...
May. 6, 2016 06:15 PM EDT Reads: 1,430
As enterprises around the world struggle with their digital transformation efforts, many are finding that innovative digital teams are moving much faster than their hidebound IT organizations. Rather than struggling to convince traditional IT to get with the digital program, executives are taking advice from IT research firm Gartner, and encouraging existing IT to continue in their desultory ways. However, many CIOs are realizing the dangers of following Gartner’s advice. The central challenge ...
May. 6, 2016 06:00 PM EDT Reads: 354
Much of the value of DevOps comes from a (renewed) focus on measurement, sharing, and continuous feedback loops. In increasingly complex DevOps workflows and environments, and especially in larger, regulated, or more crystallized organizations, these core concepts become even more critical. In his session at @DevOpsSummit at 18th Cloud Expo, Andi Mann, Chief Technology Advocate at Splunk, will show how, by focusing on 'metrics that matter,' you can provide objective, transparent, and meaningfu...
May. 6, 2016 06:00 PM EDT Reads: 1,261
trust and privacy in their ecosystem. Assurance and protection of device identity, secure data encryption and authentication are the key security challenges organizations are trying to address when integrating IoT devices. This holds true for IoT applications in a wide range of industries, for example, healthcare, consumer devices, and manufacturing. In his session at @ThingsExpo, Lancen LaChance, vice president of product management, IoT solutions at GlobalSign, will teach IoT developers how t...
May. 6, 2016 06:00 PM EDT Reads: 868
A critical component of any IoT project is the back-end systems that capture data from remote IoT devices and structure it in a way to answer useful questions. Traditional data warehouse and analytical systems are mature technologies that can be used to handle large data sets, but they are not well suited to many IoT-scale products and the need for real-time insights. At Fuze, we have developed a backend platform as part of our mobility-oriented cloud service that uses Big Data-based approache...
May. 6, 2016 05:30 PM EDT Reads: 843
The pace of innovation, vendor lock-in, production sustainability, cost-effectiveness, and managing risk… In his session at 18th Cloud Expo, Dan Choquette, Founder of RackN, will discuss how CIOs are challenged finding the balance of finding the right tools, technology and operational model that serves the business the best. He will discuss how clouds, open source software and infrastructure solutions have benefits but also drawbacks and how workload and operational portability between vendors...
May. 6, 2016 05:00 PM EDT Reads: 361
Cloud Object Storage is effectively infinitely scalable and boasts the lowest total costs. But cloud SLAs and T&Cs are traditionally optimized for huge customers like Netflix, so applications demanding better confidentiality or higher availability typically can’t reap the benefits of public cloud storage. In his session at 18th Cloud Expo, Don Martin, CTO of Security First Corp, will provide an overview of innovative technologies available today – secret sharing and information dispersal algori...
May. 6, 2016 05:00 PM EDT Reads: 374
We're entering the post-smartphone era, where wearable gadgets from watches and fitness bands to glasses and health aids will power the next technological revolution. With mass adoption of wearable devices comes a new data ecosystem that must be protected. Wearables open new pathways that facilitate the tracking, sharing and storing of consumers’ personal health, location and daily activity data. Consumers have some idea of the data these devices capture, but most don’t realize how revealing and...
May. 6, 2016 04:00 PM EDT Reads: 886
As the rapid adoption of containers continues, companies are finding that they lack the operational tools to understand the behavior of applications deployed in these containers, and how to identify issues in their application infrastructure. For example, how are multiple containers within an application impacting each other’s performance? If an application’s service is degraded, which container is to blame? In the case of an application outage, what was the root cause of the outage?
May. 6, 2016 04:00 PM EDT Reads: 1,214
When it comes to IoT in the enterprise, namely the commercial building and hospitality markets, a benefit not getting the attention it deserves is energy efficiency, and IoT's direct impact on a cleaner, greener environment when installed in smart buildings. Until now clean technology was offered piecemeal and led with point solutions that require significant systems integration to orchestrate and deploy. There didn't exist a 'top down' approach that can manage and monitor the way a Smart Buildi...
May. 6, 2016 03:15 PM EDT Reads: 304
Cloud-based NCLC (No-code/low code) application builder platforms empower everyone in the organization to quickly build applications and executable processes that broaden access, deepen collaboration, and enhance transparency for all team members. Line of business owners (LOBO) and operations managers know best their part of the business and their processes. IT departments are beginning to leverage NCLC platforms to empower and enable LOBOs to lead the innovation, transform the organization, an...
May. 6, 2016 03:00 PM EDT Reads: 350
Unless you don’t use the internet, don’t live in California, or haven’t been paying attention to the recent news… you should be aware that self-driving cars are on their way to becoming a reality. I have seen them – they are real. If you believe in the future reality of self-driving cars, then continue reading on. If you don’t believe in the future possibilities, then I am not sure what to do to convince you other than discuss the very real changes that will roll out with the consumer producti...
May. 6, 2016 02:15 PM EDT Reads: 921
There is an ever-growing explosion of new devices that are connected to the Internet using “cloud” solutions. This rapid growth is creating a massive new demand for efficient access to data. And it’s not just about connecting to that data anymore. This new demand is bringing new issues and challenges and it is important for companies to scale for the coming growth. And with that scaling comes the need for greater security, gathering and data analysis, storage, connectivity and, of course, the...
May. 6, 2016 01:45 PM EDT Reads: 1,379
The IETF draft standard for M2M certificates is a security solution specifically designed for the demanding needs of IoT/M2M applications. In his session at @ThingsExpo, Brian Romansky, VP of Strategic Technology at TrustPoint Innovation, will explain how M2M certificates can efficiently enable confidentiality, integrity, and authenticity on highly constrained devices.
May. 6, 2016 01:00 PM EDT Reads: 1,424