征稿已开启

查看我的稿件

注册已开启

查看我的门票

已截止
活动简介

The Advances in Software and Hardware for Big Data to Knowledge Discovery (ASH) workshop aims to connect the latest hardware and software developments with the end users of big data. It focuses on the accessibility and applicability of the latest hardware and software to practical domain problems and hence directly facilitates domain researchers' data driven discovery. The issues in discussion include performance evaluation, optimizations, accessibility and usability of new technologies. Hailed by some as the fourth paradigm in science, data-intensive science has brought a profound transformation to scientific research. Indeed, the data-driven discovery has already happened in various research fields, such as earth sciences, medical sciences, biology and physics, to name just a few. It is expected that a vast volume of scientific data captured by new instruments will be publically accessible for the purposes of continued and deeper data analysis. Big Data analytic will result in the development of many new theories and discoveries but will also require substantial computational resources in the process. However, many domain sciences still mostly rely on traditional experimental paradigms. It is often a major challenge to transform a solution obtained on a standalone server into a massively parallel one running on tens, hundreds, or even thousands of servers. It is a crucial issue to make the latest technology advancements in software and hardware accessible and usable to the domain scientists, especially those in the fields that traditionally lack computation and programming, but have nonetheless become the driving forces of scientific discovery. Fueled by the big data analytics needs, new computing and storage technologies in hardware and software are also in rapid development and pushing for new high-end hardware for big data problems. These new hardware brings new opportunities for performance improvement but also new challenges. While those technologies have the potential to greatly improve the capabilities of big data analytics, such potential are often not fully realized. Due to the cost, sophistications of those technology, and limited initial application support, the new technologies often seem remote to the end users and are not fully utilized in the academia years after their invention. It is therefore very important to make those technologies understood and accessible by data scientists in a timely manner. Meanwhile, comprehensive analytic software packages and programming environments, have become increasingly popular as open-source platforms for data analysis and need to be scaled and adapted for Big Data analysis. Those software not only provide collection of analytic methods but also have the potential to utilize new hardware transparently and reduce the efforts required of the end users. For examples, Recently members of the R and HPC communities have tried to step up to big data with R, resulting in methods for effectively adapting R to a variety of high-performance and high-throughput computing technologies. Parallel to these developments, a family of software frameworks (e.g., Apache Spark, Airavata) has been developed for executing and managing computational jobs and workflows on distributed computing resources, while providing web-based science gateways to assist domain scientists to compose, manage, execute, and monitor big data applications and workflows composed of these services.

征稿信息

重要日期

2015-08-30
初稿截稿日期
留言
验证码 看不清楚,更换一张
全部留言
重要日期
  • 10月29日

    2015

    会议日期

  • 08月30日 2015

    初稿截稿日期

  • 10月29日 2015

    注册截止日期

主办单位
International Society of Granular Computing
联系方式
移动端
在手机上打开
小程序
打开微信小程序
客服
扫码或点此咨询