TY - JOUR T1 - Automatic extraction of retinal features from colour retinal images for glaucoma diagnosis: A review JF - Computerized Medical Imaging and Graphics Y1 - 2013 A1 - Haleem, M.S. A1 - Han, L. A1 - van Hemert, J. A1 - Li, B. KW - retinal imaging AB - Glaucoma is a group of eye diseases that have common traits such as, high eye pressure, damage to the Optic Nerve Head and gradual vision loss. It affects peripheral vision and eventually leads to blindness if left untreated. The current common methods of pre-diagnosis of Glaucoma include measurement of Intra-Ocular Pressure (IOP) using Tonometer, Pachymetry, Gonioscopy; which are performed manually by the clinicians. These tests are usually followed by Optic Nerve Head (ONH) Appearance examination for the confirmed diagnosis of Glaucoma. The diagnoses require regular monitoring, which is costly and time consuming. The accuracy and reliability of diagnosis is limited by the domain knowledge of different ophthalmologists. Therefore automatic diagnosis of Glaucoma attracts a lot of attention. This paper surveys the state-of-the-art of automatic extraction of anatomical features from retinal images to assist early diagnosis of the Glaucoma. We have conducted critical evaluation of the existing automatic extraction methods based on features including Optic Cup to Disc Ratio (CDR), Retinal Nerve Fibre Layer (RNFL), Peripapillary Atrophy (PPA), Neuroretinal Rim Notching, Vasculature Shift, etc., which adds value on efficient feature extraction related to Glaucoma diagnosis. VL - 37 SN - 0895-6111 UR - http://linkinghub.elsevier.com/retrieve/pii/S0895611113001468?showall=true ER - TY - CONF T1 - Automatic Extraction of the Optic Disc Boundary for Detecting Retinal Diseases T2 - 14th {IASTED} International Conference on Computer Graphics and Imaging (CGIM) Y1 - 2013 A1 - M.S. Haleem A1 - L. Han A1 - B. Li A1 - A. Nisbet A1 - van Hemert, J. A1 - M. Verhoek ED - L. Linsen ED - M. Kampel KW - retinal imaging AB - In this paper, we propose an algorithm based on active shape model for the extraction of Optic Disc boundary. The determination of Optic Disc boundary is fundamental to the automation of retinal eye disease diagnosis because the Optic Disc Center is typically used as a reference point to locate other retinal structures, and any structural change in Optic Disc, whether textural or geometrical, can be used to determine the occurrence of retinal diseases such as Glaucoma. The algorithm is based on determining a model for the Optic Disc boundary by learning patterns of variability from a training set of annotated Optic Discs. The model can be deformed so as to reflect the boundary of Optic Disc in any feasible shape. The algorithm provides some initial steps towards automation of the diagnostic process for retinal eye disease in order that more patients can be screened with consistent diagnoses. The overall accuracy of the algorithm was 92% on a set of 110 images. JF - 14th {IASTED} International Conference on Computer Graphics and Imaging (CGIM) PB - {ACTA} Press ER - TY - CHAP T1 - DISPEL Enactment T2 - THE DATA BONANZA: Improving Knowledge Discovery for Science, Engineering and Business Y1 - 2013 A1 - Chee Sun Liew A1 - Krause, Amrey A1 - Snelling, David ED - Malcolm Atkinson ED - Rob Baxter ED - Peter Brezany ED - Oscar Corcho ED - Michelle Galea ED - Parsons, Mark ED - Snelling, David ED - van Hemert, Jano KW - Data Streaming KW - Data-Intensive Engineering KW - Dispel KW - Workflow Enactment AB - Chapter 12: "DISPEL enactment", describes the four stages of DISPEL enactment. It is targeted at the data-intensive engineers who implement enactment services. JF - THE DATA BONANZA: Improving Knowledge Discovery for Science, Engineering and Business PB - John Wiley & Sons Inc. ER - TY - JOUR T1 - Embedded systems for global e-Social Science: Moving computation rather than data JF - Future Generation Computer Systems Y1 - 2013 A1 - Ashley D. Lloyd A1 - Terence M. Sloan A1 - Antonioletti, Mario A1 - Gary McGilvary AB - There is a wealth of digital data currently being gathered by commercial and private concerns that could supplement academic research. To unlock this data it is important to gain the trust of the companies that hold the data as well as showing them how they may benefit from this research. Part of this trust is gained through established reputation and the other through the technology used to safeguard the data. This paper discusses how different technology frameworks have been applied to safeguard the data and facilitate collaborative work between commercial concerns and academic institutions. The paper focuses on the distinctive requirements of e-Social Science: access to large-scale data on behaviour in society in environments that impose confidentiality constraints on access. These constraints arise from both privacy concerns and the commercial sensitivities of that data. In particular, the paper draws on the experiences of building an intercontinental Grid–INWA–from its first operation connecting Australia and Scotland to its subsequent extension to China across the Trans-Eurasia Information Network–the first large-scale research and education network for the Asia-Pacific region. This allowed commercial data to be analysed by experts that were geographically distributed across the globe. It also provided an entry point for a major Chinese commercial organization to approve use of a Grid solution in a new collaboration provided the centre of gravity of the data is retained within the jurisdiction of the data owner. We describe why, despite this approval, an embedded solution was eventually adopted. We find that ‘data sovereignty’ dominates any decision on whether and how to participate in e-Social Science collaborations and how this might impact on a Cloud based solution to this type of collaboration. VL - 29 UR - http://www.sciencedirect.com/science/article/pii/S0167739X12002336 IS - 5 ER - TY - JOUR T1 - Exploiting Parallel R in the Cloud with SPRINT JF - Methods of Information in Medicine Y1 - 2013 A1 - Piotrowski, Michal A1 - Gary McGilvary A1 - Sloan, Terence A1 - Mewissen, Muriel A1 - Ashley Lloyd A1 - Forster, Thorsten A1 - Mitchell, Lawrence A1 - Ghazal, Peter A1 - Hill, Jon AB - Background: Advances in DNA Microarray devices and next-generation massively parallel DNA sequencing platforms have led to an exponential growth in data availability but the arising opportunities require adequate computing resources. High Performance Computing (HPC) in the Cloud offers an affordable way of meeting this need. Objectives: Bioconductor, a popular tool for high-throughput genomic data analysis, is distributed as add-on modules for the R statistical programming language but R has no native capabilities for exploiting multi-processor architectures. SPRINT is an R package that enables easy access to HPC for genomics researchers. This paper investigates: setting up and running SPRINT-enabled genomic analyses on Amazon’s Elastic Compute Cloud (EC2), the advantages of submitting applications to EC2 from different parts of the world and, if resource underutilization can improve application performance. Methods: The SPRINT parallel implementations of correlation, permutation testing, partitioning around medoids and the multi-purpose papply have been benchmarked on data sets of various size on Amazon EC2. Jobs have been submitted from both the UK and Thailand to investigate monetary differences. Results: It is possible to obtain good, scalable performance but the level of improvement is dependent upon the nature of algorithm. Resource underutilization can further improve the time to result. End-user’s location impacts on costs due to factors such as local taxation. Conclusions: Although not designed to satisfy HPC requirements, Amazon EC2 and cloud computing in general provides an interesting alternative and provides new possibilities for smaller organisations with limited funds. VL - 52 IS - 1 ER - TY - CONF T1 - Towards Addressing CPU-Intensive Seismological Applications in Europe T2 - International Supercomputing Conference Y1 - 2013 A1 - Michele Carpené A1 - I.A. Klampanos A1 - Siew Hoon Leong A1 - Emanuele Casarotti A1 - Peter Danecek A1 - Graziella Ferini A1 - Andre Gemünd A1 - Amrey Krause A1 - Lion Krischer A1 - Federica Magnoni A1 - Marek Simon A1 - Alessandro Spinuso A1 - Luca Trani A1 - Malcolm Atkinson A1 - Giovanni Erbacci A1 - Anton Frank A1 - Heiner Igel A1 - Andreas Rietbrock A1 - Horst Schwichtenberg A1 - Jean-Pierre Vilotte AB - Advanced application environments for seismic analysis help geoscientists to execute complex simulations to predict the behaviour of a geophysical system and potential surface observations. At the same time data collected from seismic stations must be processed comparing recorded signals with predictions. The EU-funded project VERCE (http://verce.eu/) aims to enable specific seismological use-cases and, on the basis of requirements elicited from the seismology community, provide a service-oriented infrastructure to deal with such challenges. In this paper we present VERCE’s architecture, in particular relating to forward and inverse modelling of Earth models and how the, largely file-based, HPC model can be combined with data streaming operations to enhance the scalability of experiments.We posit that the integration of services and HPC resources in an open, collaborative environment is an essential medium for the advancement of sciences of critical importance, such as seismology. JF - International Supercomputing Conference CY - Leipzig, Germany ER - TY - CONF T1 - V-BOINC: The Virtualization of BOINC T2 - In Proceedings of the 13th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGrid 2013). Y1 - 2013 A1 - Gary McGilvary A1 - Barker, Adam A1 - Ashley Lloyd A1 - Malcolm Atkinson AB - The Berkeley Open Infrastructure for Network Computing (BOINC) is an open source client-server middleware system created to allow projects with large computational requirements, usually set in the scientific domain, to utilize a technically unlimited number of volunteer machines distributed over large physical distances. However various problems exist deploying applications over these heterogeneous machines using BOINC: applications must be ported to each machine architecture type, the project server must be trusted to supply authentic applications, applications that do not regularly checkpoint may lose execution progress upon volunteer machine termination and applications that have dependencies may find it difficult to run under BOINC. To solve such problems we introduce virtual BOINC, or V-BOINC, where virtual machines are used to run computations on volunteer machines. Application developers can then compile their applications on a single architecture, checkpointing issues are solved through virtualization API's and many security concerns are addressed via the virtual machine's sandbox environment. In this paper we focus on outlining a unique approach on how virtualization can be introduced into BOINC and demonstrate that V-BOINC offers acceptable computational performance when compared to regular BOINC. Finally we show that applications with dependencies can easily run under V-BOINC in turn increasing the computational potential volunteer computing offers to the general public and project developers. V-BOINC can be downloaded at http://garymcgilvary.co.uk/vboinc.html JF - In Proceedings of the 13th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGrid 2013). CY - Delft, The Netherlands ER - TY - JOUR T1 - Data-Intensive Architecture for Scientific Knowledge Discovery JF - Distributed and Parallel Databases Y1 - 2012 A1 - Atkinson, Malcolm P. A1 - Chee Sun Liew A1 - Michelle Galea A1 - Paul Martin A1 - Krause, Amrey A1 - Adrian Mouat A1 - Oscar Corcho A1 - Snelling, David KW - Knowledge discovery, workflow management system AB - This paper presents a data-intensive architecture that demonstrates the ability to support applications from a wide range of application domains, and support the different types of users involved in defining, designing and executing data-intensive processing tasks. The prototype architecture is introduced, and the pivotal role of DISPEL as a canonical language is explained. The architecture promotes the exploration and exploitation of distributed and heterogeneous data and spans the complete knowledge discovery process, from data preparation, to analysis, to evaluation and reiteration. The architecture evaluation included large-scale applications from astronomy, cosmology, hydrology, functional genetics, imaging processing and seismology. VL - 30 UR - http://dx.doi.org/10.1007/s10619-012-7105-3 IS - 5 ER - TY - JOUR T1 - OMERO: flexible, model-driven data management for experimental biology JF - NATURE METHODS Y1 - 2012 A1 - Chris Allan A1 - Jean-Marie Burel A1 - Josh Moore A1 - Colin Blackburn A1 - Melissa Linkert A1 - Scott Loynton A1 - Donald MacDonald A1 - et al. AB - Data-intensive research depends on tools that manage multidimensional, heterogeneous datasets. We built OME Remote Objects (OMERO), a software platform that enables access to and use of a wide range of biological data. OMERO uses a server-based middleware application to provide a unified interface for images, matrices and tables. OMERO's design and flexibility have enabled its use for light-microscopy, high-content-screening, electron-microscopy and even non-image-genotype data. OMERO is open-source software, available at http://openmicroscopy.org/. PB - Nature Publishing Group, a division of Macmillan Publishers Limited. All Rights Reserved. VL - 9 SN - 1548-7091 UR - http://dx.doi.org/10.1038/nmeth.1896 IS - 3 ER - TY - BOOK T1 - Optimisation of the enactment of fine-grained distributed data-intensive workflows Y1 - 2012 A1 - Chee Sun Liew AB - The emergence of data-intensive science as the fourth science paradigm has posed a data deluge challenge for enacting scientific workflows. The scientific community is facing an imminent flood of data from the next generation of experiments and simulations, besides dealing with the heterogeneity and complexity of data, applications and execution environments. New scientific workflows involve execution on distributed and heterogeneous computing resources across organisational and geographical boundaries, processing gigabytes of live data streams and petabytes of archived and simulation data, in various formats and from multiple sources. Managing the enactment of such workflows not only requires larger storage space and faster machines, but the capability to support scalability and diversity of the users, applications, data, computing resources and the enactment technologies. We argue that the enactment process can be made efficient using optimisation techniques in an appropriate architecture. This architecture should support the creation of diversified applications and their enactment on diversified execution environments, with a standard interface, i.e.~a workflow language. The workflow language should be both human readable and suitable for communication between the enactment environments. The data-streaming model central to this architecture provides a scalable approach to large-scale data exploitation. Data-flow between computational elements in the scientific workflow is implemented as streams. To cope with the exploratory nature of scientific workflows, the architecture should support fast workflow prototyping, and the re-use of workflows and workflow components. Above all, the enactment process should be easily repeated and automated. In this thesis, we present a candidate data-intensive architecture that includes an intermediate workflow language, named DISPEL. We create a new fine-grained measurement framework to capture performance-related data during enactments, and design a performance database to organise them systematically. We propose a new enactment strategy to demonstrate that optimisation of data-streaming workflows can be automated by exploiting performance data gathered during previous enactments. PB - The University of Edinburgh CY - Edinburgh ER - TY - JOUR T1 - Principles of Provenance (Dagstuhl Seminar 12091) JF - Dagstuhl Reports Y1 - 2012 A1 - James Cheney A1 - Anthony Finkelstein A1 - Bertram Ludäscher A1 - Stijn Vansummeren VL - 2 ER - TY - JOUR T1 - A Generic Parallel Processing Model for Facilitating Data Mining and Integration JF - Parallel Computing Y1 - 2011 A1 - Liangxiu Han A1 - Chee Sun Liew A1 - van Hemert, Jano A1 - Malcolm Atkinson KW - Data Mining and Data Integration (DMI) KW - Life Sciences KW - OGSA-DAI KW - Parallelism KW - Pipeline Streaming KW - workflow AB - To facilitate Data Mining and Integration (DMI) processes in a generic way, we investigate a parallel pipeline streaming model. We model a DMI task as a streaming data-flow graph: a directed acyclic graph (DAG) of Processing Elements PEs. The composition mechanism links PEs via data streams, which may be in memory, buffered via disks or inter-computer data-flows. This makes it possible to build arbitrary DAGs with pipelining and both data and task parallelisms, which provides room for performance enhancement. We have applied this approach to a real DMI case in the Life Sciences and implemented a prototype. To demonstrate feasibility of the modelled DMI task and assess the efficiency of the prototype, we have also built a performance evaluation model. The experimental evaluation results show that a linear speedup has been achieved with the increase of the number of distributed computing nodes in this case study. PB - Elsevier VL - 37 IS - 3 ER - TY - CONF T1 - Optimum Platform Selection and Configuration for Computational Jobs T2 - All Hands Meeting 2011 Y1 - 2011 A1 - Gary McGilvary A1 - Malcolm Atkinson A1 - Barker, Adam A1 - Ashley Lloyd AB - The performance and cost of many scientific applications which execute on a variety of High Performance Computing (HPC), local cluster environments and cloud services could be enhanced, and costs reduced if the platform was carefully selected on a per-application basis and the application itself was optimally configured for a given platform. With a wide-variety of computing platforms on offer, each possessing different properties, all too frequently platform decisions are made on an ad-hoc basis with limited ‘black-box’ information. The limitless number of possible application configurations also make it difficult for an individual who wants to achieve cost-effective results with the maximum performance available. Such individuals may include biomedical researchers analysing microarray data, software developers running aviation simulations or bankers performing risk assessments. However in either case, it is likely that many may not have the required knowledge to select the optimum platform and setup for their application; to do so, would require extensive knowledge of their applications and various platforms. In this paper we describe a framework that aims to resolve such issues by (i) reducing the detail required in the decision making process by placing this information within a selection framework, thereby (ii) maximising an application’s performance gain and/or reducing costs. We present a set of preliminary results where we compare the performance of running the Simple Parallel R INTerface (SPRINT) over a variety of platforms. SPRINT is a framework providing parallel functions of the statistical package R, allowing post genomic data to be easily analysed on HPC resources [1]. We run SPRINT on Amazon’s Elastic Compute Cloud (EC2) to compare the performance with the results obtained from HECToR, the UK’s National Supercomputing Service, and the Edinburgh Compute and Data Facilities (ECDF) cluster. JF - All Hands Meeting 2011 CY - York ER - TY - JOUR T1 - Performance database: capturing data for optimizing distributed streaming workflows JF - Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences Y1 - 2011 A1 - Chee Sun Liew A1 - Atkinson, Malcolm P. A1 - Radoslaw Ostrowski A1 - Murray Cole A1 - van Hemert, Jano I. A1 - Liangxiu Han KW - measurement framework KW - performance data KW - streaming workflows AB - The performance database (PDB) stores performance-related data gathered during workflow enactment. We argue that by carefully understanding and manipulating this data, we can improve efficiency when enacting workflows. This paper describes the rationale behind the PDB, and proposes a systematic way to implement it. The prototype is built as part of the Advanced Data Mining and Integration Research for Europe project. We use workflows from real-world experiments to demonstrate the usage of PDB. VL - 369 IS - 1949 ER - TY - CHAP T1 - Exploiting P2P and Grid Computing Technologies for Resource Sharing to support High Performance Distributed System T2 - Handbook of Research on P2P and Grid Systems for Service-Oriented Computing: Models, Methodologies and Applications Y1 - 2010 A1 - Liangxiu Han ED - Nick Antonopoulos ED - Georgios Exarchakos ED - Maozhen Li ED - Antonio Liottac JF - Handbook of Research on P2P and Grid Systems for Service-Oriented Computing: Models, Methodologies and Applications PB - IGI Global publishing VL - 1 ER - TY - Generic T1 - Federated Enactment of Workflow Patterns T2 - Lecture Notes in Computer Science Y1 - 2010 A1 - Yaikhom, Gagarine A1 - Liew, Chee A1 - Liangxiu Han A1 - van Hemert, Jano A1 - Malcolm Atkinson A1 - Krause, Amy ED - D’Ambra, Pasqua ED - Guarracino, Mario ED - Talia, Domenico AB - In this paper we address two research questions concerning workflows: 1) how do we abstract and catalogue recurring workflow patterns?; and 2) how do we facilitate optimisation of the mapping from workflow patterns to actual resources at runtime? Our aim here is to explore techniques that are applicable to large-scale workflow compositions, where the resources could change dynamically during the lifetime of an application. We achieve this by introducing a registry-based mechanism where pattern abstractions are catalogued and stored. In conjunction with an enactment engine, which communicates with this registry, concrete computational implementations and resources are assigned to these patterns, conditional to the execution parameters. Using a data mining application from the life sciences, we demonstrate this new approach. JF - Lecture Notes in Computer Science PB - Springer Berlin / Heidelberg VL - 6271 UR - http://dx.doi.org/10.1007/978-3-642-15277-1_31 N1 - 10.1007/978-3-642-15277-1_31 ER - TY - CONF T1 - The MoSGrid Gaussian Portlet – Technologies for the Implementation of Portlets for Molecular Simulations T2 - Proceedings of the International Workshop on Science Gateways (IWSG10) Y1 - 2010 A1 - Wewior, Martin A1 - Packschies, Lars A1 - Blunk, Dirk A1 - Wickeroth, D. A1 - Warzecha, Klaus A1 - Herres-Pawlis, Sonja A1 - Gesing, Sandra A1 - Breuers, Sebastian A1 - Krüger, Jens A1 - Birkenheuer, Georg A1 - Lang, Ulrich ED - Barbera, Roberto ED - Andronico, Giuseppe ED - La Rocca, Giuseppe JF - Proceedings of the International Workshop on Science Gateways (IWSG10) PB - Consorzio COMETA ER - TY - CONF T1 - Towards Optimising Distributed Data Streaming Graphs using Parallel Streams T2 - Data Intensive Distributed Computing (DIDC'10), in conjunction with the 19th International Symposium on High Performance Distributed Computing Y1 - 2010 A1 - Chee Sun Liew A1 - Atkinson, Malcolm P. A1 - van Hemert, Jano A1 - Liangxiu Han KW - Data-intensive Computing KW - Distributed Computing KW - Optimisation KW - Parallel Stream KW - Scientific Workflows AB - Modern scientific collaborations have opened up the opportunity of solving complex problems that involve multi- disciplinary expertise and large-scale computational experiments. These experiments usually involve large amounts of data that are located in distributed data repositories running various software systems, and managed by different organisations. A common strategy to make the experiments more manageable is executing the processing steps as a workflow. In this paper, we look into the implementation of fine-grained data-flow between computational elements in a scientific workflow as streams. We model the distributed computation as a directed acyclic graph where the nodes represent the processing elements that incrementally implement specific subtasks. The processing elements are connected in a pipelined streaming manner, which allows task executions to overlap. We further optimise the execution by splitting pipelines across processes and by introducing extra parallel streams. We identify performance metrics and design a measurement tool to evaluate each enactment. We conducted ex- periments to evaluate our optimisation strategies with a real world problem in the Life Sciences—EURExpress-II. The paper presents our distributed data-handling model, the optimisation and instrumentation strategies and the evaluation experiments. We demonstrate linear speed up and argue that this use of data-streaming to enable both overlapped pipeline and parallelised enactment is a generally applicable optimisation strategy. JF - Data Intensive Distributed Computing (DIDC'10), in conjunction with the 19th International Symposium on High Performance Distributed Computing PB - ACM CY - Chicago, Illinois UR - http://www.cct.lsu.edu/~kosar/didc10/index.php ER - TY - CONF T1 - Understanding TSP Difficulty by Learning from Evolved Instances T2 - Lecture Notes in Computer Science Y1 - 2010 A1 - Smith-Miles, Kate A1 - van Hemert, Jano A1 - Lim, Xin ED - Blum, Christian ED - Battiti, Roberto AB - Whether the goal is performance prediction, or insights into the relationships between algorithm performance and instance characteristics, a comprehensive set of meta-data from which relationships can be learned is needed. This paper provides a methodology to determine if the meta-data is sufficient, and demonstrates the critical role played by instance generation methods. Instances of the Travelling Salesman Problem (TSP) are evolved using an evolutionary algorithm to produce distinct classes of instances that are intentionally easy or hard for certain algorithms. A comprehensive set of features is used to characterise instances of the TSP, and the impact of these features on difficulty for each algorithm is analysed. Finally, performance predictions are achieved with high accuracy on unseen instances for predicting search effort as well as identifying the algorithm likely to perform best. JF - Lecture Notes in Computer Science PB - Springer Berlin / Heidelberg VL - 6073 UR - http://dx.doi.org/10.1007/978-3-642-13800-3_29 N1 - 10.1007/978-3-642-13800-3_29 ER - TY - CONF T1 - A Distributed Architecture for Data Mining and Integration T2 - Data-Aware Distributed Computing (DADC'09), in conjunction with the 18th International Symposium on High Performance Distributed Computing Y1 - 2009 A1 - Atkinson, Malcolm P. A1 - van Hemert, Jano A1 - Liangxiu Han A1 - Ally Hume A1 - Chee Sun Liew AB - This paper presents the rationale for a new architecture to support a significant increase in the scale of data integration and data mining. It proposes the composition into one framework of (1) data mining and (2) data access and integration. We name the combined activity “DMI”. It supports enactment of DMI processes across heterogeneous and distributed data resources and data mining services. It posits that a useful division can be made between the facilities established to support the definition of DMI processes and the computational infrastructure provided to enact DMI processes. Communication between those two divisions is restricted to requests submitted to gateway services in a canonical DMI language. Larger-scale processes are enabled by incremental refinement of DMI-process definitions often by recomposition of lower-level definitions. Autonomous types and descriptions which will support detection of inconsistencies and semi-automatic insertion of adaptations.These architectural ideas are being evaluated in a feasibility study that involves an application scenario and representatives of the community. JF - Data-Aware Distributed Computing (DADC'09), in conjunction with the 18th International Symposium on High Performance Distributed Computing PB - ACM ER - TY - JOUR T1 - A Strategy for Research and Innovation in the Century of Information JF - Prometheus Y1 - 2009 A1 - e-Science Directors’ Forum Strategy Working Group A1 - Atkinson, M. A1 - Britton, D. A1 - Coveney, P. A1 - De Roure, D A1 - Garnett, N. A1 - Geddes, N. A1 - Gurney, R. A1 - Haines, K. A1 - Hughes, L. A1 - Ingram, D. A1 - Jeffreys, P. A1 - Lyon, L. A1 - Osborne, I. A1 - Perrott, P. A1 - Procter. R. A1 - Rusbridge, C. AB - More data will be produced in the next five years than in the entire history of human kind, a digital deluge that marks the beginning of the Century of Information. Through a year‐long consultation with UK researchers, a coherent strategy has been developed, which will nurture Century‐of‐Information Research (CIR); it crystallises the ideas developed by the e‐Science Directors’ Forum Strategy Working Group. This paper is an abridged version of their latest report which can be found at: http://wikis.nesc.ac.uk/escienvoy/Century_of_Information_Research_Strategy which also records the consultation process and the affiliations of the authors. This document is derived from a paper presented at the Oxford e‐Research Conference 2008 and takes into account suggestions made in the ensuing panel discussion. The goals of the CIR Strategy are to facilitate the growth of UK research and innovation that is data and computationally intensive and to develop a new culture of ‘digital‐systems judgement’ that will equip research communities, businesses, government and society as a whole, with the skills essential to compete and prosper in the Century of Information. The CIR Strategy identifies a national requirement for a balanced programme of coordination, research, infrastructure, translational investment and education to empower UK researchers, industry, government and society. The Strategy is designed to deliver an environment which meets the needs of UK researchers so that they can respond agilely to challenges, can create knowledge and skills, and can lead new kinds of research. It is a call to action for those engaged in research, those providing data and computational facilities, those governing research and those shaping education policies. The ultimate aim is to help researchers strengthen the international competitiveness of the UK research base and increase its contribution to the economy. The objectives of the Strategy are to better enable UK researchers across all disciplines to contribute world‐leading fundamental research; to accelerate the translation of research into practice; and to develop improved capabilities, facilities and context for research and innovation. It envisages a culture that is better able to grasp the opportunities provided by the growing wealth of digital information. Computing has, of course, already become a fundamental tool in all research disciplines. The UK e‐Science programme (2001–06)—since emulated internationally—pioneered the invention and use of new research methods, and a new wave of innovations in digital‐information technologies which have enabled them. The Strategy argues that the UK must now harness and leverage its own, plus the now global, investment in digital‐information technology in order to spread the benefits as widely as possible in research, education, industry and government. Implementing the Strategy would deliver the computational infrastructure and its benefits as envisaged in the Science & Innovation Investment Framework 2004–2014 (July 2004), and in the reports developing those proposals. To achieve this, the Strategy proposes the following actions: 1. support the continuous innovation of digital‐information research methods; 2. provide easily used, pervasive and sustained e‐Infrastructure for all research; 3. enlarge the productive research community which exploits the new methods efficiently; 4. generate capacity, propagate knowledge and develop skills via new curricula; and 5. develop coordination mechanisms to improve the opportunities for interdisciplinary research and to make digital‐infrastructure provision more cost effective. To gain the best value for money strategic coordination is required across a broad spectrum of stakeholders. A coherent strategy is essential in order to establish and sustain the UK as an international leader of well‐curated national data assets and computational infrastructure, which is expertly used to shape policy, support decisions, empower researchers and to roll out the results to the wider benefit of society. The value of data as a foundation for wellbeing and a sustainable society must be appreciated; national resources must be more wisely directed to the collection, curation, discovery, widening access, analysis and exploitation of these data. Every researcher must be able to draw on skills, tools and computational resources to develop insights, test hypotheses and translate inventions into productive use, or to extract knowledge in support of governmental decision making. This foundation plus the skills developed will launch significant advances in research, in business, in professional practice and in government with many consequent benefits for UK citizens. The Strategy presented here addresses these complex and interlocking requirements. VL - 27 ER - TY - JOUR T1 - Distributed Computing Education, Part 3: The Winter School Online Experience JF - Distributed Systems Online Y1 - 2008 A1 - Low, B. A1 - Cassidy, K. A1 - Fergusson, D. A1 - Atkinson, M. A1 - Vander Meer, E. A1 - McGeever, M. AB - The International Summer Schools in Grid Computing (ISSGC) have provided numerous international students with the opportunity to learn grid systems, as detailed in part 2 of this series (http://doi.ieeecomputersociety.org/10.1109/MDSO.2008.20). The International Winter School on Grid Computing 2008 (IWSGC 08) followed the successful summer schools, opening up the ISSGC experience to a wider range of students because of its online format. The previous summer schools made it clear that many students found the registration and travel costs and the time requirements prohibitive. The EU FP6 ICEAGE project held the first winter school from 6 February to 12 March 2008. The winter school repurposed summer school materials and added resources such as the ICEAGE digital library and summer-school-tested t-Infrastructures such as GILDA (Grid INFN Laboratory for Dissemination Activities). The winter schools shared the goals of the summer school, which emphasized disseminating grid knowledge. The students act as multipliers, spreading the skills and knowledge they acquired at the winter school to their colleagues to build strong and enthusiastic local grid communities. PB - IEEE Computer Society VL - 9 UR - http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=4659260 IS - 9 ER - TY - JOUR T1 - Distributed Computing Education, Part 5: Coming to Terms with Intellectual Property Rights JF - Distributed Systems Online Y1 - 2008 A1 - Boon Low A1 - Kathryn Cassidy A1 - Fergusson, David A1 - Malcolm Atkinson A1 - Elizabeth Vander Meer A1 - Mags McGeever AB - In part 1 of this series on distributed computing education, we introduced a list of components important for teaching environments. We outlined the first three components, which included development of materials for education, education for educators and teaching infrastructures, identifying current practice, challenges, and opportunities for provision. The final component, a supportive policy framework that encourages cooperation and sharing, includes the need to manage intellectual property rights (IPR). PB - IEEE Computer Society VL - 9 UR - http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=4755177 IS - 12 ER - TY - JOUR T1 - A Grid infrastructure for parallel and interactive applications JF - Computing and Informatics Y1 - 2008 A1 - Gomes, J. A1 - Borges, B. A1 - Montecelo, M. A1 - David, M. A1 - Silva, B. A1 - Dias, N. A1 - Martins, JP A1 - Fernandez, C. A1 - Garcia-Tarres, L. , A1 - Veiga, C. A1 - Cordero, D. A1 - Lopez, J. A1 - J Marco A1 - Campos, I. A1 - Rodríguez, David A1 - Marco, R. A1 - Lopez, A. A1 - Orviz, P. A1 - Hammad, A. VL - 27 IS - 2 ER - TY - JOUR T1 - The interactive European Grid: Project objectives and achievements JF - Computing and Informatics Y1 - 2008 A1 - J Marco A1 - Campos, I. A1 - Coterillo, I. A1 - Diaz, I. A1 - Lopez, A. A1 - Marco, R. A1 - Martinez-Rivero, C. A1 - Orviz, P. A1 - Rodríguez, David A1 - Gomes, J. A1 - Borges, G. A1 - Montecelo, M. A1 - David, M. A1 - Silva, B. A1 - Dias, N. A1 - Martins, JP A1 - Fernandez, C. A1 - Garcia-Tarres, L. VL - 27 IS - 2 ER - TY - CONF T1 - Data Integration in eHealth: A Domain/Disease Specific Roadmap T2 - Studies in Health Technology and Informatics Y1 - 2007 A1 - Ure, J. A1 - Proctor, R. A1 - Martone, M. A1 - Porteous, D. A1 - Lloyd, S. A1 - Lawrie, S. A1 - Job, D. A1 - Baldock, R. A1 - Philp, A. A1 - Liewald, D. A1 - Rakebrand, F. A1 - Blaikie, A. A1 - McKay, C. A1 - Anderson, S. A1 - Ainsworth, J. A1 - van Hemert, J. A1 - Blanquer, I. A1 - Sinno ED - N. Jacq ED - Y. Legr{\'e} ED - H. Muller ED - I. Blanquer ED - V. Breton ED - D. Hausser ED - V. Hern{\'a}ndez ED - T. Solomonides ED - M. Hofman-Apitius KW - e-Science AB - The paper documents a series of data integration workshops held in 2006 at the UK National e-Science Centre, summarizing a range of the problem/solution scenarios in multi-site and multi-scale data integration with six HealthGrid projects using schizophrenia as a domain-specific test case. It outlines emerging strategies, recommendations and objectives for collaboration on shared ontology-building and harmonization of data for multi-site trials in this domain. JF - Studies in Health Technology and Informatics PB - IOPress VL - 126 SN - 978-1-58603-738-3 ER - TY - CONF T1 - Interaction as a Grounding for Peer to Peer Knowledge Sharing T2 - Advances in Web Semantics Y1 - 2007 A1 - Robertson, D. A1 - Walton, C. A1 - Barker, A. A1 - Besana, P. A1 - Chen-Burger, Y. A1 - Hassan, F. A1 - Lambert, D. A1 - Li, G. A1 - McGinnis, J A1 - Osman, N. A1 - Bundy, A. A1 - McNeill, F. A1 - van Harmelen, F. A1 - Sierra, C. A1 - Giunchiglia, F. JF - Advances in Web Semantics PB - LNCS-IFIP VL - 1 ER - TY - CONF T1 - Optimization and evaluation of parallel I/O in BIPS3D parallel irregular application T2 - IPDPS Y1 - 2007 A1 - Rosa Filgueira A1 - David E. Singh A1 - Florin Isaila A1 - Jesús Carretero A1 - Antonio Garcia Loureiro JF - IPDPS ER - TY - CHAP T1 - Knowledge and Data Management in Grids, CoreGRID T2 - Euro-Par'06 Proceedings of the CoreGRID 2006, UNICORE Summit 2006, Petascale Computational Biology and Bioinformatics conference on Parallel processing Y1 - 2006 A1 - Chue Hong, N. P. A1 - Antonioletti, M. A1 - Karasavvas, K. A. A1 - Atkinson, M. ED - Lehner, W. ED - Meyer, N. ED - Streit, A. ED - Stewart, C. JF - Euro-Par'06 Proceedings of the CoreGRID 2006, UNICORE Summit 2006, Petascale Computational Biology and Bioinformatics conference on Parallel processing T3 - Lecture Notes in Computer Science PB - Springer CY - Berlin, Germany VL - 4375 SN - 978-3-540-72226-7 UR - http://www.springer.com/computer/communication+networks/book/978-3-540-72226-7 ER - TY - CONF T1 - Complexity Transitions in Evolutionary Algorithms: Evaluating the impact of the initial population T2 - Proceedings of the Congress on Evolutionary Computation Y1 - 2005 A1 - Defaweux, A. A1 - Lenaerts, T. A1 - van Hemert, J. I. A1 - Parent, J. KW - constraint satisfaction KW - transition models AB - This paper proposes an evolutionary approach for the composition of solutions in an incremental way. The approach is based on the metaphor of transitions in complexity discussed in the context of evolutionary biology. Partially defined solutions interact and evolve into aggregations until a full solution for the problem at hand is found. The impact of the initial population on the outcome and the dynamics of the process is evaluated using the domain of binary constraint satisfaction problems. JF - Proceedings of the Congress on Evolutionary Computation PB - {IEEE} Press ER - TY - JOUR T1 - The design and implementation of Grid database services in OGSA-DAI JF - Concurrency - Practice and Experience Y1 - 2005 A1 - Antonioletti, Mario A1 - Atkinson, Malcolm P. A1 - Baxter, Robert M. A1 - Borley, Andrew A1 - Hong, Neil P. Chue A1 - Collins, Brian A1 - Hardman, Neil A1 - Hume, Alastair C. A1 - Knox, Alan A1 - Mike Jackson A1 - Krause, Amrey A1 - Laws, Simon A1 - Magowan, James A1 - Pato VL - 17 ER - TY - CONF T1 - The Digital Curation Centre: a vision for digital curation T2 - 2005 IEEE International Symposium on Mass Storage Systems and Technology Y1 - 2005 A1 - Rusbridge, C. A1 - P. Burnhill A1 - S. Ross A1 - P. Buneman A1 - D. Giaretta A1 - Lyon, L. A1 - Atkinson, M. AB - We describe the aims and aspirations for the Digital Curation Centre (DCC), the UK response to the realisation that digital information is both essential and fragile. We recognise the equivalence of preservation as "interoperability with the future", asserting that digital curation is concerned with "communication across time". We see the DCC as having relevance for present day data curation and for continuing data access for generations to come. We describe the structure and plans of the DCC, designed to support these aspirations and based on a view of world class research being developed into curation services, all of which are underpinned by outreach to the broadest community. JF - 2005 IEEE International Symposium on Mass Storage Systems and Technology PB - IEEE Computer Society CY - Sardinia, Italy SN - 0-7803-9228-0 ER - TY - CONF T1 - Evolutionary Transitions as a Metaphor for Evolutionary Optimization T2 - LNAI 3630 Y1 - 2005 A1 - Defaweux, A. A1 - Lenaerts, T. A1 - van Hemert, J. I. ED - M. Capcarrere ED - A. A. Freitas ED - P. J. Bentley ED - C. G. Johnson ED - J. Timmis KW - constraint satisfaction KW - transition models AB - This paper proposes a computational model for solving optimisation problems that mimics the principle of evolutionary transitions in individual complexity. More specifically it incorporates mechanisms for the emergence of increasingly complex individuals from the interaction of more simple ones. The biological principles for transition are outlined and mapped onto an evolutionary computation context. The class of binary constraint satisfaction problems is used to illustrate the transition mechanism. JF - LNAI 3630 PB - Springer-Verlag SN - 3-540-28848-1 ER - TY - Generic T1 - Experience with the international testbed in the crossgrid project T2 - Advances in Grid Computing-EGC 2005 Y1 - 2005 A1 - Gomes, J. A1 - David, M. A1 - Martins, J. A1 - Bernardo, L. A1 - A García A1 - Hardt, M. A1 - Kornmayer, H. A1 - Marco, Jesus A1 - Marco, Rafael A1 - Rodríguez, David A1 - Diaz, Irma A1 - Cano, Daniel A1 - Salt, J. A1 - Gonzalez, S. A1 - J Sánchez A1 - Fassi, F. A1 - Lara, V. A1 - Nyczyk, P. A1 - Lason, P. A1 - Ozieblo, A. A1 - Wolniewicz, P. A1 - Bluj, M. A1 - K Nawrocki A1 - A Padee A1 - W Wislicki ED - Peter M. A. Sloot, Alfons G. Hoekstra, Thierry Priol, Alexander Reinefeld ED - Marian Bubak JF - Advances in Grid Computing-EGC 2005 T3 - LNCS PB - Springer Berlin/Heidelberg CY - Amsterdam VL - 3470 ER - TY - CONF T1 - A New Architecture for OGSA-DAI T2 - UK e-Science All Hands Meeting Y1 - 2005 A1 - Atkinson, M. A1 - Karasavvas, K. A1 - Antonioletti, M. A1 - Baxter, R. A1 - Borley, A. A1 - Hong, N. C. A1 - Hume, A. A1 - Jackson, M. A1 - Krause, A. A1 - Laws, S. A1 - Paton, N. A1 - Schopf, J. A1 - Sugden, T. A1 - Tourlas, K. A1 - Watson, P. JF - UK e-Science All Hands Meeting ER - TY - CONF T1 - OGSA-DAI Status and Benchmarks T2 - All Hands Meeting 2005 Y1 - 2005 A1 - Antonioletti, Mario A1 - Malcolm Atkinson A1 - Rob Baxter A1 - Andrew Borle A1 - Hong, Neil P. Chue A1 - Patrick Dantressangle A1 - Hume, Alastair C. A1 - Mike Jackson A1 - Krause, Amy A1 - Laws, Simon A1 - Parsons, Mark A1 - Paton, Norman W. A1 - Jennifer M. Schopf A1 - Tom Sugden A1 - Watson, Paul AB - This paper presents a status report on some of the highlights that have taken place within the OGSADAI project since the last AHM. A description of Release 6.0 functionality and details of the forthcoming release, due in September 2005, is given. Future directions for this project are discussed. This paper also describes initial results of work being done to systematically benchmark recent OGSADAI releases. The OGSA-DAI software distribution, and more information about the project, is available from the project website at www.ogsadai.org.uk. JF - All Hands Meeting 2005 CY - Nottingham, UK ER - TY - CONF T1 - Organization of the International Testbed of the CrossGrid Project T2 - Cracow Grid Workshop 2005 Y1 - 2005 A1 - Gomes, J. A1 - David, M. A1 - Martins, J. A1 - Bernardo, L. A1 - Garcia, A. A1 - Hardt, M. A1 - Kornmayer, H. A1 - Marco, Rafael A1 - Rodríguez, David A1 - Diaz, Irma A1 - Cano, Daniel A1 - Salt, J. A1 - Gonzalez, S. A1 - Sanchez, J. A1 - Fassi, F. A1 - Lara, V. A1 - Nyczyk, P. A1 - Lason, P. A1 - Ozieblo, A. A1 - Wolniewicz, P. A1 - Bluj, M. JF - Cracow Grid Workshop 2005 ER - TY - CONF T1 - Transition Models as an incremental approach for problem solving in Evolutionary Algorithms T2 - Proceedings of the Genetic and Evolutionary Computation Conference Y1 - 2005 A1 - Defaweux, A. A1 - Lenaerts, T. A1 - van Hemert, J. I. A1 - Parent, J. ED - H.-G. Beyer ED - et al KW - constraint satisfaction KW - transition models AB - This paper proposes an incremental approach for building solutions using evolutionary computation. It presents a simple evolutionary model called a Transition model. It lets building units of a solution interact and then uses an evolutionary process to merge these units toward a full solution for the problem at hand. The paper provides a preliminary study on the evolutionary dynamics of this model as well as an empirical comparison with other evolutionary techniques on binary constraint satisfaction. JF - Proceedings of the Genetic and Evolutionary Computation Conference PB - {ACM} Press ER - TY - CONF T1 - Dynamic Routing Problems with Fruitful Regions: Models and Evolutionary Computation T2 - LNCS Y1 - 2004 A1 - van Hemert, J. I. A1 - la Poutré, J. A. ED - Xin Yao ED - Edmund Burke ED - Jose A. Lozano ED - Jim Smith ED - Juan J. Merelo-Guerv\'os ED - John A. Bullinaria ED - Jonathan Rowe ED - Peter Ti\v{n}o Ata Kab\'an ED - Hans-Paul Schwefel KW - dynamic problems KW - evolutionary computation KW - vehicle routing AB - We introduce the concept of fruitful regions in a dynamic routing context: regions that have a high potential of generating loads to be transported. The objective is to maximise the number of loads transported, while keeping to capacity and time constraints. Loads arrive while the problem is being solved, which makes it a real-time routing problem. The solver is a self-adaptive evolutionary algorithm that ensures feasible solutions at all times. We investigate under what conditions the exploration of fruitful regions improves the effectiveness of the evolutionary algorithm. JF - LNCS PB - Springer-Verlag CY - Birmingham, UK VL - 3242 SN - 3-540-23092-0 ER - TY - CONF T1 - OGSA-DAI Status Report and Future Directions T2 - All Hands Meeting 2004 Y1 - 2004 A1 - Antonioletti, Mario A1 - Malcolm Atkinson A1 - Rob Baxter A1 - Borley, Andrew A1 - Hong, Neil P. Chue A1 - Collins, Brian A1 - Jonathan Davies A1 - Desmond Fitzgerald A1 - Hardman, Neil A1 - Hume, Alastair C. A1 - Mike Jackson A1 - Krause, Amrey A1 - Laws, Simon A1 - Paton, Norman W. A1 - Tom Sugden A1 - Watson, Paul A1 - Mar AB - Data Access and Integration (DAI) of data resources, such as relational and XML databases, within a Grid context. Project members also participate in the development of DAI standards through the GGF DAIS WG. The standards that emerge through this effort will be adopted by OGSA-DAI once they have stabilised. The OGSA-DAI developers are also engaging with a growing user community to gather their data and functionality requirements. Several large projects are already using OGSA-DAI to provide their DAI capabilities. This paper presents a status report on OGSA-DAI activities since the last AHM and announces future directions. The OGSA-DAI software distribution and more information about the project is available from the project website at http://www.ogsadai.org.uk/. JF - All Hands Meeting 2004 CY - Nottingham, UK ER - TY - CONF T1 - OGSA-DAI: Two Years On T2 - GGF10 Y1 - 2004 A1 - Antonioletti, Mario A1 - Malcolm Atkinson A1 - Rob Baxter A1 - Borley, Andrew A1 - Neil Chue Hong A1 - Collins, Brian A1 - Jonathan Davies A1 - Hardman, Neil A1 - George Hicken A1 - Ally Hume A1 - Mike Jackson A1 - Krause, Amrey A1 - Laws, Simon A1 - Magowan, James A1 - Jeremy Nowell A1 - Paton, Norman W. A1 - Dave Pearson A1 - To AB - The OGSA-DAI project has been producing Grid-enabled middleware for almost two years now, providing data access and integration capabilities to data resources, such as databases, within an OGSA context. In these two years, OGSA-DAI has been tracking rapidly evolving standards, managing changes in software dependencies, contributing to the standardisation process and liasing with a growing user community together with their associated data requirements. This process has imparted important lessons and raised a number of issues that need to be addressed if a middleware product is to be widely adopted. This paper examines the experiences of OGSA-DAI in implementing proposed standards, the likely impact that the still-evolving standards landscape will have on future implementations and how these affect uptake of the software. The paper also examines the gathering of requirements from and engagement with the Grid community, the difficulties of defining a process for the management and publishing of metadata, and whether relevant standards can be implemented in an efficient manner. The OGSA-DAI software distribution and more details about the project are available from the project Web site at http://www.ogsadai.org.uk/. JF - GGF10 CY - Berlin, Germany ER - TY - CONF T1 - Phase transition properties of clustered travelling salesman problem instances generated with evolutionary computation T2 - LNCS Y1 - 2004 A1 - van Hemert, J. I. A1 - Urquhart, N. B. ED - Xin Yao ED - Edmund Burke ED - Jose A. Lozano ED - Jim Smith ED - Juan J. Merelo-Guerv\'os ED - John A. Bullinaria ED - Jonathan Rowe ED - Peter Ti\v{n}o Ata Kab\'an ED - Hans-Paul Schwefel KW - evolutionary computation KW - problem evolving KW - travelling salesman AB - This paper introduces a generator that creates problem instances for the Euclidean symmetric travelling salesman problem. To fit real world problems, we look at maps consisting of clustered nodes. Uniform random sampling methods do not result in maps where the nodes are spread out to form identifiable clusters. To improve upon this, we propose an evolutionary algorithm that uses the layout of nodes on a map as its genotype. By optimising the spread until a set of constraints is satisfied, we are able to produce better clustered maps, in a more robust way. When varying the number of clusters in these maps and, when solving the Euclidean symmetric travelling salesman person using Chained Lin-Kernighan, we observe a phase transition in the form of an easy-hard-easy pattern. JF - LNCS PB - Springer-Verlag CY - Birmingham, UK VL - 3242 SN - 3-540-23092-0 UR - http://www.vanhemert.co.uk/files/clustered-phase-transition-tsp.tar.gz ER - TY - JOUR T1 - The Research of Relationship between Self-similar of TCP and Network Performance JF - Journal on communications Y1 - 2004 A1 - yan Liu A1 - Liangxiu Han VL - 25 IS - 4 ER - TY - CONF T1 - Dependable Grid Services T2 - UK e-Science All Hands Meeting 2003, 2-4th September, Nottingham, UK Y1 - 2003 A1 - Stuart Anderson A1 - Yin Chen A1 - Glen Dobson A1 - Stephen Hall A1 - Conrad Hughes A1 - Yong Li A1 - Sheng Qu A1 - Ed Smith A1 - Ian Sommerville A1 - Ma Tiejun ED - Proceedings of UK e-Science All Hands Meeting 2003 AB - The provision of dependable computer systems by deploying diverse, redundant components in order to mask or provide recovery from component failures has mostly been restricted to systems with very high criticality. In this paper we present an architecture and prototype implementation of an approach to providing such redundancy at low cost in service-based infrastructures. In particular we consider services that are supplied by composing a number of component services and consider how service discovery, automatic monitoring and failure detection have the potential to create composed services that are more dependable than might be possible using a straightforward approach. The work is still in its early stages and so far no evaluation of the approach has been carried out. JF - UK e-Science All Hands Meeting 2003, 2-4th September, Nottingham, UK CY - Nottingham, UK ER - TY - CONF T1 - The Design and Implementation of Grid Database Services in OGSA-DAI T2 - All Hands Meeting 2003 Y1 - 2003 A1 - Ali Anjomshoaa A1 - Antonioletti, Mario A1 - Malcolm Atkinson A1 - Rob Baxter A1 - Borley, Andrew A1 - Hong, Neil P. Chue A1 - Collins, Brian A1 - Hardman, Neil A1 - George Hicken A1 - Ally Hume A1 - Knox, Alan A1 - Mike Jackson A1 - Krause, Amrey A1 - Laws, Simon A1 - Magowan, James A1 - Charaka Palansuriya A1 - Paton, Norman W. AB - This paper presents a high-level overview of the design and implementation of the core components of the OGSA-DAI project. It describes the design decisions made, the project’s interaction with the Data Access and Integration Working Group of the Global Grid Forum and provides an overview of implementation characteristics. Further details of the implementation are provided in the extensive documentation available from the project web site. JF - All Hands Meeting 2003 CY - Nottingham, UK ER - TY - JOUR T1 - A network traffic model based on the cascade process JF - Journal of Mini-Micro Computer System Y1 - 2002 A1 - Liangxiu Han A1 - yan Liu A1 - Zhiwei Cen VL - 23 IS - 12 ER - TY - CONF T1 - Adaptive Genetic Programming Applied to New and Existing Simple Regression Problems T2 - Springer Lecture Notes on Computer Science Y1 - 2001 A1 - Eggermont, J. A1 - van Hemert, J. I. ED - J. Miller ED - Tomassini, M. ED - P. L. Lanzi ED - C. Ryan ED - A. G. B. Tettamanzi ED - W. B. Langdon KW - data mining AB - In this paper we continue our study on adaptive genetic pro-gramming. We use Stepwise Adaptation of Weights to boost performance of a genetic programming algorithm on simple symbolic regression problems. We measure the performance of a standard GP and two variants of SAW extensions on two different symbolic regression prob-lems from literature. Also, we propose a model for randomly generating polynomials which we then use to further test all three GP variants. JF - Springer Lecture Notes on Computer Science PB - Springer-Verlag, Berlin SN - 9-783540-418993 ER - TY - CONF T1 - An Engineering Approach to Evolutionary Art T2 - Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001) Y1 - 2001 A1 - van Hemert, J. I. A1 - Jansen, M. L. M. ED - Lee Spector ED - Erik D. Goodman ED - Annie Wu ED - W. B. Langdon ED - Hans-Michael Voigt ED - Mitsuo Gen ED - Sandip Sen ED - Marco Dorigo ED - Shahram Pezeshk ED - Max H. Garzon ED - Edmund Burke KW - evolutionary art AB - We present a general system that evolves art on the Internet. The system runs on a server which enables it to collect information about its usage world wide; its core uses operators and representations from genetic program-ming. We show two types of art that can be evolved using this general system. JF - Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2001) PB - Morgan Kaufmann Publishers, San Francisco ER - TY - CONF T1 - Evolutionary Computation in Constraint Satisfaction and Machine Learning --- An abstract of my PhD. T2 - Proceedings of the Brussels Evolutionary Algorithms Day (BEAD-2001) Y1 - 2001 A1 - van Hemert, J. I. ED - Anne Defaweux ED - Bernard Manderick ED - Tom Lenearts ED - Johan Parent ED - Piet van Remortel KW - constraint satisfaction KW - data mining JF - Proceedings of the Brussels Evolutionary Algorithms Day (BEAD-2001) PB - Vrije Universiteit Brussel (VUB) ER - TY - CONF T1 - A ``Futurist'' approach to dynamic environments T2 - Proceedings of the Workshops at the Genetic and Evolutionary Computation Conference, Dynamic Optimization Problems Y1 - 2001 A1 - van Hemert, J. I. A1 - Van Hoyweghen, C. A1 - Lukschandl, E. A1 - Verbeeck, K. ED - J. Branke ED - Th. B{\"a}ck KW - dynamic problems AB - The optimization of dynamic environments has proved to be a difficult area for Evolutionary Algorithms. As standard haploid populations find it difficult to track a moving target, diffKerent schemes have been suggested to improve the situation. We study a novel approach by making use of a meta learner which tries to predict the next state of the environment, i.e. the next value of the goal the individuals have to achieve, by making use of the accumulated knowledge from past performance. JF - Proceedings of the Workshops at the Genetic and Evolutionary Computation Conference, Dynamic Optimization Problems PB - Morgan Kaufmann Publishers, San Francisco ER - TY - CONF T1 - Adapting the Fitness Function in GP for Data Mining T2 - Springer Lecture Notes on Computer Science Y1 - 1999 A1 - Eggermont, J. A1 - Eiben, A. E. A1 - van Hemert, J. I. ED - R. Poli ED - P. Nordin ED - W. B. Langdon ED - T. C. Fogarty KW - data mining KW - genetic programming AB - In this paper we describe how the Stepwise Adaptation of Weights (SAW) technique can be applied in genetic programming. The SAW-ing mechanism has been originally developed for and successfully used in EAs for constraint satisfaction problems. Here we identify the very basic underlying ideas behind SAW-ing and point out how it can be used for different types of problems. In particular, SAW-ing is well suited for data mining tasks where the fitness of a candidate solution is composed by `local scores' on data records. We evaluate the power of the SAW-ing mechanism on a number of benchmark classification data sets. The results indicate that extending the GP with the SAW-ing feature increases its performance when different types of misclassifications are not weighted differently, but leads to worse results when they are. JF - Springer Lecture Notes on Computer Science PB - Springer-Verlag, Berlin SN - 3-540-65899-8 ER - TY - JOUR T1 - Neural network applied to the prediction of the failure stress for pressurized cylinders containing defects JF - International Journal of Pressure Vessels and Piping Y1 - 1999 A1 - Lianghao Han A1 - Liangxiu Han A1 - zengdian Liu PB - Elsevier VL - 76 IS - 4 ER -