name
stringlengths
7
10
title
stringlengths
13
125
abstract
stringlengths
67
3.02k
fulltext
stringclasses
1 value
keywords
stringlengths
17
734
train_1183
Evolving robust asynchronous cellular automata for the density task
In this paper the evolution of three kinds of asynchronous cellular automata are studied for the density task. Results are compared with those obtained for synchronous automata and the influence of various asynchronous update policies on the computational strategy is described. How synchronous and asynchronous cellular...
asynchronous cellular automata;synchronous automata;discrete dynamical systems;random noise;cellular automata;fault tolerance
train_1184
Measuring return: revealing ROI
The most critical part of the return-on-investment odyssey is to develop metrics that matter to the business and to measure systems in terms of their ability to help achieve those business goals. Everything must flow from those key metrics. And don't forget to revisit those every now and then, too. Since all systems wi...
key metrics;roi;automation investment;technology purchases;return-on-investment
train_1185
Trading exchanges: online marketplaces evolve
Looks at how trading exchanges are evolving rapidly to help manufacturers keep up with customer demand
supply chain management;manufacturers;online marketplaces;customer demand;xml standards;enterprise platforms;core software platform;private exchanges;integration technology;middleware;enterprise resource planning;trading exchanges;content management capabilities
train_1186
Implementing: it's all about processes
Looks at how the key to successful technology deployment can be found in a set of four basic disciplines
implementation;manufacturers;third-party integration;vendor-supplied hardware integration services;technology deployment;incremental targets;vendor-supplied software integration services
train_1187
Ethernet networks: getting down to business
While it seems pretty clear that Ethernet has won the battle for the mindshare as the network of choice for the factory floor, there's still a war to be won in implementation as cutting-edge manufacturers begin to adopt the technology on a widespread basis
ethernet;supervisory level;cutting-edge manufacturers;factory floor
train_1188
It's time to buy
There is an upside to a down economy: over-zealous suppliers are willing to make deals that were unthinkable a few years ago. That's because vendors are experiencing the same money squeeze as manufacturers, which makes the year 2002 the perfect time to invest in new technology. The author states that when negotiating t...
exit strategy;money squeeze;bargaining power;vendor;buyers market;shared risk;negotiation;suppliers;unexpected costs
train_1189
CRM: approaching zenith
Looks at how manufacturers are starting to warm up to the concept of customer relationship management. CRM has matured into what is expected to be big business. As CRM software evolves to its second, some say third, generation, it's likely to be more valuable to holdouts in manufacturing and other sectors
manufacturers;customer relationship management;manufacturing;crm
train_119
JPEG2000: standard for interactive imaging
JPEG2000 is the latest image compression standard to emerge from the Joint Photographic Experts Group (JPEG) working under the auspices of the International Standards Organization. Although the new standard does offer superior compression performance to JPEG, JPEG2000 provides a whole new way of interacting with compre...
image compression;client-server systems;international standards organization;joint photographic experts group;interoperable compression;review;interactive imaging;scalable compression;jpeg2000
train_1190
Buying into the relationship [business software]
Choosing the right software to improve business processes can have a huge impact on a company's efficiency and profitability. While it is sometimes hard to get beyond vendor hype about software features and functionality and know what to realistically expect, it is even more difficult to determine if the vendor is the ...
business software;software evaluation;management;functionality;vendor relationship
train_1191
On the monotonicity conservation in numerical solutions of the heat equation
It is important to choose such numerical methods in practice that mirror the characteristic properties of the described process beyond the stability and convergence. The investigated qualitative property in this paper is the conservation of the monotonicity in space of the initial heat distribution. We prove some state...
monotonicity conservation;characteristic properties;qualitative property;necessary and sufficient condition;heat equation;one-step vector-iterations;numerical solutions
train_1192
Construction of two-sided bounds for initial-boundary value problems
This paper extends the bounding operator approach developed for boundary value problems to the case of initial-boundary value problems (IBVPs). Following the general principle of bounding operators enclosing methods for the case of partial differential equations are discussed. In particular, continuous discretization m...
partial differential equations;two-sided bounds;bounding operators;parabolic problems;bounding operator approach;initial-boundary value problems
train_1193
Operator splitting and approximate factorization for taxis-diffusion-reaction
models In this paper we consider the numerical solution of 2D systems of certain types of taxis-diffusion-reaction equations from mathematical biology. By spatial discretization these PDE systems are approximated by systems of positive, nonlinear ODEs (Method of Lines). The aim of this paper is to examine the numerical...
numerical integration;mathematical biology;approximate matrix factorization;taxis-diffusion-reaction models;spatial discretization;runge-kutta methods;numerical solution;nonlinear odes;approximate factorization;pde systems;operator splitting;linearly implicit runge-kutta-rosenbrock methods
train_1194
New methods for oscillatory problems based on classical codes
The numerical integration of differential equations with oscillatory solutions is a very common problem in many fields of the applied sciences. Some methods have been specially devised for this kind of problem. In most of them, the calculation of the coefficients needs more computational effort than the classical codes...
numerical integration;oscillatory problems;oscillatory solutions;perturbed oscillators;classical codes;differential equations
train_1195
Sharpening the estimate of the stability constant in the maximum-norm of the
Crank-Nicolson scheme for the one-dimensional heat equation This paper is concerned with the stability constant C/sub infinity / in the maximum-norm of the Crank-Nicolson scheme applied. to the one-dimensional heat equation. A well known result due to S.J. Serdyukova is that C/sub infinity / < 23. In the present paper,...
one-dimensional heat equation;cauchy formula;discrete laplacian;dirichlet boundary conditions;stability constant;sharp resolvent estimate;crank-nicolson scheme;neumann boundary conditions
train_1196
Multiple shooting using a dichotomically stable integrator for solving
differential-algebraic equations In previous work by the first author, it has been established that a dichotomically stable discretization is needed when solving a stiff boundary-value problem in ordinary differential equations (ODEs), when sharp boundary layers may occur at each end of the interval. A dichotomically s...
implicit runge-kutta method;stiff boundary-value problem;ordinary differential equations;lobatto iiia formulae;initial-value integrator;differential-algebraic equations;dichotomically stable integrator;multiple shooting
train_1197
Numerical behaviour of stable and unstable solitary waves
In this paper we analyse the behaviour in time of the numerical approximations to solitary wave solutions of the generalized Benjamin-Bona-Mahony equation. This equation possesses an important property: the stability of these solutions depends on their velocity. We identify the error propagation mechanisms in both the ...
numerical methods;stable solitary waves;numerical approximations;numerical behaviour;unstable solitary waves;error propagation mechanisms;generalized benjamin-bona-mahony equation
train_1198
Post-projected Runge-Kutta methods for index-2 differential-algebraic equations
A new projection technique for Runge-Kutta methods applied to index-2 differential-algebraic equations is presented in which the numerical approximation is projected only as part of the output process. It is shown that for methods that are strictly stable at infinity, the order of convergence is unaffected compared to ...
post-projected runge-kutta methods;index-2 differential-algebraic equations;order of convergence;projected methods;numerical approximation
train_1199
Quasi stage order conditions for SDIRK methods
The stage order condition is a simplifying assumption that reduces the number of order conditions to be fulfilled when designing a Runge-Kutta (RK) method. Because a DIRK (diagonally implicit RK) method cannot have stage order greater than 1, we introduce quasi stage order conditions and derive some of their properties...
quasi stage order conditions;sdirk methods;differential-algebraic systems;numerical tests;embedded error estimator;diagonally implicit runge-kutta method
train_12
National learning systems: a new approach on technological change in late
industrializing economies and evidences from the cases of Brazil and South Korea The paper has two intertwined parts. The first one is a proposal for a conceptual and theoretical framework to understand technical change in late industrializing economies. The second part develops a kind of empirical test of the usefulne...
national learning systems;brazil;late industrializing economies;technological change;national innovation system;south korea
train_120
Self-organized critical traffic in parallel computer networks
In a recent paper, we analysed the dynamics of traffic flow in a simple, square lattice architecture. It was shown that a phase transition takes place between a free and a congested phase. The transition point was shown to exhibit optimal information transfer and wide fluctuations in time, with scale-free properties. I...
internet traffic;parallel computer networks;wide fluctuations;phase transition;scale-free properties;packet release;square lattice architecture;statistical features;self-organized critical traffic;congested phase;optimal information transfer;traffic flow dynamics;packet emission;free phase;transition point;generalizati...
train_1200
From continuous recovery to discrete filtering in numerical approximations of
conservation laws Modern numerical approximations of conservation laws rely on numerical dissipation as a means of stabilization. The older, alternative approach is the use of central differencing with a dose of artificial dissipation. In this paper we review the successful class of weighted essentially non-oscillatory...
conservation laws;continuous recovery;discrete filter operation;image processing;numerical approximations;central differencing;artificial dissipation;numerical dissipation;finite volume schemes;discrete filtering;highly nonlinear artificial dissipation terms
train_1201
Moving into the mainstream [product lifecycle management]
Product lifecycle management (PLM) is widely recognised by most manufacturing companies, as manufacturers begin to identify and implement targeted projects intended to deliver return-on investment in a timely fashion. Vendors are also releasing second-generation PLM products that are packaged, out-of-the-box solutions
product data management;product lifecycle management;manufacturing companies;enterprise resource planning;product development
train_1202
More than the money [software project]
Experiences creating budgets for large software projects have taught manufacturers that it is not about the money - it is about what one really needs. Before a company can begin to build a budget for a software. project, it has to have a good understanding of what business issues need to be addressed and what the busin...
budgeting;management;software requirements;software projects;manufacturing industry
train_1203
Technology decisions 2002
The paper looks at the critical hardware, software, and services choices manufacturers are making as they begin to emerge from the recession and position themselves for the future
manufacturing industries;customer relationship management;information technology;management of change;enterprise resource planning;services choices
train_1204
Design and prototype of a performance tool interface for OpenMP
This paper proposes a performance tools interface for OpenMP, similar in spirit to the MPI profiling interface in its intent to define a clear and portable API that makes OpenMP execution events visible to runtime performance tools. We present our design using a source-level instrumentation approach based on OpenMP dir...
opari;openmp directive rewriting;source-to-source translation tool;expert automatic event trace analyzer;mpi profiling interface;directive semantics;performance tool interface;source-level instrumentation approach;arbitrary code regions;tau performance analysis framework;parallel programming;api;performance libraries
train_1205
HPCVIEW: a tool for top-down analysis of node performance
It is increasingly difficult for complex scientific programs to attain a significant fraction of peak performance on systems that are based on microprocessors with substantial instruction-level parallelism and deep memory hierarchies. Despite this trend, performance analysis and tuning tools are still not used regularl...
peak performance;binary analysis;software tools;instruction-level parallelism;performance analysis;data layout;commodity web browser;source code;deep memory hierarchies;complex scientific programs;optimizing compilers;node performance;hpcview;top-down analysis
train_1206
The MAGNeT toolkit: design, implementation and evaluation
The current trend in constructing high-performance computing systems is to connect a large number of machines via a fast interconnect or a large-scale network such as the Internet. This approach relies on the performance of the interconnect (or Internet) to enable fast, large-scale distributed computing. A detailed und...
internet;interconnects;virtual supercomputing;traffic characterization;optimized networks;monitor for application-generated network traffic toolkit;high-performance computing;magnet;computational grids;high-performance computing systems;network protocol
train_1207
Packet spacing: an enabling mechanism for delivering multimedia content in
computational grids Streaming multimedia with UDP has become increasingly popular over distributed systems like the Internet. Scientific applications that stream multimedia include remote computational steering of visualization data and video-on-demand teleconferencing over the Access Grid. However, UDP does not posses...
udp;internet;transport protocols;inter-packet spacing;streaming multimedia;udp-based streaming;distributed systems;remote computational steering;visualization data;network protocol
train_1208
A Virtual Test Facility for the simulation of dynamic response in materials
The Center for Simulating Dynamic Response of Materials at the California Institute of Technology is constructing a virtual shock physics facility for studying the response of various target materials to very strong shocks. The Virtual Test Facility (VTF) is an end-to-end, fully three-dimensional simulation of the deto...
problem solving environment;compressible turbulence;python scripting language;shock physics simulation;pressure loading;parallel fluid solver;pyre;parallel solid mechanics;solid material response;virtual test facility;data visualization;high explosives;virtual shock physics facility;shock wave propagation
train_1209
High-level language support for user-defined reductions
The optimized handling of reductions on parallel supercomputers or clusters of workstations is critical to high performance because reductions are common in scientific codes and a potential source of bottlenecks. Yet in many high-level languages, a mechanism for writing efficient reductions remains surprisingly absent....
clusters of workstations;reductions;parallel supercomputers;language construct;parallel programming;scientific computing
train_121
Formula-dependent equivalence for compositional CTL model checking
We present a polytime computable state equivalence that is defined with respect to a given CTL formula. Since it does not attempt to preserve all CTL formulas, like bisimulation does, we can expect to compute coarser equivalences. This equivalence can be used to reduce the complexity of model checking a system of inter...
interacting fsm;formula-dependent equivalence;computation tree logic;ctl model checking;automatic method;compositional minimization;formal design verification;complexity reduction;coarse equivalence;ctl formula;polytime computable state equivalence
train_1210
Adaptive optimizing compilers for the 21st century
Historically, compilers have operated by applying a fixed set of optimizations in a predetermined order. We call such an ordered list of optimizations a compilation sequence. This paper describes a prototype system that uses biased random search to discover a program-specific compilation sequence that minimizes an expl...
compilers;configurable compilers;compilation sequence;biased random search;optimizing compilers;adaptive compiler;optimizations
train_1211
Hybrid decision tree
In this paper, a hybrid learning approach named hybrid decision tree (HDT) is proposed. HDT simulates human reasoning by using symbolic learning to do qualitative analysis and using neural learning to do subsequent quantitative analysis. It generates the trunk of a binary HDT according to the binary information gain ra...
data sets;hypothesis-driven constructive induction;storage requirements;neural learning;incremental learning;fannc;hybrid learning approach;qualitative analysis;quantitative analysis;reasoning;symbolic learning;feedforward neural network;hybrid decision tree;binary information gain ratio criterion
train_1212
TCRM: diagnosing tuple inconsistency for granulized datasets
Many approaches to granularization have been presented for knowledge discovery. However, the inconsistent tuples that exist in granulized datasets are hardly ever revealed. We developed a model, tuple consistency recognition model (TCRM) to help efficiently detect inconsistent tuples for datasets that are granulized. T...
tuple inconsistency;processing time;relational database;granularization;granulized datasets;large database;knowledge discovery;largest binning size algorithm;tcrm;sql;equal width interval technique;s-plus histogram binning algorithm;tuple consistency recognition model
train_1213
A knowledge intensive multi-agent framework for cooperative/collaborative
design modeling and decision support of assemblies Multi-agent modeling has emerged as a promising discipline for dealing with the decision making process in distributed information system applications. One of such applications is the modeling of distributed design or manufacturing processes which can link up various d...
agent negotiation;knowledge intensive multi-agent framework;collaborative design modeling;concurrent intelligent design;distributed information system applications;cooperative framework;knowledge intensive petri nets;design for assembly;knowledge object cooperation;assembly simulation;ai protocol;assembly planning;dist...
train_1214
Multi-agent collaboration for B2B workflow monitoring
Business-to-business (B2B) application environments are exceedingly dynamic and competitive. This dynamism is manifested in the form of changing process requirements and time constraints. However, current workflow management technologies have difficulties trying to solve problems, such as: how to deal with the dynamic ...
business rules;internet;electronic commerce;b2b workflow monitoring;workflow management;changing process requirements;business-to-business applications;time constraints;transaction risk;multi-agent collaboration
train_1215
A knowledge-based approach for business process reengineering, SHAMASH
We present an overview of SHAMASH, a process modelling tool for business process reengineering. The main features that differentiate it from most current related tools are its ability to define and use organisation standards, functional structure, and develop automatic model simulation and optimisation. SHAMASH is a kn...
process modelling tool;automatic model simulation;knowledge acquisition;knowledge-based approach;shamash;conceptual model;business process reengineering;knowledge-based system;functional structure;optimisation;organisation standards
train_1216
Knowledge flow management for distributed team software development
Cognitive cooperation is often neglected in current team software development processes. This issue becomes more important than ever when team members are globally distributed. This paper presents a notion of knowledge flow and the related management mechanism for realizing an ordered knowledge sharing and cognitive co...
knowledge grid platform;internet;knowledge flow management;software development management;knowledge flow representation;workflow process;cognitive cooperation;ordered knowledge sharing;cooperative work;distributed team software development
train_1217
A knowledge-based approach for managing urban infrastructures
This paper presents a knowledge e-based approach dedicated to the efficient management, regulation, interactive and dynamic monitoring of urban infrastructures. This approach identifies the data and related treatments common to several municipal activities and defines the requirements and functionalities of the compute...
intelligent decision support system;global planning system;sigiu;municipal activities;regulation;video;multi-agent systems;mobile agent;sygec;knowledge-based approach;urban infrastructure management;dynamic monitoring;real-time management;alarm control agent;sydex;cooperative system;coordination system;integrated opera...
train_1218
Knowledge acquisition for expert systems in accounting and financial problem
domains Since the mid-1980s, expert systems have been developed for a variety of problems in accounting and finance. The most commonly cited problems in developing these systems are the unavailability of the experts and knowledge engineers and difficulties with the rule extraction process. Within the field of artificia...
problem domain taxonomy;finance;knowledge acquisition;expert systems;artificial intelligence;accounting;rule extraction process
train_1219
Knowledge organisation of product design blackboard systems via graph
decomposition Knowledge organisation plays an important role in building a knowledge-based product design blackboard system. Well-organised knowledge sources will facilitate the effectiveness and efficiency of communication and data exchange in a blackboard system. In a previous investigation, an approach for construct...
case study;knowledge-based product design;product design blackboard systems;knowledge organisation;graph decomposition;data exchange
train_122
A formal framework for viewpoint consistency
Multiple viewpoint models of system development are becoming increasingly important. Each viewpoint offers a different perspective on the target system and system development involves parallel refinement of the multiple views. Viewpoint related approaches have been considered in a number of different guises by a spectr...
viewpoint consistency;open distributed processing;formal description technique;iso/itu standardisation framework;system development;odp;consistency checking;development models;formal framework;lotos;multiple viewpoint models;process algebra
train_1220
Modeling discourse in collaborative work support systems: a knowledge
representation and configuration perspective Collaborative work processes usually raise a lot of intricate debates and negotiations among participants, whereas conflicts of interest are inevitable and support for achieving consensus and compromise is required. Individual contributions, brought up by parties with differ...
human-computer interaction;compromise;consensus;conflicts of interest;hermes;knowledge representation;knowledge communication;collaborative work support systems;discourse modeling
train_1221
An approach to developing computational supports for reciprocal tutoring
This study presents a novel approach to developing computational supports for reciprocal tutoring. Reciprocal tutoring is a collaborative learning activity, where two participants take turns to play the role of a tutor and a tutee. The computational supports include scaffolding tools for the tutor and a computer-simula...
collaborative learning;computer-simulated virtual participant;intelligent tutoring system;system architecture;reciprocal tutoring computational support;scaffolding tools
train_1222
Mining the optimal class association rule set
We define an optimal class association rule set to be the minimum rule set with the same predictive power of the complete class association rule set. Using this rule set instead of the complete class association rule set we can avoid redundant computation that would otherwise be required for mining predictive associati...
dense databases;data mining;experimental results;upward closure property;predictive association rules;optimal class association rule set mining;predictive power;relational database;redundant computation;weak rule pruning;minimum rule set
train_1223
Formalising optimal feature weight setting in case based diagnosis as linear
programming problems Many approaches to case based reasoning (CBR) exploit feature weight setting algorithms to reduce the sensitivity to distance functions. We demonstrate that optimal feature weight setting in a special kind of CBR problems can be formalised as linear programming problems. Therefore, the optimal weig...
classification;heuristics;linear programming;searching;polynomial time;case based diagnosis;optimal feature weight setting;case based reasoning;distance functions;exponential weight space
train_1224
Formalization of weighted factors analysis
Weighted factors analysis (WeFA) has been proposed as a new approach for elicitation, representation, and manipulation of knowledge about a given problem, generally at a high and strategic level. Central to this proposal is that a group of experts in the area of the problem can identify a hierarchy of factors with posi...
organization;weighted factors analysis;significance;wefa graph;directed arc;logic-based argumentation;knowledge manipulation;reasoning;knowledge representation;knowledge management;knowledge elicitation;decision-support;profitability;binary wefa;directed weighted graph
train_1225
BT voices its support for IP
BTexact's chief technology officer, Mick Reeve, gives his views on the future for voice over DSL services and virtual private networks, and defends the slow rollout of public access WLANs
virtual private networks;btexact;public access wlans;voice over dsl
train_1226
Temp IT chief rallies troops [Mori]
The appointment of a highly qualified interim IT manager enabled market research company Mori to rapidly restructure its IT department. Now the resulting improvements are allowing it to support an increasing role for technology in the assimilation and analysis of market research
interim it manager;mori;market research company
train_1227
Will new Palms win laurels.?
PalmSource's latest operating system for mobile devices harnesses the ARM architecture to support more powerful business software, but there are concerns over compatibility with older applications
compatibility;palmsource;arm architecture;mobile devices;palm os 5.0;operating system
train_1228
Outsourced backup saves time
To increase the efficiency of its data backup and to free staff to concentrate on core business, The Gadget Shop is relying on a secure, automated system hosted by a third party
the gadget shop;outsourced;data backup;e-business
train_1229
Dot-Net makes slow progress
Microsoft's Windows .Net Enterprise Server Release Candidate I, which was released at the end of last month, provides an early glimpse of the system that will eventually replace Windows 200 Advanced Server. The software has been improved so that Active Directory is more flexible and easier to deploy; and security, scal...
scalability;windows .net enterprise server;security;active directory
train_123
A new identification approach for FIR models
The identification of stochastic discrete systems disturbed with noise is discussed in this brief. The concept of general prediction error (GPE) criterion is introduced for the time-domain estimate with optimal frequency estimation (OFE) introduced for the frequency-domain estimate. The two estimation methods are combi...
identification approach;stochastic discrete systems;optimal frequency estimation;frequency-domain estimate;fir models;empirical frequency-domain optimal parameter estimate;time-domain estimate;general prediction error criterion
train_1230
Server safeguards tax service
Peterborough-based tax consultancy IE Taxguard wanted real-time failover protection for important Windows-based applications. Its solution was to implement a powerful failover server from UK supplier Neverfail in order to provide real-time backup for three core production servers
ie taxguard;neverfail;backup;tax consultancy;failover server
train_1231
Efficient parallel programming on scalable shared memory systems with High
Performance Fortran OpenMP offers a high-level interface for parallel programming on scalable shared memory (SMP) architectures. It provides the user with simple work-sharing directives while it relies on the compiler to generate parallel programs based on thread parallelism. However, the lack of language features for ...
scalable hardware;scalable shared memory;multiprocessor architectures;shared memory multiprocessor;high performance fortran;parallel programming
train_1232
Techniques for compiling and implementing all NAS parallel benchmarks in HPF
The NAS parallel benchmarks (NPB) are a well-known benchmark set for high-performance machines. Much effort has been made to implement them in High-Performance Fortran (HPF). In previous attempts, however, the HPF versions did not include the complete set of benchmarks, and the performance was not always good. In this ...
hpf compiler;compiler;nas parallel benchmarks;distributed-memory parallel supercomputers;high-performance machines
train_1233
Advanced optimization strategies in the Rice dHPF compiler
High-Performance Fortran (HPF) was envisioned as a vehicle for modernizing legacy Fortran codes to achieve scalable parallel performance. To a large extent, today's commercially available HPF compilers have failed to deliver scalable parallel performance for a broad spectrum of applications because of insufficiently po...
legacy fortran codes;compiler optimization;automatic parallelization;rice dhpf compiler;multipartitioning;parallel performance;compiler analysis;hpf compilers;mgh-performance fortran
train_1234
Achieving performance under OpenMP on ccNUMA and software distributed shared
memory systems OpenMP is emerging as a viable high-level programming model for shared memory parallel systems. It was conceived to enable easy, portable application development on this range of systems, and it has also been implemented on cache-coherent Non-Uniform Memory Access (ccNUMA) architectures. Unfortunately, i...
hpf;shared memory parallel systems;cache-coherent non-uniform memory access;programming model;openmp;parallel programming;single program multiple data
train_1235
Finding performance bugs with the TNO HPF benchmark suite
High-Performance Fortran (HPF) has been designed to provide portable performance on distributed memory machines. An important aspect of portable performance is the behavior of the available HPF compilers. Ideally, a programmer may expect comparable performance between different HPF compilers, given the same program and...
compiler optimizations;benchmark suite;portable performance;parallel compilers;high-performance fortran;performance portability;hpf compilers;distributed memory machines
train_1236
Compatibility comparison and performance evaluation for Japanese HPF compilers
using scientific applications The lack of compatibility of High-Performance Fortran (HPF) between vender implementations has been disheartening scientific application users so as to hinder the development of portable programs. Thus parallel computing is still unpopular in the computational science community, even thoug...
hpf;compilers;high-performance fortran;source compatability;parallel performance;portable programs;parallel programming
train_1237
High-performance numerical pricing methods
The pricing of financial derivatives is an important field in finance and constitutes a major component of financial management applications. The uncertainty of future events often makes analytic approaches infeasible and, hence, time-consuming numerical simulations are required. In the Aurora Financial Management Syst...
financial management;aurora financial management system;numerical pricing kernel;backward induction methods;investment strategies;finance;monte carlo simulation;derivative pricing;stochastic processes;pricing
train_1238
Optimization of element-by-element FEM in HPF 1.1
In this study, Poisson's equation is numerically evaluated by the element-by-element (EBE) finite-element method in a parallel environment using HPF 1.1 (High-Performance Fortran). In order to achieve high parallel efficiency, the data structures have been altered to node-based data instead of mixtures of node- and ele...
hpf;message passing;hpf compiler;conjugate gradient solver;finite element method;parallel programs;poisson equation;element-by-element
train_1239
Three-dimensional global MHD simulation code for the Earth's magnetosphere
using HPF/JA We have translated a three-dimensional magnetohydrodynamic (MHD) simulation code of the Earth's magnetosphere from VPP Fortran to HPF/JA on the Fujitsu VPP5000/56 vector-parallel supercomputer and the MHD code was fully vectorized and fully parallelized in VPP Fortran. The entire performance and capability...
parallel computation;hpf mhd code;mhd simulation;fujitsu vpp5000/56;vector-parallel supercomputer;magnetohydrodynamic simulation
train_124
High-speed CMOS circuits with parallel dynamic logic and speed-enhanced skewed
static logic In this paper, we describe parallel dynamic logic (PDL) which exhibits high speed without charge sharing problem. PDL uses only parallel-connected transistors for fast logic evaluation and is a good candidate for high-speed low-voltage operation. It has less back-bias effect compared to other logic styles,...
32 bit;parallel-connected transistors;logic synthesis;parallel dynamic logic;low-voltage operation;speed-enhanced skewed static logic;0.25 micron;back-bias effect;stacked transistors;power-delay product;delay;high-speed cmos circuits;carry lookahead adders
train_1240
Implementation and evaluation of HPF/SX V2
We are developing HPF/SX V2, a High Performance Fortran (HPF) compiler for vector parallel machines. It provides some unique extensions as well as the features of HPF 2.0 and HPF/JA. In particular, this paper describes four of them: (1) the ON directive of HPF 2.0; (2) the REFLECT and LOCAL directives of HPF/JA; (3) ve...
benchmark;compiler;hpf/sx v2;parallelization;vector parallel machines;high performance fortran compiler
train_1241
Code generator for the HPF Library and Fortran 95 transformational functions
One of the language features of the core language of HPF 2.0 (High Performance Fortran) is the HPF Library. The HPF Library consists of 55 generic functions. The implementation of this library presents the challenge that all data types, data kinds, array ranks and input distributions need to be supported. For instance,...
hpf;data types;parallel languages;code generation;library generator;matrix multiplication;parameterized templates;library functions;generic functions;high performance fortran;hpf library;parallel computing
train_1242
VPP Fortran and the design of HPF/JA extensions
VPP Fortran is a data parallel language that has been designed for the VPP series of supercomputers. In addition to pure data parallelism, it contains certain low-level features that were designed to extract high performance from user programs. A comparison of VPP Fortran and High-Performance Fortran (HPF) 2.0 shows th...
high performance;benchmark;data parallel language;vpp fortran;asynchronous communication;data locality;asynchronous interprocessor communication;explicit shadow;data parallelism
train_1243
HPF/JA: extensions of High Performance Fortran for accelerating real-world
applications This paper presents a set of extensions on High Performance Fortran (HPF) to make it more usable for parallelizing real-world production codes. HPF has been effective for programs that a compiler can automatically optimize efficiently. However, once the compiler cannot, there have been no ways for the user...
hpf;compiler;data parallel language;parallel processing;high performance fortran;parallel programming;parallelization of loops;supercomputer
train_1244
Applied ethics in business information units
The primary thesis of this paper is that business information professionals commonly overlook ethical dilemmas in the workplace. Although the thesis remains unproven, the author highlights, by way of real and hypothetical case studies, a number of situations in which ethical tensions can be identified, and suggests tha...
moral context;ethical dilemmas;business information professionals;applied ethics;business information units
train_1245
A brief guide to competitive intelligence: how to gather and use information on
competitors The author outlines the processes involved in competitive intelligence, and discusses what it is, how to do it and gives examples of what happens when companies fail to monitor their competitive environment effectively. The author presents a case study, showing how the company that produced the pre-cursor t...
intelligence analysis;intelligence collection;intelligence utilization;intelligence communication;barbie doll;competitor information;competitive intelligence;business environment
train_1246
Why information departments are becoming academic
This article outlines the increasing convergence between academia and business over the last decade or so, and the mutual benefits that this closer association has brought. It also looks at the growing importance of the information profession, suggesting that this is leading to a greater need for specialist skills, as ...
information departments;academic courses;information science;universities;business;academia;information profession;specialist skills
train_1247
The changing landscape for multi access portals
Discusses the factors that have made life difficult for consumer portal operators in recent years causing them, like others in the telecommunications, media and technology sector, to take a close look at their business models following the dot.com crash and the consequent reassessment of Internet-related project financ...
multi-access portals;blue-chip sponsorship;fee-based services;consumer portal operators;revenue streams;advertising
train_1248
Public business libraries: the next chapter
Traces the history of the provision of business information by Leeds Public Libraries, UK, from the opening of the Public Commercial and Technical Library in 1918 to the revolutionary impact of the Internet in the 1990s. Describes how the Library came to terms with the need to integrate the Internet into its mainstream...
trade journal articles;press cuttings;steel companies;market research;leeds public libraries;business information services;sheffield universities;budgets;sinto;government;it use;public business libraries;internet;public commercial and technical library;patents information;british standards information;resources;priorit...
train_1249
Aggregators versus disintermediators: battling it out in the information
superhighstreet Perhaps the future of large-scale content aggregators is now no longer in doubt but this was not the case 10 years ago, when many leading industry experts were much more pessimistic in their predictions. In the year that Dialog celebrates its thirtieth anniversary as the world's oldest and largest profe...
information supply chain;disintermediators;large-scale content aggregators;online information service
train_125
A fast implementation of correlation of long data sequences for coherent
receivers Coherent reception depends upon matching of phase between the transmitted and received signal. Fast convolution techniques based on fast Fourier transform (FFT) are widely used for extracting time delay information from such matching. The latency in processing a large data window of the received signal is a s...
computation;mission critical real time applications;transmitted signal;received signal;parallel algorithm;time delay information;latency;multiprocessor environment;correlation;fast fourier transform;coherent receivers;long data sequences
train_1250
The impact and implementation of XML on business-to-business commerce
This paper discusses the impact analysis of the Extensible Markup Language (XML). Each business partner within a supply chain will be allowed to generate its own data exchange format by adopting an XML meta-data management system in the local side. Followed after a brief introduction of the information technology for B...
business to customer;electronic commerce;xml schema builder;business to business;xml;electronic data interchange;enterprise resources planning;extensible markup language;biztalk
train_1251
A synergic analysis for Web-based enterprise resources planning systems
As the central nervous system for managing an organization's mission and critical business data, Enterprise Resource Planning (ERP) system has evolved to become the backbone of e-business implementation. Since an ERP system is multimodule application software that helps a company manage its important business functions...
enterprise resource planning;customer relationship management;web-based enterprise resources planning;erp;synergic analysis;e-business
train_1252
An agent-oriented and service-oriented environment for deploying dynamic
distributed systems This paper presents JASE, a Java-based Agent-oriented and Service-oriented Environment for deploying dynamic distributed systems. JASE utilizes two important concepts in the field of distributed computing: the concept of services and remote programming with mobile agents. In JASE, mobile agents are ...
service-oriented environment;remote programming;dynamic distributed systems;performance evaluation;mobile agents;agent-oriented environment;jase;java-based agent-oriented and service-oriented environment
train_1253
Application of XML for neural network exchange
This article introduces a framework for the interchange of trained neural network models. An XML-based language (Neural Network Markup Language) for the neural network model description is offered. It allows to write down all the components of neural network model which are necessary for its reproduction. We propose to...
network output interpretation;data dictionary;preprocessing methods;xml;neural network exchange;network structure;neural network markup language
train_1254
Supporting unified interface to wrapper generator in integrated information
retrieval Given the ever-increasing scale and diversity of information and applications on the Internet, improving the technology of information retrieval is an urgent research objective. Retrieved information is either semi-structured or unstructured in format and its sources are extremely heterogeneous. In consequenc...
internet;integrated information retrieval;corba;meta-search engine;unified interface;wrapper generator;automatic extensible markup language
train_1255
Succession in standardization: grafting XML onto SGML
Succession in standardization is often a problem. The advantages of improvements must be weighed against those of compatibility. If compatibility considerations dominate, a grafting process takes place. According to our taxonomy of succession, there are three types of outcomes. A Type I succession, where grafting is su...
type i succession;sgml;standard generalized markup language;grafting process;standardization;xml;extensible markup language
train_1256
High-speed consistency checking for hypothetical reasoning systems using
inference path network Hypothetical reasoning is popular in fault diagnostics and design systems, but slow reasoning speed is its drawback. The goal of the current study is developing hypothetical reasoning based on an inference path network, which would overcome this drawback. In hypothetical reasoning systems based o...
fault diagnostics;hypothesis composition;hypothetical reasoning;reasoning speed;inconsistencies;high-speed consistency checking;speed up;inference path network
train_1257
Definition of a similarity measure between cases based on auto/cross-fuzzy
thesauri A similarity measure between cases is needed in order to evaluate the degree of similarity when using past similar cases in order to resolve current problems. In similar case retrieval, multiple indices are set up in order to characterize the queries and individual cases, then terms are given as values to each...
case similarity measure;problem solving;relationship indices;case-based retrieval;mutually distinct indices;similar case retrieval;auto-fuzzy thesaurus;decision making support system;corresponding indices;cross-fuzzy thesaurus
train_1258
Implementation and performance evaluation of a FIFO queue class library for
time warp The authors describe the implementation, use, and performance evaluation of a FIFO queue class library by means of a high-performance, easy-to-use interface employed for queuing simulations in parallel discrete simulations based on the time warp method. Various general-purpose simulation libraries and languag...
object oriented method;embedded data structures;general-purpose simulation libraries;fifo queue;simulation languages;state management;operational history;performance evaluation;first-in-first-out characteristics;state vectors;time warp simulation;queuing simulations;dynamic data structures;garbage collection;easy-to-us...
train_1259
A mechanism for inferring approximate solutions under incomplete knowledge
based on rule similarity This paper proposes an inference method which can obtain an approximate solution even if the knowledge stored in the problem-solving system is incomplete. When a rule needed for solving the problem does not exist, the problem can be solved by using rules similar to the existing rules. In an imp...
word knowledge base;problem solving;inference method;subgoal atom;reliability;sld procedure;reasoning;consequence atom;rule similarity;incomplete knowledge;representative solution;approximate solution;common sense knowledge
train_126
A new architecture for implementing pipelined FIR ADF based on classification
of coefficients In this paper, we propose a new method for implementing pipelined finite-impulse response (FIR) adaptive digital filter (ADF), with an aim of reducing the maximum delay of the filtering portion of conventional delayed least mean square (DLMS) pipelined ADF. We achieve a filtering section with a maximum ...
pre-upsampled filter;pipelined fir adf;coefficient classification;adaptive digital filter;convergence speed;post-downsampled filter;maximum delay;delayed least mean square filter
train_1260
A dataflow computer which accelerates execution of sequential programs by
precedent firing instructions In the dataflow machine, it is important to avoid degradation of performance in sequential processing, and it is important from the viewpoint of hardware scale to reduce the number of waiting operands. This paper demonstrates that processing performance is degraded by sequential processing...
hardware scale;processing performance;execution time;execution acceleration;precedent firing control;precedent firing instructions;sequential programs;computer architecture;switching process;waiting operands;parallel processing;dataflow computer
train_1261
Topology-adaptive modeling of objects using surface evolutions based on 3D
mathematical morphology Level set methods were proposed mainly by mathematicians for constructing a model of a 3D object of arbitrary topology. However, those methods are computationally inefficient due to repeated distance transformations and increased dimensions. In the paper, we propose a new method of modeling fast...
energy minimization;arbitrary topology;surface evolutions;topology-adaptive modeling;initial approximation;3d mathematical morphology;repeated distance transformations;level set methods;3d object;pseudo curvature flow;refined approximation;range data
train_1262
The development and evaluation of SHOKE2000: the PCI-based FPGA card
This paper describes a PCI-based FPGA card, SHOKE2000, which was developed in order to study reconfigurable computing. Since the latest field programmable gate arrays (FPGA) consist of input/output (I/O) configurable blocks as well as internal configurable logic blocks, they not only realize various user logic circuits...
user logic circuits;computer education;image recognition application;interfaces;dsp;i/o standard;pci;analog devices;reconfigurable computing;fpga;digital signal processors;fpga card;teaching materials;intellectual property;hybrid reconfigurable systems;shoke2000;field programmable gate arrays;distributed shared memory ...
train_1263
Super high definition image (WHD: Wide/Double HD) transmission system
This paper describes a WHD image transmission system constructed from a display projector, CODECs, and a camera system imaging a super high definition image (WHD: Wide/Double HD) corresponding to two screen portions of common high-vision images. This system was developed as a transmission system to communicate with or ...
codecs;whd image transmission system;camera system imaging;reality-enhanced feeling;super high definition image transmission system
train_1264
Estimation of the vanishing point for automatic driving system using a cross
ratio This paper proposes a new method to estimate the vanishing point used as the vehicle heading, which is essential in automatic driving systems. The proposed method uses a cross ratio comprised of a ratio of lengths from four collinear points for extracting the edges that shape the vanishing point. Then, lines that...
automatic driving system;automatic driving systems;hough space;lane markings;vanishing point estimation;collinear points;cross ratio;real road scene
train_1265
Optimization of requantization parameter for MPEG transcoding
This paper considers transcoding in which an MPEG stream is converted to a low-bit-rate MPEG stream, and proposes a method in which the transcoding error can be reduced by optimally selecting the quantization parameter for each macroblock. In transcoding with a low compression ratio, it is crucial to prohibit transcodi...
twofold recompression;requantization parameter optimization;low-bit-rate mpeg stream;error propagation effect;macroblock;rate conversion;simulation;rate control;transcoding error;psnr;6 mbit/s;interframe prediction;compression ratio;rate-distortion curve
train_1266
An intelligent information gathering method for dynamic information mediators
The Internet is spreading into our society rapidly and is becoming one of the information infrastructures that are indispensable for our daily life. In particular, the WWW is widely used for various purposes such as sharing personal information, academic research, business work, and electronic commerce, and the amount ...
internet;electronic commerce;academic research;business work;dynamic information mediators;real-world flight information service;information infrastructures;artificial information integration problem;intelligent information gathering method;www
train_1267
3D reconstruction from uncalibrated-camera optical flow and its reliability
evaluation We present a scheme for reconstructing a 3D structure from optical flow observed by a camera with an unknown focal length in a statistically optimal way as well as evaluating the reliability of the computed shape. First, the flow fundamental matrices are optimally computed from the observed flow. They are th...
covariance matrices;3d reconstruction;real-image experiments;epipolar equation;reliability evaluation;normalization;flow fundamental matrices;motion parameters;uncalibrated-camera optical flow
train_1268
Reachability in contextual nets
Contextual nets, or Petri nets with read arcs, are models of concurrent systems with context dependent actions. The problem of reachability in such nets consists in finding a sequence of transitions that leads from the initial marking of a given contextual net to a given goal marking. The solution to this problem that ...
linear programming;concurrent systems;contextual nets reachability;finite prefix;context dependent actions;goal marking;petri nets
train_1269
Minimizing the number of successor states in the stubborn set method
Combinatorial explosion which occurs in parallel compositions of LTSs can be alleviated by letting the stubborn set method construct on-the-fly a reduced LTS that is CFFD- or CSP-equivalent to the actual parallel composition. This article considers the problem of minimizing the number of successor states of a given sta...
combinatorial explosion;weighted vertices;csp-equivalence;stubborn set method;low-degree polynomial time
train_127
Asymptotical stability in discrete-time neural networks
In this work, we present a proof of the existence of a fixed point and a generalized sufficient condition that guarantees the stability of it in discrete-time neural networks by using the Lyapunov function method. We also show that for both symmetric and asymmetric connections, the unique attractor is a fixed point whe...
connection weight matrix;unique attractor;symmetric connections;asymptotical stability;discrete-time neural networks;lyapunov function method;equilibrium stability;fixed point;asymmetric connections;interval matrix;generalized sufficient condition;stability
train_1270
A comparison of different decision algorithms used in volumetric storm cells
classification Decision algorithms useful in classifying meteorological volumetric radar data are discussed. Such data come from the radar decision support system (RDSS) database of Environment Canada and concern summer storms created in this country. Some research groups used the data completed by RDSS for verifying t...
support vector machine;rough set algorithms;meteorological volumetric radar data;decision algorithms;summer storms;radar decision support system;object oriented reducts;volumetric storm cells classification;classification accuracy
train_1271
Verification of non-functional properties of a composable architecture with
Petri nets In this paper, we introduce our concept of composability and present the MSS architecture as an example for a composable architecture. MSS claims to be composable with respect to timing properties. We discuss, how to model and prove properties in such an architecture with time-extended Petrinets. As a result...
composable architecture;mss architecture;timing properties;petri nets;non-functional properties verification;proof of composability
train_1272
Global action rules in distributed knowledge systems
Previously Z. Ras and J.M. Zytkow (2000) introduced and investigated query answering system based on distributed knowledge mining. The notion of an action rule was introduced by Z. Ras and A. Wieczorkowska (2000) and its application domain e-business was taken. In this paper, we generalize the notion of action rules in...
distributed knowledge mining;query answering system;attributes;global action rules;e-commerce;action rules