name
stringlengths
7
10
title
stringlengths
13
125
abstract
stringlengths
67
3.02k
fulltext
stringclasses
1 value
keywords
stringlengths
17
734
train_1839
The dynamic aspect of land administration: an often-forgotten component in
system design Although the establishment of a land administration system is enough of a challenge as it is, the task of keeping the system up to date with developments in society is even more challenging. Initial adjudication and cadastral mapping basically record land tenure as it exists at a given moment, i.e. the static situation. The paper aims to analyse the developments that might occur in a society with respect to tenure, value and use of land. These developments constitute a dynamic component of land administration. As land administration systems have to serve society on a long-term basis and normally have a long-term return on investment, the author recommends taking into account both the static and dynamic component when designing land administration systems
land administration system;land tenure;dynamic component;static situation;system design;dynamic aspect;cadastral mapping;return on investment;adjudication;societal developments;static component
train_184
On the expected value of the minimum assignment
The minimum k-assignment of an m*n matrix X is the minimum sum of k entries of X, no two of which belong to the same row or column. Coppersmith and Sorkin conjectured that if X is generated by choosing each entry independently from the exponential distribution with mean 1, then the expected value of its minimum k-assignment is given by an explicit formula, which has been proven only in a few cases. In this paper we describe our efforts to prove the Coppersmith-Sorkin conjecture by considering the more general situation where the entries x/sub ij/ of X are chosen independently from different distributions. In particular, we require that x/sub ij/ be chosen from the exponential distribution with mean 1/r/sub i/c/sub j/. We conjecture an explicit formula for the expected value of the minimum k-assignment of such X and give evidence for this formula
m * n matrix;minimum k-assignment;exponential distribution;rational function;bipartite graph
train_1840
Too much middleware
The movement from client-server computing to multi-tier computing has created a potpourri of so-called middleware systems, including application servers, workflow products, EAI systems, ETL systems and federated data systems. We argue that the explosion in middleware has created a myriad of poorly integrated systems with overlapping functionality. The world would be well served by considerable consolidation, and we present some of the ways this might happen. Some of the points covered in the article have been previously explored by P. Bernstein (1996)
multi-tier computing;middleware systems;eai systems;federated data systems;client-server computing;etl systems;poorly integrated systems;overlapping functionality;application servers;workflow products
train_1841
The XML typechecking problem
When an XML document conforms to a given type (e.g. a DTD or an XML schema type) it is called a valid document. Checking if a given XML document is valid is called the validation problem, and is typically performed by a parser (hence, validating parser), more precisely it is performed right after parsing, by the same program module. In practice however, XML documents are often generated dynamically, by some program: checking whether all XML documents generated by the program are valid WRT a given type is called the typechecking problem. While a validation analyzes an XML document, a type checker analyzes a program, and the problem's difficulty is a function of the language in which that program is expressed. The XML typechecking problem has been investigated recently and the XQuery Working Group adopted some of these techniques for typechecking XQuery. All these techniques, however, have limitations which need to be understood and further explored and investigated. We define the XML typechecking problem, and present current approaches to typechecking, discussing their limitations
xml document;xquery;xml typechecking problem;validation
train_1842
The role of B2B engines in B2B integration architectures
Semantic B2B integration architectures must enable enterprises to communicate standards-based B2B events like purchase orders with any potential trading partner. This requires not only back end application integration capabilities to integrate with e.g. enterprise resource planning (ERP) systems as the company-internal source and destination of B2B events, but also a capability to implement every necessary B2B protocol like electronic data interchange (EDI), RosettaNet as well as more generic capabilities like Web services (WS). This paper shows the placement and functionality of B2B engines in semantic B2B integration architectures that implement a generic framework for modeling and executing any B2B protocol. A detailed discussion shows how a B2B engine can provide the necessary abstractions to implement any standard-based B2B protocol or any trading partner specific specialization
edi;semantic b2b integration architectures;b2b engines;trading partner;rosettanet;purchase orders;erp systems;modeling;standards-based b2b event communication;web services
train_1843
Supply chain infrastructures: system integration and information sharing
The need for supply chain integration (SCI) methodologies has been increasing as a consequence of the globalization of production and sales, and the advancement of enabling information technologies. In this paper, we describe our experience with implementing and modeling SCIs. We present the integration architecture and the software components of our prototype implementation. We then discuss a variety of information sharing methodologies. Then, within the framework of a multi-echelon supply chain process model spanning multiple organizations, we summarize research on the benefits of intraorganizational knowledge sharing, and we discuss performance scalability
system integration;sales;performance scalability;software components;supply chain infrastructures;globalization;intraorganizational knowledge sharing;information sharing;multiple organizations;production;multi-echelon supply chain process model;supply chain integration
train_1844
A multi-agent system infrastructure for software component marketplace: an
ontological perspective In this paper, we introduce a multi-agent system architecture and an implemented prototype for a software component marketplace. We emphasize the ontological perspective by discussing ontology modeling for the component marketplace, UML extensions for ontology modeling, and the idea of ontology transfer which makes the multi-agent system adapt itself to dynamically changing ontologies
ontology modeling;uml extensions;dynamically changing ontologies;multi-agent system architecture;ontology transfer;software component marketplace;adaptation
train_1845
Business data management for business-to-business electronic commerce
Business-to-business electronic commerce (B2B EC) opens up new possibilities for trade. For example, new business partners from around the globe can be found, their offers can be compared, even complex negotiations can be conducted electronically, and a contract can be drawn up and fulfilled via an electronic marketplace. However, sophisticated data management is required to provide such facilities. In this paper, the results of a multi-national project on creating a business-to-business electronic marketplace for small and medium-sized enterprises are presented. Tools for information discovery, protocol-based negotiations, and monitored contract enactment are provided and based on a business data repository. The repository integrates heterogeneous business data with business communication. Specific problems such as multilingual nature, data ownership, and traceability of contracts and related negotiations are addressed and it is shown that the present approach provides efficient business data management for B2B EC
business data repository;multilingual system;electronic marketplace;business communication;data ownership;protocol-based negotiations;traceability;information discovery;multi-national project;small and medium-sized enterprises;heterogeneous business data;business-to-business electronic commerce;monitored contract enactment;business data management
train_1846
Semantic B2B integration: issues in ontology-based approaches
Solving queries to support e-commerce transactions can involve retrieving and integrating information from multiple information resources. Often, users don't care which resources are used to answer their query. In such situations, the ideal solution would be to hide from the user the details of the resources involved in solving a particular query. An example would be providing seamless access to a set of heterogeneous electronic product catalogues. There are many problems that must be addressed before such a solution can be provided. In this paper, we discuss a number of these problems, indicate how we have addressed these and go on to describe the proof-of-concept demonstration system we have developed
e-commerce transactions;multiple information resources;semantic b2b integration;information integration;queries;information retrieval;ontology-based approaches;heterogeneous electronic product catalogues
train_1847
Conceptual modeling and specification generation for B2B business processes
based on ebXML In order to support dynamic setup of business processes among independent organizations, a formal standard schema for describing the business processes is basically required. The ebXML framework provides such a specification schema called BPSS (Business Process Specification Schema) which is available in two standalone representations: a UML version, and an XML version. The former, however, is not intended for the direct creation of business process specifications, but for defining specification elements and their relationships required for creating an ebXML-compliant business process specification. For this reason, it is very important to support conceptual modeling that is well organized and directly matched with major modeling concepts. This paper deals with how to represent and manage B2B business processes using UML-compliant diagrams. The major challenge is to organize UML diagrams in a natural way that is well suited to the business process meta-model and then to transform the diagrams into an XML version. This paper demonstrates the usefulness of conceptually modeling business processes by prototyping a business process editor tool called ebDesigner
ebdesigner;business process specification schema;conceptual modeling;ebxml;uml-compliant diagrams;specification generation;business process editor;meta model;formal standard schema;b2b business processes
train_1848
Contracting in the days of ebusiness
Putting electronic business on a sound foundation-model theoretically as well as technologically-is a central challenge for research as well as commercial development. This paper concentrates on the discovery and negotiation phase of concluding an agreement based on a contract. We present a methodology for moving seamlessly from a many-to-many relationship in the discovery phase to a one-to-one relationship in the contract negotiation phase. Making the content of contracts persistent is achieved by reconstructing contract templates by means of mereologic (logic of the whole-part relation). Possibly nested sub-structures of the contract template are taken as a basis for negotiation in a dialogical way. For the negotiation itself the contract templates are extended by implications (logical) and sequences (topical)
contract negotiation phase;sequences;contract templates;many-to-many relationship;discovery phase;nested sub-structure;mereologic;implications;one-to-one relationship;electronic business;contracting
train_1849
An active functionality service for e-business applications
Service based architectures are a powerful approach to meet the fast evolution of business rules and the corresponding software. An active functionality service that detects events and involves the appropriate business rules is a critical component of such a service-based middleware architecture. In this paper we present an active functionality service that is capable of detecting events in heterogeneous environments, it uses an integral ontology-based approach for the semantic interpretation of heterogeneous events and data, and provides notifications through a publish/subscribe notification mechanism. The power of this approach is illustrated with the help of an auction application and through the personalization of car and driver portals in Internet-enabled vehicles
business rules;personalized driver portals;personalized car portals;heterogeneous environments;auction application;ontology based approach;internet-enabled vehicles;software;publish/subscribe notification mechanism;service-based middleware architecture;event detection;e-business applications;semantic interpretation;active functionality service
train_185
Property testers for dense Constraint Satisfaction programs on finite domains
Many NP-hard languages can be "decided" in subexponential time if the definition of "decide" is relaxed only slightly. Rubinfeld and Sudan introduced the notion of property testers, probabilistic algorithms that can decide, with high probability, if a function has a certain property or if it is far from any function having this property. Goldreich, Goldwasser, and Ron constructed property testers with constant query complexity for dense instances of a large class of graph problems. Since many graph problems can be viewed as special cases of the Constraint Satisfaction Problem on Boolean domains, it is natural to try to construct property testers for more general cases of the Constraint Satisfaction Problem. In this paper, we give explicit constructions of property testers using a constant number of queries for dense instances of Constraint Satisfaction Problems where the constraints have constant arity and the variables assume values in some domain of finite size
dense instances;constraint satisfaction problem;subexponential time;randomized sampling;constraint satisfaction;graph problems;np-hard languages;property testers;constant query complexity;probabilistic algorithms
train_1850
The n-tier hub technology
During 2001, the Enterprise Engineering Laboratory at George Mason University was contracted by the Boeing Company to develop an eHub capability for aerospace suppliers in Taiwan. In a laboratory environment, the core technology was designed, developed, and tested, and now a large first-tier aerospace supplier in Taiwan is commercializing the technology. The project objective was to provide layered network and application services for transporting XML-based business transaction flows across multi-tier, heterogeneous data processing environments. This paper documents the business scenario, the eHub application, and the network transport mechanisms that were used to build the n-tier hub. In contrast to most eHubs, this solution takes the point of view of suppliers, pushing data in accordance with supplier requirements; hence, enhancing the probability of supplier adoption. The unique contribution of this project is the development of an eHub that meets the needs of small and medium enterprises (SMEs) and first-tier suppliers
multi-tier heterogeneous data processing environments;business scenario;boeing company;n-tier hub technology;xml-based business transaction flows;taiwan;network transport mechanisms;small and medium enterprises;supplier adoption;aerospace suppliers;first-tier suppliers
train_1851
Supporting global user profiles through trusted authorities
Personalization generally refers to making a Web site more responsive to the unique and individual needs of each user. We argue that for personalization to work effectively, detailed and interoperable user profiles should be globally available for authorized sites, and these profiles should dynamically reflect changes in user interests. Creating user profiles from user click-stream data seems to be an effective way of generating detailed and dynamic user profiles. However, a user profile generated in this way is available only on the computer where the user accesses his browser, and is inaccessible when the same user works on a different computer. On the other hand, integration of the Internet with telecommunication networks has made it possible for the users to connect to the Web with a variety of mobile devices as well as desktops. This requires that user profiles should be available to any desktop or mobile device on the Internet that users choose to work with. In this paper, we address these problems through the concept of "trusted authority". A user agent at the client side that captures the user click stream, dynamically generates a navigational history 'log' file in Extensible Markup Language (XML). This log file is then used to produce 'user profiles' in a resource description framework (RDF). A user's right to privacy is provided through the Platform for Privacy Preferences (P3P) standard. User profiles are uploaded to the trusted authority and served next time the user connects to the Web
trusted authorities;privacy;namespace qualifier;user click stream;telecommunication networks;internet;platform for privacy preferences standard;resource description framework;user agent;navigational history log file;globally unique user id/password identification;xml;personalization;global user profiles;web site;mobile device
train_1852
The design and performance evaluation of alternative XML storage strategies
This paper studies five strategies for storing XML documents including one that leaves documents in the file system, three that use a relational database system, and one that uses an object manager. We implement and evaluate each approach using a number of XQuery queries. A number of interesting insights are gained from these experiments and a summary of the advantages and disadvantages of the approaches is presented
file system;xml document storage;xquery queries;relational database system;performance evaluation;object manager
train_1853
CherylAnn Silberer: all about process [accounting technologist]
Silberer's company, CompLete, is making a specialty of workflow process analysis
complete;workflow process analysis;accounting technologist
train_1854
Software Technology: looking for quality accountants
Software Technology wants to turn 23 years of reselling experience in the legal business into an asset in the accounting market
software technology;accounting market;reselling
train_1855
Distribution software: ROI is king
Middle-market accounting software vendors are taking to the open road, by way of souped-up distribution suites that can track product as it wends its way from warehouse floor to customer site. Integration provides efficiencies, and cost savings
distribution;accounting software;warehouse management
train_1856
Tax forms: CD or not CD?
The move from CD to the Web looks unstoppable. Besides counting how many thousands of electronic tax forms they offer, vendors are rapidly moving those documents to the Web
web;cch perform plus h;electronic tax forms;ria eform;universal tax systems forms complete;stf services superform;nelco laserlibrarian ii;atx forms zillion forms;kleinrock forms library plus
train_1858
Account aggregation: shaping up portfolios
CPA providers of financial planning services are providing clients with a unified view of their investments
financial planning services;account aggregation
train_1859
Practice management goes remote [accounting]
There's a lot of life in accounting practice management software, a valuable category that has been subject to much change in the last few years. Web-based time tracking grows in popularity. Looks at CCH ProSystem fx Practice, CMS Open Solutions 6, Creative Solutions Practice, Time Matters, CPASoftware Visual Practice Management, and Abak
web-based time tracking;cms open solutions 6;cpasoftware visual practice management;abak;creative solutions practice;cch prosystem fx practice;accounting practice management software;time matters
train_186
The diameter of a long-range percolation graph
We consider the following long-range percolation model: an undirected graph with the node set {0, 1, . . . , N}/sup d/, has edges (x, y) selected with probability approximately= beta /||x - y||/sup s/ if ||x - y|| > 1, and with probability 1 if ||x - y|| = 1, for some parameters beta , s > 0. This model was introduced by who obtained bounds on the diameter of this graph for the one-dimensional case d = 1 and for various values of s, but left cases s = 1, 2 open. We show that, with high probability, the diameter of this graph is Theta (log N/log log N) when s = d, and, for some constants 0 < eta /sub 1/ < eta /sub 2/ < 1, it is at most N/sup eta 2/ when s = 2d, and is at least N/sup eta 1/ when d = 1, s = 2, beta < 1 or when s > 2d. We also provide a simple proof that the diameter is at most log/sup O(1)/ N with high probability, when d < s < 2d, established previously in Benjamini and Berger (2001)
positive probability;undirected graph;random graph;probability;networks;percolation;long-range percolation model
train_1860
The art of the cross-sell [accounting software]
With the market for accounting software nearing saturation, vendors are training resellers in the subtleties of the cross-sell. The rewards can be great. The key is knowing when to focus, and when to partner
resellers;cross-selling;accounting software
train_1861
Technology in distance education: a global perspective to alternative delivery
mechanisms Technology is providing a positive impact on delivery mechanisms employed in distance education at the university level. Some institutions are incorporating distance education as a way to extend the classroom. Other institutions are investigating new delivery mechanisms, which support a revised perspective on education. These latter institutions are revising their processes for interacting with students, and taking a more "learner centered" approach to the delivery of education. This article discusses the impact of technology on the delivery mechanisms employed in distance education. A framework is proposed here, which presents a description of alternative modes of generic delivery mechanisms. It is suggested that those institutions, which adopt a delivery mechanism employing an asynchronous mode, can gain the most benefit from technology. This approach seems to represent the only truly innovative use of technology in distance education. The approach creates a student-oriented environment while maintaining high levels of interaction, both of which are factors that contribute to student satisfaction with their overall educational experience
student satisfaction;distance education;university education;learner centered approach;global perspective;asynchronous mode;educational technology
train_1862
Global comparison of stages of growth based on critical success factors
With increasing globalization of business, the management of IT in international organizations is faced with the complex task of dealing with the difference between local and international IT needs. This study evaluates, and compares, the level of IT maturity and the critical success factors (CSFs) in selected geographic regions, namely, Norway, Australia/New Zealand, North America, Europe, Asia/Pacific, and India. The results show that significant differences in the IT management needs in these geographic regions exist, and that the IT management operating in these regions must balance the multiple critical success factors for achieving an optimal local-global mix for business success
asia/pacific;business success;india;critical success factors;local it needs;australia;it maturity;europe;international it needs;north america;optimal local-global mix;business globalization;new zealand;it management;norway
train_1863
Information systems project failure: a comparative study of two countries
Many organizations, regardless of size, engage in at least one, and often many information system projects each year. Many of these projects consume massive amounts of resources, and may cost as little as a few thousand dollars to ten, and even hundreds of millions of dollars. Needless to say, the investment of time and resources into these ventures are of significant concern to chief information officers (CIOs), executives staff members, project managers, and others in leadership positions. This paper describes the results of a survey performed between Australia and the United States regarding factors leading to IS project failure. The findings suggest that, among other things, end user involvement and executive management leadership are key indicators influencing IS project failure
australia;information systems project failure;end user involvement;executive management leadership;united states
train_1864
Cultural differences in developers' perceptions of information systems success
factors: Japan vs. the United States The study examined the perceptions of information systems (IS) developers from Japan and the United States regarding the strategies that are considered most important for successful implementation of an IS. The results of principal component analysis revealed that the IS strategies could be reduced to five components: (1) characteristics of the team members, (2) characteristics of the project leader, (3) management/user input, (4) proper technology, and (5) communication. The results indicated that there was a significant difference in the perceptions of Japanese and US developers with respect to the importance of the five components. Japanese developers perceived the project leader as the most crucial component for determining the success of an IS project. Team member characteristics was viewed as the least important by Japanese developers. On the other hand, developers from the US viewed communications as the most critical component. Project leader characteristics were perceived to be the least important by US developers. The results were discussed in terms of cultural differences
communication;united states;principal component analysis;japan;cultural differences;information systems developer perceptions;management/user input;proper technology;project leader characteristics;is project;team member characteristics;information systems success factors
train_1865
Towards the globalisation of the IS/IT function
The IS/IT function has recently emerged from the peripheral aspects of the finance department to the centre of critical organisational change. There is an increasing dependency on its activities as systems extend beyond supporting the internal efficiency of the organisation to augmenting global performance. The growth of wide and local networks has resulted in communication possibilities that were not possible a few years ago. E-commerce challenges the achievements of the IS/IT function and is very prominent in the globalisation of modern organisations. The complexity and diversity of electronic exchange is also well documented (Hackney et al., 2000). This has a number of impacts on the development and implementation of IS/IT solutions for organisations involved in international trade. It is a conjecture that the IS/IT function is critically important for the alignment of the business to meet the demands of global competition, through building internal marketing strategies and creating knowledge based communities. There is clear evidence that IS/IT can lead to improved business performance and potentially for sustained competitive advantage. This is obviously true through the advent of new and emerging technologies such as the Internet
internet;wide area networks;knowledge based communities;electronic exchange;internal marketing strategies;international trade;is/it function;e-commerce;globalisation;local area networks
train_1866
Tracking with sensor failures
Studies the reliability with sensor failures of the asymptotic tracking problem for linear time invariant systems using the factorization approach. The plant is two-output and the compensator is two-degree-of-freedom. Necessary and sufficient conditions are presented for the general problem and a simple solution is given for problems with stable plants
linear time invariant systems;necessary and sufficient conditions;asymptotic tracking problem;two-output plant;sensor failures;factorization approach;two-degree-of-freedom compensator;reliability
train_1867
Optimization-based design of fixed-order controllers for command following
For discrete-time scalar systems, we propose an approach for designing feedback controllers of fixed order to minimize an upper bound on the peak magnitude of the tracking error to a given command input. The work makes use of linear programming to design over a class of closed-loop systems proposed for the rejection of non-zero initial conditions and bounded disturbances. We incorporate performance robustness in the form of a guaranteed upper bound on the peak magnitude of the tracking error under plant coprime factor uncertainty
closed-loop systems;coprime factor uncertainty;guaranteed upper bound;linear programming;fixed-order controllers;command following;tracking error;performance robustness;optimization-based design;feedback controllers;discrete-time scalar systems
train_1868
Estimation of an N-L-N Hammerstein-Wiener model
Estimation of a single-input single-output block-oriented model is studied. The model consists of a linear block embedded between two static nonlinear gains. Hence, it is called an N-L-N Hammerstein-Wiener model. First, the model structure is motivated and the disturbance model is discussed. The paper then concentrates on parameter estimation. A relaxation iteration scheme is proposed by making use of a model structure in which the error is bilinear-in-parameters. This leads to a simple algorithm which minimizes the original loss function. The convergence and consistency of the algorithm are studied. In order to reduce the variance error, the obtained linear model is further reduced using frequency weighted model reduction. A simulation study is used to illustrate the method
bilinear-in-parameters error;nonlinear process;n-l-n hammerstein-wiener model;disturbance model;linear block;convergence;frequency weighted model reduction;model structure;relaxation iteration scheme;variance error;parameter estimation;static nonlinear gains;single-input single-output block-oriented model;consistency
train_1869
Stability and L/sub 2/ gain properties of LPV systems
Stability and L/sub 2/ gain properties of linear parameter-varying systems are obtained under assumed bounds on either the maximum or average value of the parameter rate
parameter rate;linear parameter-varying systems;gain scheduled control;l/sub 2/ gain properties;gromwall-bellman inequality;stability
train_187
Control of a thrust-vectored flying wing: a receding horizon - LPV approach
This paper deals with the application of receding horizon methods to hover and forward flight models of an experimental tethered flying wing developed at Caltech. The dynamics of the system are representative of a vertical landing and take off aircraft, such as a Harrier around hover, or a thrust-vectored aircraft such as F18-HARV or X-31 in forward flight. The adopted control methodology is a hybrid of receding horizon techniques and control Lyapunov function (CLF)-based ideas. First, a CLF is generated using quasi-LPV methods and then, by using the CLF as the terminal cost in the receding horizon optimization, stability is guaranteed. The main advantage of this approach is that stability can be guaranteed without imposing constraints in the on-line optimization, allowing the problem to be solved in a more efficient manner. Models of the experimental set-up are obtained for the hover and forward flight modes. Numerical simulations for different time horizons are presented to illustrate the effectiveness of the discussed methods. Specifically, it is shown that a mere upper bound on the cost-to-go is not an appropriate choice for a terminal cost, when the horizon length is short. Simulation results are presented using experimentally verified model parameters
x-31;caltech;stability guarantee;f18-harv;nonlinear system;forward flight models;thrust-vectored aircraft;numerical simulations;receding horizon-lpv approach;vertical landing take off aircraft;quasi-lpv methods;tethered flying wing;hover flight models;terminal cost;harrier around hover;thrust-vectored flying wing control;control lyapunov function-based ideas;receding horizon optimization;receding horizon techniques
train_1870
Robust control of nonlinear systems with parametric uncertainty
Probabilistic robustness analysis and synthesis for nonlinear systems with uncertain parameters are presented. Monte Carlo simulation is used to estimate the likelihood of system instability and violation of performance requirements subject to variations of the probabilistic system parameters. Stochastic robust control synthesis searches the controller design parameter space to minimize a cost that is a function of the probabilities that design criteria will not be satisfied. The robust control design approach is illustrated by a simple nonlinear example. A modified feedback linearization control is chosen as controller structure, and the design parameters are searched by a genetic algorithm to achieve the tradeoff between stability and performance robustness
performance requirements violation;probabilistic robustness synthesis;probabilistic robustness analysis;modified feedback linearization control;stochastic control synthesis;monte carlo simulation;uncertain parameters;parametric uncertainty;system instability;genetic algorithm;robust control;nonlinear systems;input-to-state stability
train_1871
Strong and weak points of the MUSCADET theorem prover-examples from CASC-JC
MUSCADET is a knowledge-based theorem prover based on natural deduction. It has participated in CADE Automated theorem proving System Competitions. The results show its complementarity with regard to resolution-based provers. This paper presents some of its crucial methods and gives some examples of MUSCADET proofs from the last competition (CASC-JC in IJCAR 2001)
muscadet;resolution-based provers;knowledge-based theorem prover;casc-jc;natural deduction;cade automated theorem proving system competitions
train_1872
TPTP, CASC and the development of a semantically guided theorem prover
The first-order theorem prover SCOTT has been through a series of versions over some ten years. The successive provers, while retaining the same underlying technology, have used radically different algorithms and shown wide differences of behaviour. The development process has depended heavily on experiments with problems from the TPTP library and has been sharpened by participation in CASC each year since 1997. We outline some of the difficulties inherent in designing and refining a theorem prover as complex as SCOTT, and explain our experimental methodology. While SCOTT is not one of the systems which have been highly optimised for CASC, it does help to illustrate the influence of both CASC and the TPTP library on contemporary theorem proving research
casc;first-order theorem prover;tptp library;semantically guided theorem prover;experimental methodology;proof searches;scott;semantically constrained otter
train_1873
A phytography of WALDMEISTER
The architecture of the WALDMEISTER prover for unit equational deduction is based on a strict separation of active and passive facts. After an inspection of the system's proof procedure, the representation of each of the central data structures is outlined, namely indexing for the active facts, compression for the passive facts, successor sets for the hypotheses, and minimal recording of inference steps for the proof object. In order to cope with large search spaces, specialized redundancy criteria are employed, and the empirically gained control knowledge is integrated to ease the use of the system. The paper concludes with a quantitative comparison of the WALDMEISTER versions over the years, and a view of the future prospects
future prospects;passive facts;indexing;redundancy;cade atp system competition;hypotheses;phytography;unit equational deduction;inference;waldmeister;data structures;active facts;large search spaces;theorem prover
train_1874
E - a brainiac theorem prover
We describe the superposition-based theorem prover E. E is a sound and complete prover for clausal first order logic with equality. Important properties of the prover include strong redundancy elimination criteria, the DISCOUNT loop proof procedure, a very flexible interface for specifying search control heuristics, and an efficient inference engine. We also discuss the strengths and weaknesses of the system
search control heuristics;casc;e automatic theorem prover;completeness;soundness;superposition-based theorem prover;discount;strong redundancy elimination criteria;clausal first order logic;loop proof procedure;cade atp system competitions;inference engine;rewriting;brainiac theorem prover;equality
train_1875
The design and implementation of VAMPIRE
We describe VAMPIRE: a high-performance theorem prover for first-order logic. As our description is mostly targeted to the developers of such systems and specialists in automated reasoning, it focuses on the design of the system and some key implementation features. We also analyze the performance of the prover at CASC-JC
first-order logic;resolution theorem proving;performance evaluation;casc-jc;automated reasoning;high-performance theorem prover;vampire
train_1876
The development of CASC [automated theorem proving]
Researchers who make theoretical advances also need some way to demonstrate that an advance really does have general, overall positive consequences for system performance. For this it is necessary to evaluate the system on a set of problems that is sufficiently large and diverse to be somehow representative of the intended application area as a whole. It is only a small step from system evaluation to a communal system competition. The CADE ATP System Competition (CASC) has been run annually since 1996. Any competition is difficult to design and organize in the first instance, and to then run over the years. In order to obtain the full benefits of a competition, a thoroughly organized event, with an unambiguous and motivated design, is necessary. For some issues relevant to the CASC design, inevitable constraints have emerged. For other issues there have been several choices, and decisions have had to be made. This paper describes the evolution of CASC, paying particular attention to its design, design changes, and organization
automated theorem proving;casc;ai;automated deduction;system performance;cade atp system competition;artificial intelligence;system evaluation;classical first order logic
train_1877
Strong completeness of lattice-valued logic
This paper shows strong completeness of the system L for lattice valued logic given by S. Titani (1999), in which she formulates a lattice-valued set theory by introducing the logical implication which represents the order relation on the lattice. Syntax and semantics concerned are described and strong completeness is proved
order relation;strong completeness;semantics;lattice-valued set theory;lattice-valued logic;syntax
train_1878
Max and min limiters
If A contained in omega , n>or=2, and the function max({x/sub 1/,...,x/sub n/} intersection A) is partial recursive, it is easily seen that A is recursive. In this paper, we weaken this hypothesis in various ways (and similarly for "min" in place of "max") and investigate what effect this has on the complexity of A. We discover a sharp contrast between retraceable and co-retraceable sets, and we characterize sets which are the union of a recursive set and a co-r.e., retraceable set. Most of our proofs are noneffective. Several open questions are raised
max limiters;recursive set;complexity;retraceable sets;min limiters
train_1879
On the distribution of Lachlan nonsplitting bases
We say that a computably enumerable (c.e.) degree b is a Lachlan nonsplitting base (LNB), if there is a computably enumerable degree a such that a>b, and for any c.e. degrees w, v<or=a, if a<or=wVvV b then either a<or=wV b or a<or=vV b. In this paper we investigate the relationship between bounding and nonbounding of Lachlan nonsplitting bases and the high/low hierarchy. We prove that there is a non-Low/sub 2/ c.e. degree which bounds no Lachlan nonsplitting base
turing degrees;computably enumerable degree;lachlan nonsplitting bases distribution
train_188
Sampled-data implementation of a gain scheduled controller
A continuous-time gain-scheduled controller must be transformed to a corresponding discrete-time controller for sampled-data implementation. We show that certain linearization properties of a continuous-time gain scheduled controller are inherited by its sampled-data implementation. We also show that a similar relationship exists for multi-rate gain scheduled controllers arising in flight control applications
discrete-time controller;linearization properties;gain scheduled controller;flight control applications;continuous-time gain-scheduled controller;sampled-data implementation;multi-rate gain scheduled controllers
train_1880
Simulation and transient testing of numerical relays
A hybrid and practical solution for relay evaluation is presented. Two main issues are taken into account: power system simulation and relay simulation, both of which consist of different stages. System simulation is carried out by means of EMTP and is complemented by additional features, such as filtering for location and determination of fault parameters that allow comparing simulated and actual fault records to improve and guarantee a correct system simulation. Relay simulation includes filtering algorithms, all the relaying units, and the decision logic. Playing simulated or real faults over the actual relay and comparing simulated and real responses can check for correct relay simulation
filtering algorithms;decision logic;emtp;real faults;relay evaluation;numerical relays;relay simulation;filtering;simulated faults;fault location;relay testing;power system simulation
train_1881
Stabilization of positive systems with first integrals
Positive systems possessing first integrals are considered. These systems frequently occur in applications. The paper is devoted to two stabilization problems. The first is concerned with the design of feedbacks to stabilize a given level set. Secondly, it is shown that the same feedback allows us to globally stabilize an equilibrium point if it is asymptotically stable with respect to initial conditions in its level set. Two examples are provided and the results are compared with those in the literature
feedbacks design;asymptotic stability;chemical reactors;global stabilization;first integrals;positive systems;equilibrium point
train_1882
Bandwidth vs. gains design of H/sub infinity / tracking controllers for
current-fed induction motors Describes a systematic procedure for designing speed and rotor flux norm tracking H/sub infinity /. controllers with unknown load torque disturbances for current-fed induction motors. A new effective design tool is developed to allow selection of the control gains so as to adjust the disturbances' rejection capability of the controllers in the face of the bandwidth requirements of the closed-loop system. Application of the proposed design procedure is demonstrated in a case study, and the results of numerical simulations illustrate the satisfactory performance achievable even in presence of rotor resistance uncertainty
closed-loop system;current-fed induction motors;observers;unknown load torque disturbances;rotor flux norm controllers;h/sub infinity / tracking controllers;feedback linearization;design tool;disturbances rejection capability;bandwidth requirements;speed controllers
train_1883
Analysis of exclusively kinetic two-link underactuated mechanical systems
Analysis of exclusively kinetic two-link underactuated mechanical systems is undertaken. It is first shown that such systems are not full-state feedback linearizable around any equilibrium point. Also, the equilibrium points for which the system is small-time locally controllable (STLC) is at most a one-dimensional submanifold. A concept less restrictive than STLC, termed the small-time local output controllability (STLOC) is introduced, the satisfaction of which guarantees that a chosen configuration output can be controlled at its desired value. It is shown that the class of systems considered is STLOC, if the inertial coupling between the input and output is nonzero. Also, in such a case, the system is nonminimum phase. An example section illustrates all the results presented
equilibrium points;small-time local output controllability;nonminimum phase;small-time locally controllable system;exclusively kinetic two-link underactuated mechanical systems;one-dimensional submanifold
train_1884
Observer-based strict positive real (SPR) feedback control system design
Presents theory for stability analysis and design for a class of observer-based feedback control systems. Relaxation of the controllability and observability conditions imposed in the Yakubovich-Kalman-Popov lemma can be made for a class of nonlinear systems described by a linear time-invariant system with a feedback-connected cone-bounded nonlinear element. It is shown how a circle-criterion approach can be used to design an observer-based state feedback control which yields a closed-loop system with specified robustness characteristics. The approach is relevant for design with preservation of stability when a cone-bounded nonlinearity is introduced in the feedback loop. Important applications are to be found in nonlinear control with high robustness requirements
linear time-invariant system;closed-loop system;robustness characteristics;yakubovich-kalman-popov lemma;cone-bounded nonlinearity;control system design;state feedback control;circle-criterion approach;nonlinear systems;stability analysis;observer-based strict positive real feedback control system;feedback-connected cone-bounded nonlinear element
train_1885
Analysis of nonlinear time-delay systems using modules over non-commutative
rings The theory of non-commutative rings is introduced to provide a basis for the study of nonlinear control systems with time delays. The left Ore ring of non-commutative polynomials defined over the field of a meromorphic function is suggested as the framework for such a study. This approach is then generalized to a broader class of nonlinear systems with delays that are called generalized Roesser systems. Finally, the theory is applied to analyze nonlinear time-delay systems. A weak observability is defined and characterized, generalizing the well-known linear result. Properties of closed submodules are then developed to obtain a result on the accessibility of such systems
modules;meromorphic function;left ore ring;noncommutative polynomials;noncommutative rings;nonlinear control systems;weak observability;nonlinear time-delay systems;generalized roesser systems
train_1886
Non-asymptotic confidence ellipsoids for the least-squares estimate
We consider the finite sample properties of least-squares system identification, and derive non-asymptotic confidence ellipsoids for the estimate. The shape of the confidence ellipsoids is similar to the shape of the ellipsoids derived using asymptotic theory, but unlike asymptotic theory, they are valid for a finite number of data points. The probability that the estimate belongs to a certain ellipsoid has a natural dependence on the volume of the ellipsoid, the data generating mechanism, the model order and the number of data points available
least-squares system identification;data points;finite sample properties;least-squares estimate;nonasymptotic confidence ellipsoids;model order;data generating mechanism;probability
train_1887
Doubly invariant equilibria of linear discrete-time games
The notion of doubly invariant (DI) equilibrium is introduced. The concept extends controlled and robustly controlled invariance notions to the context of two-person dynamic games. Each player tries to keep the state in a region of state space independently of the actions of the rival player. The paper gives existence conditions, criteria and algorithms for the determination of DI equilibria of linear dynamic games in discrete time. Two examples illustrate the results. The first one is in the area of fault-tolerant controller synthesis. The second is an application to macroeconomics
macroeconomics;two-person dynamic games;fault-tolerant controller synthesis;doubly invariant equilibria;state space;existence conditions;robustly controlled invariance;linear discrete-time games
train_1888
L/sub 2/ model reduction and variance reduction
We examine certain variance properties of model reduction. The focus is on L/sub 2/ model reduction, but some general results are also presented. These general results can be used to analyze various other model reduction schemes. The models we study are finite impulse response (FIR) and output error (OE) models. We compare the variance of two estimated models. The first one is estimated directly from data and the other one is computed by reducing a high order model, by L/sub 2/ model reduction. In the FIR case we show that it is never better to estimate the model directly from data, compared to estimating it via L/sub 2/ model reduction of a high order FIR model. For OE models we show that the reduced model has the same variance as the directly estimated one if the reduced model class used contains the true system
variance reduction;identification;l/sub 2/ model reduction;fir models;output error models;finite impulse response models
train_1889
Sliding mode dynamics in continuous feedback control for distributed
discrete-event scheduling A continuous feedback control approach for real-time scheduling of discrete events is presented motivated by the need for control theoretic techniques to analyze and design such systems in distributed manufacturing applications. These continuous feedback control systems exhibit highly nonlinear and discontinuous dynamics. Specifically, when the production demand in the manufacturing system exceeds the available resource capacity then the control system "chatters" and exhibits sliding modes. This sliding mode behavior is advantageously used in the scheduling application by allowing the system to visit different schedules within an infinitesimal region near the sliding surface. In the paper, an analytical model is developed to characterize the sliding mode dynamics. This model is then used to design controllers in the sliding mode domain to improve the effectiveness of the control system to "search" for schedules with good performance. Computational results indicate that the continuous feedback control approach can provide near-optimal schedules and that it is computationally efficient compared to existing scheduling techniques
control theoretic techniques;highly nonlinear discontinuous dynamics;resource capacity;distributed discrete-event scheduling;continuous feedback control;distributed manufacturing applications;production demand;real-time scheduling;sliding mode dynamics
train_189
Identification of linear parameter varying models
We consider identification of a certain class of discrete-time nonlinear systems known as linear parameter varying system. We assume that inputs, outputs and the scheduling parameters are directly measured, and a form of the functional dependence of the system coefficients on the parameters is known. We show how this identification problem can be reduced to a linear regression, and provide compact formulae for the corresponding least mean square and recursive least-squares algorithms. We derive conditions on persistency of excitation in terms of the inputs and scheduling parameter trajectories when the functional dependence is of polynomial type. These conditions have a natural polynomial interpolation interpretation, and do not require the scheduling parameter trajectories to vary slowly. This method is illustrated with a simulation example using two different parameter trajectories
least mean square algorithms;identification;parameter trajectories;persistency of excitation conditions;discrete-time nonlinear systems;scheduling parameter trajectories;recursive least-squares algorithms;functional dependence;polynomial interpolation interpretation;system coefficients;scheduling parameters;linear regression;time-varying systems;linear parameter varying models
train_1890
Robustness of trajectories with finite time extent
The problem of estimating perturbation bounds of finite trajectories is considered. The trajectory is assumed to be generated by a linear system with uncertainty characterized in terms of integral quadratic constraints. It is shown that such perturbation bounds can be obtained as the solution to a nonconvex quadratic optimization problem, which can be addressed using Lagrange relaxation. The result can be used in robustness analysis of hybrid systems and switched dynamical systems
robustness analysis;uncertainty;nonconvex quadratic optimization problem;trajectories robustness;perturbation bounds;switched dynamical systems;hybrid systems;integral quadratic constraints;lagrange relaxation;linear system;finite time extent
train_1891
On trajectory and force tracking control of constrained mobile manipulators
with parameter uncertainty Studies the trajectory and force tracking control problem of mobile manipulators subject to holonomic and nonholonomic constraints with unknown inertia parameters. Adaptive controllers are proposed based on a suitable reduced dynamic model, the defined reference signals and the mixed tracking errors. The proposed controllers not only ensure the entire state of the system to asymptotically converge to the desired trajectory but also ensure the constraint force to asymptotically converge to the desired force. A detailed numerical example is presented to illustrate the developed methods
force tracking control;holonomic constraints;asymptotic convergence;mixed tracking errors;trajectory control;nonholonomic constraints;position control;constrained mobile manipulators;reduced dynamic model;mobile robots;adaptive controllers;parameter uncertainty
train_1892
Closed-loop persistent identification of linear systems with unmodeled dynamics
and stochastic disturbances The essential issues of time complexity and probing signal selection are studied for persistent identification of linear time-invariant systems in a closed-loop setting. By establishing both upper and lower bounds on identification accuracy as functions of the length of observation, size of unmodeled dynamics, and stochastic disturbances, we demonstrate the inherent impact of unmodeled dynamics on identification accuracy, reduction of time complexity by stochastic averaging on disturbances, and probing capability of full rank periodic signals for closed-loop persistent identification. These findings indicate that the mixed formulation, in which deterministic uncertainty of system dynamics is blended with random disturbances, is beneficial to reduction of identification complexity
stochastic disturbances;closed-loop persistent identification;identification accuracy;upper bounds;time complexity;full rank periodic signals;unmodeled dynamics;lower bounds;linear time-invariant systems;probing signal selection
train_1893
Closed-loop model set validation under a stochastic framework
Deals with probabilistic model set validation. It is assumed that the dynamics of a multi-input multi-output (MIMO) plant is described by a model set with unstructured uncertainties, and identification experiments are performed in closed loop. A necessary and sufficient condition has been derived for the consistency of the model set with both the stabilizing controller and closed-loop frequency domain experimental data (FDED). In this condition, only the Euclidean norm of a complex vector is involved, and this complex vector depends linearly on both the disturbances and the measurement errors. Based on this condition, an analytic formula has been derived for the sample unfalsified probability (SUP) of the model set. Some of the asymptotic statistical properties of the SUP have also been briefly discussed. A numerical example is included to illustrate the efficiency of the suggested method in model set quality evaluation
stabilizing controller;complex vector;unstructured uncertainties;stochastic framework;asymptotic statistical properties;euclidean norm;closed-loop model set validation;probabilistic model set validation;mimo plant;closed-loop frequency domain experimental data;robust control;multi-input multi-output plant;necessary and sufficient condition;unstructured uncertainty
train_1894
Switching controller design via convex polyhedral Lyapunov functions
We propose a systematic switching control design method for a class of nonlinear discrete time hybrid systems. The novelty of the adopted approach is in the fact that unlike conventional control the control burden is shifted to a logical level thus creating the need for the development of new analysis/design methods
convex polyhedral lyapunov functions;nonlinear discrete time hybrid systems;switching controller design;systematic design method
train_1895
An algorithm combining neural networks with fundamental parameters
An algorithm combining neural networks with the fundamental parameters equations (NNFP) is proposed for making corrections for non-linear matrix effects in x-ray fluorescence analysis. In the algorithm, neural networks were applied to relate the concentrations of components to both the measured intensities and the relative theoretical intensities calculated by the fundamental parameter equations. The NNFP algorithm is compared with the classical theoretical correction models, including the fundamental parameters approach, the Lachance-Traill model, a hyperbolic function model and the COLA algorithm. For an alloy system with 15 measured elements, in most cases, the prediction errors of the NNFP algorithm are lower than those of the fundamental parameters approach, the Lachance-Traill model, the hyperbolic function model and the COLA algorithm separately. If there are the serious matrix effects, such as matrix effects among Cr, Fe and Ni, the NNFP algorithm generally decreased predictive errors as compared with the classical models, except for the case of Cr by the fundamental parameters approach. The main reason why the NNFP algorithm has generally a better predictive ability than the classical theoretical correction models might be that neural networks can better calibrate the non-linear matrix effects in a complex multivariate system
algorithm;complex multivariate system;fe;ni;x-ray fluorescence analysis;fundamental parameters;lachance-traill model;intensities;alloy system;cr;cola algorithm;neural networks;nnfp algorithm;hyperbolic function model;theoretical correction models;fundamental parameters equations;nonlinear matrix effects
train_1896
The dynamics of a railway freight wagon wheelset with dry friction damping
We investigate the dynamics of a simple model of a wheelset that supports one end of a railway freight wagon by springs with linear characteristics and dry friction dampers. The wagon runs on an ideal, straight and level track with constant speed. The lateral dynamics in dependence on the speed is examined. We have included stick-slip and hysteresis in our model of the dry friction and assume that Coulomb's law holds during the slip phase. It is found that the action of dry friction completely changes the bifurcation diagram, and that the longitudinal component of the dry friction damping forces destabilizes the wagon
lateral dynamics;hysteresis;stick-slip;linear characteristics;coulomb law;longitudinal component;dry friction damping;railway freight wagon wheelset;dynamics;bifurcation diagram
train_1897
User-appropriate tyre-modelling for vehicle dynamics in standard and limit
situations When modelling vehicles for the vehicle dynamic simulation, special attention must be paid to the modelling of tyre forces and -torques, according to their dominant influence on the results. This task is not only about sufficiently exact representation of the effective forces but also about user-friendly and practical relevant applicability, especially when the experimental tyre-input-data is incomplete or missing. This text firstly describes the basics of the vehicle dynamic tyre model, conceived to be a physically based, semi-empirical model for application in connection with multi-body-systems (MBS). On the basis of tyres for a passenger car and a heavy truck the simulated steady state tyre characteristics are shown together and compared with the underlying experimental values. The possibility to link the tyre model TMeasy to any MBS-program is described, as far as it supports the 'Standard Tyre Interface'. As an example, the simulated and experimental data of a heavy truck doing a standardized driving manoeuvre are compared
semi-empirical model;multi-body-systems;simulated steady state tyre characteristics;tmeasy;standard situations;vehicle dynamics;standardized driving manoeuvre;tyre modelling;tyre torques;standard tyre interface;heavy truck;passenger car;limit situations
train_1898
Design of a stroke dependent damper for the front axle suspension of a truck
using multibody system dynamics and numerical optimization A stroke dependent damper is designed for the front axle suspension of a truck. The damper supplies extra damping for inward deflections rising above 4 cm. In this way the damper should reduce extreme suspension deflections without deteriorating the comfort of the truck. But the question is which stroke dependent damping curve yields the best compromise between suspension deflection working space and comfort. Therefore an optimization problem is defined to minimize the maximum inward suspension deflection subject to constraints on the chassis acceleration for three typical road undulations. The optimization problem is solved using sequential linear programming (SLP) and multibody dynamics simulation software. Several optimization runs have been carried out for a small two degree of freedom vehicle model and a large full-scale model of the truck semi-trailer combination. The results show that the stroke dependent damping can reduce large deflections at incidental road disturbances, but that the optimum stroke dependent damping curve is related to the acceleration bound. By means of vehicle model simulation and numerical optimization we have been able to quantify this trade-off between suspension deflection working space and truck comfort
sequential linear programming;damping;acceleration bound;inward deflections;multibody system dynamics;full-scale model;truck comfort;numerical optimization;extreme suspension deflections;truck semi-trailer combination;front axle suspension;incidental road disturbances;chassis acceleration;vehicle model simulation;road undulations;stroke dependent damper
train_1899
Human face detection in visual scenes using neural networks
This paper presents a neural network based face detection system. Our objective is to design a system that can detect human faces in visual scenes at high searching speed and accuracy. We used a neural network with a simple structure but trained using face and non-face samples preprocessed by several methods (position normalization, histogram equalization, etc.) to attain high accuracy, then pruned the size of the neural network so that it could run faster and reduced the total search area of a target visual scene using the skin color detector. Skin color detection assumes that faces reside only in skin color regions. The system design is made up of two parts: the face detecting system that detects the faces, and the searching speed improving system. Speed improvement is achieved by reducing the face locator network size using the structural learning with knowledge and by reducing the face search area using the skin color detection system. Faster training of the neural networks was also achieved using variable step sizes
neural network;visual scene;down-sampling;face detecting system;self organizing maps;merging overlapping detections;skin color detector;human face recognition;backpropagation
train_19
Decentralized adaptive output feedback stabilization for a class of
interconnected systems with unknown bound of uncertainties The problem of adaptive decentralized stabilization for a class of linear time-invarying large-scale systems with nonlinear interconnectivity and uncertainties is discussed. The bounds of uncertainties are assumed to be unknown. For such uncertain dynamic systems, an adaptive decentralized controller is presented. The resulting closed-loop systems are asymptotically stable in theory. Moreover, an adaptive decentralized control scheme is given. The scheme ensures the closed-loop systems exponentially practically stable and can be used in practical engineering. Finally, simulations show that the control scheme is effective
closed-loop systems;large scale systems;uncertain dynamic systems;adaptive decentralized stabilization;robust control
train_190
On the design of gain-scheduled trajectory tracking controllers [AUV
application] A new methodology is proposed for the design of trajectory tracking controllers for autonomous vehicles. The design technique builds on gain scheduling control theory. An application is made to the design of a trajectory tracking controller for a prototype autonomous underwater vehicle (AUV). The effectiveness and advantages of the new control laws derived are illustrated in simulation using a full set of non-linear equations of motion of the vehicle
nonlinear equations of motion;control laws;gain scheduling control theory;gain-scheduled trajectory tracking controller design;autonomous underwater vehicle;autonomous vehicles
train_1900
Robust L/sub 2/ disturbance attenuation for nonlinear systems with input
dynamical uncertainty Deals with the problem of robust L/sub 2/ disturbance attenuation for nonlinear systems with input dynamical uncertainty. The input dynamical uncertainty is restricted to be minimum-phase and relative degree zero. A sufficient condition. is given such that the nonlinear system satisfies the L/sub 2/ gain performance and input-to-state stable property. Using this condition, a design approach is given for a smooth state feedback control law that solves the robust L/sub 2/ disturbance attenuation problem, and the approach is extended to a more general case where the nominal system has higher relative degree. Finally, a numerical example is given to demonstrate the proposed approach
nominal system;sufficient condition;smooth state feedback control law;robust l/sub 2/ disturbance attenuation;design approach;input-to-state stable property;robust control;nonlinear systems;input dynamical uncertainty;l/sub 2/ gain performance
train_1901
FC++: Functional tools for object-oriented tasks
FC++ is a library for programming functionally in C++. Compared to other C++ functional programming libraries, FC++ is distinguished by its powerful type system which allows the manipulation of parametrically polymorphic functions (e.g., passing them as arguments to other functions and returning them as results). In this paper, we show how FC++ can be used in common object-oriented programming tasks. We demonstrate FC++ implementations of several common design patterns (Adapter, Builder, Command, and more). Compared to conventional C++ implementations of these patterns, our implementations are either simpler (in that fewer classes/dependencies are needed), more efficient, or more type-safe (thanks to parametric polymorphism and type inference)
functional programming;c++;parametric polymorphism;library;object-oriented programming;fc++
train_1902
Engineering plug-in software components to support collaborative work
Many software applications require co-operative work support, including collaborative editing, group awareness, versioning, messaging and automated notification and co-ordination agents. Most approaches hard-code such facilities into applications, with fixed functionality and limited ability to reuse groupware implementations. We describe our recent work in seamlessly adding such capabilities to component-based applications via a set of collaborative work-supporting plug-in software components. We describe a variety of applications of this technique, along with descriptions of the novel architecture, user interface adaptation and implementation techniques for the collaborative work-supporting components that we have developed. We report on our experiences to date with this method of supporting collaborative work enhancement of component-based systems, and discuss the advantages of our approach over conventional techniques
co-operative work support;collaborative work tools;plug-in software components;collaborative editing;versioning;automated notification;groupware;software applications;group awareness;messaging
train_1903
The BLISS programming language: a history
The BLISS programming language was invented by William A. Wulf and others at Carnegie-Mellon University in 1969, originally for the DEC PDP-10. BLISS-10 caught the interest of Ronald F. Brender of DEC (Digital Equipment Corporation). After several years of collaboration, including the creation of BLISS-11 for the PDP-11, BLISS was adopted as DEC's implementation language for use on its new line of VAX computers in 1975. DEC developed a completely new generation of BLISSs for the VAX, PDP-10 and PDP-11, which became widely used at DEC during the 1970s and 1980s. With the creation of the Alpha architecture in the early 1990s, BLISS was extended again, in both 32- and 64-bit flavors. BLISS support for the Intel IA-32 architecture was introduced in 1995 and IA-64 support is now in progress. BLISS has a number of unusual characteristics: it is typeless, requires use of an explicit contents of operator (written as a period or 'dot'), takes an algorithmic approach to data structure definition, has no goto, is an expression language, and has an unusually rich compile-time language. This paper reviews the evolution and use of BLISS over its three decade lifetime. Emphasis is on how the language evolved to facilitate portable programming while retaining its initial highly machine-specific character. Finally, the success of its characteristics are assessed
system implementation language;machine-oriented language;compile-time language;data structure definition;bliss programming language;portable programming
train_1904
Component support in PLT scheme
PLT Scheme (DrScheme and MzScheme) supports the Component Object Model (COM) standard with two pieces of software. The first piece is MzCOM, a COM class that makes a Scheme evaluator available to COM clients. With MzCOM, programmers can embed Scheme code in programs written in mainstream languages such as C++ or Visual BASIC. Some applications can also be used as MzCOM clients. The other piece of component-support software is MysterX, which makes COM classes available to PLT Scheme programs. When needed, MysterX uses a programmable Web browser to display COM objects. We describe the technical issues encountered in building these two systems and sketch some applications
mzcom;web browser;reuse;plt scheme;component object model
train_1905
Tool and process improvements from MFC control system technology
A new approach to MFC calibration links the physical parameters of nitrogen to the physical characteristics of various process gases. This precludes the conventional need for surrogate gases. What results is a physics-based tuning algorithm and enhanced digital control system that enables rearranging and gas change of digital MFCs. The end result should be better process control through more accurate gas flow. The new method also decreases the number of MFC spare parts required to back up a fab
process control;tuning algorithm;digital control;calibration;tool technology;mfc control system technology;process gas;semiconductor fab;n/sub 2/;gas flow;mass flow controller;surrogate gas
train_1906
Integrated process control using an in situ sensor for etch
The migration to tighter geometries and more complex process sequence integration schemes requires having the ability to compensate for upstream deviations from target specifications. Doing so ensures that-downstream process sequences operate on work-in-progress that is well within control. Because point-of-use visibility of work-in-progress quality has become of paramount concern in the industry's drive to reduce scrap and improve yield, controlling trench depth has assumed greater importance. An integrated, interferometric based, rate monitor for etch-to-depth and spacer etch applications has been developed for controlling this parameter. This article demonstrates that the integrated rate monitor, using polarization and digital signal processing, enhances control etch-to-depth processes and can also be implemented as a predictive endpoint in a wafer manufacturing environment for dual damascene trench etch and spacer etch applications
complex process sequence integration schemes;upstream deviation compensation;point-of-use visibility;digital signal processing;interferometry;interferometric in situ etch sensor;downstream process sequences;wafer manufacturing environment;work-in-progress quality;yield improvement;scrap reduction;target specifications;integrated process control;ic geometry;spacer etch applications;dual damascene trench etch;process predictive endpoint;integrated etch rate monitor;polarization;trench depth control
train_1907
Multiple comparison methods for means
Multiple comparison methods (MCMs) are used to investigate differences between pairs of population means or, more generally, between subsets of population means using sample data. Although several such methods are commonly available in statistical software packages, users may be poorly informed about the appropriate method(s) to use and/or the correct way to interpret the results. This paper classifies the MCMs and presents the important methods for each class. Both simulated and real data are used to compare the methods, and emphasis is placed on a correct application and interpretation. We include suggestions for choosing the best method. Mathematica programs developed by the authors are used to compare MCMs. By taking the advantage of Mathematica's notebook structure, all interested student can use these programs to explore the subject more deeply
error rate;single-step procedures;step-down procedures;sales management;population means;multiple comparison procedures;pack-age design
train_1908
Explicit solutions for transcendental equations
A simple method to formulate an explicit expression for the roots of any analytic transcendental function is presented. The method is based on Cauchy's integral theorem and uses only basic concepts of complex integration. A convenient method for numerically evaluating the exact expression is presented. The application of both the formulation and evaluation of the exact expression is illustrated for several classical root finding problems
transcendental equations;complex integration;singularity;analytic functions;cauchy integral theorem;polynomial;fourier transform;root finding
train_1909
Breast MR imaging with high spectral and spatial resolutions: preliminary
experience The authors evaluated magnetic resonance (MR) imaging with high spectral and spatial resolutions (HSSR) of water and fat in breasts of healthy volunteers (n=6) and women with suspicious lesions (n=6). Fat suppression, edge delineation, and image texture were improved on MR images derived from HSSR data compared with those on conventional MR images. HSSR MR imaging data acquired before and after contrast medium injection showed spectrally inhomogeneous changes in the water resonances in small voxels that were not detectable with conventional MR imaging
women;small voxels;image texture;magnetic resonance imaging data;suspicious lesions;high spectral spatial resolutions;edge delineation;magnetic resonance images;breast magnetic resonance imaging;contrast medium injection;fat suppression;healthy volunteers;water resonances
train_191
Linear, parameter-varying control and its application to a turbofan engine
This paper describes application of parameter-dependent control design methods to a turbofan engine. Parameter-dependent systems are linear systems, whose state-space descriptions are known functions of time-varying parameters. The time variation of each of the parameters is not known in advance, but is assumed to be measurable in real-time. Three linear, parameter-varying (LPV) approaches to control design are discussed. The first method is based on linear fractional transformations which relies on the small gain theorem for bounds on performance and robustness. The other methods make use of either a single (SQLF) or parameter-dependent (PDQLF) quadratic Lyapunov function to bound the achievable level of performance. The latter two techniques are used to synthesize controllers for a high-performance turbofan engine. A LPV model of the turbofan engine is constructed from Jacobian linearizations at fixed power codes for control design. The control problem is formulated as a model matching problem in the H/sub infinity / and LPV framework. The objective is decoupled command response of the closed-loop system to pressure and rotor speed requests. The performance of linear, H/sub infinity / point designs are compared with the SQLF and PDQLF controllers. Nonlinear simulations indicate that the controller synthesized using the SQLF approach is slightly more conservative than the PDQLF controller. Nonlinear simulations with the SQLF and PDQLF controllers show very robust designs that achieve all desired performance objectives
h/sub infinity / framework;nonlinear simulations;closed-loop system;parameter-dependent quadratic lyapunov function;small gain theorem;decoupled command response;state-space descriptions;model matching problem;linear parameter-varying control;turbofan engine;robustness bounds;parameter-dependent control design methods;time-varying parameters;linear fractional transformations;single quadratic lyapunov function;jacobian linearizations;very robust designs;performance bounds
train_1910
Breast cancer: effectiveness of computer-aided diagnosis-observer study with
independent database of mammograms Evaluates the effectiveness of a computerized classification method as an aid to radiologists reviewing clinical mammograms for which the diagnoses were unknown to both the radiologists and the computer. Six mammographers and six community radiologists participated in an observer study. These 12 radiologists interpreted, with and without the computer aid, 110 cases that were unknown to both the 12 radiologist observers and the trained computer classification scheme. The radiologists' performances in differentiating between benign and malignant masses without and with the computer aid were evaluated with receiver operating characteristic (ROC) analysis. Two-tailed P values were calculated for the Student t test to indicate the statistical significance of the differences in performances with and without the computer aid. When the computer aid was used, the average performance of the 12 radiologists improved, as indicated by an increase in the area under the ROC curve (A/sub z/) from 0.93 to 0.96 (P<.001), by an increase in partial area under the ROC curve (/sub 0.9/0A'/sub z/) from 0.56 to 0.72 (P<.001), and by an increase in sensitivity from 94% to 98% (P=.022). No statistically significant difference in specificity was found between readings with and those without computer aid ( Delta +-0.014; P=.46; 95% Cl: -0.054, 0.026), where Delta is difference in specificity. When we analyzed results from the mammographers and community radiologists as separate groups, a larger improvement was demonstrated for the community radiologists. Computer-aided diagnosis can potentially help radiologists improve their diagnostic accuracy in the task of differentiating between benign and malignant masses seen on mammograms
performances;community radiologists;statistical significance;two-tailed p values;student t test;mammographers;trained computer classification scheme;breast cancer;malignant masses;computer-aided diagnosis;diagnostic accuracy;radiologist observers;benign masses;average performance;clinical mammograms;receiver operating characteristic curve;receiver operating characteristic analysis;observer study;independent database;computerized classification method;computer aid
train_1911
Pulmonary perfusion patterns and pulmonary arterial pressure
Uses artificial intelligence methods to determine whether quantitative parameters describing the perfusion image can be synthesized to make a reasonable estimate of the pulmonary arterial (PA) pressure measured at angiography. Radionuclide perfusion images were obtained in 120 patients with normal chest radiographs who also underwent angiographic PA pressure measurement within 3 days of the radionuclide study. An artificial neural network (ANN) was constructed from several image parameters describing statistical and boundary characteristics of the perfusion images. With use of a leave-one-out cross-validation technique, this method was used to predict the PA systolic pressure in cases on which the ANN had not been trained. A Pearson correlation coefficient was determined between the predicted and measured PA systolic pressures. ANN predictions correlated with measured pulmonary systolic pressures (r=0.846, P<.001). The accuracy of the predictions was not influenced by the presence of pulmonary embolism. None of the 51 patients with predicted PA pressures of less than 29 mm Hg had pulmonary hypertension at angiography. All 13 patients with predicted PA pressures greater than 48 mm Hg had pulmonary hypertension at angiography. Meaningful information regarding PA pressure can be derived from noninvasive radionuclide perfusion scanning. The use of image analysis in concert with artificial intelligence methods helps to reveal physiologic information not readily apparent at visual image inspection
48 pa;normal chest radiographs;quantitative parameters;statistical characteristics;boundary characteristics;physiologic information;29 pa;radionuclide perfusion images;pulmonary perfusion patterns;accuracy;leave-one-out cross-validation technique;visual image inspection;pearson correlation coefficient;image parameters;artificial intelligence methods;perfusion image;noninvasive radionuclide perfusion scanning;image analysis;angiography;pulmonary embolism;angiographic pulmonary arterial pressure measurement;artificial neural network predictions;pulmonary arterial systolic pressure;pulmonary hypertension;patients
train_1912
A novel preterm respiratory mechanics active simulator to test the performances
of neonatal pulmonary ventilators A patient active simulator is proposed which is capable of reproducing values of the parameters of pulmonary mechanics of healthy newborns and preterm pathological infants. The implemented prototype is able to: (a) let the operator choose the respiratory pattern, times of apnea, episodes of cough, sobs, etc., (b) continuously regulate and control the parameters characterizing the pulmonary system; and, finally, (c) reproduce the attempt of breathing of a preterm infant. Taking into account both the limitation due to the chosen application field and the preliminary autocalibration phase automatically carried out by the proposed device, accuracy and reliability on the order of 1% is estimated. The previously indicated value has to be considered satisfactory in light of the field of application and the small values of the simulated parameters. Finally, the achieved metrological characteristics allow the described neonatal simulator to be adopted as a reference device to test performances of neonatal ventilators and, more specifically, to measure the time elapsed between the occurrence of a potentially dangerous condition to the patient and the activation of the corresponding alarm of the tested ventilator
pressure transducer;pneumatic/mechanical unit;accuracy;ventilatory support;healthy newborns;intensive care equipment;double compartment model;neonatal pulmonary ventilators;artificial trachea;preterm pathological infants;upper airway resistance;patient active simulator;electronic unit;apnea times;variable clamp resistance;respiratory diseases;compliance;software control;preterm respiratory mechanics active simulator;autocalibration phase;reliability
train_1913
A six-degree-of-freedom precision motion stage
This article presents the design and performance evaluation of a six-degree-of-freedom piezoelectrically actuated fine motion stage that will be used for three dimensional error compensation of a long-range translation mechanism. Development of a single element, piezoelectric linear displacement actuator capable of translations of 1.67 mu m with 900 V potential across the electrodes and under a 27.4 N axial load and 0.5 mm lateral distortion is presented. Finite element methods have been developed and used to evaluate resonant frequencies of the stage platform and the complete assembly with and without a platform payload. In general, an error of approximately 10.0% between the finite element results and the experimentally measured values were observed. The complete fine motion stage provided approximately +or-0.93 mu m of translation and +or-38.0 mu rad of rotation in all three planes of motion using an excitation range of 1000 V. An impulse response indicating a fundamental mode resonance at 162 Hz was measured with a 0.650 kg payload rigidly mounted to the top of the stage
single element piezoelectric linear displacement actuator;platform payload;650.0 gm;finite element methods;162 hz;1.67 micron;0.93 to -0.93 micron;1000 v;3d error compensation;performance evaluation;long-range translation mechanism;fundamental mode resonance;impulse response;design;stage platform;resonant frequency;piezoelectrically actuated fine motion stage;six-degree-of-freedom precision motion stage;900 v
train_1914
Vacuum-compatible vibration isolation stack for an interferometric
gravitational wave detector TAMA300 Interferometric gravitational wave detectors require a large degree of vibration isolation. For this purpose, a multilayer stack constructed of rubber and metal blocks is suitable, because it provides isolation in all degrees of freedom at once. In TAMA300, a 300 m interferometer in Japan, long-term dimensional stability and compatibility with an ultrahigh vacuum environment of about 10/sup -6/ Pa are also required. To keep the interferometer at its operating point despite ground strain and thermal drift of the isolation system, a thermal actuator was introduced. To prevent the high outgassing rate of the rubber from spoiling the vacuum, the rubber blocks were enclosed by gas-tight bellows. Using these techniques, we have successfully developed a three-layer stack which has a vibration isolation ratio of more than 10/sup 3/ at 300 Hz with control of drift and enough vacuum compatibility
gas-tight bellows;metal blocks;ground strain;interferometric gravitational wave detectors;vibration isolation stack;300 m;thermal actuator;ultrahigh vacuum environment;long-term dimensional stability;thermal drift;10/sup -6/ pa;rubber blocks;rubber outgassing;multilayer stack;300 hz;operating point;tama300 interferometer;vacuum compatibility
train_1915
Multichannel scaler for general statistical analysis of dynamic light
scattering A four channel scaler for counting applications has been designed and built using a standard high transfer rate parallel computer interface bus parallel data card. The counter section is based on standard complex programmable logic device integrated circuits. With a 200 MHz Pentium based host PC a sustained counting and data transfer with channel widths as short as 200 ns for a single channel is realized. The use of the multichannel scaler is demonstrated in dynamic light scattering experiments. The recorded traces are analyzed with wavelet and other statistical techniques to obtain transient changes in the properties of the scattered light
dynamic light scattering;pentium based host pc;standard high transfer rate parallel computer interface;optical spectroscopic techniques;200 mhz;correlation spectroscopy;200 ns;interface bus parallel data card;photon signal statistical properties;complex programmable logic device;standard cpld ics;windowed fourier transform;general statistical analysis;four channel scaler;multichannel scaler
train_1916
Changes in the entropy and the Tsallis difference information during
spontaneous decay and self-organization of nonextensive systems A theoretical-information description of self-organization processes during stimulated transitions between stationary states of open nonextensive systems is presented. S/sub q/- and I/sub q/-theorems on changes of the entropy and Tsallis difference information measures in the process of evolution in the space of control parameters are proved. The entropy and the Tsallis difference information are derived and their new extreme properties are discussed
tsallis difference information;control parameters;nonextensive statistical mechanics;stimulated transitions;information measures;spontaneous decay;entropy;nonextensive systems;self-organization
train_1917
Design and modeling of an interval-based ABR flow control protocol
A novel flow control protocol is presented for availability bit rate (ABR) service in asynchronous transfer mode (ATM) networks. This scheme features periodic explicit rate feedback that enables precise allocation of link bandwidth and buffer space on a hop-by-hop basis to guarantee maximum throughput, minimum cell loss, and high resource efficiency. With the inclusion of resource management cell synchronization and consolidation algorithms, this protocol is capable of controlling point-to-multipoint ABR services within a unified framework. The authors illustrate the modeling of single ABR connection, the interaction between multiple ABR connections, and the constraints applicable to flow control decisions. A loss-free flow control mechanism is presented for high-speed ABR connections using a fluid traffic model. Supporting algorithms and ATM signaling procedures are specified, in company with linear system modeling, numerical analysis, and simulation results, which demonstrate its performance and cost benefits in high-speed backbone networking scenarios
linear system modeling;buffer space allocation;periodic explicit rate feedback;design;signaling;high-speed backbone networking scenarios;atm networks;maximum throughput;resource management cell consolidation algorithms;link bandwidth allocation;high-speed abr connections;point-to-multipoint services;resource management cell synchronization algorithms;availability bit rate service;interval-based abr flow control protocol;high resource efficiency;simulation;minimum cell loss;flow control decisions;numerical analysis;modeling;loss-free flow control mechanism;fluid traffic model
train_1918
Negotiating the semantics of agent communication languages
This article presents a formal framework and outlines a method that autonomous agents can use to negotiate the semantics of their communication language at run-time. Such an ability is needed in open multi-agent systems so that agents can ensure they understand the implications of the utterances that are being made and so that they can tailor the meaning of the primitives to best fit their prevailing circumstances. To this end, the semantic space framework provides a systematic means of classifying the primitives along multiple relevant dimensions. This classification can then be used by the agents to structure their negotiation (or semantic fixing) process so that they converge to the mutually agreeable semantics that are necessary for coherent social interactions
multi-agent systems;semantic space;communication language;autonomous agents;semantic fixing;social interactions
train_1919
Toward a formalism for conversation protocols using joint intention theory
Conversation protocols are used to achieve certain goals or to bring about certain states in the world. Therefore, one may identify the landmarks or the states that must be brought about during the goal-directed execution of a protocol. Accordingly, the landmarks, characterized by propositions that are true in the state represented by that landmark, are the most important aspect of a protocol. Families of conversation protocols can be expressed formally as partially ordered landmarks after the landmarks necessary to achieve a goal have been identified. Concrete protocols represented as joint action expressions can, then, be derived from the partially ordered landmarks and executed directly by joint intention interpreters. This approach of applying Joint Intention theory to protocols also supports flexibility in the actions used to get to landmarks, shortcutting protocol execution, automatic exception handling, and correctness criterion for protocols and protocol compositions
finite state machines;automatic exception handling;conversation protocols;protocol execution;multi-agent interactions;correctness criterion
train_192
New Jersey African American women writers and their publications: a study of
identification from written and oral sources This study examines the use of written sources, and personal interviews and informal conversations with individuals from New Jersey's religious, political, and educational community to identify African American women writers in New Jersey and their intellectual output. The focus on recognizing the community as an oral repository of history and then tapping these oral sources for collection development and acquisition purposes is supported by empirical and qualitative evidence. Findings indicate that written sources are so limited that information professionals must rely on oral sources to uncover local writers and their publications
new jersey african american women writers;written sources;informal conversations;local writers;intellectual output;special collections;personal interviews;collection development;history;oral repository
train_1920
To commit or not to commit: modeling agent conversations for action
Conversations are sequences of messages exchanged among interacting agents. For conversations to be meaningful, agents ought to follow commonly known specifications limiting the types of messages that can be exchanged at any point in the conversation. These specifications are usually implemented using conversation policies (which are rules of inference) or conversation protocols (which are predefined conversation templates). In this article we present a semantic model for specifying conversations using conversation policies. This model is based on the principles that the negotiation and uptake of shared social commitments entail the adoption of obligations to action, which indicate the actions that agents have agreed to perform. In the same way, obligations are retracted based on the negotiation to discharge their corresponding shared social commitments. Based on these principles, conversations are specified as interaction specifications that model the ideal sequencing of agent participations negotiating the execution of actions in a joint activity. These specifications not only specify the adoption and discharge of shared commitments and obligations during an activity, but also indicate the commitments and obligations that are required (as preconditions) or that outlive a joint activity (as postconditions). We model the Contract Net Protocol as an example of the specification of conversations in a joint activity
social commitments;rules of inference;speech acts;conversation protocols;conversation templates;autonomous agents;specifications;interacting agents;software agents
train_1921
An ACL for a dynamic system of agents
In this article we present the design of an ACL for a dynamic system of agents. The ACL includes a set of conversation performatives extended with operations to register, create, and terminate agents. The main design goal at the agent-level is to provide only knowledge-level primitives that are well integrated with the dynamic nature of the system. This goal has been achieved by defining an anonymous interaction protocol which enables agents to request and supply knowledge without considering symbol-level issues concerning management of agent names, routing, and agent reachability. This anonymous interaction protocol exploits a distributed facilitator schema which is hidden at the agent-level and provides mechanisms for registering capabilities of agents and delivering requests according to the competence of agents. We present a formal specification of the ACL and of the underlying architecture, exploiting an algebra of actors, and illustrate it with the help of a graphical notation. This approach provides the basis for discussing dynamic primitives in ACL and for studying properties of dynamic multi agent systems, for example concerning the behavior of agents and the correctness of their conversation policies
agents;actors;anonymous interaction protocol;acl;dynamic system;agent communication languages;dynamic system of agents;distributed facilitator;system of agents
train_1922
Trends in agent communication language
Agent technology is an exciting and important new way to create complex software systems. Agents blend many of the traditional properties of AI programs - knowledge-level reasoning, flexibility, proactiveness, goal-directedness, and so forth - with insights gained from distributed software engineering, machine learning, negotiation and teamwork theory, and the social sciences. An important part of the agent approach is the principle that agents (like humans) can function more effectively in groups that are characterized by cooperation and division of labor. Agent programs are designed to autonomously collaborate with each other in order to satisfy both their internal goals and the shared external demands generated by virtue of their participation in agent societies. This type of collaboration depends on a sophisticated system of inter-agent communication. The assumption that inter-agent communication is best handled through the explicit use of an agent communication language (ACL) underlies each of the articles in this special issue. In this introductory article, we will supply a brief background and introduction to the main topics in agent communication
teamwork;semantics;agent communication language;agent societies;conversations;distributed software engineering;negotiation;social sciences;machine learning;agent technology;ai programs;kqml;inter-agent communication
train_1923
Predictive control of a high temperature-short time pasteurisation process
Modifications on the dynamic matrix control (DMC) algorithm are presented to deal with transfer functions with varying parameters in order to control a high temperature-short time pasteurisation process. To control processes with first order with pure time delay models whose parameters present an exogenous variable dependence, a new method of free response calculation, using multiple model information, is developed. Two methods, to cope with those nonlinear models that allow a generalised Hammerstein model description, are proposed. The proposed methods have been tested, both in simulation and in real cases, in comparison with PID and DMC classic controllers, showing important improvements on reference tracking and disturbance rejection
free response calculation;time delay models;disturbance rejection;nonlinear models;generalised hammerstein model description;predictive control;exogenous variable dependence;multiple model information;dynamic matrix control algorithm;high temperature-short time pasteurisation process;first order processes;reference tracking;transfer functions
train_1924
Existence theorems for nonconvex problems of variational calculus
A solution to a variational calculus problem is studied under the conditions of integrant convexity. The existence theorem is proved. As an example, a function is given, which satisfies all the conditions of the theorem but is not convex
nonconvex problems;variational calculus;integrant convexity;existence theorems
train_1925
On the accuracy of polynomial interpolation in Hilbert space with disturbed
nodal values of the operator The interpolation accuracy of polynomial operators in a Hilbert space with a measure is estimated when nodal values of these operators are given approximately
polynomial interpolation;hilbert space;polynomial operators;disturbed nodal values
train_1926
Simulation of ecological and economical structural-type functions
An optimization approach to the simulation of ecological and economical structural-type functions is proposed. A methodology for construction of such functions is created in an explicit analytical form
simulation;optimisation approach;explicit analytical form;economical structural-type functions
train_1927
Optimal strategies for a semi-Markovian inventory system
Control for a semi-Markovian inventory system is considered. Under general assumptions on system functioning, conditions for existence of an optimal nonrandomized Markovian strategy are found. It is shown that under some additional assumptions on storing conditions for the inventory, the optimal strategy has a threshold (s, S)-frame
semi-markovian inventory system;optimal nonrandomized markovian strategy;optimal strategy;optimal strategies;system functioning
train_1928
Solution of a Euclidean combinatorial optimization problem by the
dynamic-programming method A class of Euclidean combinatorial optimization problems is selected that can be solved by the dynamic programming method. The problem of allocation of servicing enterprises is solved as an example
dynamic programming method;euclidean combinatorial optimization problem
train_1929
Optimal time of switching between portfolios of securities
Optimal time of switching between several portfolios of securities are found for the purpose of profit maximization. Two methods of their determination are considered. The cases with three and n portfolios are studied in detail
profit maximization;portfolios of securities;optimal time