III.
Symposium on Data, Information and Knowledge: Earth and Physical
Sciences, Engineering and Industry
Prof. Alexei Gvishiani
CGDS, IPE RAS
Centre of Geophysical Computer Data Studies and Telematics Applications
Molodezhnaya Str. 3
117296 Moscow, Russia
Phone: +7 095 133 4339
Fax: +7 095 930 5559
Email: gvi@wdcb.ru
Theme
III-1: Volcanological, Seismological and Other Geophysical
Data Management and Studies in the 21st Century Prof.
Claude Jaupart (IPG, France) and Prof. A. Gvishiani (UIPE RAS,
Russia)
Earth Science in general
and volcanological, seismological and other geophysical studies
in particular, is one of the fields, where modern informatics
and computer networking techniques of data handling, management
and disssemination is becoming nowadays crucial. Since data sets
under consideration become more and more huge, new methods of
teleworking, emerging techniques for the information visualization
and pattern recognition methodologies are becoming critical issues
in geophysical data studies and management.
The session
"Volcanological, Seismological and Other Geophysical Data Management
and Studies in the 21st Century" will focus on a number of aspects
of application of the different kinds of new methods of data processing
to Earth science data. It will foster an exchange of ideas and
methodologies on different aspects of volcanological, seismological
and other geophysical data handling, management and processing.
The principle objective of the session is to bring together active
researchers, data and knowledge base developers and administrators
along with applied mathematicians and software developers to present
their papers and to discuss the following themes:
- Geomagnetic and
gravity data centers and databases;
- Seismological and
engineering seismology data acquisition, handling and dissemination;
- Volcanological
data management;
- Identifying crucial
issues in Earth Sciences data management in the 21st century;
Possible topics include:
- Seismological data
centers and databases;
- Volcanalogical
data management;
- Geomagnetic and
gravity databases and data centers;
- Teleworking in
geophysical data handling and management;
- Virtual laboratory
approach to geophysical data handling;
- Vizualization technique;
- Algorithms of processing;
- Mathematical models;
- Pattern recognition
approach;
- Fuzzy logic approach
Theme
III-2: Environmental Risk Management: Earthquakes, Floods and
other Natural Calamities Prof. Jean Bonnin (IPG, Strasbourg,
France)
Along with Earth Science
, environmental risk management, such as management of data produced
by observations of earthquakes, floods, and other natural calamities
is one of the fields, where modern informatics and computer networking
technique of data handling, management and dissemination is becoming
nowadays crucial. Since environmental and natural risk studies
and corresponding practical operations require fast decision making,
new methods of teleworking and pattern recognition classification
techniques play more and more important role in this field.
The session "Environmental
risk management: earthquakes, floods, other natural calamities"
will focus on a number of aspects of the application of the different
kinds of new methods of data processing to evaluation and prevention
of environmental risk. It will foster an exchange of ideas and
methodologies in handling, managment and processing of environmental
risk data. The principle objective of the session is to bring
together active researchers, data and knowledge base developers
and administrators along with applied mathematicians and software
developers to present their papers.
The session will consist
of invited papers, short papers, posters and computer demonstrations,
possibly on the following topics:
- Floods data management;
- Seismological and
engineering seismology data acquisition, handling and dissemination;
- Internet in environmental
risk;
- Identifying crucial
issues in environmental data management in the 21st Century;
- Real time operations;
- Multidisciplinary
approach to environmental data handling;
- Artificial intelligence
approach;
- Virtual observatories
for environmental data handling
-
Theme
III-3: Oceans: Marine Geology and Marine Geophysics Data Acquisition,
Management and Processing Prof. Jean Bonnin and Prof. J. O.
Dubois
Oceanic, marine geology
and marine geophysics .studies is one of the fields where modern
informatics and computer networking techniques of data handling
and management are really important nowadays The data sets in
the above studies become more and more huge,. Therefore, new methods
of teleworking, emerging techniques for information visualization
and artificial intelligence approach are becoming useful and important
in data studies, management and processing. The session "Oceans:
marine geology and marine geophysics data acquisition, management
and processing" will focus on a number of aspects of the applications
of new methods of data processing to marine geology and geophysics
data. It will foster an exchange of methodologies and techniques
to study different aspects of the oceanic crust and the lithosphere.
The goal of the session
is to provide a forum for theoretical and experimental researchers
and data and knowledge base developers and administrators along
with applied mathematicians and software developers. The session
will consist of invited papers, short papers, posters and computer
demonstrations. Possible topics for this session include:
- Multi-beams bathymetry
data acquisition and databases;
- Gravity and magnetic
field mapping in the oceans (surface and sea-bottom data);
- Seismic survey
reflection and refraction techniques;
- Sea-bottom observatories
operated by submarines;
- Spatial approach
including altimetry, gravity and geomagnetism;
- Sampling of rocks
and sediments from the sea-bottom
Theme
III-4: Physico-Chemical Data Standards and Databases Dr.
H. Kehiaian (France)
The main topic of this
session will be the development of new approaches for the
electronic publication, collection and dissemination of physico-chemical
data, of data input, storage and exchange standards. The goal
of the session is to demonstrate and discuss how data can be made
easily and widely accessible in the Internet age.
The session will consist
of invited papers, short papers, posters and computer demonstrations,
possibly in the following areas:
- Actual needs
for critically evaluated data of thermodynamical, transport,
and interfacial properties of pure substances and mixtures;
- New models for
the calculation, correlation, evaluation, and prediction of
thermophysical properties of pure substances and mixtures;
- Use of thermophysical
data in process simulation packages;
- Databases in chemical
engineering including database demonstrations
Theme
III-5: Modern materials design: Databases, combinatorial design,
virtual materials and knowledge discovery Prof. S. Iwata (Japan)
and Dr. K. Rajan (US)
Information technology
(IT) forms the foundation on which the next metamorphosis of Materials
Science will be built . In the last decade the pharmaceutical
industry has adopted IT as its primary tool in synthesizing and
screening new drugs. "Libraries" of 103 to
106 distinct compounds are routinely created and tested
for biological activity. This is now practical because of the
convergence of low cost computer systems, reliable robotic systems,
sophisticated molecular modeling, statistical experimental strategies
and datebase software tools, all the various components of an
IT infrastructure.
This extraordinary
explosion in information and data however has now lead to concerns
of how to manage and interpret that data / information. The science
of managing this information using IT tools or what one may term
"informatics" is as critical as the gathering of new
information itself. The impact of information technology is nowhere
more apparent than in the increasing proliferation and complexity
of materials science data. This wealth of information provides
a powerful resource for industry to accelerate the development
of new materials and new processing techniques for emerging technologies.
In order to harness that information , one needs to develop strategies
based on sound experimental and theoretical foundations. The integration
of materials science fundamentals into the science of information
processing and transmission or Materials Informatics is
the theme for this symposium.
The
objective of this symposium is to address the issue of how one
goes beyond the mere storage and retrieval of data sets in materials
science. Traditionally, computers have been used as storage mediums
for large volumes of data and as tools for carrying out extensive
numerical computations and simulations. Recently, computers have
started to take a more active role in guiding the scientist through
the research and discovery process with the help of data mining
methods for automatic discovery of patterns in large volumes of
data. The challenge is now to make these data mining methods ubiquitous
and an integral part of the data collection and verification process.
Materials
Science offers a unique challenge in data mining due to the variety
of data types, and their complex interconnections. During the
material discovery process, there is a need to integrate multiple,
heterogeneous databases to reach new and even unexpected conclusions
as well as to use databases actively to design new processing
strategies. This complex coupling of data models, data analysis
methods and physical methods offer a unique computing challenge
that has not yet been addressed sufficiently in information technology
research. This symposium aims at establishing an interdisciplinary
dialogue bringing together domain specialists in materials science,
computer science, mathematics and statistics. The symposium will
be directed towards discussing the methodology for Materials Science
to accelerate the discovery of new materials and materials
based design strategies, which we term as Materials Informatics.
Some
of the topics include:
- Databases and information
systems in materials science
- Virtual experimentation
- Intelligent processing
of materials
- Informatics strategies
in combinatorial materials science
- Informatics based
utilization of materials
Theme
III-6: Tools for Data Validation and Quality Control Dr.
P. Mezey (Canada) and Dr. D. Lide (US)
This session will deal
with methodologies and tools for checking the validity of scientific
data and selecting the most reliable values for recommended use.
The concept of data evaluation and selection dates to the International
Critical Tables project in the 1920s. The process was refined
for thermodynamic data in the 1930s and for atomic and nuclear
physics data in the 1940s and 1950s, and it was extended to many
other fields in the ensuing decades. This classical approach was
highly labor intensive and often somewhat subjective; experienced
scientists appraised the data taken from the primary literature,
using their own judgment to discard questionable values and make
the final selection of values to recommend. Varying degrees of
automation have been introduced in the last 25 years as computers
became available; nevertheless, human intervention by experts
is still needed in most fields, and the expense of this human
effort becomes more and more prohibitive as the amount of data
generated by experiment and observation increases.
The talks in this session
will illustrate the growing use of automated approaches to data
validation. Such approaches attempt to capture the scientific
judgment of humans who are intimately familiar with the particular
measurement process, but the details differ from field to field.
The areas to be discussed
include:
- Crystallographic
data
- Nuclear structure
and interaction data
- Fundamental physical
constants
- Thermophysical properties
of fluids
- Physical constants
of organic compounds
Theme
III-7: Interoperability and metadata standards Dr. J. Rumble
(US) and Dr. Peter Murray-Rust (UK)
The World Wide Web
is revolutionizing our acess to scientific and technical data.
Today and in the future, scientists and engineers will be able
to access virtually every data resource they will need from their
own desk. Databases, computational packages and other software
tools will be combined by the individual user as appropriate for
his or her needs. This scenario, however, cannot succeed unless
the needed databases and other software can work together harmoniously
with minimal efforts by the user.
The key to this integration
is interoperability, that is, the ability of database and software
to work together with little or no modification. Interoperability
works at many levels, but three levels present the most siginficant
challenges: accessiblity standards; physical format standards;and
metadataa (content) standards. Many accessibility standards are
already in place (TCP/IP) and have led to the power of the world
wide web. Physical format standards, that is, standards addressing
the structure of the bits and bytes transmitting information are
constantly evolving as new technology develops. Today emphasis
is being given to development of powerful mark-up languages that
provide a simple yet powerful physical structure to data and which
can be extended to numerous disciplines and content domains.
Metadata standards
are concerned with the content of databases and the flow of data
to and from software. These standards address the need for common
vocabulary and nomenclature, so that when different databases
and software packages can work together without the need for specialized
programming to translate data from one representation to another.
At the same time, the development of metadata standards must take
into account the existence of many different nomenclature systems
and database structures, each of which has validity in their own
context. Metadata standards must capture the full essence of each
scientific discipline, but be respectful of the needs of specific
applications.
This session of CODATA
2000 contains several presentations that discuss the state-of-the-art
of interoperability and metadata standards development. It will
present an overview of the major components of interoperability
and different approaches to achieve it. The session then presents
talks by leading experts on specific eactivities so that the attendee
will be exposed to the latest developments in this area.
Prospective talks may
include:
- Interoperability:
An overview and why it is important
- Mark-up languages
- Developing DTDs
- Interoperability
in Bioinformatics - USA
- Interoperability
in Bioinformatics - Europe
- Interoperability
and industrial data - Oil and Gas
- Interoperability
in Regulatory Affairs
- How does a community
develop a DTD (Schema, etc.)?
- Materials MatML
Theme
III-8: Databases and new analytical chemistry systems Dr.
G. Kramer (US) and Dr. R. Schaefer (Germany)
Theme
III-9: Industrial Coatings Prof. A. Revillon (France)
Surface coatings have
a double purpose: to protect and decorate objects of various
natures. This means using treatments adapted to the systems and
the use expected of them. Similarly to other industries, the paint
industry has changed in line with the Computing and Communications
Revolution. There is a better understanding of the nature of the
paints and the use of new components. Better modelizations based
on Computer Assisted Information Handling and Computer Assisted
Production Managing (CAPM) are essential informatic tools and
link users needs to professional expertise and basic knowledge.
Current concerns of the paint industry are:
- technical (improving
products the applications of which are subject to scientific
constraints e.g., temperature, medium, or achieving
new products and new applications)
- economic (low cost
products or those with high added value, manpower costs, sales
improvement by impact of products, strong international competition)
- ecological
(harmful effects, diminishing percent of solvents, elimination
pigments of heavy metals).
The session will consist
of invited papers, short papers, posters and computer demonstrations.
Some topics that may be discussed are:
- Certification of
quality, labeling constraints (standards, approvals), multiplication
and distribution of diverse instruments (for composition, structure,
property measurements) that requires and insures a good definition
of the product;
- Replacing known,
but toxic components, involves searching for new solutions (new
chemical structures, water based, high solid content, powders,
radiation cure) to ensure at least the same properties and the
same lasting qualities;
- Merging firms means
site closures and real time management of factories (reagents,
products, stock, adjustment adaptability and flexibility of
products, controls) helped by data measurement, acquisition
and handling, clever procedures;
- Extending markets
and transport facilities make it possible to closely calculate
cost (manpower, energy, raw material, rulings) and markets (either
specialised or ordinary use products) and encourage international
flow.
Theme
III-10: Modern Construction Industries Session Dr. L. Kaetzel
(US)
This session will discuss
the design, construction or management of constructed facilities
and materials, products, and systems for construction industry
application, with an emphasis on the management of scientific
and technical data, information and knowledge. Papers will address
activities and issues beyond project-specific goals and will follow
this criteria:
Theme
III-11: Flavor and Food Industry Issues : Analytical Tools
and Public Expectation Prof. C. Bicchi (Italy), Prof. R. Fellous
and Prof. P. Schreier (Germany)
Although the food industry
has always been perceived as a complex science visible in daily
life, the new information society, with its media components,
gives it new and exceptional dimensions. New strategies to improve
food security are discussed before the public at large, and suggested
solutions create new challenges for the food industry and its
experts. This industry is faced with cognitive factor
which are clearly becoming part of a new, more complex situation
involving consumer reactions.
The session will present
and discuss different approaches to this complexity and may cover:
Quality and authenticating
natural products or extracts
- Latest analytical
challenges;
- Human descriptor
assistance;
- Need for standards
and references;
- Case studies and
modeling;
Consumer protection
- Food security and
sanitary context;
- Producer/consumer
chains;
- Specific health
claims and public demand;
- Security and traceability
concepts;
- Emergence of new
legislation;
The session will consist
of selected talks, short papers, posters and eventually, computer
demonstrations.
|