Knowledge Discovery Nuggets Index
To KDNuggets Directory
| Here is how to subscribe to KD Nuggets |
This Year |
Past Issues
Knowledge Discovery Nuggets(tm) 98:24, e-mailed 98-11-08
News:
(text)
Won Kim, ACM SIGKDD update
(text)
Fiona Neil, Quadstone KDD-Cup-98 story
(text)
Maria Zemankova, NSF KDI: New announcement; deadlines 2/1/99, 5/17/99
(text)
Amit Seth, Yahoo Data Mining Forum/Club formed
Publications:
(text)
Xindong Wu, Re: Incremental Data Mining
(text)
Mohammed Zaki, Ph.D. Thesis: Scalable Data Mining for Rules
Tools/Services:
(text)
Iztok Savnik, Siftware: New version of FDEP
(text)
James Newton, Request for comments on SEARCH program
Positions:
(text)
Ronny Kohavi, San Mateo, CA: Blue Martini Software looking for people
(text)
Letty Orozco, Austin, Texas: ITC Data Mining Specialist
(text)
Christophe Giraud, 2 RA Posts in Machine Learning at Bristol, UK
(text)
D. R. Mani, Mountain View, CA: KDD Job at GTE Govt Systems
Courses:
(text)
Bill Goodin, UCLA short course: Data Mining Techniques and
Applications, January 25-28, 1999, Los Angeles
Meetings:
(text)
Belur Dasarathy, SPIE Conference on Data Mining and Knowledge
Discovery, Orlando, April 1999
(text)
David Heckerman, Workshop on AI and Statistics,
Jan 3-6, 1999, Ft. Lauderdale, Florida
http://uncertainty99.microsoft.com/
--
Knowledge Discovery Nuggets (TM) or KDNuggets for short, is an
electronic newsletter focusing on the latest news, publications, tools,
meetings, and other relevant items in the Data Mining and Knowledge Discovery
field. KDNuggets is currently reaching over 5900 readers in 75+ countries
2-3 times a month.
Relevant items are welcome and should be emailed to gps
in ASCII text or HTML format.
An item should have a subject line which clearly describes
what is it about to KDNuggets readers.
Please keep calls for papers and meeting announcements
short (50 lines or less of up to 80-characters), and provide a web site for
details, such as papers submission guidelines.
All items may be edited for size.
To subscribe, see http://www.kdnuggets.com/subscribe.html
Back issues of KD Nuggets, a catalog of data mining tools
('Siftware'), pointers to data mining companies, relevant websites,
meetings, etc are available at KDNuggets Directory at
http://www.kdnuggets.com/
-- Gregory Piatetsky-Shapiro (editor)
gps
********************* Official disclaimer ***************************
All opinions expressed herein are those of the contributors and not
necessarily of their respective employers (or of KD Nuggets)
*********************************************************************
~~~~~~~~~~~~ Quotable Quote ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Experience is that marvellous thing that enables you to recognise a mistake
when you make it again -- F. P. Jones.
(thanks to http://www.geocities.com/Colosseum/3505/quotations.htm
Previous
1 Next Top
Date: Wednesday, November 04, 1998 11:27 AM
From: Won Kim [won.kim@cyberdb.com]
Subject: ACM SIGKDD update
Web:www.acm.org/sigkdd
Here is a brief update on SIGKDD news.
1. Membership drive -- please join ACM SIGKDD -- seewww.acm.org/sigkdd
for details.
2. KDD-99: In addition to KDD-99 main hotel (Marriott Mission Valley),
ACM signed a contract with an overflow hotel for the KDD-99
conference -- Double Tree Hotel in Mission Valley, San Diego.
It is within a one trolley stop from the Marriott Mission Valley.
The guestroom in each hotel will be offered to KDD-99 attendees
at $129 a night.
3. The SIGKDD Awards Committee will consist of
Jiawei Han (Chair), David Hand, Ross Quinlan, Tom Ditterich, Steve Marron,
Heikki Mannila, and Hiroshi Motoda.
The committee will be solciting nominations for
a) Innovations Award:
for significant technical innovations in the field
that have been reduced to practice in significant ways
or that have significantly influenced direction of research
and development in the field
b) Services Award:
for significant services to the field, that include
professional volunteer services in disseminating technical
information to the field, education (i.e., teaching),
research funding, etc.
(current ACM SIGKDD officers are not eligible for Services award).
4. The Editorial Board of the ACM Transactions on Database Systems
journal has agreed to have TODS publish one 'best TODS-relevant' paper
nominated by the KDD conference Program Committee each year.
Previous
2 Next Top
Date: Mon, 02 Nov 1998 13:38:34 +0000
From: Fiona Neil fkn@quadstone.com
Subject: Quadstone KDD Cup story
[GPS: Full description of how Quadstone worked on KDD-Cup-98 is available
at www.kdnuggets.com/meetings/kdd98/quadstone/
]
KDD Cup 98:
Quadstone Take Bronze Miner Award
Decisionhouse
Quadstone's Decisionhouse is a complete and integrated suite of software for customer behaviour modelling.
The Decisionhouse software incorporates all the necessary elements of
database connectivity, analysis, statistics, visualisation, and data mining for
business-focused customer modelling. Decisionhouse is used to
understand customer behaviour for applications including:
- propensity and response modelling
- customer retention and churn prediction
- database marketing
- cross-selling target identification
- customer profitability analysis
- credit scoring.
For more information about Decisionhouse visit the Quadstone web pages at www.quadstone.com
Task
The dataset for this year's competition was provided by the Paralysed Veterans of America (PVA).
The PVA provides programs and services for US veterans and is one of the largest
direct mail fund raisers in the US.
The competition dataset consisted of 191,779 lapsed donor records who received a mailing as part of
larger campaign sent to a total of 3.5 million donors. Lapsed donors were an important group as the longer
someone goes without donating the less likely they are to donate again. It was therefore vital for the PVA
to reactivate these donors.
Competitors were tasked with developing a model that would help the PVA maximize the net
revenue generated from future renewal mailings to these lapsed donors. Competitors were assessed on
the predicted net donations from the campaign.
Full story of Quadstone methodology is at
at www.kdnuggets.com/meetings/kdd98/quadstone/
Previous
3 Next Top
Date: Wed, 4 Nov 1998 09:24:39 -0500
From: Maria Zemankova mzemanko@nsf.gov
Subject: NSF KDI: New announcement; deadlines 2/1/99, 5/17/99
NATIONAL SCIENCE FOUNDATION
KDI: Knowledge and Distributed Intelligence in the Information Age
Proposal Solicitation, NSF 99-29 (replaces NSF 98-55)
http://www.nsf.gov/cgi-bin/getpub?nsf9929
DEADLINES:
PREPROPOSALS (REQUIRED): FEBRUARY 1, 1999
FULL PROPOSALS: MAY 17, 1999
Anticipated date of award: September 1999
INTRODUCTION:
Recent advances in computer power and connectivity are reshaping
relationships among people and organizations, and transforming the
processes of discovery, learning, and communication.These advances create
unprecedented opportunities for providing rapid and efficient access to
enormous amounts of knowledge and information; for studying vastly more
complex systems than was hitherto possible; and, for increasing in
fundamental ways our understanding of learning and intelligence in living
and engineered systems.NSF's Knowledge and Distributed Intelligence (KDI)
theme is a Foundation-wide effort to promote the realization of these
opportunities.Proposals are solicited from individuals or groups for
research that is inherently multidisciplinary or that, while lying within a
single discipline, has clear impact on at least one other discipline.With a
budget of approximately $50 million, KDI anticipates funding 40-50
proposals of varying size and duration.
To achieve the aims of KDI, proposals are solicited from individuals or
groups for research that is inherently multidisciplinary or that, while
lying within a single discipline, has clear impact on at least one other
discipline. (Throughout, the term multidisciplinary is intended to include
interdisciplinary and cross-disciplinary research.)
In FY 1999, KDI will have three foci:
* Knowledge Networking (KN);
* Learning and Intelligent Systems (LIS); and
* New Computational Challenges (NCC).
This document describes the three KDI foci, and serves as a solicitation
for proposals in all three areas. We anticipate that research on many
important problems will span the foci of KN, LIS, and NCC.
CONTACTS FOR ADDITIONAL INFORMATION
Further information on KDI can be found at www.nsf.gov/kdi.
General inquiries should be made via email to kdi@nsf.gov.
For questions related to use of FastLane, contact FastLane Project Officer,
703-306-1145, e-mail: fastlane@nsf.gov.
Cognizant Program Officer(s): Dr. Richard Hilderbrandt, Program Officer,
Room 1055S, MPS/CHE, telephone 703-306-1844, e-mail: rhilderb@nsf.gov.
Previous
4 Next Top
Date: Sun, 1 Nov 1998 13:17:01 EST
From: Amit Seth, AmitSeth@aol.com
Subject: Yahoo Data Mining Forum/Club formed
Where do you go if you have a question on Data Mining? A lot of newsgroups
are present today that handle a variety of areas, but none devoted to this
topic.
The forum called 'Data Mining Club' would help by being a public place
to exchange ideas and/or ask questions on the topic of Data Mining.
You can take a look at the club by going to:
http://clubs.yahoo.com/clubs/datamining
Please register for the club and also please help answer some questions
asked here.
Previous
5 Next Top
Date: Sat, 31 Oct 1998 10:29:27 -0700 (MST)
From: xwu@gauss.Mines.EDU
(Xindong Wu)
Subject: Re: Incremental Data Mining
Further to Foster Provost's posting in KDNuggets 98:23 on incremental
data mining, I have some general comments and a few references to add.
Incremental data mining is not the same as incremental learning. The
latter generally handles training data items one by one, while the
former handles subsets (or samples) of a database one by one
sequentially. Quinlan's windowing technique (as mentioned in Foster
Provost's posting) is along this line. There have been discussions
about the costs vs benefits of windowing, and [Wirth and Catlett 1998]
has been cited in many places in this regard.
- J. Wirth and J. Catlett, Experiments on the Costs and Benefits of
Windowing in ID3, Proceedings of the Fifth International Conference
on Machine Learning, J. Laird (Ed.), Morgan Kauffman, 1988.
Incremental batch learning [Clearwater et al 1989] and multi-layer
incremental induction [Wu & Lu 1998] construct classifications through
data sampling/partitioning and knowledge refinement across more than
one learning run. Some research efforts have concentrated on the
run-time complexity (as mentioned in Foster Provost's posting), and
others aim to achieve better predictive accuracy in noisy domains,
because data partitioning is expected to dilute the effects of noise
into data subsets.
- S.H. Clearwater, T.P. Cheng, H. Hirsh, and B.G. Buchanan,
Incremental Batch Learning, Proceedings of the Sixth International
Workshop on Machine Learning, Morgan Kaufmann, 1989.
- X. Wu and W. Lo, Multi-Layer Incremental Induction, Proceedings of
the 5th Pacific Rim International Conference on Artificial
Intelligence, Springer-Verlag, 1998.
Recent research efforts on learning ``ensembles of classifiers'' are
relevant to but significantly different from incremental data mining.
Learning ``ensembles of classifiers'' such as bagging and boosting
generates multiple versions of a predictor by running an existing
learning algorithm many times on a set of re-sampled data, and
combines them to get an aggregated prediction. Each version of the
predictor is generated from a sample of the original database, and a
data item in the original database can be used in many samples for
generating different versions of the predictor. In incremental data
mining (at least in multi-layer incremental induction), each data item
in the original database is partitioned into only one subset, and used
only once in the induction process.
Similar to bagging and boosting, stacked generalization [Wolpert 1992,
Ting & Witten 1997] uses a high-level model to combine lower-level
models to achieve greater predictive accuracy. These lower-level
models are generated also by sampling the original database, and
cross-validation is used to generate higher-level data for the
high-level model.
- D.H. Wolpert, Stacked Generalization, Neural Networks, 5 (1992),
241--259.
- K.M. Ting & I.H. Witten, Stacked Generalization: When Does It Work?
Proceedings of the 15th International Joint Conference on
Artificial Intelligence (IJCAI-97), 866--871.
Previous
6 Next Top
Date: Fri, 30 Oct 1998 19:28:35 -0500 (EST)
From: Mohammed Zaki zaki@cs.rpi.edu
Subject: Ph.D. Thesis: Scalable Data Mining for Rules
Web: http://www.cs.rpi.edu/~zaki.
My thesis is now available on-line, at http://www.cs.rpi.edu/~zaki.
It describes scalable and parallel solutions for association rules,
sequence mining and decision tree classification. Abstract follows.
- Mohammed Zaki
Asst. Prof., CS Dept., Rensselaer Polytechnic Inst., Troy NY
===================================================================
Scalable Data Mining for Rules
Mohammed J. Zaki
Ph.D. Thesis, University of Rochester, July 1998
Data Mining is the process of automatic extraction of novel, useful,
and understandable patterns in very large databases. High-performance
scalable and parallel computing is crucial for ensuring system
scalability and interactivity as datasets grow inexorably in size and
complexity. This thesis deals with both the algorithmic and systems
aspects of scalable and parallel data mining algorithms applied to
massive databases. The algorithmic aspects focus on the design of
efficient, scalable, disk-based parallel algorithms for three key rule
discovery techniques --- Association Rules, Sequence Discovery, and
Decision Tree Classification. The systems aspects deal with the
scalable implementation of these methods on both sequential machines
and popular parallel hardware ranging from shared-memory systems (SMP)
to hybrid hierarchical clusters of networked SMP workstations.
The association and sequence mining algorithms use lattice-theoretic
combinatorial properties to decompose the original problem into small
independent sub-problems that can be solved in main memory. Using
efficient search techniques and simple intersection operations all
frequent patterns are enumerated in a few database scans. The parallel
algorithms are asynchronous, requiring no communication or
synchronization after an initial set-up phase. Furthermore, the
algorithms are based on a hierarchical parallelization, utilizing both
shared-memory and message-passing primitives. In classification rule
mining, we present disk-based parallel algorithms on shared-memory
multiprocessors, the first such study. Extensive experiments have
been conducted for all three problems, showing immense improvement
over previous approaches, with linear scalability in database size.
Previous
7 Next Top
Date: Mon, 2 Nov 1998 09:09:22 +0100 (MET)
From: Iztok Savnik, Iztok.Savnik@fri.uni-lj.si
Subject: Siftware: New version of FDEP
Web: http://infolab.fri.uni-lj.si/~savnik/fdep.html
We would like to announce thru the KDD-Nuggets distribution
list a new version of FDEP - program for the discovery of
functional dependencies from relations. In comparison to
the previous version (v1.0), FDEP 1.1 includes the following
improvements:
- novel algorithm for the discovery of functional dependencies,
- support for unknown values,
- method for handling noisy data,
- sample domains are included in the distribution,
- increased maximal number of relation tuples, and,
- increased maximal number of relation attributes.
FDEP distribution (including source code, documentation
and sample relations) can be obtained at
http://infolab.fri.uni-lj.si/~savnik/fdep.html
Previous
8 Next Top
Date: Mon, 2 Nov 1998 09:54:49 -0500 (EST)
From: james newton morgan jnmorgan@umich.edu
Subject: Request for comments on some proposed enhancements to SEARCH program
The SEARCH program from OSIRIS (formerly Automatic Interaction
Detector or AID) is available in MicrOsiris for the PC from
vaneckn@philacol.edu.
We are thinking of some enhancements to it and
would appreciate comments. It is designed for analyzing surveys or
subsamples with 1,000-30,000 cases where there is a criterion variable
or classification to be accounted for. It also allows a covariance
search where there is one dominant explanatory factor like income,
gender or race. We propose the following enhancements:
1. In addition to a tree diagram, a hierarchical summary table with
one line for each group, indented for subgroups and with final groups
indicated. This would allow more information about each group (cases,
mean, distribution for a categorical criterion, or simple regression
for covariance analysis). It should be simple for the user to alter
the table to make the definitions of each split more understandable.
2. For a ranked dependent 'variable', a split criterion based on rank
correlation, probably Kendall's Tau-b. Using the likelihood ratio chi
square designed for a categorical criterion loses information about
rank order.
3. Extend the facility to produce the recode for expected values or
residuals to include categorical dependent variables (chi or rank),
where the expected values would be a distribution rather than a mean.
4. Make the use of a 'runfile' easier, with the whole analysis
strategy including data source, filter, recode, and analysis, in a
separate file so there is a record and one can alter for reruns and
save for the files.
5. Provide a fourth possible stopping rule based on a miniumu t-ratio
or maximum null probability, even though such tests are improper after
extensive searching and with survey data with design effects. (And
exclusive reliance on significance can isolate very small homogenous
groups which would be useless in predicting back to the population.)
Currently one can specify minimum final group size, maximum number of
splits, and/or minimum reduction in unexplained variance relative to
original total.
6. Since datafiles are reasonably standard rectangular asci files, but
dictionaries are not, stay with separate data and dictionary files,
but provide the recode to convert SAS or SPSS and perhaps some other
dictionaries to an OSIRIS dictionary.
james n. morgan e-mail: jnmorgan@umich.edu
institute for social research phone 313 764 8388
po box 1248 home 313 668 8304
ann arbor, mi 48109 fax 313 747 4575
Previous
9 Next Top
Date: Mon, 02 Nov 1998 13:22:38 -0800
From: Ronny Kohavi ronnyk@bluemartini.com
Subject: Ron Kohavi at Blue Martini Software looking for people
Web: http://www.bluemartini.com/
Blue Martini Software, a startup combining two of the hottest
technologies: data mining and e-commerce, is hiring.
Ronny Kohavi, director of data mining applications at Blue Martini,
is looking for at least two engineers to join the company and
work on the data mining component of the system.
Responsibilities include developing, modifying, interfacing machine
learning and statistical algorithms, and integrating them into the larger
e-commerce solution.
Applicants should have extensive programming experience with C++, an
MS or Ph.D. degree in Computer Science (or equivalent), and prior knowledge
of machine learning/data mining.
Interested? Send e-mail to ronnyk@bluemartini.com
and CC hr@bluemartini.com
More information:
- Before joining Blue Martini Software, Ronny kohavi managed the
MineSet(TM) data mining and visualization project at Silicon Graphics
http://mineset.sgi.com,
using MLC++ http://www.sgi.com/Technology/mlc/,
the Machine Learning library in C++, which he headed at Stanford University
and then at SGI.
- Information about Blue Martini can be found at http://www.bluemartini.com/
Look at the board of directors and management and you'll be impressed.
We are located in San Mateo, California.
-- Ronny Kohavi
Previous
10 Next Top
Date: Thu, 5 Nov 1998 12:01:39 -0600
From: Letty Orozco letty@itctx.com
Subject: Job: TX-US-Austin Data Mining Specialist
Intelligent Technologies Corporation (ITC) provides fraud detection
solutions for healthcare and insurance industries. ITC has developed
end-to-end solutions for Medicaid fraud management and tax compliance
management based on advanced pattern recognition technologies. ITC's
products and related services have saved businesses and consumers hundreds
of millions of dollars annually by detecting and helping prevent fraudulent
transactions. ITC is looking for key individuals to become part of a highly
motivated team working with hot technologies in a company with great growth
potential. Be a part of the success.
DATA MINING SPECIALIST
The Austin, Texas office of ITC is looking for a candidate with applied
mathematics/engineering/CS degree and 2-5 years of experience in data
mining. Candidates not requiring relocation to Texas are preferred. The
ideal candidate will have a broad background and the following skill sets.
* Applications of neural networks, decision trees, fuzzy logic and statistics
* Experience in commercial data mining tools (Clementine, SAS, etc.)
* Working knowledge of relational data bases and SQL
* Good communication skills
We offer excellent compensation plus a complete benefits package including
an employee stock option plan. For consideration, send your resume
including salary history to: ITC, Job Code E, 9015 Mountain Ridge Drive,
Suite 350, Austin, TX, 78759. Or fax to (512) 343-1608. Or e-mail to:
hello@itctx.com.
Check out our Web site at: http://www.itctx.com
Previous
11 Next Top
Date: Wed, 21 Oct 1998 13:34:43 +0100 ()
From: Christophe Giraud-Carrier cgc@cs.bris.ac.uk
Subject: 2 RA Posts in Machine Learning at Bristol
The Machine Learning Research Group at the University of Bristol
offers two Research Assistantships. One assistantship (RA1) is for a
36-month duration and must be filled as soon as possible. The other
(RA2) is for a 26-month duration and would start by 1 May 1999.
The 2 assistantships focus on machine learning and data mining, and
are part of a European Long-term Reactive Research Project entitled
'METAL: A Meta-Learning Assistant for Providing User Guidance in
Machine Learning and Data Mining'. The main research themes of the
project are: model selection (choice of an appropriate learning
model/algorithm for a given application task), data screening and
data transformation prior to knowledge extraction, meta-learning
approaches to model selection and data transformation. The project
involves 6 partners (2 industrial and 4 academic) and its results
will be implemented in the commercial system Clementine.
Ideally, candidates should have: (1) a Ph.D. in Computer Science;
(2) a strong background in artificial intelligence, machine learning
(both symbolic and numeric), and data modelling/analysis/mining;
and (3) experience in software development (e.g., implementing
prototypes).
Starting salary for both posts will be on the Grade 1A scale, up to
point 5 depending on qualifications (between 15,462 and 18,864 sterling
pounds per annum). If you wish to be considered for any of these
positions, please send your curriculum vitae, together with the names
and addresses of 2 references, to the following contact person:
Dr. Christophe Giraud-Carrier cgc@cs.bris.ac.uk
Electronic applications are encouraged for reasons of speed; please
specify the reference of the post for which you are applying.
For those with no e-mail facilities, applications can be sent by
fax or regular mail to the above contact person (address and
fax number below):
Department of Computer Science
University of Bristol
Merchant Venturers Building
Woodland Rd
Bristol, BS8 1UB
United Kingdom
Fax: +44-117-954-5208
Applications will be accepted until the positions are filled.
Previous
12 Next Top
Date: Tue, 03 Nov 1998 15:26:16 -0500
From: 'D. R. Mani' mani@gte.com
Subject: KDD Job Opening at GTE Govt Systems, Mountain View, CA
An Applied Researcher/Developer needed to lead the
Knowledge Discovery initiative at GTE Government Systems
RESPONSIBILITIES: Lead the Knowledge Discovery initiative at GTE
Government Systems in the Electronic Systems Division (ESD).
Participate in the design and development of a wide range of
state-of-the-art systems for data mining and knowledge discovery --
including systems for information assurance, intrusion detection and
security, electronic commerce and multi-media data mining. Develop
and grow in-house knowledge discovery expertise through additional
training and hiring.
QUALIFICATIONS: The ideal candidate will have a Master's or Ph.D
Degree in Computer Science or a related field with proven technical
abilities in the areas of knowledge discovery and datamining (KDD),
machine learning (neural networks, decision trees, genetic
algorithms, clustering, etc.) and artificial intelligence
(rule-based systems, expert systems, etc.). In addition to
technical excellence, specific experience in dealing with complex
real-world problems and large data sets is important. Good
communication and presentation skills for technical, semi-technical
and non-technical audiences is a strong asset.
United States citizenship is required in order to obtain the highest
levels of security clearance.
WORK ENVIRONMENT: The Electronics Systems Division (ESD) at GTE
Government Systems engineers specialized, state-of-the-art
telecommunications and data communication systems, and develops
knowledge management and assurance infrastructure for these systems
over intranet, extranet and Internet environments. ESD subscribes to
a customer-centric business model, and uses technical expertise,
talent and innovation to support customer needs. Projects involve
use of a wide range of cutting-edge technologies, contributing to a
challenging (and sometimes demanding), team-oriented and proactive
work environment.
ESD promotes a professional, task oriented environment with the
option for flexible hours/telecommuting and excellent benefits
(including every other Friday off). ESD is head quartered in
Mountain View CA, with locations in Washington DC, Baltimore MD,
Thousand Oaks CA, and Colorado Springs CO.
CONTACT INFORMATION:
Please send a resume and a cover letter to
Wade F. Schott
Director, Technology and Product Development
100 Ferguson Drive
P. O. Box 7188 MS 7316
Mountain View, CA 94039
Electronic submissions should be directed to:
wade.schott@gsc.gte.com
Previous
13 Next Top
Date: Fri, 30 Oct 1998 17:41:23 -0800
From: Bill Goodin, bgoodin@unex.ucla.edu
Subject: UCLA short course on 'Data Mining Techniques and Applications'
Web: http://www.unex.ucla.edu/shortcourses
On January 25-28, 1999, UCLA Extension will present the short course,
'Data Mining Techniques and Applications' on the UCLA campus in
Los Angeles.
The instructors are Wei-Min Shen, PhD, USC Information Sciences
Institute; Rakesh Agrawal, PhD, IBM Almaden Research Center; and
Jiawei Han, PhD, Simon Fraser University.
This course is intended for scientists, engineers, and information
managers who need to learn and apply data mining techniques (tools
for discovering valuable knowledge from very large data sets) to their
scientific research, system design, business management, or any
other related applications. The lecturers are among the world-leading
experts in the field with extensive experience in basic research as
well as in real industrial application.
This course should enable participants to understand and have hands-on
experience in:
o Basic concepts of data mining
o The overall process of data mining
o Critical steps in the data mining process
o Relationships between data mining and other scientific disciplines
o Formalizing data mining problems
o Data preprocessing
o Data classification
o Data clustering
o Database structures and their operations
o Time serial data analysis
o Visualization
o Prediction and forecasting
The course fee is $1395, which includes extensive course materials.
These materials are for participants only, and are not for sale.
For additional information and a complete course description, please
contact Marcus Hennessy at:
(310) 825-1047
(310) 206-2815 fax
mhenness@unex.ucla.edu
http://www.unex.ucla.edu/shortcourses
This course may also be presented on-site at company locations.
Previous
14 Next Top
Date: Fri, 30 Oct 1998 17:24:21 -0800 (PST)
From: Dr. Belur V. Dasarathy belur@yahoo.com
Subject: SPIE Conference on Data Mining and Knowledge Discovery, Orlando, April 1999
Web: http://members.tripod.com/~belur/kddprog.html
The Final program, including dates/times etc., for the SPIE Conference
on Data Mining and Knowledge Discovery: Theory, Tools, and Technology,
being held in Orlando, in April 1999 has now been published on the web
at http://members.tripod.com/~belur/kdd.html
Click on 'Final Program' to see the details.
Belur V. Dasarathy
Conference Chairman
Previous
15 Next Top
Date: Fri, 30 Oct 1998 13:24:35 -0800
From: David Heckerman heckerma@MICROSOFT.com
Subject: Call for participation: Workshop on AI and Statistics, Jan 1999, Florida
Web: http://uncertainty99.microsoft.com/
Call For Participation
Uncertainty 99: Seventh International Workshop on
Artificial Intelligence and Statistics
January 3-6, 1999
Ft. Lauderdale, Florida
http://uncertainty99.microsoft.com/
PURPOSE:
This is the seventh in a series of workshops that has brought together
researchers in Artificial Intelligence and in Statistics to discuss
problems of mutual interest. The exchange has broadened research in
both fields and has strongly encouraged interdisciplinary work.
Attendance at the workshop is *not* limited to paper presenters.
TECHNICAL PROGRAM:
To encourage interaction and a broad exchange of ideas, there will be
20 discussion papers in single session meetings over three days
(Jan. 4-6). A poster session and brief plenary summaries of
posters prior to these sessions will provide the means for presenting
and discussing the remaining research papers. Papers accepted for
plenary and poster presentation are listed at the web site listed above.
[...edited. GPS]
For more information, including
Registration forms, a provisional program, a list of the program
committee, and other details about the conference and the Society for
Artificial Intelligence and Statistics can be found at
http://uncertainty99.microsoft.com/.
David Heckerman and Joe Whittaker,
Conference Chairs
Previous
16 Next Top