1. Trang chủ
  2. » Công Nghệ Thông Tin

Data Mining and Knowledge Discovery Handbook, 2 Edition part 47 potx

10 374 0

Đang tải... (xem toàn văn)

THÔNG TIN TÀI LIỆU

Thông tin cơ bản

Định dạng
Số trang 10
Dung lượng 363,24 KB

Nội dung

440 G. Peter Zhang Dai Y., Nakano Y. (1998), Recognition of facial images with low resolution using a Hopfield memory model. Pattern Recognition ; 31:159-167. Dasu T., Johnson T. (2003), Exploratory Data Mining and Data Cleaning. New Jersey: Wiley . De Groot D., Wurtz D. (1991), Analysis of univariate time series with connectionist nets: A case study of two classical examples. Neurocomputing ;3:177-192. Deboeck G., Kohonen T. (1998), Visual Explorations in Finance with Self-organizing Maps. London: Springer-Verlag. Delen D., Sharda R., Bessonov M. (2006), Identifying significant predictors of injury severity in traffic accidents using a series of artificial neural networks Accident Analysis and Prevention ; 38:434-444. Dhar V., Chou D. (2001), A comparison of nonlinear methods for predicting earnings sur- prises and returns. IEEE Transactions on Neural Networks ; 12:907-921. Dia H. (2001), An object-oriented neural network approach to short-term traffic forecasting. European Journal of Operation Research ; 131:253-261. Dittenbach M., Rauber A., Merkl, D. (2002), Uncovering hierarchical structure in data using the growing hierarchical self-organizing map. Neurocompuing ; 48:199-216. Doganis P., Alexandridis A., Patrinos P., Sarimveis H. (2006), Time series sales forecasting for short shelf-life food products based on artificial neural networks and evolutionary computing. Journal of Food Engineering ; 75:196-204. Dutot A.L., Rynkiewicz J., Steiner F.E., Rude J. (2007), A 24-h forecast of ozone peaks and exceedance levels using neural classifiers and weather predictions Modelling and Software; 22:1261-1269. Dutta S., Shenkar S. (1993), “Bond rating: a non-conservative application of neural net- works.” In Neural Networks in Finance and Investing, Trippi, R., and Turban, E., eds. Chicago: Probus Publishing Company. Enke D., Thawornwong S. (2005), The use of data mining and neural networks for forecast- ing stock market returns. Expert Systems with Applications ; 29:927-940. Evans O.V.D. (1997), Discovering associations in retail transactions using neural networks. ICL Systems Journal ; 12:73-88. Fahlman S., Lebiere C. (1990), “The cascade-correlation learning architecture.” In Advances in Neural Information Processing Systems, Touretzky, D., ed. . Fletcher R. (1987), Practical Methods of Optimization 2 nd . Chichester: John Wiley & Sons . Frean M. (1990), The Upstart algorithm: a method for constructing and training feed-forward networks. Neural Computations ; 2:198-209. Funahashi K. (1998), Multilayer neural networks and Bayes decision theory. Neural Net- works ; 11:209-213. Gallinari P., Thiria S., Badran R., Fogelman-Soulie, F. (1991), On the relationships between discriminant analysis and multilayer perceptrons. Neural Networks ; 4:349-360. Geman S., Bienenstock E., Doursat T. (1992), Neural networks and the bias/variance dilemma. Neural Computation ; 5:1-58. Gorr L. (1994), Research prospective on neural network forecasting. International Journal of Forecasting ; 10:1-4. He H., Wang J., Graco W., Hawkins S. (1997), Application of neural networks to detection of medical fraud. Expert Systems with Applications ; 13:329-336. Hebb D.O. (1949), The Organization of Behavior. New York: Wiley. Hinton G.E. (1992), How neural networks learn from experience. Scientific American; 9:145-151. 21 Neural Networks For Data Mining 441 Hornik K., Stinchcombe M., White H. (1989), Multilayer feedforward networks are universal approximators. Neural Networks ; 2:359–366. Hopfield J.J. (2558), (1982), Neural networks and physical systems with emergent collective Hopfield J.J., Tank D.W. (1985), Neural computation of decisions in optimization problems. Biological Cybernetics ; 52:141-152. Hu J.Q., Rose, E. (1995), On-line fuzzy modeling by data clustering using a neural network. Advances in Process Control. , 4, 187-194. Huang J.S., Liu H.C. (2004), Object recognition using genetic algorithms with a Huang Z. Chen, H., Hsu, C.J. Chen, W.H. and Wu, S., Credit rating analysis with support vector machines and neural networks: a market comparative study. Decision Support Systems ; 37:543-558. Hopfield’s neural model (1997). Expert Systems with Applications 1997; 13:191-199. Jain L.C., Vemuri V.R. (1999), Industrial Applications of Neural Networks. Boca Raton: CRC Press. Kiang M.Y., Hu, M.Y., Fisher D.M. (2006), An extended self-organizing map network for market segmentation—a telecommunication example Decision Support Systems ; 42:36- 47. Kiang M.Y., Kulkarni U.R., Tam K.Y. (1995), Self-organizing map network as an interactive clustering tool-An application to group technology. Decision Support Systems ; 15:351- 374. Kim T., Kumara S.R.T., (1997), Boundary defect recognition using neural networks. Inter- national Journal of Production Research; 35:2397-2412. Kim T.Y., Oh K.J., Sohn K., Hwang C. (2004), Usefulness of artificial neural networks for early warning system of economic crisis. Expert Systems with Applications ; 26:583- 590. Kirkos E., Spathis C., Manolopoulos Y., (2007), Data Mining techniques for the detection of fraudulent financial statements. Expert Systems with Applications ; 32: 995-1003. Kiviluoto K. (1998), Predicting bankruptcy with the self-organizing map. Neurocomputing ; 21:203-224. Klein B.D., Rossin D. F. (1999), Data quality in neural network models: effect of error rate and magnitude of error on predictive accuracy. Omega ; 27:569-582. Kohonen T. (1982), Self-organized formation of topologically correct feature maps. Biolog- ical Cybernetics ; 43:59-69. Kolehmainen M., Martikainen H., Ruuskanen J. (2001), Neural networks and periodic com- ponents used in air quality forecasting. Atmospheric Environment ; 35:815-825. Law R. (2000), Back-propagation learning in improving the accuracy of neural network- based tourism demand forecasting. Tourism Management ; 21:331-340. Lee D.L. (2002), Pattern sequence recognition using a time-varying Hopfield network. IEEE Transactions on Neural Networks ; 13:330-343. Lewis O.M., Ware J.A., Jenkins D. (1997), A novel neural network technique for the valua- tion of residential property. Neural Computing and Applications ; 5:224-229. Li W.J., Lee T., (2002), Object recognition and articulated object learning by accumulative Hopfield matching. Pattern Recognition; 35:1933-1948. Lim G.S., Alder M., Hadingham P. (1992), Adaptive quadratic neural nets. Pattern Recogni- tion Letters ; 13: 325-329. Lisboa P.J.G., Edisbury B., Vellido A. (2000), Business Applications of Neural Networks : The State-of-the-art of Real-world Applications. River Edge: World Scientific. computational abilities. Proceedings of National Academy of Sciences; 79(8):2554-2558. 442 G. Peter Zhang McCulloch W., Pitts W. (1943), A logical calculus of the ideas immanent in nervous activity. Bulletin of Mathematical Biophysics; 5:115-133. Min S.H., Lee J., Han I. (2006), Hybrid genetic algorithms and support vector machines for bankruptcy prediction. Expert Systems with Applications ; 31: 652-660. Minsky M. L., Papert S. A. (1969), Perceptrons. MA: MIT press. Miyake S., Kanaya F. (1991), A neural network approach to a Bayesian statistical decision problem. IEEE Transactions on Neural Networks ; 2:538-540. Mozer M.C., Wolniewics R. (2000), Predicting subscriber dissatisfaction and improving re- tention in the wireless telecommunication. IEEE Transactions on Neural Networks; 11:690-696. Nag A.K., Mitra A. (2002), Forecasting daily foreign exchange rates using genetically opti- mized neural networks. Journal of Forecasting ; 21:501-512. Nelson M., Hill T., Remus T., O’Connor, M. (1999), Time series forecasting using neural networks: Should the data be deseasonalized first? Journal of Forecasting ; 18:359-367. O’Connor N., Madden M.G. (2006), A neural network approach to predicting stock exchange movements using external factors. Knowledge-Based Systems ; 19:371-378. Paik J.K., Katsaggelos, A.K. (1992), Image restoration using a modified Hopfield neural network. IEEE Transactions on Image Processing ; 1:49-63. Pajares G., Cruz J.M., Aranda, J. (1998), Relaxation by Hopfield network in stereo image matching. Pattern Recognition ; 31:561-574. Panda C., Narasimhan V. (2007), Forecasting exchange rate better with artificial neural net- work. Journal of Policy Modeling ; 29:227-236. Parker D.B. (1985), Learning-logic: Casting the cortex of the human brain in silicon, Tech- nical Report TR-47, Center for Computational Research in Economics and Management Science, MIT. Palmer A., Monta ˜ no J.J., Ses ´ e, A. (2006), Designing an artificial neural network for forecast- ing tourism time series. Tourism Management ; 27: 781-790. Partovi F.Y., Anandarajan M. (2002), Classifying inventory using an artificial neural network approach. Computers and Industrial Engineering ; 41:389-404. Petersohn H. (1998), Assessment of cluster analysis and self-organizing maps. International Journal of Uncertainty Fuzziness and Knowledge-Based Systems. ; 6:139-149. Prybutok V.R., Yi J., Mitchell D. (2000), Comparison of neural network models with ARIMA and regression models for prediction of Houston’s daily maximum ozone concentrations. European Journal of Operational Research ; 122:31-40. Qi M. (2001), Predicting US recessions with leading indicators via neural network models. International Journal of Forecasting ; 17:383-401. Qi M., Zhang G.P. (2001), An investigation of model selection criteria for neural network time series forecasting. European Journal of Operational Research ; 132:666-680. Qiao F., Yang H., Lam, W.H.K. (2001), Intelligent simulation and prediction of traffic flow dispersion. Transportation Research, Part B ; 35:843-863. Raudys S. (1998), Evolution and generalization of a single neuron: I., Single-layer perceptron as seven statistical classifiers Neural Networks ; 11:283-296. Raudys S. (1998), Evolution and generalization of a single neuron: II., Complexity of statis- tical classifiers and sample size considerations. Neural Networks ; 11:297-313. Raviwongse R. Allada V., Sandidge T. (2000), Plastic manufacturing process selection methodology using self-organizing map (SOM)/fuzzy analysis. International Journal of Advanced Manufacturing Technology; 16:155-161. Reed R. (1993), Pruning algorithms-a survey. IEEE Transactions on Neural Networks ; 4:740-747. 21 Neural Networks For Data Mining 443 Remus W., O’Connor M. (2001), “Neural networks for time series forecasting.” In Princi- ples of Forecasting: A Handbook for Researchers and Practitioners, Armstrong, J. S. ed. Norwell:Kluwer Academic Publishers, 245-256. Reutterer T., Natter M. (2000), Segmentation based competitive analysis with MULTICLUS and topology representing networks. Computers and Operations Research; 27:1227- 1247. Richard, M. (1991), D., Lippmann, R., Neural network classifiers estimate Bayesian aposteriori probabilities. Neural Computation ; 3:461-483. Ripley A. (1993), “Statistical aspects of neural networks.” In Networks and Chaos - Statisti- cal and Probabilistic Aspects, Barndorff-Nielsen, O. E., Jensen J. L. and Kendall, W. S. eds. London: Chapman and Hall, 40-123. Ripley A. (1994), Neural networks and related methods for classification. Journal of Royal Statistical Society, Series B ; 56:409-456. Roh T. H. (2007), Forecasting the volatility of stock price index. Expert Systems with Ap- plications ; 33:916-922. Rosenblatt F. (1958), The perceptron: A probabilistic model for information storage and organization in the brain. Psychological Review ; 65:386-408. Rout S., Srivastava, S.P., Majumdar, J. (1998), Multi-modal image segmentation using a modified Hopfield neural network. Pattern Recognition ; 31:743-750. Rumelhart D.E., Hinton G.E., Williams R.J. (1986), “Learning internal representation by back-propagating errors.” In Parallel Distributed Processing: Explorations in the Mi- crostructure of Cognition Press, Rumelhart, D.E., McCleland, J.L. and the PDP Research Group, eds. MA: MIT. Saad E.W., Prokhorov D.V., Wunsch, D.C. II. (1998), Comparative study of stock trend pre- diction using time delay, recurrent and probabilistic neural networks. IEEE Transactions on Neural Networks; 9:456-1470. Salcedo-Sanz S., Santiago-Mozos R.,Bousono-Calzon, C. (2004), A hybrid Hopfield network-simulated annealing approach for frequency assignment in satellite communi- cations systems. IEEE Transactions on System, Man and Cybernetics, Part B:108-116. Sarle W.S. (1994), Neural networks and statistical models. Poceedings of the Nineteenth Annual SAS Users Group International Conference, Cary, NC: SAS Institute. Schumacher M., Robner R., Vach W. (1996), Neural networks and logistic regression: Part I., Computational Statistics and Data Analysis ; 21:661-682. Smith K.A., Ng, A. (2003), Web page clustering using a self-organizing map of user naviga- tion patterns. Decision Support Systems ; 35:245-256. Smith K.A., Willis R.J., Brooks M. (2000), An analysis of customer retention and insur- ance claim patterns using data mining: a case study. Journal of the Operational Research Society; 51:532-541. Soulie F.F., Gallinari P. (1998), Industrial Applications of Neural Networks. River Edge, NJ: World Scientific. Suganthan P.N., Teoh E.K., Mital D.P. (1995), Self-organizing Hopfield network for at- tributed relational graph matching. Image and Vision Computing; 13:61-71. Sun Z.Z., Yu S. (1995), Improvement on performance of modified Hopfield neural network for image restoration. IEEE Transactions on Image processing; 4:683-692. Suykens J.A.K., Vandewalle J.P.L., De Moor B.L.R. (1996), Artificial Neural Networks for Modeling and Control of Nonlinear Systems. Boston: Kluwer. Swanson N.R., White H. (1995), A model-selection approach to assessing the information in the term structure using linear models and artificial neural networks. Journal of Business and Economic Statistics; 13;265-275. 444 G. Peter Zhang Tatem A.J., Lewis H.G., Atkinson P.M., Nixon M.S. (2002), Supre-resolution land cover pattern prediction using a Hopfield neural network. Remote Sensing of Environment; 79:1-14. Temponi C., Kuo Y.F., Corley H.W. (1999), A fuzzy neural architecture for customer satis- faction assessment. Journal of Intelligent & Fuzzy Systems; 7:173-183. Thieme R.J., Song M., Calantone R.J. (2000), Artificial neural network decision support systems for new product developement project selection. Journal of Marketing Research; 37:543-558. Vach W., Robner R., Schumacher M. (1996), Neural networks and logistic regression: Part I. Computational Statistics and Data Analysis; 21:683-701. Wang T., Zhuang X., Xing X. (1992), Robust segmentation of noisy images using a neural network model. Image Vision Computing; 10:233-240. Webb A.R., Lowe D., (1990), The optimized internal representation of multilayer classifier networks performs nonlinear discriminant analysis. Neural Networks; 3:367-375. Werbos P.J., (1974), Beyond regression: New tools for prediction and analysis in the behav- ioral sciences. Ph.D. thesis, Harvard University, 1974. West D., (2000), Neural network credit scoring models. Computers and Operations Research; 27:1131-1152. West P.M., Brockett P.L., Golden L.L., (1997), A comparative analysis of neural networks and statistical methods for predicting consumer choice. Marketing Science; 16:370-391. Widrow B., Hoff M.E., (1960), Adaptive switching circuits, 1960 IRE WESCON Convention Record, New York: IRE Part 4 1960:96-104. Widrow B., Rumelhart D.E., Lehr M.A., (1994), Neural networks: applications in industry, business and science, Communications of the ACM; 37:93-105. Wong B.K., Bodnovich T.A., Selvi Y., (1997), Neural network applications in business: A review and analysis of the literature (1988-1995). Decision Support Systems; 19:301- 320. Young S.S., Scott P.D., Nasrabadi, N.M., (1997), Object recognition using multilayer Hop- field neural network. IEEE Transactions on Image Processing; 6:357-372. Zhang G.P., (2007), Avoiding Pitfalls in Neural Network Research. IEEE Transactions on Systems, Man, and Cybernetics; 37:3-16. Zhang G.P., Hu M.Y., Patuwo B.E., Indro D.C., (1999), Artificial neural networks in bankruptcy prediction: general framework and cross-validation analysis. European Jour- nal of Operational Research; 116:16-32. Zhang G.P., Keil M., Rai A., Mann J., (2003), Predicting information technology project escalation: a neural network approach. European Journal of Operational Research 2003; 146:115–129. Zhang G.P., Qi M. (2002), “Predicting consumer retail sales using neural networks.” In Neu- ral Networks in Business: Techniques and Applications, Smith, K. and Gupta, J.eds. Hershey: Idea Group Publishing, 26-40. Zhang G.P., Patuwo E.P., Hu M.Y., (1998), Forecasting with artificial neural networks: the state of the art. International Journal of Forecasting; 14:35-62. Zhang W., Cao Q., Schniederjans M.J., (2004), Neural Network Earnings per Share Fore- casting Models: A Comparative Analysis of Alternative Methods. Decision Sciences; 35: 205–237. Zhu Z., He H., Starzyk J.A., Tseng, C., (2007), Self-organizing learning array and its appli- cation to economic and financial problems. Information Sciences; 177:1180-1192. 22 Granular Computing and Rough Sets - An Incremental Development Tsau Young (’T. Y.’) Lin 1 and Churn-Jung Liau 2 1 Department of Computer Science San Jose State University San Jose, CA 95192 tylin@cs.sjsu.edu 2 Institute of Information Science Academia Sinica, Taipei 115, Taiwan liaucj@iis.sinica.edu.tw Summary. This chapter gives an overview and refinement of recent works on binary granular computing. For comparison and contrasting, granulation and partition are examined in parallel from the prospect of rough Set theory (RST).The key strength of RST is its capability in representing and processing knowledge in table formats. Even though such capabilities, for general granulation, are not available, this chapter illustrates and refines some such capability for binary granulation. In rough set theory, quotient sets, table representations, and concept hierarchy trees are all set theoretical, while in binary granulation, they are special kind of pretopological spaces, which is equivalent to a binary relation Here a pretopological space means a space that is equipped with a neighborhood system (NS). A NS is similar to the classical NS of a topological space, but without any axioms attached to it 3 . Key words: Granular computing, rough set, binary relation, equivalence relation 22.1 Introduction Though the label, granular computing is relatively recent, the notion of granulation has in fact been appeared, under different names, in many related fields, such as pro- gramming, divide and conquer, fuzzy and rough set theories, pretopological spaces, interval computing, quantization, data compression, chunking, cluster analysis, be- lief functions, machine learning, databases, and many others. In the past few years, we have seen a renewed and fast growing interest in Granular Computing (GrC). Many applications of granular computing have appeared in fields, such as medicine, economics, finance, business, environment, electrical and computer engineering, a number of sciences, software engineering, and information science. 3 This is an expansion of the article (Lin, 2005) in IEEE connections, the news letter of the IEEE Computational Intelligence Society O. Maimon, L. Rokach (eds.), Data Mining and Knowledge Discovery Handbook, 2nd ed., DOI 10.1007/978-0-387-09823-4_22, © Springer Science+Business Media, LLC 2010 446 Tsau Young (’T. Y.’) Lin and Churn-Jung Liau Granulation seems to be a natural problem-solving methodology deeply rooted in human thinking. Many daily ”things” have been routinely granulated into sub”things;” human body has been granulated into head, neck, and so forth; geo- graphic features into mountains, planes, and others. The notion is intrinsically fuzzy, vague and imprecise. Mathematicians idealized it into the notion of partitions, and developed it into a fundamental problem-solving methodology; it has played major roles throughout the entire history of mathematics. Nevertheless, the notion of partitions, which absolutely does not permit any over- lapping among its granules, seems to be too restrictive for real world problems. Even in natural science, classification does permit small degree of overlapping; there are beings that are both appropriate subjects of zoology and botany. A more general theory is needed. Based on Zadeh’s grand project on granular mathematics, during his sabbati- cal leave (l996/l997) at Berkeley, Lin focused on a subset of granular mathematics, which he called granular computing (Zadeh, 1998). To stimulate research on granu- lar computing, a special interest group, with T. Y. Lin as its Chair, was formed within BISC (Berkeley Initiative in Soft Computing). Since then, granular computing has evolved into an active research area, generating many articles, books and presen- tations at conferences, workshops and special sessions. This chapter is devoted to present some of such development over the past few years. There are two possible approaches: (1) One is starting from fuzzy side and mov- ing down, and (2) the other one is from extreme crisp side and moving up. In this chapter, we take the second approach incrementally. Recall that algebraically a parti- tion is an equivalence relation, so a natural next step is the binary granulation defined by a binary relation. For contrasting, we may call a partition A-granulation and the more general granulation B-granulation. 22.2 Naive Model for Problem Solving An obvious approach to a large-scaled computing problem is: (1) To divide the prob- lem into subtasks, might be point by point and level by level. (2) To elevate or abstract the problem into concept/knowledge spaces, could be in multilevels. (3) To integrate the solutions of subtasks and quotient tasks (knowledge spaces) of several levels 22.2.1 Information Granulations/Partitions In the first step, we select an appropriate system of granulation/partition so that only the summaries of granules/equivalence classes may enter into the higher level computing. The information in data space is transformed to a concept space, pos- sibly in levels, which may be locally at each point or globally at eh whole uni- verse (Lin, 2003b). Classically, we granulate by partitioning (no overlapping on gran- ules). Such examples are plentiful: in mathematics (quotient groups, quotient rings and etc. (Birkhoff and MacLane, 1977)), in theoretical computer science (divide-and- conquer (Aho et al., 1974)), in software engineering (the structural, object oriented, 22 Granular Computing and Rough Sets - An Incremental Development 447 and component based design and programming (Szyperski, 2002)), in artificial intel- ligence (Hobbs, 1985, Zhang and Zhang, 1992), in rough set theory (Pawlak, 1991) among others. However, these are all partition based, where no overlapping of gran- ules is permitted. As we have observed, even in biology, classification does allow some overlapping. The focus of this presentation will be on non-partition theory,but only in an epsilon step away from partitioning method. 22.2.2 Knowledge Level Processing and Computing with Words The information in each granule is summarized and the original problem is re- expressed in terms of symbols, words, predicates or linguistic variables. Such re- expressing is often referred to as knowledge representations. Its processing has been termed computing with symbols (table processing, computing with words, knowl- edge level processing, even precisiated natural language, depending on the complex- ity of the representations. In this chapter, we are computing on the space of granules or ”quotient space.” in which each granule is represented by a word that carries different degree of seman- tics. For partition theory, the knowledge representation is in table format (Pawlak, 1991) and its computation is syntactic in nature. For binary granulation, that we have focused here, is semantic oriented. We expand and streamline the previous works (Lin, 1998a,Lin, 1998b,Lin, 2000); the main idea is to transfer the computing with words into computing with symbols. Loosely speaking computing with symbols or symbolic computing is an “ax- iomatic” Computing: all rules of computing symbols are determined by the axioms. The computation follows the formal specifications. Such computing occurs only in an ideal situation. In many real world applications, unfortunately, such as non-linear computing, the formal specifications are often unavailable. So computing with words are needed; it can be processed informally. Semantics of words often may not be completely or precisely formalized. Their semantic computing is often carried out in the systems with human helps (the semantics of symbols are not implemented). Human enforced semantic computing are common in data processing environment. 22.2.3 Information Integration and Approximation Theory Most applications require the solutions be presented in the same level as input data. So the solutions often need to be integrated from subtasks (solutions in granules) and quotient tasks (solutions in the spaces of granules). For some applications, such as Data Mining and some rough set theory, are aimed at high level information; in such cases this step can be skipped. In general, the integration is not easy. In partition world, many theories have been developed in mathematics; e. g., extension functors. The approximation theory of pretopological spaces and rough set theory can be regarded as in this step. 448 Tsau Young (’T. Y.’) Lin and Churn-Jung Liau 22.3 A Geometric Models of Information Granulations For understanding the general idea, in this section, we recall and refine a previous formalization in (Lin, 1998a). The goal is to formalize Zadeh’s informal notion of granulation mathematically. As original thesis is informal, the best we could do is to present, hopefully, con- vincing arguments. We believe our formal theory is very close to the informal one. According to Zadeh (1996): Information granulation involves partitioning a class of objects(points) into granules, with a granule being a clump of objects (points) which are drawn together by indistinguishability, similarity or functionality. We will literally take Zadeh’s informal words as a formal definition of granula- tion. We observe that: 1. A granule is a group of objects that are draw together (by indistinguishability, similarity or functionality). The phrase ”drawn together” implicitly implies certain level of symmetry among the objects in a granule. Namely, if p is drawn towards q, then q is also drawn towards p. Such symmetry, we believe, is imposed by imprecise-ness of natural language. To avoid such an implications, we will rephrase it to ”drawn towards an object p,” so that it is clear the reverse may or may not be true. So we have first revision: 2. A granule is a group B(p) of objects that are draw toward an object p. Here p varies through every object in the universe. 3. Such an association between object p and a granule B(p) induces a map from the object space to power set of object space. This map has been called a binary granulation (BG). 4. Geometric View: We may use geometric terminology and refer to the granule as a neighborhood of p, and the collection {B(p)} a binary neighborhood system (BNS). It is possible that B(p) is an empty set. In this case we will simply say p has no neighborhood (abuse of language; to be very correct, we should say p has an empty neighbor- hood). Also it is possible that different points may have the same neighborhood (granule) B(p)=B(q). The set of all q, where B(q) is equal to B(p), is called the centers C(p) of B( p). 5. Algebraic View: Consider the set R = {(p, u)}, where u in B(p) and p in U. It is clear that R is a subset of U ×U, hence defines a binary relation (BR), and vice versa. Proposition 1 A binary neighborhood system (BNS), A binary granulation (BG), and a binary relation (BR) are equivalent. From the analysis given above, we propose the following mathematical model for information granulation. 22 Granular Computing and Rough Sets - An Incremental Development 449 Definition 1 By a (single level) information granulation defined on a set U we mean a binary granulation (binary neighborhood system, binary relation) defined on U. Let us goes a little bit further. Note that the binary relation is a mathematical expression of Zadeh’s ”indistinguishability, similarity or functionality.” We abstract the three properties into a list of abstract binary relations {B j | j run through some index set }, where each B j is a binary relation. Note that at each point p, each B j induces a neighborhood B j (p). Some may be empty, or identical. By removing empty set and duplications, the family have been we re-indexed N i (p). As in the single level case, we will define directly the granulation N : U → 2 2 U ; p →{B i (p) | i run through some index set }. The collection {B i (p)} is called a neighborhood system(NS)or (LNS); the latter one is used to distinguish itself from the neighborhood system (TNS) of a topological space (Lin, 1989a, Lin, 1992). Definition 2 By a local multi-level information granulation defined on U, we mean a neighborhood system (NS) is defined on U. By a global multi-level information granulation defined on U, we mean a set of BG is defined on U. All notions can be fuzzified. The right way to look at this section is to assume implicitly there is a modifier ”crisp/fuzzy” to all notions presented above. 22.4 Information Granulations/Partitions Technically, granular computing is actually computing with constraints. Especially in “infinite world”, granulation is often given in terms of constraints. In this chapter, we concerns primarily with constraints that are mathematically represented as binary relations 22.4.1 Equivalence Relations(Partitions) Partition is a decomposition of the universe into a family of disjoint subsets. They are called equivalence classes, because a partition induces an equivalence relation and vice versa. In this chapter, we will view the equivalence class in a special way. Let A ⊆U ×U be an equivalence relation (a reflexive, symmetric and transitive binary relation). For each p, let A p = {v ∈U : pAv} (22.1) A p is the equivalence class containing p, and will be called A-granule for the purpose of contrasting with general cases. Elements in A p are equivalent to each other. Let us summarize the discussions in: A : U →2 U : p →A p (22.2) . Rokach (eds.), Data Mining and Knowledge Discovery Handbook, 2nd ed., DOI 10.1007/978-0-387-09 823 -4 _22 , © Springer Science+Business Media, LLC 20 10 446 Tsau Young (’T. Y.’) Lin and Churn-Jung. Publishers, 24 5 -25 6. Reutterer T., Natter M. (20 00), Segmentation based competitive analysis with MULTICLUS and topology representing networks. Computers and Operations Research; 27 : 122 7- 124 7. Richard,. ; 21 :20 3 -22 4. Klein B.D., Rossin D. F. (1999), Data quality in neural network models: effect of error rate and magnitude of error on predictive accuracy. Omega ; 27 :569-5 82. Kohonen T. (19 82) ,

Ngày đăng: 04/07/2014, 05:21

TỪ KHÓA LIÊN QUAN