Efficient image region and shape detection by perceptual contour grouping Huiqiong Chen Qigang Gao Faculty of Computer Science Dalhousie University, Halifax, Canada [email protected] Faculty of Computer Science Dalhousie University, Dalhousie University [email protected] Abstract - Image region detection aims to extract meaningful regions from image. This task may be achieved equivalently by finding the interior or boundaries of regions. The advantage of the second strategy is that once a closure is detected not only its shape information is available, but also the interior property can be estimated with a minimum effort. In this paper, we present a novel method that detects region though region contour grouping based on Generic Edge Token (GET). GETs are a set of perceptually distinguishable edge segment types including linear and non-linear features. In our method, an image is first transformed into GET space on the fly represented by a GET graph. A GET graph presents perceptual organization of GET associations. Two types of perceptual closures, basic contour closure and object contour closure, based upon which all meaningful regions are conducted, are defined and then detected. The detection is achieved by tracking approximate adjacent edges along the GET graph to group the contour closures. Because of the descriptive nature of GET representation, the perceptual structure of detected region shape can be estimated easily based its contour GET types. By using our method, all and only perceptual closures can be extracted quickly. The proposed method is useful for image analysis applications especially real time systems like robot navigation and other vision based automation tasks. Experiments are provided to demonstrate the concept and potential of the method. Index Terms – region detection, perceptual closure, GET data I. Introduction An image region is defined as a set of connected pixels with homogenous properties like color or texture, etc, which has obvious contrast to its surroundings. Although region detection plays a significant role in many image analysis applications, it is a challenging task to efficiently obtain meaningful and accurate regions from images, particularly for real-time applications. Region detection methods can be mainly classified into region-based and boundary-based methods [1]. Region-based methods are based on merging similar pixels together to form coherent regions. Region membership decision, under/over segmentation and inaccurate boundaries are their common problems. In [2], detection groups regions for over segmented image by properties of continuity, region ration, etc. To reduce over-segmentation in watershed method, Gaussian convolution with different deviations is used in [3]. Edge information can be used here for refinement, such as seed placement selection, homogeneity criterion control, accuracy improvement [4]. Region-based methods usually treat region as a set of pixels therefore it is difficult to get accurate region shape or object structure from image through these methods. Boundary-based methods employ edge information. Proper edge pixels are grouped into region boundaries. The advantage of these methods is that once a closure forms both region boundaries and its interior properties can be get easily. However, when edge data is provided on pixel level, the method effect will be largely affected by the existence of edge gaps and boundary pixel recognition strategy. Perceptual origination can be used in detection to find perceptual regions/objects [5] [6] but and it always suffers from intensive computation and detecting arbitrary shape regions. In this paper we present a novel method that can detect meaningful regions with arbitrary shape efficiently and speedily. By using innovative perceptual edge features called Generic Edge Token (GET), meaningful regions can be detected promptly through perceptual region contour detection. Perceptual region contour closures are defined first to represent meaningful regions and then be detected through closure detection in GET space. Because of the simplicity of GET representation, and inherent structure information carried by each feature, the perceptual region shape can be obtained easily from region as well as region boundary and interior attributes. Therefore, our method is suitable for image analysis applications especially for real-time systems. This paper is organized as follows: Section II introduces Generic Edge Tokens. Section III presents perceptual region hierarchy and perceptual region representation. Contour closure detection method is provided in section IV. Experiments are given in section V. Section VI provides conclusions and future work. II. Generic Edge Tokens Generic Edge Tokens (GETs) are perceptually significant image primitives which represent classes of qualitatively equivalent structure elements. A complete set of GETs includes both Generic Segments (GS) and curve partition points (CPP), Each GS is perceptually distinguishable edge segment with linear or nonlinear features while each CPP is junction of GSs. Gao and Wang proposed a curve partition method in a very generic form which performs partition following the similar objective in human visual perception. The image is scanned horizontally and vertically with an interval to find strong edge pixels as tracking starting pixels then edge pixels on the traces are selected according to the GS definitions. This method is robust with low time expense. Detail explanation of GET detection can be found in [7]. Fig. 1 (a) shows curve partition examples. TABLE 1. Definitions of GS types; (M+ and M- represent monotonic increasing and decreasing properties while c mean a constant) GS type Properties f(x) j(y) f’(x) j’(y) Fig. 1. Curve partition examples. A GS is a segment in image with monotonic increasing or decreasing. GSs can be perceptually classified into eight categories as Fig. 1 (b) shows. Each type of GSs should satisfy properties described by the monotonicities of tangent function y= f(x) and its inverse function x = j(y), as Table 1 illustrates. The monotonicity is checked along each GS. A CPP is a perceptually significant point where adjacent GSs meet and curve turning takes place [8]. No intersection among GSs exists besides CPPs. CPP can group GSs into perceptual structures. There are four basic categories of CPPs according to the conjunction types as Fig. 1 (c) shows. The intrinsic perceptual features of GETs will facilitate further detection of perceptual closures formed by GETs and ease the extraction of perceptual shape from regions described by GET closures. III. Perceptual region hierarchy and representation According to MPEG-7 standard, region or object shape can be described by Contour Shape descriptors based on its contour [9]. Therefore contour closure, formed by region boundaries, is desirable representation for meaningful regions. It possesses both low-level features and perceptual concepts: closure interior pixels contain low level attributes; closure boundary contains structure/shape information at higher level. A. Perceptual region concept hierarchy From perceptual view, meaningful regions arise from objects, and then can be represented by object outline and object basic inner parts (called basic regions). An object is constituted by one or multiple basic regions. As Fig. 2 shows, object A is built by 3 basic regions, whose contours can be represented by basic contour closures e1e2e3e4, e2e5e6e7, e3e7e8e9 respectively, all of which are formed by GETs. Other closures are constituted by basic contour closures, among which e1e5e6e8e9e4 represents the object outline. Therefore all closures in image formed by GETs can be classified into three types according to region contour properties: basic contour closure which represents the contour of basic region; object contour closure which represents the contour of object outline or separated object component outline; other composite closure. All meaningful regions in an object can be represented by first two types of GET closures while GET closures can be described by GET organizations. As Fig. 2 illustrates, object B has a separate hole inside, which can be regarded as a separate object component constituted by one LS1 LS2 c n/a 0 ∞ n/a c ∞ 0 LS3 LS4 CS1 CS2 CS3 CS4 M+ M+ c c MM+ c c M+ M+ MM+ MMMM+ MMM+ M- M+ MM+ M- single basic region. The remaining part of B other than the hole is another basic region, whose contour can be described by the basic contour closure e1e2e3e4 along with the inside object contour closure e 5e6e7e8, which represents outline of the object component hole. From analysis above we can conclude that, basic contour closures and object contour closures contain most perceptual information of meaningful regions while other composite closures do not have much meaning. Definition 1: Meaningful regions in image can be represented by two types of perceptual contour closures: basic contour closures and object contour closures. A basic contour closure is a basic GET closure which can not be split into other closures; an object contour closure is a GET closure representing the outline of an object or object individual component. (Assume no occlusion occurs). This definition has several advantages. First, it leads to a hierarchical descriptor for image content: Image content can be represented by objects, while an object can be described by the hierarchy in Fig. 2. Second, both types of perceptual closures are supported in MPEG-7 frame by contour shape descriptors [10]. Last, it makes sure that useful perceptual closures will be picked up in detection and meaningless composite closures will not be detected anymore. A region/closure concept hierarchy is constructed in Fig. 2. In this hierarchy, GETs can be organized perceptually into contour closures, which can build meaningful regions in objects. To find perceptual contour closures from all GET data, we use a GET graph to code perceptual organization of GET associations in image. Fig. 2. Perceptual region/closure concept hierarchy B. Perceptual organization representation by GET graph A graph named GET graph can be derived from preextracted GET data to code the perceptual structure of image. An image is transformed into GET space represented by a special graph named GET graph. Therefore, perceptual contour closures, i.e., basic contour closures and object contour closures, can be detected by perceptual cycle search in GET graph when exploring the correspondence between the GET closures and graph cycles [11]. For the purpose of real time processing, GET graph is reduced and search starting edges are selected before search to avoid duplicate detections and reduce computation burden. Starting from selected edges, novel perceptual contour closure detection method is applied to get all perceptual closures from GET graph then results are classified into perceptual types. Taking advantage of the perceptual GET class data, the image can be transformed into GET space represented by a GET graph. With all GETs extracted from image, GET graph can be constructed by converting GSs and CPPs to graph edges and vertices respectively. Each edge in GET graph presents a perceptually distinguishable edge segment in image while the organization of GET graph represents the perceptual structures in image. If no occlusion occures, each connected component in GET graph corresponds to an object/individual object component in image constituted by basic inner parts, which can be represented by basic contour closures. Therefore basic contour closures in image one-to-one correspond to basic cycles in GET graph while object contour closures oneto-one correspond to outlines of graph connected components. For convenience sometimes we do not distinguish between the terms in the context of the paper. GET graph has some unique characters: (1) Besides representing graph structure, the edges and vertices in GET graph also represent real segments and their junction points in image. (2) According to GET properties, no intersection among edges should exist besides edge endpoints. GET graph can be constructed as follows: let G= (V, E) be an undirected graph, where V is the set of vertices and E is the set of edges in G. Initially V=Æ and E=Æ. For each GSi (0 ≤ i< m, m is number of GSs in image), insert it into E as edge ei. For each CPPj (0 ≤ j < n, where n is number of CPPs in image), insert it to V as vertex vj. In practice, there always exist edge gaps in image as Fig. 3 (a) shows. The gaps may be caused by image noise or broken edges derived from edge detector. To bridge possible gaps, we amend the concept of “adjacent” for GET graph and regard close edges within a small common region as adjacent: For edges ei and ej, if ei’s endpoint vm is close to ej’s endpoint vn, normalized distance between vm and vn is defined l (v m ) + l (v n ) (1) dist (v m , v n ) = ´ d (v m , v n ) [l (vm ) + l (vn )]2 + d where l(vm), l(vn) are length of longest edges among all edges connecting to vm, vn respectively; d is mean of length of all edges Î E, d(vm, vn) is the distance between vm and vn. If dist(vm, vn)£ threshold Τ, the gap of ei and ej should be bridged. A virtual vertex m is used as combination of real vertices vm, vn. m is virtual intersection of ei and ej obtained by extending endpoints vm and vn from ei and ej respectively. Therefore, m is a substitute for vm, vn as endpoint of ei and ej. Definition 2: Graph vertex vj (no matter vj is a real vertex or virtual vertex) is incident with edge ei iff: vj is ei’s endpoint. Edges ei, ej are adjacent iff ei and ej: (a) share a real vertex as common endpoint; or (b) share a virtual vertex as common endpoint within threshold T. we denote this as ei ~ ej. The degree of vertex vj, denoted as deg(vj), is the number of edges incident with vj within T. Fig. 3 (b) illustrates GET graph construction. If dist (v 0, v1) £ T, the edge gap between e1 and e10/e12 is so small that it can be bridged by virtual vertex m1. e1, e10, e12 are adjacent by m1, deg(m1)= 3; otherwise, this gap can not be ignored. e1 is not adjacent to e10/e12. deg(v0)=2 and deg(v1)= 1. Similarly, gap between e4 and e5 can be bridged by m2 if dist(v5, v6) £ T. GET space is reduced by graph reduction, which remove noise edges not belonging to graph cycle. It decreases future search burden and false closure detections. An edge e i with endpoints vm and vn is a noise edge iff: (a) deg (vm) = 1; or (b) besides ei, vm is incident with only noise edges; or (c) deg (vn) =1; or (d) besides ei, vn is incident with only noise edges. Fig. 4(b) shows a GET graph and (c) shows graph reduction. IV. Perceptual contour closure detection Perceptual closure detection can be carried out using GETs pre-extracted from image, as Fig. 5 shows. After graph reduction, every edge in GET graph must be element of some closures. A perceptual closure in image can be detected by finding corresponding cycle in GET graph. Fig. 4 (a) Original image; (b) GET graph; (c) GET graph reduction Closure Definition Image GET Detection GET graph Construction & Reduction Search starting edge selection (a) GETs in image with edge gaps; (b) GET graph with virtual vertices Fig. 3. Construct graph in GET space using GET data. Closure Detection Closure detection by cycle search Perceptual GET closures (regions) Result closure classification Fig. 5. Perceptual contour closure detection architecture Staring from one endpoint of an edge then tracking along adjacent edges, a graph cycle can be formed if a path back to starting point is found in graph. All closures may be detected from GET graph by this method when starting from closure edges. However, this method has two problems. First, proper starting edges should be selected instead of starting search from every graph edge. This can reduce computation cost and duplicate detections, avoid time expense increasing greatly with graph edge number. Second, search strategy is needed in tracking to guarantee the results be perceptual closures. Spanning trees are employed for the first problems. For the second problem, a new perceptual search strategy is proposed. A. Starting edge selection Spanning tree is a connected sub-graph with tree structure including all vertices of a graph connected component. Lemma 1: all perceptual closures can be obtained by tracking in graph starting from edges not belonging to spanning trees. That is, edges not belonging to spanning trees should be selected as starting edges. Proof of Lemma 1: Let Gi =(Vi, Ei) be the ith connected component of GET graph, (V i’, Ei’) be set of vertex and edges for the spanning tree of Gi, Ni = Ei - Ei’. We need to prove (a) "eÎNi, e must be element of a basic contour closure; (b) each basic contour closure in Gi must have at least one edge e such that eÎNi; (c) the object contour closure corresponding to component outline must have at least one edge e so that Î Ni. (a) As spanning tree, Vi’=Vi; (Vi’, Ei’) contains no cycle. "eÎNi with endpoints vx vy, there have vxÎVi’, vyÎVi’ and one single path p between (vx, vy) in the tree. If e is added to the tree, e combines with path p forming a cycle. This cycle either is a basic cycle or is built by basic cycles. e belongs to one of basic cycles, i.e., e belongs to a basic contour closure. (b) Since spanning tree contains no cycle, any basic cycle in (Vi, Ei) must have at least one edge e excluded from (Vi’, Ei’) to avoid forming this cycle, i.e., eÎNi. (c) The proof of (c) is similar with (b). From (a)-(c), we can conclude that Lemma 1 is true. B. Perceptual Cycle search algorithm In connected component i, basic cycle is cycle without sub-cycle inside and component outline is a cycle without any cycle outside. Starting from "eiÎNi, which must belong to some basic cycle of Gi, basic contour closure can be formed by basic cycle search in graph. The search tracks along adjacent edges by always selecting the innermost edge in given direction (anticlockwise/clockwise) from all candidates in each step of path selection. If ei is on component outline, object contour closure can also be extracted by selecting outmost edge in given direction (which can be considered as selecting innermost edge in the other direction) in each step. A new closure detection method is proposed to find all perceptual closures by perceptual cycle search algorithm. This algorithm is based on the following hypothesis which will be proven by lemma 2: "eiÎNi with endpoints vx and vy, starting from vx of ei, a perceptual closure (object contour closure or basic contour closure) can be found by cycle search in graph, which always selects innermost edge in some direction from all adjacent edges as next edge in each step of selection. Since innermost edge is direction sensitive, the cycle search must keep the same direction in each step. Fig. 6 (a) illustrates cycle search. If a clockwise search starts from v7 of e7, e7 is current edge while v7 is current vertex. e5, e6 ~ e7 by current vertex v7. As the clockwise innermost edge for e7, e5 is selected to be next cycle edge. In next step, e5 is current edge while e6’s another endpoint m2, is the current vertex. e4 ~ e5 by m2 so e4 is selected. Search stops when new selected e10 meets e7 at v9 and basic contour closure e7e5e4e3e2e1e10 forms. Another basic contour closure e 7e6e8e9 can also be formed by anticlockwise search starting from v 7 of e7. Object contour closure can be extracted as well. Starting from v10 of object contour edge e11, object contour closure e11e8e6e5e4e3e2e1e12, which is constituted by clockwise outermost edges, can be got by anticlockwise cycle search. Algorithm 1: let ei with endpoints vx, vy be starting edge, the cycle search algorithm in direction d can be presented as: Step 1: initially, current edge ce:=ei, current vertex cv:=vx. Let C be set of closure edges, C={ei}. Step 2: if ce= ei and cv=vy, search stops and perceptual closure forms; otherwise perform step 3~4; Step 3: for each edge ej adjacent to ce by cv, calculate the included angle from cv to ej in direction d. Select the adjacent edge with minimal included angle as new selected edge ne, C=CÈ{ne}; Step 4: update cv and ce; cv:= ne’s another endpoint which is not incident with ce; ce:=ne; go to step 2. Lemma 2: the closure extracted using perceptual cycle search algorithm must be a perceptual closure. Proof of Lemma 2: let C be the cycle search result in direction d starting from ei. Here we prove C must represent either a basic contour closure or an object contour closure. (1) If ei is not on component outline, C can not be an object contour closure. There must exist more than one basic cycles sharing ei; assume that C is not a basic cycle, there must exist basic cycles C’ inside C and C’’ outside of C sharing ei with C. If C C’ and C’’ have identity edges ei ~ ej at first several steps of search, let em , en and el be next edges selected in C, C’ and C’’ respectively. (a) If em en and el are different edges, em should be located inside the area between ej and em. Thus em can not be the innermost edge against ej in whatever search direction, which contradicts the fact that edges in C should be the innermost edge in direction d. (b) If (a) Detect perceptual closures (b) Simulate closure by polygon Fig. 6. (a) shows the process of perceptual closures detection. Dashes in (a) are starting edges and arrows indicate search direction. Polygon in (b) simulates the closure formed by clockwise search starting from v 7 of e7. Dashes in (b) indicate polygon edges different from graph edges. em is same to en but different with el; let ek be the last identical edge for C and C’, ep and eq be next edges for C and C’. eq is inside C and el is outside C. No matter d is clockwise or anticlockwise, either ep is not the innermost edge against e k (eq is the innermost in stead) or em is not the innermost edge against ej (en is the innermost). This contradicts to previous fact. (c) If em = el but different with en; proof is similar to (b). Summarizing (a)~(c), if ei is on component outline, the prior assumption can not be true. C must be a basic contour closure. (2) If ei is on component outline; proof is similar to (1). From (1)-(2), we can conclude that Lemma 2 is true. C. Perceptual closure detection via cycle search Perceptual closure detection can be described as follows: Step 1: Construct GET graph and remove noise graph edges. Step 2: Find all connected components in the GET graph. For each component, perform step 3. Step 3: Extract a spanning tree. Starting form each edge ei not in the tree, perform step 4. Step 4: perform both clockwise and anticlockwise closure searches by applying perceptual cycle search algorithm. It can be easily concluded from lemma 2 that using this detection method, two closures can be extracted by clockwise and anticlockwise searches starting from a given edge e, and there are two possibilities for extracted closures: (1) If e is not on component outline, both of the detected closures are basic contour closures; (2) If e is on component outline, the two closures have one basic contour closure and one object contour closure. Lemma 3: the perceptual closure detection method is valid because by using this method, (1) all extracted closures are perceptual closures; (2) all perceptual closures are extracted. Proof of lemma 3: (1) has been represented in lemma 2. (2) First we prove that object contour closure must be detected. Based on lemma 1, there must have an object contour closure edge ejÎNi as starting edge. Starting from this edge, the two results must have one object contour closure. Second we prove that all basic contour closures must also be extracted. Assume there is a basic contour closure C that cannot be detected by our method, C must have an edge ejÎNi. If ej is not on the component outline, C must share ej with two other basic contour closures C’ and C’’, both of which are detected by cycle search starting from ej. No intersection exist among the three closures, so C must be outside the area of (C’’È C’) while ej is an edge inside the area of (C’’ÈC’). This contradicts with the fact that ej is an edge of C. If ej is on component outline; C shares e j with basic contour closure C’ and object contour closure C’’. The areas of C and C’ do no intersect while both of them are inside C’’. Therefore C should be inside the area of (C’’-C’), in which ej can not be included. This also conflicts with previous fact. Thus this kind of C can not exist no matter e j’s type. D. Closure classification Perceptual closure detection method above finds all perceptual closures without concerning closure types. The result closures should be classified into two types: basic contour closure and object contour closure. The classification can be achieved by summing up all included angles of a polygon which simulates the extracted closure. The polygon can be formed during cycle search by simulating the selected edge in each step as a straight line between its two endpoints. Included angles between simulation lines with search direction are recorded. Fig. 6 (b) shows an example: in clockwise search starting from v7 of e7, the first selected edge is e5 so the clockwise included angle from e7 to e5 is recorded as q v v ®v m . The next 7 9 7 2 selected edge e4 is simulated by line from m2 to v4 so q m v ®m v 2 7 2 4 is recorded. The polygon forms when q v m ®v v is recorded. 9 1 9 7 For an n-edge closure, let qTotal be sum of all n included angles recorded during simulation. If the closure is basic contour closure, recorded included angles should be interior angles of polygon since the simulation passes inside closure in each step, as arrows in Fig. 6(b) shows. qTotal = ∑(all interior angles of polygon) = 180*(n-2). If the closure is object contour closure, the included angles should be exterior angles of polygon thus qTotal = 180*(n+2). Therefore we conclude: If qTotal = 180*(n-2), the closure is basic contour closure; and If qTotal = 180*(n+2), the closure is an object contour closure. where n is the number of edges in the closure. Fig. 7 illustrates the process of closures detection and classification. V. Experiment Results To evaluate validity of our method, algorithms are implemented using C++. Sun Fire 4800, a multi-user UNIX server running Solaris 8 with 16 GB RAM, 4 UltraSPARC-III processors at 900 MHz, is used in test. Server load average = 2.64 during test. The test images, whose sizes range from 256*256 to 512*512 pixels, include both synthetic image and real images. The experiment first extract GETs from image; then detect regions based on GET data. In our test, average times of GET extraction and region detection are 418 and 447 millisecond. The number of perceptual closures in image and correct/error/ missing detections in the results are determined manually with human perception. Tiny regions are omitted as noises. Experiment results are listed in Table 2. Average detection correctness = 91.13%. Experiments show that our method can get arbitrary region shapes with high correctness and short process time, which exhibits its potential in image analysis applications. Fig. 8 gives detection examples. The error/missing detections may be caused by false edge selection in search: (1) although normalized threshold is used to bridge gaps between closure edges, it can not distinguish edge gaps between relevant edges with cracks in irrelevant edges by perceptual view. Larger threshold, although bridges larger gaps, raises the risk of taking unrelated edges as adjacent, thus increases false detection. (2) Detection depends heavily on GET data got from GET tracker. It can not find a region contour if one contour edge is missing detected in GET tracking, i.e., it can not recovery lost data by knowledge in closure detection. (3) The GETs derived from image noises may also mix up with valid GETs thus mislead search. Fig. 7. Closure extraction and classification for Fig. 4(a); left-most figure shows all contours; other small figures show all basic contour closures. Fig. 8. Detection result examples. Each row from left to right: original image; extracted GETs; region contour formed by GETs; filled regions. TABLE 2. Experiment results. Average correctness= 91.13%. Number of Number of Image correct Error No detections detections 1 22 0 2 33 2 3 23 3 4 15 1 5 32 1 6 18 0 7 27 1 Number of missing detections 0 1 1 1 7 0 5 Number of closures in image 22 36 26 16 39 18 33 Percentage of correct detections 100% 91.7% 88.5% 93.8% 82.1% 100% 81.8% VI. Conclusions and Future Work In this paper, we propose a GET based method for extracting meaningful regions from image through contour closure detection. Perceptual closures are defined to represent meaningful regions and then be extracted from GET data. The proposed method has following advantages. (1) It provides real-time region segmentation in which region shape structure can be extracted as well. (2) It achieves high accuracy of detected regions with low computation cost. (3) The method is suitable for segmenting objects with arbitrary shapes. (4) It can detect both object contours and their components, based upon which all meaningful regions and objects can be conducted hierarchically as Fig. 2 shows. The possible extension of this method may include the follows. Other features, such as color and texture, can be added in the process to increase the robustness of detection in order to cover a wider range of applications. Meanwhile the added features would provide with a complete set of region descriptions. A region may be encoded with all attributes, i.e. shape, color, and texture. The segmentation technique may support various application domains include robot navigation, surveillance motion analysis, image retrieval, etc. References [1] A. Narendra, “On Detection and Structure Representation of Multiscale Low-Level Image” ACM Computmg Surveys, vol. 27, no. 3, pp. 304-306, 1995 [2] T. Tamaki, T. Yamamura, and N. Ohniski, “Image Segmentation and Object Extraction based on Geometric features of Regions”, SPIE Conf. on Visual communication & image processing, vol. 3653, pp. 937-945, 1999 [3] J. M. Gauch, “Image Segmentation and Analysis via Multi-scale gradient watershed hierarchies”, IEEE Trans. on Image Processing, vol. 8, No. 1, pp. 69-79, 1999 [4] J. Freixenet, X. Munoz, D. Raba, J. Mart, and X. Cuf, “Yet Another Survey on Image Segmentation: Region and Boundary Information Integration”, Proc. of the 7th European Conf. on Computer Vision, pp.408-422, 2002 [5] A. L. Ralesc., and J. G. Shananhan, “Pereptual Organisation for inferring Object Boundaries in an Image”, Pattern Recognition, Vol. 32, pp. 1923~1933, 1999 [6] S.Sarkar and K. Boyer, “Perceptual Organization in Computer Vision: A review and a proposal for classificatory structure”, IEEE Trans. on Systems, Man & Cybernetics, Vol.23, No. 2, pp. 382~398, 1993 [7] Q. Gao, and A. Wong, “Curve Detection based on Perceptual Organization” Pattern Recognition, Vol. 26, No. 1, pp. 1039~ 1046, 1993 [8] X. Zheng, and Q. Gao, “Generic Edge Tokens, Representation, Segmentation and Grouping”, Proc. of the 16th International Conf. on Vision Interface, pp. 388-394, 2003 [9] MPEG-7 Overview, appears in http://www.chiariglione.org/mpeg /standards/mpeg-7/mpeg-7.htm, Mar, 2003, access at Oct, 2004. [10] T. Sikora, “the MPEG-7 Visual Standard for Content Description - An Overview”, IEEE TRANS. on Circuits and Systems for Video Technology, Vol. 11, No. 6, pp. 716~719, 2001 [11] Q. Iqbal and J. K., “Retrieval by Classification of Images containing large Manmade Objects using Perceptual Grouping”, Pattern Recognition , Vol. 35, No. 7, pp. 1463-1479, 2002
© Copyright 2026 Paperzz