Você está na página 1de 7

Content Based Image Retrieval (CBIR) Software Requirements Specification Document

Table of Contents

1. Introduction 1.1 Purpose In last few years the potential growth in digitization of images has occurred, with immense amount of information flowing and stored in the database of World Wide Web. The number of users exploiting the WWW has increased tremendously while accessing and manipulate remotely-stored images in all kinds of new and exciting ways. However, they are also discovering that the process of locating a desired image in a large and varied collection can be a source of considerable frustration. The problems of image retrieval are becoming widely recognized, and the search for solutions an increasingly active area for research and development. Traditional way to search image in database is to create a textual description of all the images in the database and use the methods from text-based information retrieval to search based on the textual descriptions. Unfortunately, this method is not feasible. On the one hand annotating images has to be done manually and is a very time-consuming task and on the other hand images may have contents that words cannot convey. This has given rise in interest of techniques for retrieving images on the basis of automatically-derived features such as colour, texture and shape a technology now generally referred to as Content-Based Image Retrieval (CBIR) 1.2 Scope The software product is content based image retrieval (CBIR) is about developing an image search engine, not only by using the text annotated to the image by an end user (as traditional image search engines), but also using the visual contents available into the images itself. Initially, CBIR system should have a database, containing several images to be searched. Then, it should derive the feature vectors of these images, and stores them into a data structure like on of the Tree Data Structures (these structures will improve searching efficiency). A CBIR system gets a query from user, whether an image or the specification of the desired image. Then it searches the whole database in order to find the most similar images to the input or desired image. CBIR usually deals with large image collection of low level and high level features, which directly influence indexing and retrieval complexity, memory and disk space requirment. due to high memory and processing power requirment,cbir has not widely been applied on platforms having limited resources, such as mobile devices 1.3 Definitions, Acronyms, and Abbreviations. CBIR: content based image retrieval

QBIC: query by image content CBVIR :content-based visual information retrieval 1.4 References
[1]. H.B.Kekre, Sudeep D. Thepade, Boosting Block Truncation Coding using Kekres LUV Color Space for Image Retrieval, WASET International Journal of Electrical, Computer and System Engineering (IJECSE), Volume 2, Number 3, pp. 172-180, Summer 2008. Available online at http://www.waset.org/ijecse/v2/v2-3-23.pdf [2]. H.B.Kekre, Sudeep D. Thepade, Image Retrieval using Augmented Block Truncation Coding Techniques, ACM International Conference on Advances in Computing, Communication and Control (ICAC3-2009), pp. 384-390, 23-24 Jan 2009, Fr. Conceicao Rodrigous College of Engg., Mumbai. Is uploaded on online ACM portal. [3]. H.B.Kekre, Sudeep D. Thepade, Rendering Futuristic Image Retrieval System, National Conference on Enhancements in Computer, Communication and Information Technology, EC2IT-2009, 20-21 Mar 2009, K.J.Somaiya College of Engineering, Vidyavihar, Mumbai-77. [4]. H.B.kekre, Sudeep D. Thepade, Improving Color to Gray and Back using Kekres LUV Color Space, IEEE International Advanced Computing Conference 2009 (IACC09), Thapar University, Patiala, INDIA, 6-7 March 2009. Is uploaded and available online at IEEE Xplore. [5]. H.B.Kekre, Sudeep D. Thepade, Color Traits Transfer to Grayscale Images, In Proc.of IEEE First International Conference on Emerging Trends in Engg. & Technology, (ICETET08), G.H.Raisoni COE, Nagpur, INDIA. Uploaded on online IEEE Xplore. [6]. Image database - http://wang.ist.psu.edu/docs/related/Image.orig (Last referred on 23 Sept 2008)

1.5 Overview The sections below describe common methods for extracting content from images so that they can be easily compared. The methods outlined are not specific to any particular application domain. Color - Retrieving images based on color similarity is achieved by computing a color histogram for each image that identifies the proportion of pixels within an image holding specific values (that humans express as colors). Current research is attempting to segment color proportion by region and by spatial relationship among several color regions. Examining images based on the colors they contain is one of the most widely used techniques because it does not depend on image size or orientation. Color searches will usually involve comparing color histograms, though this is not the only technique in practice.

Texture - Texture measures look for visual patterns in images and how they are spatially defined. Textures are represented by texels which are then placed into a number of sets, depending on how many textures are detected in the image. These sets not only define the texture, but also where in the image the texture is located. Texture is a difficult concept to represent. The identification of specific textures in an image is achieved primarily by modeling texture as a two-dimensional gray level variation. The relative brightness of pairs of pixels is computed such that degree of contrast, regularity, coarseness and directionality may be estimated. However, the problem is in identifying patterns of co-pixel variation and associating them with particular classes of textures such as ``silky, or ``rough. Shape - Shape does not refer to the shape of an image but to the shape of a particular region that is being sought out. Shapes will often be determined first applying segmentation or edge detection to an image. Other methods use shape filters to identify given shapes of an image. In some case accurate shape detection will require human intervention because methods like segmentation are very difficult to completely automate. 2. Overall Description 2.1 Product Perspective Existing system In the existing system, people carry out search using text based search. For example, in Google, if we need to search for an image of a rose, and we type in the keyword rose, we get all images which are tagged with the word rose which may include a woman named Rose or a building named Rose or any object of such type. Hence there is a lot of redundancy in the search results obtained. Proposed System To remove the redundancies, the proposed system applies various techniques to retrieve images based on the shape content of the image. The user would have to provide an image to the system and the system matches it with its database and retrieves all those images that are similar to it in shape features hence refining the search results. 2.2 Product Functions The functioning of the system is divided into two parts feature extraction and query execution. Feature Extraction In this step, a database of feature vectors is created. Originally a database of around 1000 images is formed spanning 11 categories. On these images, various techniques are applied to extract the edges and these are stored as feature vectors. Query extraction In this stage, an image is taken as input by the system and its shape feature is extracted using the respective technique and the feature vector is then matched with the database of feature vectors on the basis of mean square error. The performance of various techniques is compared on the basis of 2 parameters

precision and recall. Precision measures the correctness of the technique and recall measures the completeness of the technique. 2.3 User Characteristics The users of the system need to understand computers and need to be well versed with the use of computers. They should have a clear understanding of their requirement from the system. They should understand the tradeoff between accuracy of the retrieval technique and the time required to retrieve the results and hence decide what is more important to them. Depending on that the technique to be applied can be decided. 2.4 Constraints Constraints of the system are that the images of the database are of JPEG format only. Hence the query image must also be of JPEG format. Also each technique has different requirements for the sizes of the query image. 2.5 Assumptions and Dependencies Assumptions for this system are that the images entered as query image are of JPEG formats for accurate results and the users know what they exactly require as output from the system. Accordingly, the user enters their query image. The entire system depends on the input image format. 2.6 Apportioning of requirements The requirements that can be delayed according to later versions of the project could be techniques independent of image sizes and the image formats so that the techniques can be scaled to generic data bases. 3. Specific Requirements 3.1 External Interfaces The system contains an interface for the user to enter a query image for extracting images with similar features. The image can be in any format like JPG, BMP, etc. The rest of the functioning happens at the backend. Output of the system will be the images whose shape content matches the best with the images stored in the database. The accuracy of the output images depends on the techniques used for matching the shape content of the images. Every images has its own shape content, thus the output will vary depending on the input image. 3.2 Functions The system shall take input in the form of an image, which could be any format. For this a validity check is performed that the input given has to be an image. Incase it is text or some other content then an error message will be thrown back stating that Please enter images as input! The conversion of the input image format to the format required for processing shall be done by the system. Then the system shall calculate the mean square error for all the database images w.r.t the

input image. Then the mean square error is sorted in descending order to find out the best matching images. After sorting the mean square error, the best matches are displayed as the output of the system. It might be the case that the shape content of the input does not match well with the shape content of the database image. This means that the category in which the image lies is not included in the database. 3.3 Performance Requirements

Você também pode gostar