Free Essay

Tic Tac Toe

In:

Submitted By andreeeea
Words 14368
Pages 58
EBizPort: Collecting and Analyzing Business Intelligence Information

Byron Marshall, Daniel McDonald, Hsinchun Chen, and Wingyan Chung Artificial Intelligence Lab, Management Information Systems Department, University of Arizona, Tucson, AZ 85721. E-mail: {byronm, dmm, hchen, wchung}@eller.arizona.edu

To make good decisions, businesses try to gather good intelligence information. Yet managing and processing a large amount of unstructured information and data stand in the way of greater business knowledge. An effective business intelligence tool must be able to access quality information from a variety of sources in a variety of forms, and it must support people as they search for and analyze that information. The EBizPort system was designed to address information needs for the business/IT community. EBizPort’s collection-building process is designed to acquire credible, timely, and relevant information. The user interface provides access to collected and metasearched resources using innovative tools for summarization, categorization, and visualization. The effectiveness, efficiency, usability, and information quality of the EBizPort system were measured. EBizPort significantly outperformed Brint, a business search portal, in search effectiveness, information quality, user satisfaction, and usability. Users particularly liked EBizPort’s clean and user-friendly interface. Results from our evaluation study suggest that the visualization function added value to the search and analysis process, that the generalizable collection-building technique can be useful for domain-specific information searching on the Web, and that the search interface was important for Web search and browse support.

load (Bowman, Danzig, Manber, & Schwartz, 1994), and the questionable quality of many resources available on the Web. Appropriately adapted collection building and metasearching techniques developed in the digital library community hold promise to help address the challenges of the BI domain. In this report, we review the characteristics of existing techniques from a business intelligence perspective, explore a method of collecting quality resources, and test an interface designed to support BI activities. Once quality resources have been gathered, users need to interpret, assess, and exploit the knowledge embedded in those resources. While intelligent analysis tools and other forms of post-retrieval processing may support these processes, the human-computer interaction aspects need to be studied. Analysis is a complex human task but most search tools provide relatively simple search interfaces. This work seeks to identify the kind of user interface elements that can enhance human analysis of document lists returned from a query, and to assess the usefulness of a proposed document search interface in supporting Web information seeking in the business/IT domain.

Introduction Business intelligence (BI) has been defined as the acquisition, interpretation, collation, assessment, and exploitation of business-related information (Chung, Chen, & Nunamaker, 2003a).The primary objective of a BI system is to support sound decision making. However, gathering business intelligence from the vast set of available resources can be a challenge. Key issues include the rapid rate of change in the business and information technology (IT) environment, the often-referenced problems of information over-

Literature Review Business organizations need to stay on top of current information and developments in their field. Top executives spend a large portion of their time scanning for strategic information and rely on external sources for the information they need (Vedder, Vanecek, Guynes, & Cappel, 1999). Interestingly, smaller organizations scan the Internet more frequently than larger ones. Increased frequency of Internet scanning increases the effectiveness of environmental scanning (Tan, Teo, Tan, & Wei, 1998). Even when frequent scanning is conducted, low recall rates, outdated indexes, and limited analysis tools have been identified as impediments to effective information search (Chau, Zeng, & Chen, 2001). The use of improved information tools and collections would increase the effectiveness of the time spent on these scanning activities.

Accepted November 7, 2003 © 2004 Wiley Periodicals, Inc. ● Published online 25 March 2004 in Wiley InterScience (www.interscience.wiley.com). DOI: 10.1002/asi.20037

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY, 55(10):873– 891, 2004

Fuld & Company prepare an annual review of competitive intelligence (CI) software (Fuld, Sawka, Carmichael, Kim, & Hynes, 2002). In their 2002 Intelligence Software Report, they evaluated 13 Competitive Intelligence tools using a range of criteria and came to several conclusions. Good CI tools should handle multiple information repositories stored in various formats. Tools should be able to take advantage of various information repositories through metasearch, where a single query is passed to multiple search engines in multiple languages on the Internet. Tools should support comprehensive search capabilities such as phrase searching and searching with date ranges. Other functional effectiveness criteria for CI tools include filtering out noisy or irrelevant content through controlled Web-spidering, automatic filtering, relevance ranking, and dynamic summarization of articles and documents. Fuld considers support for information analysis to be the most lacking functionality in currently available competitive intelligence tools. These criteria identify important features and characteristics appropriate for the gathering of competitive intelligence. Digital library and information retrieval researchers have studied techniques that address many of the challenges identified for the business intelligence domain. The following section of this literature review focuses on the gathering of appropriate resources. Information Seeking for Business Intelligence provides background material on the information search process.

Gathering Appropriate Resources Content quality is a major contributor to overall system quality in a business intelligence tool. In this section, we review vertical search engines, describe techniques used to collect relevant documents, and present background information related to the measurement of information quality. Several techniques have been developed to build searchable, domain-specific, or vertical collections. Some of these approaches are described and considered in Building Vertical Collections. One alternative to collection-building is metasearch. A metasearch tool accepts user queries and passes them to external information sources for processing. Metasearch reviews the metasearch approach to gathering quality resources. Vertical collection building and metasearch processes aim to gather relevant resources. Relevance evaluation is usually based on some measure of topicality. Resources are selected because they are about the topic of interest. However, other measures of information quality may also be employed to gather useful resources. Information Quality and Business/IT Information Providers expands the notion of information quality and surveys existing types of business/IT information providers. Building vertical collections. Vertical search engines provide access to information on specific segments of the documents available on the Web. Vertical search engines provide customized functions and more precise results when

compared to general search engines like Google or AltaVista. One research group estimates the number of special purpose search engines at a quarter of a million sites (Bergman, 2001). Numerous small collections have been created using human-selected documents or content from published works. Human selection of documents is an expensive and increasingly inadequate process as the number of pages found on the Internet is counted in billions. Many techniques have been developed to collect documents for vertical search engines. Vertical collections can be created using a number of different automated approaches. Automated approaches use document collection mechanisms, often called spiders, to access pages. Spiders fetch documents from a list of starting or “seed” URLs. They identify additional page references in the document and often save the resources for further processing. This kind of “spidering” or Web-crawling process uses the structural information embedded in the Web to identify additional, potentially appropriate resources. Spiders follow links found in one page to additional pages and additional links. Sometimes, these spiders are guided by algorithms that assess the potential relevance of new links before fetching the pages. Collection algorithms can exploit content and/or structural information found in the documents to identify appropriate pages. Examples of the use of structural information to evaluate resources include PageRank (Brin & Page 1998), which uses inlink analysis to rank results, and the HITS algorithm (Kleinberg, 1999), which identifies hubs and authorities. Although some use of structure is common in collection-building tools, the use of document content to recognize relevant documents is crucial to most vertical collection building approaches. A few examples of vertical collection building systems are listed here. The Deadliner system uses formatting and key word heuristics to recognize desirable research documents (Kruger et al., 2000). The MERCATOR system allows for various controlling functions to be implemented (Najork & Heydon, 2001). It was used to test a focused-crawling collection-building approach that simultaneously creates a set of collections by finding documents clustered around term and document vector centroids (Bergmark, 2002). Chau and Chen (2003) compare breadth-first search, Page Rank, and Hopfield Net spiders’ effectiveness in building vertical collections in a medical domain. These and other collection techniques are intended to identify focused lists of documents related to a topic or domain of interest. Metasearch. Metasearching identifies relevant documents through unified access to multiple search engines. A metasearch tool processes queries from users by formatting the requests and passing them on to other search engines. The results are collated and presented to a user. Metasearch tools are intended to increase search coverage and improve search scalability (Meng, Yu, & Liu, 2002). There are billions of documents available on the Internet. Because even the larg-

874

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

est of the general purpose search engines index only a fraction of the World Wide Web (Kobayashi & Takeda, 2000), many search engines have been created and are available on the Internet. Metasearching aims to link distributed collections of information. Meng et al. (2002) identify a number of challenges to be faced when developing metasearch systems, including integrating local systems that use different indexing techniques and query types, discovering knowledge about component search engines, developing effective merging methods, and deciding where to place the software components of a metasearch engine. Stated simply, metasearch tool developers have to address the difficulties of accessing and merging information from disparate sources. In order to provide access to external information, a metasearch system developer must acquire an understanding of the functionality of the target information search engines. Search engine functions can be grouped into three categories (Huang, Ulrich, Hemmje, & Neuhold, 2001). Classification selection controls manage the sources or classes of documents to consider. Result display controls customize the format, size, or sorting of results. Query input controls accept terms, term modifiers, and logical operators. Metasearch tools must be flexible enough to match the variety of different search interfaces. One interesting system intended to help in establishing metasearch interfaces is the ACQUIRE system, which moves towards automatically extracting and maintaining a metasearch interface to search tools. It used the information it saves to generate its own user interface (Huang et al., 2001). The National Science Digital Library (NSDL) project aims to create widely accepted standards for interfacing with search portals (Lagoze et al., 2002). But for now, metasearch system developers are faced with a tremendous variety of search functions, input formats, and output formats. The second challenge for a metasearch system, integration of results, has been addressed in the literature using source selection and result merging algorithms. Systems have been developed using several classifications of algorithms to select the target search sites for a query. Site selection approaches have included rough representative, statistical representative, and learning-based approaches. The information about available collections is used to direct a query to appropriate sources and to select the number of documents to return from each of those sources. Meng, Wu, Yu, & Li (2001) review a number of these approaches. As an alternative to merging metasearch results based on heuristics and metadata, the results lists from several sources can be merged by fetching, indexing, and ranking each document. One example of this alternative approach is seen in MetaSpider (Chen, Fan, Chau, & Zeng, 2001). Metaspider uses a metasearch, fetch, and categorize technique to present integrated metasearch results to users. Merging results from various sources is a difficult task. Information about the local ranking of a document relative to a query for a particular database can be used but is not always available. When it is available, it is not in a consis-

tent format from source to source. No metasearch result merging approach has emerged in the literature or in industry practice as a clear favorite. Identifying a superior merging algorithm depends on identifying appropriate resource ranking measures. Information quality and business/IT information providers. Additional measures beyond traditional term relevance rankings can be considered when gathering results to present in response to a user query. For example, credibility is important when a system acts as a knowledge source or provides decision aids (Fogg & Tseng, 1999). Many factors may influence a user’s credibility decision, but researchers generally agree that trustworthiness and expertise are key components. Trustworthiness relates to the degree to which a Web site is truthful and unbiased while expertise relates to the experience and competence of those publishing on the Web site. Credibility, as demonstrated in trustworthiness and expertise, is important in this application because one of the main objectives of BI systems is providing support for decision making. While assessing a Web page’s information quality is considered an inexact science, useful metrics have been used in past research to evaluate a site’s information. Based on the framework of information quality (Wang & Strong, 1996), Katerattanakul and Siau (1999) define several useful characteristics of high-quality pages. Different information sources can be rated by their intrinsic quality and contextual quality. Intrinsic quality is interpreted to mean the accuracy of the contents and the accuracy of the hyperlinks contained in the Web page. This is measured by the lack of errors in the page content and the meaningful placement of working hyperlinks. The contextual quality of a Web site relates to the parties of the communication and the effect of the communication. Contextual quality improves when there is a clear identification of the author (Hlynka & Welsh, 1996). Contextual quality is also benefited when the author or Web site has some level of authority. These notions of information quality are seen in varying degrees in the products of existing business/IT information providers. Business/IT information providers can be loosely grouped into 4 categories: (1) content-focused sites, (2) news magazine portals, (3) information management tool builders, and (4) general search engines.




Content-focused sites: These sites create and collect quality information resources and provide access to them, often for a fee. Factiva, a Reuters company, is one example of this category of provider. It provides an interface to Reuters’ news. Much of the value of these sites comes from the special information it makes available to subscribers, information that is not available to the general public. These sites are popular sources of information in the business IT community. In Fogg and Tseng’s (1999) terms, perceived credibility of information found on these sites comes from the expertise and special knowledge of the content creators. News magazine portals: Several IT news magazines have

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

875





publicly available, searchable Web portals, which provide access to the content published in their magazines. Computerworld and IDG are two examples of this category. News magazine sites provide search interfaces to content generated by a staff of reporters. Articles provided by these sites are generally attributed to an author and presented with a publication date. These characteristics contribute to the contextual quality of the resources. The trustworthiness of these sites is sometimes questioned by readers who believe that there is a bias towards certain vendors. Information management tool builders: These software companies provide tools to help users or companies deal with information search needs. Copernic (www.copernic.com), for instance, sells enterprise searching and indexing software solutions for companies as well as desktop software to support user searches. Copernic also provides a dynamic summarization tool intended to help users more quickly identify relevant documents. Some of the features described in Fuld et al. (2002) are available in this provider category. General search engines: Google, Yahoo, and AltaVista are representatives of this category. These well-recognized service providers build very large collections and provide a search interface. While they process millions of requests for users all over the world daily, they have several significant limitations in the business intelligence domain. They are susceptible to search engine persuasion and spamming (Bergmark, 2002). They cover only a fraction of the Web and can contain many low-quality documents (Meng et al., 2001). They also do not provide interface elements to support analysis of returned lists of documents. In spite of these limitations, these resources form an important tool for business/IT managers because they provide access to a larger selection of resources than those found in specialized sites.

interact with the documents to perform the complex searching and analysis tasks that make the information useful for decision making. Thus, it is useful to review past research into human information seeking. Information Seeking for Business Intelligence Information seeking, as studied in previous research, typically adopts a process model. The process consists of various stages of problem identification, problem definition, problem resolution, and solution presentation (Wilson, 1999). Variations of the process model also can be found in the literature (Bates, 1989; Kuhlthau, 1991; Sutcliffe & Ennis, 1998). Bates’ model for information search, called “berrypicking,” captures the idea of an evolving multistep search process as opposed to a system that supports submitting single queries alone. Kuhlthau found that High School students began research assignments by conducting general browsing and performed more directed search as their understanding of the subject increased. Sutcliffe and Ennis succinctly summarized four main activities in their process model of information searching: problem identification, need articulation, query formulation, and results evaluation. In addition to directed searching, browsing the Internet is another strategy that information seekers frequently employ. Marchionini and Shneiderman (1988) defined browsing as “an exploratory, information seeking strategy that depends upon serendipity.” Spence (1999) defined “browse” as the registration of content into a human mental model. Having compared various definitions, Chung, Chen, and Nunamaker (2003a) defined “browsing” as an exploratory information seeking process characterized by the absence of planning, with a view to forming a mental model of the content being browsed. The Internet has recently evolved into a major information-seeking platform and was found to be one of the top five information sources for business analysis (The Futures Group, 1995). Information seeking on the Web has been characterized by its different depths of analyses. Chung et al. (in press) conducted an experiment to study the use of meta-searching, summarization, and categorization to help business analysts obtain business intelligence. They found that summarization and categorization were helpful in searching and browsing business information, and that their system could significantly augment existing search engines. In an intensive two-week study of Web-use activities, Choo, Detlor, and Turnbull (2000) found that knowledge workers engaged in a range of complementary modes of information seeking in their daily work. The 34 study participants came from 7 companies and held jobs as IT technical specialists or analysts, marketing staff, consultants, etc. They primarily utilized the Web for business purposes. The study confirmed that knowledge workers performed such multiple analyses as browsing, differentiating, monitoring, and extracting in the business domain. Although the study suggests a behavioral framework for Web-based information

Content gathering for BI. While the collection and integration of resources is a difficult task in any domain, the BI domain highlights some weaknesses in existing documentgathering techniques. High quality resources are maintained in a number of news magazine and content-focused portals on the Internet. However, because many of the content providers compete for readers or subscribers, they are not adequately cross-linked. The vertical collection building literature shows that most vertical collection algorithms rely on the structural information provided in embedded crosslinks to arrive at new pages. Therefore the lack of crosslinking between BI sources impacts the effectiveness of most of the previously described spidering techniques. Also, we observe that IT/Business news magazine articles are not frequently returned when using general search engines. Perhaps this is because competing information resource providers rarely link to each other so that the structural information used by general search engines does not tend to favor these credible, high quality resources. The competitive environment also reduces the effectiveness of many metasearch algorithms. Commercial sites do not provide the kind of detailed collection statistics that are used by many result-merging algorithms. Once appropriate methods have been constructed to identify and access quality information, people need to

876

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

seeking, it did not provide a system-based solution to domain-specific information seeking on the Web. Both a searchable, high-quality, domain-specific collection and browsing supports such as summarization, categorization, and visualization are needed to effectively facilitate multiple analyses. Moreover, an advanced document search interface is needed to support more efficient information seeking on the Web. The next section reviews search interface literature to understand how interfaces can be adapted to provide the above mentioned supports.

User Interfaces for Information Seeking In information-seeking environments today, “designers often fail to provide appropriate views of materials to give an overall sense of the structure and materials available” (Greene, Marchionini, Plaisant, & Shneiderman, 2000). Browsing tools should be provided along with query interfaces to help users navigate the information space and do so in a timelier manner. Many examples of browsing tools have been studied in the literature. Greene et al. suggest the use of previews and overviews. A preview acts as a surrogate for a single document of interest while an overview represents a number of documents. Overviews assist users in extracting the “gist more accurately and rapidly than traditional hit lists.” An effective overview gives the information seeker an idea of the size and extent of the information represented by the overview, how documents in the overview relate, and what kinds of documents are missing. Overviews. Browsing, or “overview,” tools in information visualization aim to increase users’ knowledge of their search space and increase the effectiveness and efficiency with which they search (Borner, Chen, & Boyack, 2003). ¨ Visualizing information is thought to improve the process of interacting with large volumes of data (Gershon, Eick, & Card, 1998; Card, Mackinlay, & Schneiderman, 1999; Chen, 1999; Spence, 2001). Documents (e.g., articles, Web pages, or patents) are the most common unit used in creating knowledge maps (Borner et al., 2003). Maps created from ¨ co-citation (Chen, 1999; Chen, Paul, et al. 2001), co-authorship (Mahlck & Persson, 2000), and co-word analysis (Bhattacharya & Basu, 1998) are also relatively common. Co-citation maps aid in inferring the intellectual structure of a field, co-authorship maps convey the social network of a discipline, and co-word analysis is used to understand the cognitive structure of a field (Borner et al., 2003). Docu¨ ment-level maps, however, have been used in tasks commonly associated with business intelligence, such as document retrieval, domain analysis (Small, 1999, 2000), and competitive intelligence analysis (Boyack, Wylie, & Davidson, 2002). Because document-level visualization is the most common in the area of information retrieval, we focus our review on visualization techniques that facilitate document visualization and analysis.

The Self-Organizing Map (SOM) classifies documents based on their text content using a neural network (Kohonen, 1995). The SOM algorithm automatically places documents into different regions on a map based on their similarity to one other. Adjacent regions are more similar than those regions far away on the map. The SOM is a document-based visualization tool that adds additional semantics by labeling the map. Lin, Soergel, and Marchionini (1991) first adopted the SOM for information visualization to document spaces. Their work visualized important concepts queried from a database. Chen, Houston, Sewell, & Schatz (1998) present a study using an SOM to visually categorize the Entertainment subcategory of Yahoo’s directory. The usefulness of the SOM in browsing is compared to the usefulness of the human-generated Yahoo categories. The results show the tools performed comparably. The SOM has also been studied in general Internet search tasks and the SOM has been shown to work with very large collections (Chen, Schufels, et al., 1996). Clustering techniques have also been used to reduce the dimensionality of the vector space model and group similar documents together. When producing the same number of clusters, Hearst (1999) showed that no one clustering algorithm outperformed the alternatives. Pirolli, Schank, Hearst, and Diehl (1996) use clustering to create groups of documents with similar topics. Automatic summaries are then generated at varying levels of granularity on the documents within the same cluster. The tree map approach (Shneiderman, 1992) has been used to determine the placement of clustered documents. Feng and Borner (2002) proposed ¨ semantic tree maps that rely on forced directed placement to calculate the placing of clusters. Pathfinder networks have also been used to display the content similarities among documents. A pathfinder network uses pairwise similarity measures to place nodes in a graph. Distance measures can come from subjective computation or numerical analysis. Chen (1997) developed a generalized framework for visualizing information sources called generalized similarity analysis (GSA). GSA constructs pathfinder networks based on content similarity between pages. The time to generate a pathfinder network makes it impractical to use in a dynamic Web search setting. Previews. Different from overviews, effective previews shown at the appropriate time give users enough information about a document to make an accurate relevance assessment (Greene et al., 2000). Usually, this information includes a list containing information about documents sorted by their relevance to the query. These lists usually contain the title of the document, followed by metadata, such as the publishing date of the article, and a short summary of the document. This type of information has been called a document surrogate (Witten, Nevill-Manning, McNab, & Cunningham, 1998). Document surrogates have been represented visually as well as in text. Hearst’s (1995) tile bars visually represent a document’s relation to the

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

877

query terms using rectangular bars. The rectangles are displayed in varying shades of gray to highlight the existence of query terms (Hearst, 1995). Other tools that visually represent document similarities to query terms include VIBE (Korfhage, 1991) and Lyberworld (Hemmje, Clemens, & Willett, 1994). Document summarization tools create textual document previews. Indicative summaries are those intended to assist the user in making relevance decisions (Firmin & Chrzanowski, 1999). Query-biased summaries use the queryterms entered by the user to determine what piece of the document to use as a preview for the user (Sanderson, 1998). Such summaries, sometimes referred to as key-wordin-context (KWIC), can include full sentences around query terms or only sentence fragments. Query-biased summaries were found to be more useful as previews than the first two sentences of a document (Tombros & Sanderson, 1998). Different from query-biased summaries, generic summaries use document-specific heuristics, such as tf idf, and cue phrases to pick sentences that best represent the essence of the document (McDonald & Chen, 2002). Summaries are often combined with other metadata to provide document previews. Research Questions Our review of previous research and existing tools raises two important issues: (1) How can quality resources be identified and organized for retrieval? and (2) What kinds of tools can assist with the kind of searching and analysis done in this process? The BI domain highlights some important issues related to vertical collection building and metasearching. Vertical collection-building methods generally rely on structural cross-link information to direct spiders to potentially relevant information. Effective metasearching, on the other hand, involves the merging of results from several sites. The information to do this merging can be difficult to obtain. Because cross-linking within the BI domain is limited and statistical collection information is largely unavailable, the effectiveness of traditional vertical collection techniques and metasearch methods is impaired. Support for analysis functions has been identified as a weakness in existing BI tools (Fuld et al., 2002). One part of the BI process is scanning, which can be seen as a kind of browsing task. Overviews and previews have been identified as important browsing support tools. However, the particular usefulness of various kinds of search support interface modules should be evaluated in the BI context. The BI process would be more effective if there were a smaller gap between the complexities of human analysis and the inadequacies of Web document search interfaces. To explore these issues we formed three research questions:
1. What kind of user interface elements can enhance human analysis of document lists returned in response to a query?

2. Can we develop an adaptable approach to building a high-quality domain-specific collection of Web documents? 3. To what extent is the document search interface and the high-quality, domain-specific collection useful to human beings who are engaged in Web information seeking in the business/IT domain?

Research Testbed The EBizPort (Electronic Business Intelligence Portal) system was developed to provide a platform for exploring these research questions. EBizPort System Design and Implementation describes the design and implementation of the EBizPort system. EBizPort implements metasearching as a user interface and as a method of harvesting Web documents. It also integrates analysis, visualization, and summarization functionality, which may assist users as they seek for relevant knowledge that can be gleaned from available information. Subsequently, Brint.com is discussed as a benchmark that will be compared to EBizPort (also see Evaluation Methodology and Experimental Results and Discussion). Brint is a vertical search engine focused on the Business/IT domain, which provides unified access to a variety of information sources.

EBizPort System Design and Implementation EBizPort was designed to support IT managers and other IT professionals. We reviewed available online portals and identified desirable information characteristics and search tool features from an IT manager’s point of view. Among other tasks, IT managers negotiate with vendors and scan the environment for upcoming threats and opportunities. Based on the review of the tasks and existing tools (see Literature Review), we have identified the following requirements for our system:








Credible sources: Our system attempts to direct users to credible sources. Information retrieved should be reliable and credible to support good decision making. Timeliness: Our system identifies the timeliness of the information retrieved. We want users to be able to control the currency of the information when searching. Efficient browsing: We want to save users time in browsing for information by providing only material relevant to the domain. Also, we want to provide tools to help users sift through documents, identifying those that are pertinent to the task at hand. Integration of financial information: We want to integrate some of the financial information available on the Internet. We want them to be able to easily connect companies in documents to financial and market information. Well-made IT investment decisions require a variety of types of information. Product reviews and industry trends are well covered in the magazines selected as primary sources in EBizPort. However, the financial strength of a potential vendor is often important. IT investments often amount to long-term part-

878

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

FIG. 1.

EBizPort architecture.

nerships. Organizations want to choose partners who will be around in the future.

These requirements are focused on business IT users but we believe the general techniques identified will be applicable to other domains. EBizPort architecture and components. With the constantly changing nature and relevance of business intelligence information, the architecture for a business intelligent system must be flexible and modular. These two ideals guided the design of the EBizPort architecture. The EBizPort site is built on a three-tier, client/server architecture as shown in Figure 1. The architecture is data centric with most modules interacting with the database. The three main functions provided in the architecture are collection creation, searching support, and browsing support. The local collection portion of the EBizPort is created off-line. Content is collected by gathering URLs from sitespecific search engines using a metaharvest program. The metaharvester written in Java connects to several Information Technology (IT) news sources and queries the collections using key words listed in a text file. A stored procedure stores the resulting URLs in the database. After harvesting the URLs, a Java indexer retrieves unindexed or out-of-date URLs from the database. The pages are fetched and parsed for phrases, key words, and publication date. The indexed data is saved in a text file, which is later added to the database using a stored procedure.

The user interacts with the search and browse interface elements. The search interface dispatches the query to all the search engines selected by the user. This includes both the local index and the metasearch modules. The middleware then combines the results and presents them to the user sorted by search engine and relevance. Browsing support is provided once search results have been returned. All browsing tools are dynamic and are generated when selected by the user. Browsing modules are called by the middleware allowing a clean interface for additional browsing modules in the future. The user can connect using any browser and submit searching and browsing requests to the system. Interactions with the database are conducted through stored procedures and utilize a database pool of reusable connections. Collection creation. To address the limitations of existing collection building and metasearching techniques, we decided to implement a hybrid metasearch approach. The lack of structural links to help guide spiders to additional pages is addressed by passing queries to existing search portals. The items returned from these portals are then spidered and parsed to create a local index. The local index of many potentially relevant documents allows us to integrate results based on the term frequency characteristics seen in a large collection of documents. Access to a full index allows the use of traditional, effective ranking algorithms and heuristics. This kind of algorithm does not require the kind of

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

879

TABLE 1. Source

EBizPort content sources. Description Contains “IT technologies, trends, career topics and management issues.” Includes articles from Jan. 1998 through Sept. 2001. Delivers “analysis and resonant storytelling on high-tech and business topics.” Searched through the Business2.0 interface Includes content on supply chains, web development, security, and IT services. Includes TechWeb articles. Searched through the Business2.0 interface “Source for Computing and Technology” Offers content from over 300 technology magazines and newspapers. Provides news, educational content, and interactive media. Reports on major news and events that affect IT. Contains “articles on the smartest, most innovative business practices.” Searched through 1DG interface Analyzes the “driving forces affecting innovation, technology, and financial markets.” Harvested URLs 71,730 38,881 40,682 26,863 33,500 30,696 26,775 37,955 15,846 20,200 23,276 20,200 10,503 pages pages pages pages pages pages pages pages pages pages pages pages pages

Computerworld Industry Standard Wired PC World InternetWeek Infoworld CNet IDG ITWorld CIO Business 2.0 Informationweek RedHerring

statistical information that is needed by many metasearch result merging algorithms. That is important because such information is very difficult to collect at best. At the same time, this approach avoids some of the issues associated with other vertical collection building techniques. Notably, the lack of cross-linking between sites does not affect the collection process because the spiders do not follow any links found in the pages. Based on conversation with IT professionals and business school professors and an extensive Web search, we identified 13 different sources for news and commentary in the IT domain, shown in Table 1. Contextual quality of these resources is enhanced because they provide author names and publication dates. Credibility is enhanced when the reporters and columnists are perceived to have substantial expertise. The inclusion of several different providers is intended to help address potential concerns about bias. While a user may perceive a particular magazine as biased, perhaps providing results from several sources will increase the chance that a user will encounter believable information. After initially including general search engines, we limited the content providers to commercial sites with name recognition in the IT industry. The content provided by these sites is usually managed using content management software and is not heavily linked to pages outside the site’s domain. The small number of in-links to these pages means they do not rank high in the relevance rankings of search engines like Google. The high quality of these pages and their low representation in general search engines makes them good candidates for a domain-specific search engine. To build our local collection, we arranged access to the 13 sites, generated a list of 635 queries, and implemented several document conversion modules. The queries contain words extracted from approximately 3 years of Computerworld back issues. A portion of the queries contain non-IT words in an attempt to expand our list of unique URLs. For example, certain key words that provide long lists of results in target systems are combined with the names of each of the 12 months. The 12 resulting queries return a larger set of unique URLs. Using these queries and 13 sources of IT news and commentary, we were able to gather and index

over 400,000 URLs to provide broad coverage of business IT topics. Because business intelligence information is often in a format other than HTML on the Internet, our collection also includes Word, Excel, PDF, and text format files. Although our site does index the content, it does not actually serve up the articles. Instead, EBizPort directs the user to the commercial site to access the content, thereby not infringing on copyrights. Web pages, including those from commercial content providers, are noisy. Advertisements, menu bars, and dynamic news content are mixed with the relevant business intelligence content on each page. In order to filter out the unwanted text, we used HTML parsing and sentence recognition routines to index only the content that appeared in a full sentence. This allowed us to ignore nearly all the irrelevant content without missing important information about a page. In addition, the collection index included the tag source of each term to record whether it appeared in a page’s header, meta, title, or body tags. These tags are used to improve retrieval accuracy (Arasu, Cho, et al. 2001). Finally, we extracted the publish date of the article where that date was available in the meta tags. We also used date recognition techniques to extract the date from the page content itself when not available in the meta tags. Dates extracted were checked against the current date to avoid erroneous publication dates. In all, we extracted publication dates for nearly 50% of the pages contained in our local index. Searching support. The EBizPort searching interface is shown in Figure 2. A search for the words “Dell” and “Expansion” is entered in the search box for 10 hits within the last year. The user can choose to search other sites as well as searching the local database. Figure 2 also shows the interface for selecting a few of the search sources. Results that are returned from the metasearch modules are ordered and listed by source, while the results from the local index have the advantage of having been merged and ranked together. As users metasearch the news sites, new URLs returned that are not in the local collection are added to the

880

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

FIG. 2.

Enter search parameters.

database to keep the site up-to-date. The searching interface includes several “advanced” options. All these options are available when searching the local index and are available in different combinations when metasearching. Users can search for phrases by putting several words on the same line and use an and/or radio button option to request Boolean processing of the phrases. To support IT managers’ need for searching by date, nearly 50% of the indexed pages contain dates. These pages are searched when the user selects a date parameter and the entire collection is searched when the option “all documents” is chosen. The TF*IDF equation orders the list of results. Individual terms are weighted differently based on their position within different HTML tags. The summary or snippet provided by the information source is presented with each search result. The user can

also elect to see a generic summary with a user-defined number of sentences. In addition to the searching features provided to the user, a human-generated collection of links was added to the left column of the search screen, which can be reviewed by IT managers. The links are placed in general categories and are representative of the relevant IT resources available on the Internet. The information categories represented are “IT Resources,” “Business News,” “StockMarket,” “TechTips,” and “Events.” In addition to the metasearch modules that gather IT news articles, there are also two that gather financial data (see Figure 3). As users scan the competitive landscape, they are able to obtain more specific company descriptions as well as the financial positions of companies relative to their main competitors by using the company-related meta-

FIG. 3.

EBizPort news article and financial data search results.

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

881

FIG. 4.

Browse tools.

search engines. Modules access SmartMoney.com and MoneyCentral.com to obtain financial information. In the present implementation, the stock symbol is entered in the first line of the query box. Financial results are presented along with search results in the results page. Browsing support. EBizPort implements three analysis tools to help users sift through available data. The categorizer and visualizer provide overviews while the summarizer provides previews. The categorizer provides a folder display based on document clusters. It leverages a common interface metaphor to help users find relevant documents and browse the results space. The visual organization produced by the SOM algorithm in the visualizer allows users

to see the major clusters of documents found in a search result. The summarizer tool lets users see the main themes in a document while showing where those themes occur in the source text. These tools are shown in Figure 4. The first tool is a summarizer that generates generic summaries of single documents found in the list of results (McDonald & Chen, 2002). Summary sentences are generated and given a color code on the left side of the screen. On the right side of the screen, the original document is loaded with the summary sentences highlighted according to the color scheme. By clicking on the summary sentence, users are taken to the location in the document that contains the sentence of interest. By including a summary along with the original, users can quickly go to the areas of the document

882

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

that seem the most relevant. This feature seems especially useful with longer documents. The second browsing tool built into the EBizPort is an analyzer that creates a folder view of the retrieved content. In this view, all the pages returned in a search, whether metasearch or local, are sent to the Arizona Noun Phraser (Tolle & Chen, 2000). Phrases that occur frequently in the results collection are selected as the folder labels. If a document contains the same phrase as one of the folder labels, it is included underneath that folder. Documents can appear in several folders if they share many important phrases with other documents. The third browsing tool we incorporated in the EBizPort is the Self-Organizing Map (SOM) (Chen, Houston, Sewell, & Schatz, 1998). The map classifies all the pages returned in the list of results onto a 2-dimensional display. Users can click on the different regions of the map and see the URLs under the selected region in the right-hand column of the display. A larger region means more documents are in that classification. Adjacent regions are more similar in content than nonadjacent regions. EBizPort design summary. EBizPort creates value for users in two main ways. It provides access to a focused and credible set of resources and it implements a set of interesting analysis tools. In addition, the techniques used to create EBizPort are relatively scalable and modular with a built-for-maintenance paradigm. This combination of characteristics may be useful to others who want to create vertical search portals or digital libraries. Because the business information landscape changes rapidly, the architecture and database implementation of the system support site maintainability. The ranking algorithm was altered slightly so that documents could be added without re-ranking the entire database. This was done by altering the standard tf*idf calculations to use a constant of 1 million in place of the total number of documents in the collection. This has little impact on results but eases the computational burden of adding incremental results. The clear separation of functions between the front and back ends of the system makes it easier to implement changes. The JSP and servlet model used to create the user interface allow services like the market and financial information lookup module to be added based on other resources on the Internet without disrupting the existing functionality. The collection-building technique also is designed for maintainability. Using the metaharvesting approach to collect URLs allows the collection to be easily “refreshed.” We have refreshed our collection twice by rerunning the queries. Resulting URLs are checked for duplication as they are loaded into the database. This combined with the databasedriven indexing procedure allows easy implementation of various heuristics to identify pages that need to be respidered on a frequent basis. Metasearching for URLs and then creating a local index based on those documents is a promising method for building vertical search engines. The technique leverages the

value created by other vertical collection builders. It does not replace, but leverages focused crawling and other collection-building techniques. Sending queries to general search engines and implementing unguided crawling produce noisy results. Competitive intelligence system developers may be able to use the metaharvesting approach described here to reduce the complexity of collection-building processes. Sending queries to selected sites allows the builder to address the perceived credibility of the documents in a collection although it requires site-specific adjustments to accomplish the harvesting of valuable documents. Once pages are collected, site-specific heuristics are required to do a creditable job of content date extraction and filtering of unwanted information from Web pages. This is not an easy process but it does create significant value in the usefulness of the index. Brint.com Brint.com, the BizTech Network, is a Web portal that provides information from many sources in various formats regarding general business, information technology, and knowledge management. It was founded by Dr. Yogesh Malhotra, a Management faculty at Syracuse University, with co-founder Mini Malhotra. Before its 1996 launch, Brint had existed as the site “A Business Researcher’s Interest” dating back to 1994. Brint’s free content consists of two main information access methods: “Topic Specific Portals” and the “Portal on Demand” search. “Topic Specific Portals” provide directory-style access to information and can be reached through the menu of Channels, Resources, and Community. “Portal on Demand” search, the signature service of Brint, aggregates and consolidates content from multiple back-end databases. “Portal on Demand” uses a MetaSearch engine that gathers information from some of the best business and technology publications on the Web. Content from external sources are merged with local resources and ranked according to query relevance in real time. Brint.com is a good information portal benchmark for the EBizPort for three reasons. First, Brint is a popular site that is frequently visited. Brint currently has a 4:1 lead in link popularity statistics over prominent Web publications and Web sites on the topic of knowledge management (www.brint.com). Furthermore, Brint holds a 10:1 lead in link popularity over most of the top ten Web sites returned by search engines for a “knowledge management” search. In addition, Brint has approximately 100,000 registered users, even though registration is not required to use the site. Among the registered users are Fortune 500 and Global 2000 companies. Second, Brint offers a lot of information about the Information Technology industry, including current news headlines. InfoWorld recognized Brint as one of the “best Web sites for keeping up with hi-tech industry innovations. . . .” In March 1997, Computerworld’s Annual Forecast declared Brint a “Best” site “especially for its relevance to the multifaceted needs and concerns of IS

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

883

professionals. ” Brint’s large volume of IT news and information makes its content comparable to that of the EBizPort. Finally, Brint’s collection-building techniques of integrating a local index with metasearching are analogous to the strategies incorporated by EBizPort. The popularity of Brint along with its similarity to EBizPort in content and collection-building technique makes it a good benchmark for testing the additional functionality of the EBizPort. Evaluation Methodology In this section, we describe our methodology used to evaluate EBizPort. We present the objectives, experimental tasks, hypotheses, and design as follows. Objectives and Experimental Tasks Our evaluation objectives were threefold: (1) to evaluate the effectiveness, efficiency, and usability of EBizPort; (2) to evaluate the quality of the collection built using our proposed approach; and (3) to evaluate how the tools (summarizer, categorizer, visualizer, navigation side-bar) of EBizPort are used to help browsing for business information on the Web. In all the above evaluation objectives, BRINT.com (BRINT), a business search portal, was selected to be the benchmark for comparison. To evaluate how the search portals assist business information seeking, scenario-based search tasks and browse tasks were designed. An example of a search task is “In December 2002, IBM announced the acquisition of software company for $2.1 billion. What is the name of this company? ” An example of a browse task is “What is the expansion plan of Dell Inc. as reported within the most recent year? Summarize your answers into a number of distinct themes. ” The theme identification method was used to evaluate performance in browse tasks (Chen et al., 2001). To achieve objective (1), we compared the performance of EBizPort with that of BRINT in search and browse tasks. We also asked the subjects to provide subjective ratings on the usability and information quality of the two portals. To achieve objective (2), we compared the search performance between using EBizPort’s domain-specific collection together with other metasearchers, and only using EBizPort’s metasearchers without searching its domain-specific collection. To achieve objective (3), we analyzed how subjects used each tool in terms of the length of time spent on it and the number of themes obtained. Hypotheses Three groups of hypotheses were tested (see Table 2). To compare the effectiveness of the systems, we used accuracy for search tasks and precision and recall for browse tasks. A single measure called F value was used to combine recall and precision (Shaw, Burgin, & Howell, 1997). Efficiency was measured by the amount of time used to finish a task.

TABLE 2. Code

Hypotheses tested in the experiment. Hypothesis

1. Effectiveness and efficiency of the portals H1.1 EBizPort is more effective than BRINT in search tasks. H1.2 EBizPort is less efficient than BRINT in search tasks. H1.3 EBizPort is more effective than BRINT in browse tasks. H1.4 EBizPort is less efficient than BRINT in browse tasks. H1.5 A combination of EBizPort and BRINT is more effective than either EBizPort or BRINT in browse tasks. 2. Users’ subjective evaluation on usability and information quality H2.1 EBizPort achieves higher usability than BRINT. H2.1a In terms of users’ subjective ratings on usefulness, EBizPort achieves higher usability than BRINT. H2.1b In terms of users’ subjective ratings on ease of use, EBizPort achieves higher usability than BRINT. H2.1c In terms of users’ subjective ratings on information display and interface design, EBizPort achieves higher usability than BRINT. H2.2 EBizPort provides a higher information quality than BRINT. H2.2a In terms of presentation quality and clarity, EBizPort provides a higher information quality than BRINT. H2.2b In terms of coverage and reliability, EBizPort provides a higher information quality than BRINT. H2.2c In terms of usability and analysis quality, EBizPort provides a higher information quality than BRINT. H2.3 EBizPort users achieve a higher overall satisfaction than BRINT users. 3. Quality of EBizPort’s domain-specific collection H3.1 Using EBizPort’s domain-specific collection together with EBizPort’s metasearchers is more effective than only using EBizPort’s metasearchers in search tasks. H3.2 Using EBizPort’s domain-specific collection together with EBizPort’s metasearchers achieves comparable efficiency to only using EBizPort’s metasearchers in search tasks.

The formulae used to calculate the above metrics are stated below. Number of correctly answered parts Total number of parts

Accuracy Precision

Number of relevant results identified by the subject Number of all results identified by the subject Recall Number of relevant results identified by the subject Number of relevant results identified by the expert 2 Recall Precision Recall Precision

F value

Hypotheses on effectiveness and efficiency. In H1.1–H1.4 (Table 2), we hypothesized that EBizPort would be more

884

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

effective than BRINT in search and browse tasks because EBizPort’s high quality collection and domain-specific metasearchers provide better search results than BRINT’s metasearching. Also, we believed that EBizPort’s browse supports would be better than BRINT’s navigation supports. However, because BRINT has professional support while EBizPort is just a prototype, users should be less familiar with EBizPort and would, therefore, spend more learning to use it. Thus, we hypothesized that EBizPort would be less efficient than BRINT. In H1.5 (Table 2), we believed that combining browse supports from both systems would achieve better effectiveness than only using the browse supports from either one of the systems. Since we expected to obtain significantly different results from the two systems, combining the results from them would significantly increase recall but create only a small change in precision. Through this arrangement, we tried to mimic a situation in which each subject was allowed to use EBizPort and BRINT together to solve the same problem. Hypotheses on usability and information quality. In H2.1 (Table 2), we hypothesized that EBizPort would achieve a higher usability than BRINT because EBizPort provides a wide range of metasearchers to choose from, supports specialized functions like company profile search, and has a clean and user-friendly interface. In H2.2 (Table 2), we believed that EBizPort would have a better information quality because EBizPort has a high quality, domain-specific collection that gives good results for searching and browsing. Also, the different tools of EBizPort add values to searching and browsing. Based on the cited advantages of EBizPort, we therefore believed that EBizPort’s users would achieve higher overall satisfaction ratings (H2.3, Table 2). Hypotheses on quality of EBizPort’s domain-specific collection. In H3.1 (Table 2), we believed that using EBizPort’s domain-specific collection would increase effectiveness of EBizPort’s metasearchers in performing search tasks because the collection was built from a wide range of high-quality information sources. In H3.2 (Table 2), we believed that using EBizPort’s domain-specific collection would not significantly increase the time required for searching, thus efficiency would not be changed significantly. Experimental Design The experiment required each subject to perform 3 search tasks and 1 browse task using each system. A time limit of 3 minutes was imposed on each search task and 8 minutes on each browse task. Among the 3 search tasks, the first two tasks asked about recent activities of a company in the IT industry and the third task asked about the company’s profile information. The fourth task was a browse task that asked subjects to find themes related to recent events of the

company. When using EBizPort, subjects were instructed not to use EBizPort’s domain-specific collection in task 1 but had to use EBizPort’s collection in task 2 (but they could use EBizPort’s metasearchers in either task), so that we could compare the effect of using EBizPort’s collection. All tasks were randomly assigned to different questions to avoid bias due to task content. A pilot test involving three subjects was conducted to evaluate the appropriateness of the tasks before they were actually used in the experiment. Thirty subjects, mainly business college students, were recruited and each of them received a fixed amount of money as an incentive for their voluntary participation. Many of the subjects also had prior industry work experience. During the experiment, a subject used each of the two systems to perform the tasks. The order in which the systems was used was randomly assigned to the subjects to avoid bias due to system sequence. As each subject was asked to perform similar tasks using the two systems, a one-factor repeated-measures design was used, because it gives greater precision than designs that employ only between-subjects factors. Verbal comments and observational notes were recorded and analyzed. After finishing the tasks with a system, a subject rated the system on: (1) the usability of the system, (2) the information quality provided by the system, and (3) his/her overall satisfaction with the system. To measure usability, we identified three categories (perceived usefulness, perceived ease of use, and information display and interface design) of items from two usability questionnaires frequently used in the information system field (Davis, 1989; Lewis, 1995). To measure information quality, we modified the 16-dimension construct developed in Wang and Strong (1996) by dropping the dimension on “security,” which is not relevant because the information provided by the systems is already public. In addition, because there are different levels of importance in the remaining 15 dimensions, we invited an expert to provide ratings on the relative importance of different dimensions. Such ratings were used to weigh the different dimensions of information quality for the business domain. His ratings as well as the definitions of the 15 dimensions categorized into three groups are shown in Table 3. The expert holds an MBA degree with extended studies in management information systems, marketing, and entrepreneurship and has 14 years of work experience in the IT industry. Upon finishing with the two systems, the subject was asked to fill in a post-study questionnaire asking about their preferences on the systems and suggestions for improvements or other comments. Answers from search tasks were graded by their correctness while answers to browse tasks were graded according to the expert’s judgment. To increase the quality of the expert’s judgment, he was first required to provide a set of answers after using both EBizPort and BRINT, and to organize the answers into themes. After the data from all subjects had been collected, the expert read subjects’ answers and modified his original answers if

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

885

TABLE 3.

Definitions of 15 dimensions of information quality and expert ratings. Definition Expert rating

Dimension Presentation quality and clarity Accessibility Concise representation Consistent representation Ease of manipulation Interpretability Coverage and reliability Appropriate amount of information Believability Completeness Free-of-error Objectivity Usability and analysis quality Relevancy Reputation Timeliness Understandability Value-Added Note. Expert rating: 3

The extent to which information The extent to which information The extent to which information The extent to which information The extent to which information and the definitions are clear

is is is is is

available, or easily and quickly retrievable compactly represented presented in the same format easy to manipulate and apply to different tasks in appropriate languages, symbols, and units,

3 1 3 1 3

The extent to which the volume of information is appropriate for the task at hand The extent to which The extent to which depth for the task The extent to which The extent to which The The The The The extent extent extent extent extent to to to to to which which which which which information information at hand information information information information information information information is regarded as true and credible is not missing and is of sufficient breadth and is correct and reliable is unbiased, unprejudiced, and impartial is is is is is applicable and helpful for the task at hand highly regarded in terms of its source or content sufficiently up-to-date for the task at hand easily comprehended beneficial and provides advantages from its use important.

2 2 3 3 3 2 3 2 2 1

extremely important, 2

very important, 1

needed. The final list of the expert’s answers was obtained after this two-step process and was used to evaluate the performance of the systems. Experimental Results and Discussion In this section we describe and analyze the results of our user evaluation study. Table 4 summarizes the system effectiveness and efficiency in search and browse tasks. Table 5 shows the mean ratings on various dimensions, Table 6
TABLE 4. BRINT. Searching and browsing performance of EBizPort and

shows the p values and results of testing various hypotheses, and Table 7 summarizes subjects’ profiles. Effectiveness and Efficiency of the Portals The result of testing hypothesis H1.1 shows that EBizPort was significantly more effective than BRINT for search tasks. We believe this could be attributed to EBizPort’s high-quality domain-specific collection and company profile search function that provide precise and relevant search results. The insignificant difference in search efficiency (H1.2) could be attributed to the fact that both portals metasearched other information sources and hence took a significant amount of time to obtain results. The results of testing H1.3 and H1.4 show no significant difference in the effectiveness and efficiency for browse tasks. We believe that both portals provided useful browse supports that were complementary to each other, because EBizPort obtained results from major business information sources while BRINT obtained results from major search engines. The results of testing H1.5 show that a combination of the two portals outperformed both EBizPort and BRINT in browse tasks, thereby confirming our belief that EBizPort could significantly augment BRINT with its browse supports. Users’ Subjective Evaluation and Verbal Comments The results from testing H2.1–H2.2 show that EBizPort achieved a significantly higher usability and information quality than BRINT along all the different dimensions. We believe that EBizPort’s superior usability was attributed to its user-friendly interface and functionalities that support

Portal EBizPort

Task Searcha Browse

Measure Accuracy Efficiencyb Precision Recall F value Efficiencyb Accuracy Efficiencyb Precision Recall F value Efficiencyb Precision Recall F value

Mean performance 72.22% 2.48 81.67% 22.06% 34.00% 5.46 47.22% 2.61 83.05% 25.67% 37.85% 5.56 87.54% 45.94% 57.97%

Std. deviation 29.14% 1.28 31.29% 11.58% 16.27% 2.60 27.36% 1.32 23.63% 14.44% 17.79% 3.19 13.09% 18.24% 17.66%

BRINT

Searcha Browse

Combination (EBizPort BRINT) a b

Browse

The performances of the three search tasks were averaged. Efficiency was measured in minutes.

886

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

TABLE 5.

Results of users’ subjective evaluations. EBizPort Dimension Mean rating 5.27a 5.30a 5.29a 5.17a 2.70b 3.03b 2.39b 2.67b 3.20b Std. deviation 1.11 1.24 1.09 1.26 1.31 1.55 1.34 1.54 1.88 Mean rating 4.15a 4.19a 4.29a 4.04a 4.19b 4.48b 3.51b 4.58b 4.90b BRINT Std. deviation 1.46 1.38 1.73 1.66 1.47 1.95 1.49 1.53 1.88

Information quality (overall) Presentation quality and clarity Coverage and reliability Usability and analysis quality Usability (overall) Perceived usefulness Perceived ease of use Information display and interface design Overall satisfaction a b

The range of rating is from 1 to 7 with 7 being the best. The range of rating is from 1 to 7 with 1 being the best.

searching and browsing high quality business information. In particular, users’ ratings on EBizPort’s information display and interface design contributed most to EBizPort’s superior usability, as reflected by the low p value ( 0.000) in testing H2.1c. It showed that EBizPort has provided a highly usable document search interface that contributes to users’ satisfaction in searching and browsing. As subject 22 commented on EBizPort: “excellent interface, gives all possible options, very flexible, very interactive, easy to learn and use. ” In addition, we believe that EBizPort’s significantly higher information quality was attributed to the use of our vertical collection-building approach that yielded a high-quality, domain-specific collection to facilitate credible, precise, and relevant searching. Subject 16 pointed out that EBizPort had “credible sources and high flexibility.”
TABLE 6. Results of hypothesis testing.

The result of testing H2.3 shows that users’ overall satisfaction with EBizPort was significantly higher than that of BRINT. This is because EBizPort provided searching and browsing for high-quality business information and a variety of other credible sources, together with such useful browse supports as navigation links, summarizer, categorizer, and visualizer. As subject 12 said, “the visualization tool is helpful in terms of its classifications. The appearance of the Web site (EBizPort) is appealing and easy to understand. The meta search function allows for detailed search within each magazine of interest.” In contrast, BRINT only provided metasearching and many hyperlinks, some of which were broken or irrelevant to the tasks at hand. Subjects complained about BRINT’s cluttered user interface. As sub-

p value H1: Effectiveness and efficiency of the portals H1.1 Search task effectiveness: EBizPort BRINT H1.2 Search task efficiency: EBizPort BRINT H1.3 Browse task effectiveness: EBizPort BRINT H1.4 Browse task efficiency: EBizPort BRINT H1.5a Browse task effectiveness: Combined EBizPort H1.5b Browse task effectiveness: Combined BRINT H2: Users’ subjective evaluations H2.1 Usability: EBizPort BRINT H2.1a Usability (usefulness): EBizPort BRINT H2.1b Usability (ease of use): EBizPort BRINT H2.1c Usability (information display and interface design): EBizPort BRINT H2.2 Information quality: EBizPort BRINT H2.2a Information quality (presentation quality and clarity): EBizPort BRINT H2.2b Information quality (coverage and reliability): EBizPort BRINT H2.2c Information quality (usability and analysis quality): EBizPort BRINT H2.3 Satisfaction: EBizPort BRINT H3: Quality of EBizPort’s domain-specific collection H3.1 Search effectiveness: (EBizPort collection meta searchers) EBizPort meta searchers H3.2 Search efficiency: (EBizPort collection meta searchers) EBizPort meta searchers Note. Alpha error 5%; For details of the hypotheses, please refer to Table 2.

Result

0.001 0.616 0.479 0.867 0.000 0.000 0.000 0.003 0.007 0.000 0.000 0.000 0.000 0.001 0.002 0.211 0.958

Confirmed Not confirmed Not confirmed Not confirmed Confirmed Confirmed Confirmed Confirmed Confirmed Confirmed Confirmed Confirmed Confirmed Confirmed Confirmed Not confirmed Confirmed

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

887

TABLE 7.

Subjects’ profiles. Subjects’ profile 2.32 2.36 (range: 1–7, with 1 being “strongest reliance”)

TABLE 9.

Subjects’ preferences along different dimensions. Number of subjects who selected the following system to be their first preference Dimension EBizPort 17 20 18 18 20 23 20 20 16 Brint 1 2 3 3 4 2 3 2 4

Attribute Average number of years in industries Degree of reliance on the Web to search for information in subjects’ work experience Time spent on using computer (hours per week)

Gender Education level

Age

5 subjects spent between 20–25 hours per week, 6 subjects spent between 25–30 hours per week, 6 subjects spent between 30–35 hours per week, 12 subjects spent more than 40 hours per week 27 subjects are male, 3 are female 1 subject is an undergraduate student, 8 subjects have earned a bachelor’s degree, 18 subjects have earned a master’s degree, 3 subjects did not provide education information 12 subjects age between 18–25, 8 subjects age between 26–30, 6 subjects age between 31–35, 1 subject ages between 36–40, 3 subjects did not provide age information

To search for specific company information To search for information from various business information sources To search for high-quality business information To browse business-related Web resources To categorize a large number of search results into meaningful groups To visualize a large number of search results To achieve effective integration of information from different sources To discover meaningful patterns among search results To analyze business information about certain issues

ject 7 commented that BRINT’s “interface (was) painful to look at, (provided) out-of-date information and dead links.” Subject 5 said that “information overload” was a major problem of BRINT and more than 10 subjects said that BRINT put too much information on the Web pages, thus making the interface too cluttered. Table 8 summarizes subjects’ verbal comments and Table 9 lists subjects’ preferences along different dimensions. From the above results, we conclude that EBizPort achieved significantly better usability, information quality, and user satisfaction than BRINT. EBizPort’s Domain-Specific Collection and Browse Supports The results of testing H3.1 and H3.2 show that the search effectiveness and efficiency of using EBizPort’s collection
TABLE 8. Portal EBizPort A summary of subjects’ verbal comments. Positive comments - Interface is simple and easy to use - Provided useful tools to enhance the searching ability

plus metasearchers were not significantly different from those of only using metasearchers. We believe this could be attributed to the way that EBizPort presented results. EBizPort presented lists of results according to different sources. While users might expect to see one integrated long list, they actually got many lists of results put on the same page in a linear manner. Some users did not pay particular attention to the list of results from EBizPort’s collection because that list was not always put at the top of the page. They might incorrectly think that those items put at the top of the page were more important than those put at the bottom. To understand how EBizPort helped browsing for business information, we recorded subjects’ actions in using different tools, namely, search function, navigation side bar, summarizer, categorizer, and visualizer. We also have defined an information-seeking episode as a subpart of the whole information seeking process that involves using a

Negative comments - The processing speed was sometimes slow - Too many results coming from different magazines might overwhelm users - Users were not familiar with categorizer and hence did not find it useful - Irrelevant results were found - The interface was too cluttered and created information overload - Many broken links were found - Not flexible in selecting information sources - Information provided was outdated and not from credible sources

BRINT

- Allowed summarization, categorization, and visualization that other search engines couldn’t provide - Provided searching of credible sources - Users were more familiar with its interface, which is similar to that of Yahoo or Google - Provided links to reputable Web sites - Provided searching of many information sources

888

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

TABLE 10.

Usage of EBizPort’s tools. Navigation side bar 0.00 1.33 2 0 0 0 0 0

Metric Proportion of themes obtained (%) Average amount of time spent (in seconds) Number of subjects who used this tool in First episode Second episode Third episode Fourth episode Fifth episode Sixth episode

Search function 68.83 111 28 9 11 5 5 4

Summarizer 0.00 3.43 0 3 2 0 2 0

Categorizer 3.70 27 0 6 1 3 1 1

Visualizer 27.47 55.1 0 7 4 2 1 2

particular tool to achieve a subgoal of the process. By identifying information-seeking episodes of each subject, we can analyze how the tools were used and how they contributed to finding results. Table 10 summarizes the usage of the tools. Each subject on average went through 3.77 episodes in performing his browse task. As expected, the search function was the most frequently used tool and contributed most to finding themes. While the visualizer and categorizer were the next two most frequently used tools, the visualizer was found to contribute much more help to users in find themes, as shown in the proportion of themes obtained. We, therefore, conclude that EBizPort’s search function and visualizer were the two major tools that contributed to subjects’ performance in browse tasks. Implications of the Results The results of this user evaluation study have three implications. First, the encouraging results suggest that our vertical collection-building approach can benefit domainspecific information seeking on the Web. Since many domain-specific communities use the Web to share and disseminate information, our approach for building searchable collections is valuable to these communities. Second, Web document search interface was found to be very important for search and browse support as shown in subjects’ favorable comments on EBizPort’s interface. EBizPort provides an example of a clean, user-friendly document search interface that reduces information overload in Web information seeking. However, the confusing presentation of search results (see EBizPort’s Domain-Specific Collection and Browse Supports) and the relatively small usage of certain interface components (e.g., navigation side bar) also suggest that further improvements of EBizPort’s interface are needed. Third, visualization was found to add value to Web searching and browsing because it amplifies human cognition and reduces information overload. Further research in developing visualization tools that can benefit information seeking on the Web is thus needed. Conclusions and Future Directions To utilize the valuable information on the Internet, Business Intelligence tools must address the searching and anal-

ysis challenges that come with extracting useful intelligence information from large volumes of information and data. Such a portal must address the challenges of content gathering and the reported lack of analysis tools. EBizPort was created with these specific needs in mind. We incorporated metasearch with local indexing to leverage different information sources and yet address the challenge of merging results. We provided a search interface with preview and overview tools to enhance the users’ ability to analyze information retrieved. To evaluate the effectiveness of our tools, we conducted a user study comparing the EBizPort to an existing and reputable knowledge management portal. Results showed that EBizPort was significantly more effective in search tasks and could augment a comparable site in browse tasks. Users also rated EBizPort as having a significantly higher information quality, usability, and overall satisfaction than the comparable site. The analysis tools in the EBizPort are the first step to helping users dynamically visualize the news and rapidly changing conditions in the IT industry. With the large amount of time-sensitive material published on the Web, we are interested in analyzing the changes of IT topics over time. The publication date information extracted by EBizPort should help in this task. In addition, a tool kit containing some of the analysis tools we have described is under development. We plan to make this tool kit available to the research community.

Acknowledgments This research is partly supported by NSF Digital Library Initiative-2, “High-performance Digital Library Systems: From Information Retrieval to Knowledge Management,” IIS-9817473, April 1999 to March 2002. We greatly appreciate the help of our expert, Mr. Cortez W. Smith, and our evaluation team members Alfonso A. Bonillas and Theodore Elhourani for their hard work in completing the user study.

References
Arasu, A., Cho, J., Garcia-Molina, H., & Raghavan, S. (2001). Searching the Web. ACM Transactions on Internet Technologies 1(1), 2– 43.

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

889

Bates, M.J. (1989). The design of browsing and berrypicking techniques for the on-line search interface. Online Review 13(5), 407– 431. Bergman, M.K. (2001). The deep Web: Surfacing hidden value. The Journal of Electronic Publishing 7(1). Bergmark, D. (2002). Collection synthesis. In G. Marchionini & W. Hersh (Eds.), The Second ACM/IEEE-CS Joint Conference on Digital Libraries. Portland, OR: ACM Press. Bhattacharya, S., & Basu, P.K. (1998). Mapping a research area at the micro level using co-word analysis. Scientometrics, 43, 359 –372. Borner, K., Chen, C., & Boyack, K.W. (2003). Visualizing knowledge ¨ domains. Annual Review of Information Science and Technology, 37, 179 –255. Bowman, C.M., Danzig, P.B., Manber, U., & Schwartz, F. (1994). Scalable Internet resource discovery: Research problems and approaches. Communications of the ACM, 37(8), 98 –107. Boyack, K.W., Wylie, B.N., & Davidson, G.S. (2002). Domain visualization using VxInsight for science and technology management. Journal of the American Society for Information Science and Technology, 53, 764 –774. Brin, S., & Page, L. (1998). The anatomy of a large-scale hypertext Web search engine. In P.H. Enslow & A. Ellis (Eds.), Proceedings of the Seventh International World Wide Web Conference, Brisbane, Australia: Elsevier Science Publishers B.V. Card, S., Mackinlay, J., & Shneiderman, B. (Eds.). (1999). Readings in information retrieval: Using vision to think. San Francisco, CA: Morgan Kaufmann. Chau, M., & Chen, H. (2003). Comparison of three vertical search spiders. Computer, 36, 56-62. Chau, M., Zeng, D., & Chen, H. (2001). First ACM/IEEE-CS Joint Conference on Digital Libraries (pp. 79 – 87). Roanoke, VA: ACM Press. Chen, C. (1997). Structuring and visualizing the WWW with generalized similarity analysis. In M. Bernstein, K. Østerbye, & L. Carr (Eds.), Proceedings of the Eighth ACM Conference on Hypertext (pp. 177– 186). Southampton, UK: ACM Press. Chen, C. (1999). Informaton visualisation and virutual environments. London: Springer-Verlag. Chen, C., Paul, R.J., & O’Keefe, B. (2001). Fitting the jigsaw of citation: Information visualization in domain analysis. Journal of the American Society for Information Science and Technology, 52, 315–330. Chen, H., Schufels, C., & Orwig, R. (1996). Internet categorization and search: A self-organizing approach. Journal of Visual Communication and Image Representation, 7(1), 88 –102. Chen, H., Houston, A.L., Sewell, R.R., & Schatz, B.R. (1998). Internet browsing and searching: User evaluations of category map and concept space techniques. Journal of the American Society for Information Science, 49(7), 582– 603. Chen, H., Fan, H., Chau, M., & Zeng, D. (2001). MetaSpider: Metasearching and categorization on the Web. Journal of the American Society for Information and Science and Technology, 52(13), 1134 – 1147. Choo, C.W., Detlor, B., & Turnbull, D. (2000). Information seeking on the Web: An integrated model of browsing and searching. First Monday, 5(2). Chung, W., Chen, H., & Nunamaker, J.F. (2003a). Business Intelligence Explorer: A knowledge map framework for discovering business intelligence on the Web. Proceedings of the 36th Hawaii International Conference on System Sciences, Hawaii, Big Island, USA, IEEE Computer Society. Chung, W., Zhang, Y., Huang, Z., Wang, G., Ong, T.H., & Chen, H. (in press). Internet searching and browsing in a multilingual world: An experiment on the Chinese Business Intelligence Portal (CBizPort). Journal of the American Society for Information and Science and Technology. Davis, F.D. (1989). Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Quarterly, 13(3), 319 –340. Feng, Y., & Borner, K. (2002). Using semantic treemaps to categorize and ¨ visualize bookmark files. In R.F. Erbacher, P.C. Chen, M. Groehn, J.C.

Roberts, & C.M. Wittenbrink (Eds.), Proceedings of the SPIE Conference on Visualization and Data Analysis. (p. 24). San Jose, CA: IS&T and SPIE. Firmin, T., & Chrzanowski, M.J. (1999). An evaluation of automatic text summarization systems. In I. Mani, & M.T. Maybury (Eds.), Advances in automatic text summarization, I (pp. 325–336). Cambridge: The MIT Press. Fogg, B.J., & Tseng, H. (1999). The elements of computer credibility. CHI 99 conference on Human factors in computing systems: The CHI is the limit. Pittsburgh: ACM Press. Fuld, L., Sawka, K., Carmichael, J., Kim, J., & Hynes, K. (2002). Intelligence software report 2002. Cambridge, MA: Fuld & Company. The Futures Group (1995). Ostriches and eagles: Competitive intelligence capabilities in U.S. companies. Glastonbury, CT: The Futures Group. Gershon, N., Eick, S., & Card, S. (1998). Design: Information visualization. Interactions, 5(2), 9 –15. Greene, S., Marchionini, G., Plaisant, C., & Schneiderman, B. (2000). Previews and overviews in digital libraries: Designing surrogates to support visual information seeking. Journal of the American Society for Information Science, 51(4), 380–393. Hearst, M. (1995). TileBars: Visualization of term distribution information in full text information access. In I.R. Katz, R. Mack, L. Marks, M.B. Rosson, & J. Nielsen (Eds.), Proceedings of the ACM SIGCHI Conference on Human Factors in Computing Systems (pp. 29 – 66), Denver, CO. Hearst, M. (1999). User interfaces and visualization. In B. Ribeiro-Neto (Ed.), Modern information retrieval (pp. 257–339). New York: ACM Press. Hemmje, M., Clemens, K., & Willett, A. (1994). LyberWorld: A visualization user interface supporting fulltext retrieval (pp. 249 –259). The 17th Annual International ACM/SIGIR Conference. Dublin, Ireland. Hlynka, D., & Welsh, J. (1996). What makes an effective home page? A critical analysis. Retrieved from http://www.umanitoba.ca/faculties/education/cmns/aect.html. Huang, L., Ulrich, T., Hemmje, M., & Neuhold, E.J. (2001). Adaptively constructing the query interface for meta-search engines. The Sixth International Conference on Intelligent User Interfaces. Santa Fe: ACM Press. Katerattanakul, P., & Siau, K. (1999). Measuring information quality of Web sites: Development of an instrument. The 20th International Conference on IS. Charlotte, NC: Association for Information Systems. Kleinberg, J. (1999). Authoritative sources in a hyperlinked environment. Journal of the ACM (JACM), 46(5), 604 – 632. Kobayashi, M., & Takeda, K. (2000). Information retrieval on the Web. ACM Computing Surveys (CSUR), 32(2), 144 –173. Kohonen, T. (1995). Self-organizing maps. Berlin: Springer-Verlag. Korfhage, R.R. (1991). To see or not to see: Is that the query? Proceedings of the 14th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (pp. 134 –141). Chicago, IL: ACM Press. Kruger A., Giles, C.L., Coetzee, F.M., Glover, E., Flake, G.W., Lawrence, S., et al. (2000). DEADLINER: Building a new niche search engine. The Ninth International Conference on Information and Knowledge Management. McLean, VA: ACM Press. Kuhlthau, C.C. (1991). Inside the search process: Information seeking from the user’s perspective. Journal of the American Society for Information Science, 42(5), 361–371. Lagoze, C., Arms, W., Gan, S., Hillman, D., Ingram, C., Krafft, D., et al. (2002). Core services in the architecture of the national science digital library (NSDL). The Second ACM/IEEE-CS Joint Conference on Digital Libraries. Portland, OR: ACM Press. Lewis, J.R. (1995). IBM computer usability satisfaction questionnaires: Psychometric evaluation and instructions for use. International Journal of Human-Computer Interaction, 7(1), 57–78. Lin, X., Soergel, D., & Marchionini, G. (1991). A self-organizing semantic map for information retrieval. Proceedings of the ACM SIGIR Conference on Research and Development in Information Retrieval (pp. 262– 269). Chicago, IL: ACM Press. Mahlck, P., & Persson, O. (2000). Socio-bibliometric mapping of intradepartmental networks. Scientometrics, 49, 81–91.

890

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

Marchionini, G., & Shneiderman, B. (1988). Finding facts vs. browsing knowledge in hypertext systems. Computer, 21(1), 70 –79. McDonald, D., & Chen, H. (2002). Using sentence-selection heuristics to rank text segments in TXTRACTOR. Proceedings of the Second ACM/ IEEE-CS Joint Conference on Digital Libraries (pp. 28 –35). Portland, OR: ACM Press. Meng, W., Yu, C., & Liu, K.-L. (2002). Building efficient and effective metasearch engines. ACM Computing Surveys (CSUR), 34(1), 48 – 89. Meng, W., Wu, Z., Yu, C., & Li, Z. (2001). A highly scalable and effective method for metasearch. ACM Transactions on Information Systems (TOIS), 19(3), 310 –335. Najork, M., & Heydon, A. (2001). High-performance Web crawling (SRC Research Report No. 173). Retrieved October 10, 2002, from http:// gatekeeper.research.compaq.com/pub/DEC/SRC/research-reports/ abstracts/scr-rr-173.html Pirolli, P., Schank, P., Hearst, M., & Diehl, C. (1996). Scatter/gather browsing communicates the topic structure of a very large text collection. In M.J. Tauber (Ed.), Proceedings of the Conference on Human Factors in Computing Systems (pp. 213–220). Vancouver, British Columbia, Canada: ACM Press. Sanderson, M. (1998). Accurate user directed summarization from existing tools. Conference on Information and Knowledge Management, Bethesda, MD. Shaw, W.M.J., Burgin, R., & Howell, P. (1997). Performance standards and evaluations in information retrieval test collections: Cluster-based retrieval models. Information Processing and Management, 33(1), 1–14. Shneiderman, B. (1992). Tree visualization with tree-maps: A 2-D space filling approach. ACM Transactions on Graphics, 11, 92–99. Small, H. (1999). A passage through science: Crossing disciplinary boundaries. Library Trends, 48, 72–108.

Small, H. (2000). Charting pathways through science: Exploring Garfield’s vision of a unified index to science. In H.B. Atkins (Ed.), The web of knowledge: A Festschrift in honor of Eugene Garfield (pp. 449 – 473). Medford, NJ: Information Today, Inc. Spence, B. (2001). Information visualization. Reading, MA: AddisonWesley. Spence, R. (1999). A framework for navigation. International Journal of Human-Computer Studies, 51(5), 919 –945. Sutcliffe, A.G., & Ennis, M. (1998). Towards a cognitive theory of information retrieval [Special issue]. Interacting with Computers, 10, 321–351. Tan, S.S.L., Teo, H.-H., Tan, B.C.Y., & Wei, W.-K. (1998). Environmental scanning on the Internet. Proceedings of the International Conference on Information Systems. Helsinki, Finland: Association for Information Systems. Tolle, K.M., & Chen, H. (2000). Comparing noun phrasing techniques for use with medical digital library tools. Journal of the American Society for Information Science, 51(4), 352–370. Tombros, A., & Sanderson, M. (1998). Advantages of query biased summaries in information retrieval. Melbourne, Australia: ACM SIGIR. Vedder, R.G., Vanecek, M.T., Guynes, C.S., & Cappel, J.J. (1999). CEO and CIO perspectives on competitive intelligence. Communications of the ACM, 42(8), 108 –116. Wang, R.Y., & Strong, D.M. (1996). Beyond accuracy: What data quality means to data consumers. Journal of Management Information Systems, 12(14), 5–34. Wilson, T.D. (1999). Models of information behavior research. Journal of Documentation, 55(3), 249 –270. Witten, I.H., Nevill-Manning, C., McNab, R., & Cunningham, S.J. (1998). A public library based on full-text retrieval. Communications of the ACM, 41(4), 71–75.

JOURNAL OF THE AMERICAN SOCIETY FOR INFORMATION SCIENCE AND TECHNOLOGY—August 2004

891

Similar Documents

Free Essay

Tic Tac Toe

...Charles Hudgins Jaun Hernandez PT1420 4/29/2014 Let's Start: Create a New project and select windows application. Give it name and then press enter.(All of us know this step). After that you see a blank form like that: Now add 9 buttons on your form using tools. Select button then delete the text from properties (i.e. button1, button2).Actually we want no text on buttons. Now, Put a label on your form set back color according to your wish (here I choose sky-blue color from properties), set text align to middle center and set text to "Turn”. Actually we are creating a box for showing the player's Turn. Add Another Label below above label. Remove text from it, leave it blank. Give it name="display turn”. Now our Application looks something like that. Now doing the same above method, I have created a Scorecard for player1 and player2.I give name playerscore1 and playerscore2 for the labels as shown in figure. I am creating this game for 2 person so, I supposed here Player1 symbol="X" and Player2 Symbol="O".Now I want when player1 click it gives "X" symbol on button. After his turn when Player2 click on button then it show "O" on another button. Double click on button1 and add the code given below. Collapse | Copy Code void Button1Click(object sender, EventArgs e) { if(click1==0) { if(turn%2!=0) { button1.Text="X"; } else { button1.Text="O"; } turn++; click1++; } else { ...

Words: 687 - Pages: 3

Premium Essay

Tic Tac Toe Programming

...Introduction Tic-Tac-Toe is a popular simple game for two players all over the world. Traditionally one person plays for “X” while another plays for “O”. The game is played on the square board of 3x3 or even more size. The aim of this game for player is to win by filling with “X” or “O” any row, column or diagonal of the main game board. The game sets players thinking because players should reflect a little bit on strategy and predict some moves of the opponent. So it teaches people to make decisions quickly. MATLAB is computer software for technical computing from the MathWorks. Used for a wide variety of scientific and engineering calculations. The functions of the MATLAB are solving equations, plotting 2D, 3D graphs, calculating statistics, etc. The aim of the report is to describe how Tic-Tac-Toe game simulation was made with the help of MATLAB. The report describes how program reads the code and computer strategy for winning/not losing in the game. Theory. The aim the individual coursework was to program 4x4 Tic-Tac-Toe game for player and computer using MATLAB with condition that computer should always win or just not loose. Programming the game basic loops such as “while”, “for”, “if” were used, where “while” means to repeat smth until some particular conditions are holding, “for” means to do smth for particular conditions, and “if” means to do smth if some conditions are true. There also were used “sum” function which helps to count “X”-s or “O”-s in rows, columns...

Words: 3911 - Pages: 16

Premium Essay

Tic Tac Toe Java Code

...OBJECT ORIENTED PROGRAMMING TIC TAC TOE Contents: Page 1. Specifications 3 2. Background 4 3. UML Design i. Use Case Diagram 5 ii. Class Diagram 6 4. Implementation i. How code is working? 7 ii. Conclusion 9 iii. Future Work 10 5. Appendix 10 6. References 15 7. Bibliography 15 1. Specifications: A requirements specification for a software system - is a complete description of the behavior of a system to be developed. It includes a set of use cases that describe all the interactions the users will have with the software. Use cases are also known as functional...

Words: 1071 - Pages: 5

Free Essay

The Ultimate Puzzle

...How to Win and More Specifically How to Not Lose at Tic Tac Toe In order to successfully complete the puzzle you may not enter the outer spaces. The outer left, outer right, center top, center bottom will all cause you to immediately lose. Your opponent will immediately see what is going on and place his marker in the center center position. Once that occurs, you will be forced to choose one of the corners, which will place you in a position to get a tic, tac, toe. However, in blocking your tic, tac, toe, your opponent will select the other corner, which means that your next move must be to block him from getting the tic, tac toe. Upon seeing you block him, your opponent will then chose the alternate corner from where his current center piece and side piece are located. In doing this, your opponent is forcing you to choose to block him in either the center center, or the alternate corner. This puts you in a situation that is impossible to win. You may try to block him in one section however you will not be able to prevent him from winning the game of tic, tac, toe, since he will place his piece in the remaining slot and defeat you. It is critical at the beginning of the game to never select these four positions as it is impossible to defend if your opponent knows what he is doing. Best of luck as you try to conquer this ancient...

Words: 256 - Pages: 2

Free Essay

Tic Tac Oe

...Computer Programming Computer Programming TIC TAC TOE PROJECT TIC TAC TOE PROJECT Flor Garcia FINAL EXAM Flor Garcia FINAL EXAM 2015 2015 DEFINE THE PROBLEM: Write a Tic Tac Toe program that inputs the player’s names and then asks if you want to play again. It displays a score board at the end of the game. OUTLINE THE SOLUTION: INPUT | PROCESSING | OUTPUT | Name of Player 1 | Names of Both players is inputted | Score board of who won, lost or any draws | Name of Player 2 | Player 1 makes the first move by choosing a number in the box then [X] will replace the chosen number | | Numbers on each box | Player 2 makes the second move by choosing a number in the box then [O] will replace the chosen number | | | If either of the players win, they receive a point in the score board | | | If it is a draw, the score board will display it | | DEVELOP AN ALGORITHM: 1. Insert name of player 1 [X] 2. Insert name of player 2 [O] 3. Players choose the number of the box they want to input their letter in. 4. Once then choose the number they inputted their designated letter will replaced the number in the box 5. These steps are repeated as it is each player’s turn 6. The game will end until someone wins or it becomes a draw 7. The Score board will be displayed at the end TEST THE ALGORITHM: 1. Name of Player 1: Flor 2. Name of Player 2: Karla 3. Flor chooses number 3 and is replaced by [X] 4. Karla chooses...

Words: 1048 - Pages: 5

Premium Essay

A Beautiful Mind

...Beautiful Mind A Beautiful Mind is a book written by Sylvia Nasar and is A Touchstone Book published by Simon and Schuster. A Beautiful Mind won the 1998 National Book Critics Award for Biography, a contract for a movie deal, and a nomination for The Pulitzer Prize for Biography. This book is based on a man by the name of John Nash who is a brilliant mathematician and also a paranoid schizophrenic. Even though Nash struggled with paranoid schizophrenia he was still able to revolutionize a concept called game theory. This work that he did with game theory greatly improved what was already established as well as created the Nash equilibrium. Game theory is the mathematical study of strategies used to win games. It began with games like tic-tac-toe and chess since they are easy to analyze because they are known as games of complete information. Complete information is when your opponent’s positions are in clear site. Game theory then went on to analyze card games where player’s cards aren’t seen, which is known as incomplete information. In card games there are elements such as bluffing that can become a variable in the analysis. Mathematicians continued to analyze more critical games such as economics, biology, philosophy, and even which girl to go after. This is around the time that John Nash developed a principle for mutual consistency of player’s strategies which is known as the Nash equilibrium. The Nash equilibrium can be applied to a wider variety of games than the original...

Words: 482 - Pages: 2

Premium Essay

Wajihaider

...|[pic] |Syed Muhammad Waji Haider | |Born 1994 | | | |Address: House no.7, Bastami road, samnabad lahore | | |Contact: 0307-0408091 | | |Email: waji58@gmail.com | | |Site: waji58.wordpress.com | | |Date of birth: February 12, 1994 | | | | |Personal Profile |Reliable and dependable in meeting objectives | | |Hard working, possess strong planning, organizing and monitoring abilities ...

Words: 513 - Pages: 3

Premium Essay

Personal Narrative: Macaroni Grill

...eight, my family and I went to the Macaroni Grill for my cousin’s birthday. This group included my uncle Marvin. Marvin personifies the ‘goofy, old uncle that gives his nieces and nephews enough candy to get cavities’ stereotype. Whenever we ate out with him, he would always sit at the border of the adult’s side of the table and the kid’s. He would make our stomachs ache with his jokes and our brains throb with his riddles. Overall, he kept us entertained through the usually boring meals. The lunch at Macaroni Grill started off like any other: Marvin sat near us kids and we proceeded to find ways to entertain ourselves as the meal dragged on. My younger brother, Tyson, challenged me to a very serious game of tic-tac-toe which would finally decide who had the title of Tic-Tac-Toe World Champion. This game would go on to help define my life. As I surpass Tyson in every way, it soon became clear to him what the outcome of the game would be. Not ready to face defeat he dashed his red crayon across the paper in an attempt to ruin our game and thus not lose his chance at the crown. I could not take such treachery. I quickly shot my arm across the table to put a stop to the double-crossing, but in my effort I accidently displaced a glass of water....

Words: 653 - Pages: 3

Premium Essay

Concept Map

...ACID-BASE DISORDERS MADE SO EASY EVEN A CAVEMAN CAN DO IT Lorraine R Franzi, MS/HSM, RD, LDN, CNSD Nutrition Support Specialist University of Pittsburgh Medical Center Pittsburgh, PA I. LEARNING OBJECTIVES The clinician after participating in the roundtable will be able to: 1) Indicate whether the pH level indicates acidosis or alkalosis. 2) State whether the cause of the pH imbalance is respiratory or metabolic. 3) Identify if there is any compensation for the acid-base imbalance. II. INTRODUCTION Acid-Base balance is an intricate concept which requires an intimate and detailed knowledge of the body’s metabolic pathways used to eliminate the H+ ion. Clinicians may find it daunting to understand when first introduced to the subject. This roundtable session will demonstrate how to analyze blood gas levels in a very elementary manner so as to diagnose any acid-base disorder in a matter of minutes. The body is in a constant state of flux delicately stabilizing the pH so as to maintain its normalcy. In order to prevent untoward effects of alkalosis or acidosis the body has three major buffering systems that it uses to adjust the pH. They are: 1) Plasma protein (Prot-) 2) Plasma hemoglobin (Hb-) 3) Bicarbonate (HCO3-) The Bicarbonate-Carbonic acid system is the most dominate buffering system and controls the majority of the hydrogen ion (H+) equilibrium. Maintaining homeostasis when these acid-base shifts occur is vital to survival. Metabolic and respiratory...

Words: 1735 - Pages: 7

Free Essay

Activites for Small Children

...Resource II - Science| Activity: What's in the air experiment |Age Group: 3 year olds | Goal/Objective : Children learn the dangers of fire | Materials Needed: · plate · candle, & matches · play dough · jar & colored | Process and Teaching Strategies · Use playdough to stand candle up in the middle of the plate, then pour colored water around the candle. Teacher then lights the candle and talks to children about the danger of matches and fire. Put the jar over the candle and watch the flames disappear| How is this activity developmentally appropriate? It teaches children they should only handle with an adult | Resource II - Language and Literacy | Activity: Learning to write your name |Age Group: 3 year olds | Goal/Objective : help child write his/ her name | Materials Needed: · chalk · sidewalk· | Process and Teaching Strategies · Let the child see his/her name. Make marks everywhere. Provide a vertical writing surface. Give fingers a workout. Hunt for letters| How is this activity developmentally appropriate? It gives the child a chance to write and see their name spelled out | Resource II - Creative Arts | Activity: Foot Tracing|Age Group: 3 year olds | Goal/Objective : Team work | Materials Needed: · drawing paper· crayons· markers | Process and Teaching Strategies · Give each child a partner and take turns tracing each others foot | How is this activity developmentally appropriate? Gives the opportunity...

Words: 613 - Pages: 3

Free Essay

Video Games

...A video game is an electronic game that involves human interaction with a user interface to generate visual feedback on a video device. The word video in video game traditionally referred to a raster display device, but it now implies any type of display device that can produce two- or three-dimensional images. The electronic systems used to play video games are known as platforms; examples of these are personal computers and video game consoles. These platforms range from large mainframe computers to small handheld devices. Specialized video games such as arcade games, while previously common, have gradually declined in use. Video games have gone on to become an art form and industry. The input device primarily used to manipulate video games is called a game controller, and varies across platforms. For example, a controller might consist of only a button and a joystick, while another may feature a dozen buttons and one or more joysticks. Early personal computer games often needed a keyboard for gameplay, or more commonly, required the user to buy a separate joystick with at least one button. Many modern computer games allow or require the player to use a keyboard and a mouse simultaneously. A few of the most common game controllers are gamepads, mouses, keyboards, and joysticks. In recent years, additional methods of input have emerged such as camera-based player observation for video game consoles and touch-sensitive screens on mobile devices. Video games typically use...

Words: 854 - Pages: 4

Free Essay

Freemasons Cipher

...Brice Matthews Dr. Linda Harned Communication 1201-240 28 Aug 2015 Learning the Freemason's Cipher Don't be threatened by the name, this is actually one of the most simple ways of writing a secret message, or cipher, that there is. Before we get into the specifics of the certain cipher I am going to be teaching you today, the first thing you need to know is the difference between the terms code and cryptography. Code is the broader term, which most of us are familiar with, simply meaning “a system of rules to convert information”. One of the earliest examples of code being the invention of language itself, because it allowed a form of communication to convert information. Today though, we are mostly going to be focused on cryptography. All cryptography is code but not all code is cryptography; codes can be a code of law, a code of conduct, or a dress code. It is simply a way of giving information in a way that is comprehensible. Cryptography is the use of ciphers and code to protect a message or secret. A cipher is a secret/disguised way of writing code. In order to effectively decode (a term we are all familiar with) a cipher, you need to possess the knowledge of the way it was originally ciphered. Ciphers can be kept solely in your mind and then spread through word of mouth, or written on a piece of paper or even implemented into mechanical device, which is then required to break the cipher, unless you can remember that information and store it. All ciphers can technically...

Words: 1006 - Pages: 5

Free Essay

Asdasdasd

...Step 1: Materials Tools * Arduino * Processing * Wire cutters * Soldering iron * Box cutter Materials * (3) 270k resistors * (3) 10k resistors * Solder * Wire * Aluminum foil * Cardboard Optional: * Tape (e.g.: scotch) * Shielded wire (e.g.: coaxial cable, ~3') * (3) alligator clips * 3-pin header * Zip-tie * Shrink wrap tubing or hot glue Step 2: Make the Plates This sensor will work using simple RC circuits, with each circuit sensing distance in one dimension. I found that the easiest way to arrange three capacitive plates for this purpose is in the corner of a cube. I cut the corner of a cardboard box into an 8.5" cube, and then cut some aluminum foil to fit as slightly smaller squares. Tape on the corners keeps them in place. Don't tape down the entire perimeter, we'll need it later for attaching the alligator clips. Step 3: Make the Connectors To connect the Arduino to the plates we need some shielded wire. If the wire isn't shielded, the wires themselves act more obviously as part of the capacitor. Also, I've found that alligator clips make it really easy to connect things to aluminum -- but there are probably plenty of other ways, too. * Cut three equal lengths of shielded cable. I chose about 12". The shorter the better. Coaxial cable works, but the lighter/more flexible the better. * Strip the last half inch or so to reveal the shielding, and the last quarter inch to reveal...

Words: 1116 - Pages: 5

Free Essay

Prg 410 Week 3 Assignment

...PRG 410 WEEK 3 ASSIGNMENT http://www.coursehomework.com/product/prg-410-week-3-assignment/ Contact us at:  +1 315-750-4434  help@coursehomework.com PRG 410 WEEK 3 ASSIGNMENT C++ Programming I PRG-410 Assignment#3 Problem 1: [marks 20] Write a program that implements simple game scenario. Game has two-player or a three-player MODE. For each MODE you have to choose among four HEROs. These are ALPHA, BRAVO, CHALIE and DELTA. You have to design and implement proper notation for each MODE and implement them accordingly using appropriate conditions e.g., First player is selecting ALPHA then Second Player can’t select ALPHA. Against each MODE ask five values for each player if: • Ask input range is between (1-100) player will win with 10 points for one value. • Ask input range is between (101-200) player will win with 20 points for one value. • Ask input range is between (201-500) player will win with 50 points for one value. Add all five values for HERO/HEROs and printout the winner of the game. Problem 2: Rock, Paper, Scissors Game [marks 20] Write a program that lets the user play the game of Rock, Paper, and Scissors against the computer. The program should work as follows. 1. When the program begins, a random number in the range of 1 through 3 is generated. If the number is 1, then the computer has chosen rock. If the number is 2, then the computer has chosen paper. If the number is 3, then the computer has chosen scissors. (Don’t display the...

Words: 933 - Pages: 4

Free Essay

Bank of Activities

...Banks activities No | Activity | Preparation | Rules | Players | Material | Time | 1 | I Spy | Write many words on a blank piece of paper in random order and positioned at all angles. Make copies of the page for each of your students to have one. To play, you read a word to your class. They race to find that word on the page. The first one to find it reads it out loud and then circles it or crosses it out. That person scores one point. Then wait until everyone in class has found the word. Read a second word, and students then search for that one on the page. The first to find it scores a point. As you call out more words, students will be reading many of the words on the page.. Play until someone reaches a certain number of points or until you have called out all the words you want your students to read. | Everyone has a paper with words written on. I will read every single word and you guys have to find that word. The first one to find it has to read it out loud and will score 1 point. Play till the last word and the one with most points wins. | 6-7 | Words paper | 20 mins | 2 | Snake and Ladder | Draw a grid 5x5 (depends on how long you want to play). Each square will have a question/ puzzle/ command/ picture. Depending on the question, who lands on the square will have to answer or do that. Otherwise, some squares will have snakes, which force to go back, and ladders which force them to go up. Play until one team reachs the final square. | We will divide into 2 teams...

Words: 1739 - Pages: 7