Off the Top: Searching Entries
Showing posts: 16-30 of 35 total posts
Flickr and the Future of the Internet
Peter's post on Flickr Wondering triggers some thoughts that have been gelling for a while, not only about what is good about Flickr, but what is missing on the internet as we try to move forward to mobile use, building for the Personal InfoCloud (allowing the user to better keep information the like attracted to them and find related information), and embracing Ubicomp. What follows is my response to Peter's posting, which I posted here so I could keep better track of it. E-mail feedback is welcome. Enjoy...
You seemed to have hit on the right blend of ideas to bring together. It is Lane's picture component and it is Nadav's integration of play. Flickr is a wonderfully written interactive tool that adds to photo managing and photo sharing in ways that are very easy and seemingly intuitive. The navigations is wonderful (although there are a few tweak that could put it over the top) and the integration of presentational elements (HTML and Flash) is probably the best on the web as they really seem to be the first to understand how to use which tools for what each does best. This leads to an interface that seems quick and responsive and works wonderfully in the hands of many. It does not function perfectly across platforms, yet, but using the open API it is completely possible that it can and will be done in short order. Imagine pulling your favorites or your own gallery onto your mobile device to show to others or just entertain yourself.
Flickr not only has done this phenomenally well, but may have tipped the scales in a couple of areas that are important for the web to move forward. One area is an easy tool to extract a person's vocabulary for what they call things. The other is a social network that makes sense.
First, the easy tool for people to add metadata in their own vocabulary for objects. One of the hinderances of digital environments is the lack of tools to find objects that do not contain words the people seeking them need to make the connection to that object they are desiring. Photos, movies, and audio files have no or limited inherent properties for text searching nor associated metadata. Flickr provides a tool that does this easily, but more importantly shows the importance of the addition of metadata as part of the benefit of the product, which seems to provide incentive to add metadata. Flickr is not the first to go down this path, but it does it in a manner that is light years ahead of nearly all that came before it. The only tools that have come close is HTML and Hyperlinks pointing to these objects, which is not as easy nor intuitive for normal folks as is Flickr. The web moving forward needs to leverage metadata tools that add text addressable means of finding objects.
Second, is the social network. This is a secondary draw to Flickr for many, but it is one that really seems to keep people coming back. It has a high level of attraction for people. Part of this is Flickr actually has a stated reason for being (web-based photo sharing and photo organizing tool), which few of the other social network tools really have (other than Amazon's shared Wish Lists and Linkedin). Flickr has modern life need solved with the ability to store, manage, access, and selectively share ones digital assets (there are many life needs and very few products aim to provide a solution for these life needs or aims to provide such ease of use). The social network component is extremely valuable. I am not sure that Flickr is the best, nor are they the first, but they have made it an easy added value.
Why is social network important? Helping to reduct the coming stench of information that is resultant of the over abundance of information in our digital flow. Sifting through the voluminous seas of bytes needs tools that provide some sorting using predictive methods. Amazon's ratings and that matching to other's similar patterns as well as those we claim as our friends, family, mentors, etc. will be very important in helping tools predict which information gets our initial attention.
As physical space gets annotated with digital layers we will need some means of quickly sorting through the pile of bytes at the location to get a handful that we can skim through. What better tool than one that leverages our social networks. These networks much get much better than they are currently, possibly using broader categories or tags for our personal relationships as well as means of better ranking extended relationships of others as with some people we consider friends we do not have to go far in their group of friends before we run into those who we really do not want to consider relevant in our life structures.
Flickr is showing itself to be a popular tool that has the right elements in place and the right elements done well (or at least well enough) to begin to show the way through the next steps of the web. Flickr is well designed on many levels and hopefully will not only reap the rewards, but also provide inspiration to guide more web-based tools to start getting things right.
Fixing Permalink to Mean Something
This has been a very busy week and this weekend it continues with the same. But, I took two minutes to see if I could solve a tiny problem bugging me. I get links to the main blog, Off the Top, from outside search engines and aggregators (Technorati, etc.) that are referencing content in specific entries, but not all of those entries live on the ever-changing blog home page. All of the entries had the same link to their permanant location. The dumb thing was every link to their permanant home was named the same damn thing, "permalink". Google and other search engines use the information in the link name to give value to the page being linked to. Did I help the cause? No.
So now every permanent link states "permalink for: incert entry title". I am hoping this will help solve the problem. I will modify the other pages most likely next week sometime (it is only a two minute fix) as I am toast.
Gmail Simplifies Email
Since I have been playing with Gmail I have been greatly enjoying the greatly improved means of labeling and archiving of e-mail as opposed to throwing them in folders. Many e-mails are hard to singularly classify with one label that folders force us to use. The ability to drive the sorting of e-mail by label that allows the e-mail to sit accessibly under a filter named with the label make things much easier. An e-mail discussing CSS, XHTML, and IA for two different projects now can be easily accessed under a filter for each of these five attributes.
Dan Brown has written a wonderful article The Information Architecture of Email that dig a little deeper. Dan ponders if users will adopt the changed interface. Hearing many user frustrations with e-mail buried in their Outlook or other e-mail application, I think the improved interface may draw quite a bit of interest. As Apple is going this way for its file structure in Tiger (the next OS upgrade) with Spotlight it seems Gmail is a peak at the future and a good means to start thinking about easier to find information that the use can actually manage.
Future of Local Search on Mac
One of the best things I found to come out of the Apple WWDC keynote preview of the next update of the OS X line, Tiger, Spotlight. Spotlight is the OS file search application. Not only does Spotlight search the file name, file contents (in applications where applicable), but in the metadata. This really is going to be wonderful for me. I, as a user, can set a project name in the metadata and then I can group files from that point. I can also set a term, like "synch" and use AppleScript and Search to batch the files together for synching with mobile devices, easily. Another nice feature is the searches can be saved and stored as a dynamic folder. This provides better control of my Personal InfoCloud.
Steven Johnson provides the history of search in Apple, which has nearly the same technology in Cosmo slated for release in 1996.
Amazon Offers Alexa Augmented Search
Adam pointed out that Amazon is offering a Web search engine A9, which uses ancillary information from Alexa. I offer vanderwal.net/random/ as your jumping off point to explore (leave a review if you wish). I am please with the related sites that are offered as similar sites, not that I am trying for anything in particular.
I agree with Adam that Amazon is offering intriguing integration of information and services, which is the position Google is working to fill. Some of the personal portal sites, like Yahoo, more so than MSN or AOL, have done a good job at innovating in this space.
Google is not my only search engine
Google has been letting me down lately. The past two months I have had too many irrelevant links or only a handful (when I narrow the terms) that do not have what I am looking for. Oddly I have Googled only my site and found the results where I mentioned what I was seeking.
I have been turning more and more to Vivisimo and DogPile for search instead. Why? Well they are both metaseach tools, Vivismo includes Google in what it searches, that search across multiple search engines and return them in one interface. These two services also have faceted filtering and/or categorical filters for the results. These facets greatly help filter out the junk. In short it solves the Paris Hilton site problem when you want a hotel room not a bimbo.
In the past I have tried Vivismo, but it did not seem to have enough depth, which has now been solved. Dogpile now offers a good breadth of search engines that seem to improve on the limited results I had been getting in the past. It is good to have options.
Keeping the Found Things Found
This weeks New York Times Circuits article: Now Where Was I? New Ways to Revisit Web Sites, which covers the Keep the Found Things Found research project at University of Washington. The program is summarized:
The classic problem of information retrieval, simply put, is to help people find the relatively small number of things they are looking for (books, articles, web pages, CDs, etc.) from a very large set of possibilities. This classic problem has been studied in many variations and has been addressed through a rich diversity of information retrieval tools and techniques.
This topic is at the heart of the Personal Information Cloud. How does a person keep the information they found attracted to themselves once they found that information. Keeping the found information at hand to use when the case to use the information arises is a regular struggle. The Personal Information Cloud is the rough cloud of information that follows the user. Users have spent much time and effort to draw information they desire close to themselves (Model of Attraction). Once they have the information, is the information in a format that is easy for the user or consumer of the information to use or even reuse.
iPIM and Chandler have a chair at the Personal Info Cloud
There are two articles that are direct hits on managing information for the individual and allowing the individual to use the information when they needed it and share it as needed. Yes, this is in line with the Personal Information Cloud.
The first article, The inter-personal information manager (iPim) by Mark Sigal about the problem with users finding information and how the can or should be able to then manage that information. There are many problems with applications (as well as the information format itself) that inhibit users reuse of information. In the comments of the article there is a link to products that are moving forward with information clients, which also fit into the Personal Information Cloud or iPIM concept. (The Personal Information Cloud tools should be easily portable or mobile device enabled or have the ability to be retrieved from anywhere sent to any device.
The second article is from the MIT Technology Review (registration required) titled Trash Your Desktop about Mitch Kapor (of founding Lotus Development fame) and his Open Source project to build Chandler. Chandler is not only a personal information manager (PIM), but the tool is a general information manager that is contextually aware. The article not only focusses on Mitch and the product (due late 2004), but the open and honest development practices of those that are building Chandler at the Open Source Application Foundation for Windows, Mac, Linux, etc. distribution.
Blogs get higher Google rankings thanks to proper HTML
Matt points out Google ranks blogs highly. This seems to be the result of Google giving strong preference to titles and other HTML elements. Tools like TypePad help the user properly develop their pages, which Google deems highly credible.
Matt's complaint is his very helpful PVR blog is turning up top results in searches for Tivo information, and other recorder info. Matt's site is relatively new and out ranking the information he is discussing.
This is something I personally run into as things I write about here often get higher Google ranking than the information I am pointing to and is the source and focus of the information. I have often had top Google ranks for items that are big news on CNN or the New York Times, which I am pointing to in my posts.
Much of the reason for this seems to be understanding proper HTML uses and not putting my branding at the forefront of the message. CNN puts their name first in the title of their pages (not the headers, which also have benefit if they are in "H&" tags). The tools and people building Web pages with attention to proper naming and labeling will get rewarded for their good work (if a top Google rank is a reward).
I have written on this in the past in Using HTML tags properly to help external search results from April, which mostly focussed on search ignoring Flash, but for the few HTML elements on a page wrapping the Flash. Fortunately there have been enough links pointing to the site that was laking the top rank to raise the site to the top Google rank.
Some of the corrected Google ranking will come over time as more sites begin to properly mark-up their content. The Google ranks will also shift as more links are processed by Google and their external linking weighting assists correcting the rankings.
Using HTML tags properly to help external search results
There are some essentials to building Web pages that get found with external search engines. Understanding the tags in HTML and how they are (rather should be) used is important. The main tags for most popular search engines are the title, heading (h1, h2, etc), paragraph (p), and anchor (a). Different search engines have given some weight in their ranking to metatags, but most do not use them or have decreased their value.
Google gives a lot of weight to the title tag, which is often what shows in the link Google gives its user to click for the entry. In the title tag the wording is important too, as the most specific information should be toward the front. A user searching for news may find a weblog toward the top of the search ahead of CNN, as CNN puts its name ahead of the title of the article. A title should echo the contents of the page as that will help the ranking of the pages, titles that are not repeated can get flagged for removal from search engines.
The headings help echo what is in the title and provide breaking points in the document. Headings not only help the user scan the page easily, but also are used by search engines to ensure the page is what it states it is. The echoing of terms are used to move an entry to the top of the rankings as the mechanical search engines get reinforcement that the information is on target for what its users may be seeking.
The paragraph tags also are used to help reinforce the text within them.
The anchor tags are used for links and this is what the search engines use to scrape and find other Web pages. The text used for the links is used by the search engines to weight their rankings also. If you want users to find information deep in your site put a short clear description between the anchor tags. The W3C standards include the ability to use a title attribute which some search tools also use. The title attribute is also used by some site readers (used by those with visual difficulties and those who want their information read aloud to them, because they may be driving or have their hands otherwise occupied) to replace the information between the anchor tags or to augment that information.
Example
The application I built to manage this weblog section is build to use each of these elements. This often results in high rankings in Google (and relatedly Yahoo), but this is not the intent, I am just a like fussy in that area. It gets to be very odd when my posting weblog posting review of a meal at Ten Penh is at the top or near the top of a Google Ten Penh search. The link for the Ten Penh restaurant is near the bottom of the first page.
Why is the restaurant not the top link? There are a few possible reasons. The restaurant page has its name at "tenpenh" in the title tag, which is very odd or sloppy. The page does not contain a heading tag nor a paragraph tag as the site is built with Flash. The semantic structure in Flash, for those search engines that scrape Flash. Equally the internal page links are not read by a search engine as they are in Flash also. A norm for many sites is having the logo of the site in the upper left corner clickable to the home page of the site, which with the use of the alt attribute in a image tag within an anchor link allow for each page to add value to the home page rant (if the alt attritute would have "Ten Penh Home" for example).
Not only does Flash hinder the scapeing of information the use of JavaScript links wipes out those as means to increase search rankings. Pages with dynamic links that are often believed to ease browsing (which may or may not prove the case depending on the site's users and the site goals in actual user testing) hurt the information in the site for being found by external search engines. JavaScript is not scrapable for links or text written out by JavaScript.
Go back
I had an early preview of a site this past week so to add comments. It is odd to me that sites are still being built with the frame of reference that the user will come through the "font door". If you read your log files the users come in at every opening. It is about even odds that a new user to the site will come there from a search engine, an external link, or from another pointer (e-mail or article). The frame of reference should always try to provide some orientation to the user, such as breadcrumbs or some other link out to related or parent information.
The item that I found a little jarring was a "Go back to the previous page" and it was not a javascript, but a link to what the developer thought was a next level up page. Pure linear navigation is a practice that is no longer a practice, if it ever was. Somebody last night at the DC-IA book club asked whether we navigated or searched, as always it seems to depend. With sites like Amazon we mostly searched, while some smaller sites we would click around. It seemed the greater volume of information lead to a greater instance of searching.
We did not talk about this for long, but it has been resonating all day. One of the things that Amazon does extremely well is end-search navigation. Most folks seem to search Amazon to find a particular item, but then Amazon's navigation and related offerings that could attract the user to the item, which they were searching for or to a similar item. The search result pages offer links to narrow the results or to ensure the user is looking for the musician Paul Young or author Paul Young. A user arriving at an Amazon book page would have all the options and information they needed to find related information and where they are in the Amazon site.
Google for your enterprise
I went to a small meeting at work today with some folks from Google who were showing their Google Appliance, which was very impressive. Having the Google search generating the search for your enterprise/organization's site would be great, but it got much better than just that bonus. The Google Appliance has the ability to augment the search with a thesaurus to offer the user the option of adding "personal safety restraint devices" when they searched for "seatbelt". This functionality works similarly to Google's spelling corrections.
The advantages did not stop with Google's great search engine, but it also comes with Google's hardware that they have specified and built with failover (if buying more than one rackmounted hardware piece). This just rocks, a software company that is responsible not only for their software, but the hardware it runs on. Apple has had success with this combination and Google's systems are renoun for their great uptime and their ability to return results very quickly. Google boasts having the hardware and software up and configured in one day (when is the last time you have seen this happen, nearly all other search engines are in the 10 to 15 day range). Color me impressed with this demo and seemingly end-to-end search hardware and software package. Google search that can be augmented to provide additional assistance to users, which could let IA's focus on providing great navigational structures for the folks that do not always search to find their information.
Improving information retrieval
Lou points to Improving Web Retrieval After the Dot-Bomb then provides a guide to information retrieval that augments the Marcia Bates article. This provides a very good combination for understanding classification systems.Findability explained
Peter Morville finally puts his findability explanation in writing for all to see (in the wonderful site called Boxes and Arrows). The idea of the term and meaning of findability is growing on me. Findability is a solid lead into the problems of information structure. The explanation of how to start fixing the problems and actions needed to help eradicate the problem can reside in the method/model of attraction (an update to the MOA should be available in two or three weeks, extenuating circumstances have slowed the updates and progress).The Hoopla saga has me trying to move from NetSol. Moving the contact information two years ago was a pain in the butt. My favorite part of this thread is MS and VeriSign (parent of NS) joining to provide better security, what a crock.
« Previous | 1 2 3 | Next »