Off the Top: InfoCloud Entries
Showing posts: 46-60 of 159 total posts
Still Thowing Out the User
There is much buzz about getting rid of the term user these days. Don Norman talks about using the term person, PeterMe picks up on this, and others are not happy with the term "user generated content", like Jon Udell who would like to use "reader-created content", Robert Scoble who believes it is screwing the Long Tail, and Jeff Veen who talks about people writing the web. I have to agree, well I did more than agree.
Throwing Out the User
More than a year ago I got fed up with the user and wrote about saying Good Bye to the User. In years prior I have watched people having painful moments in usability testing. These people felt sorry that they could not easily use what we built and designed. They had empathy for us, but we just lumped them in the category "user". User is not a good word, it is a dirty four letter word. Far too many times designers and developers blame the "user". We tried to solve the user's problems. It was not the problem of the user, it is a real person's pain.
As designers and developers we know deep inside that technology is complex and difficult to use, but we often forget it. The term user has stood in the way. But using person or people, we can see the pain and feel the pain. Many of us consider ourselves users and we do not have these problems, but we are über users, who at one point had the same pain and struggles.
People are different, we have learned this early in life. We can take some characteristics and lump groups of people together, but there are so many important facets that that make us who we are it is difficult to lump people across facets. The only way to lump people separating ourselves as designers and developers out of the equation and putting the focus on regular people. If you are reading this, you are most likely not a regular person who has problems using technology as they wish or need to. It is real people with pain. It is real people who worry about privacy, identity issues, easy access to needed info for themselves and some easy access for some people they know but impossible access for most everybody else, etc. But, the problem with this is these real people do not know this is what they want or need until they do not have it an it becomes painfully aware to them.
Generating Content
I like approach of Jeff Veen and Jon Udell who focus on person-created content. In a hip world of popularity engines like Digg where the masses or crowd bubble up information we forget that most people listen and trust individual voices. We have done this with mass media for years. We trusted certain news anchors and certain reporters on television. We read and trusted certain journalist, columnists, reviewers, and opinion writers. This trust was not always to the wrapper of the communication, like a paper or the whole network news offerings. It comes down to people trusting people. Individuals trusting individuals.
Those of us who have been blogging for nearly a dog year or more understand it is about the individual. We are individual people creating content. We are individual voices. We may be part of a collective at times, but people trust us the person and over time may come to trust people we trust, whom our readers do not know and do not trust yet.
Bringing People Together with People
So what do we need in these social computing environments? We need to see the person. We need to have the ability to find the person similar to us. We want to find those whom are near in thought to us. This may not be the most prolific person on a subject or the most linked to, but their interests match our interests and or vocabularies are similar (often a very good sign of commonality). In the popularity engines we should be able to find those who have "liked" or "dug" things similar to that which we have the same feelings and/or interests.
Doing Without the User
The past year I have been asked many times how easy it is not to use the term user. Well, at first it was hard to transition because it was a term I just used with out thinking. It was also hard because many of my clients and customers I worked with liked using the term user (they also have had many of the problems that come with the term user). But, over time I have a few clients using people and the empathy for the pain that the people who use their products feel is felt and it is reflected in their work products.
One benefit that came from focussing on the person and not the user has been being able to easily see that people have different desired uses and reuses for the data, information, media, etc. that the products I am working on or my clients are developing. I can see complexity more easily focussing on people than I could the user. Patterns are also easier to see looking at the individual people as the patterns resemble flows and not steps. When we focus on the user we try to fit what we built to pre-determined patterns, which we have broken into steps. We can determine steps that are roughly common points of task changing in the flows (changing from seeking to recognizing in a search task it part of an iterative flow, which we can determine is a separate step, but whether that leads to the next step or iterates a few more times is part of a person's information workflow.
Steps are Broken
One of the steps that is getting broken by real people is that around process. People use tools in different ways. For years we have been looking at a publish and subscribe model. But, that is missing a step or two when we look at the flows. People create content and publish it, right? Well, not quite. We are seeing people skipping the publish and pushing it straight to syndication. There is no single point where it is published and has a definable address. The old publish and subscribe model assumed publishing would syndicate the information (RSS, ATOM, RDF, etc.). But, we all know that syndication has been a really slow adoption for traditional media. It was many years after those of us blogging and syndicating information saw traditional media pick-up on the trend. But, traditional media has always understood going straight to syndication with columnists, radio, and television shows. It was the blogging community and personal content creators that were late to understanding we could just syndicate the information and skip the publishing step in the flow.
Getting to Watching People and Flows
How do we not miss things? We watch people and we need to pay attention to their flows. Each individual, each of their desires, each of their different personal information workflows, across each of their current devices, and how they wish they could have what we build inflict less pain on their person.
The person should not feel empathy for those of us building and designing tools and systems, we must feel the person's and peoples pain and feel empathy for them. Where have we stood in their way of their desired flow? Now we must get out of the way, get rid of the user, and focus on people to build and design more effectively.
Cultures of Simplicity and Information Structures
Two Conferences Draw Focus
I am now getting back to responding to e-mail sent in the last two or three weeks and digging through my to do list. As time wears I am still rather impressed with both XTech and the Microlearning conferences. Both have a focus on information and data that mirrors my approaches from years ago and are the foundation for how I view all information and services. Both rely on well structured data. This is why I pay attention and keep involved in the information architecture community. Well structured data is the foundation of what falls into the description of web 2.0. All of our tools for open data reuse demands that the underlying data is structured well.Simplicity of the Complex
One theme that continually bubbled up at Microlearning was simplicity. Peter A. Bruck in his opening remarks at Microlearning focussed on simplicity being the means to take the complex and make it understandable. There are many things in the world that are complex and seemingly difficult to understand, but many of the complex systems are made up of simple steps and simple to understand concepts that are strung together to build complex systems and complex ideas. Every time I think of breaking down the complex into the simple components I think of Instructables, which allows people to build step-by-step instructions for anything, but they make each of the steps as reusable objects for other instructions. The Instructables approach is utterly brilliant and dead in-line with the microlearning approach to breaking down learning components into simple lessons that can be used and reused across devices, based on the person wanting or needing the instruction and providing it in the delivery media that matches their context (mobile, desktop, laptop, tv, etc.).
Simple Clear Structures
This structuring of information ties back into the frameworks for syndication of content and well structured data and information. People have various uses and reuses for information, data, and media in their lives. This is the focus on the Personal InfoCloud. This is the foundation for information architecture, addressable information that can be easily found. But, in our world of information floods and information pollution due to there being too much information to sort through, findability of information is important as refindability (this is rarely addressed). But, along with refindability is the means to aggregate the information in interfaces that make sense of the information, data, and media so to provide clarity and simplicity of understanding.
Europe Thing Again
Another perspective of the two conferences was they were both in Europe. This is not a trivial variable. At XTech there were a few other Americans, but at Microlearning I was the only one from the United States and there were a couple Canadians. This European approach to understanding and building is slightly different from the approach in the USA. In the USA there is a lot of building and then learning and understanding, where as in Europe there seems to be much more effort in understanding and then building. The results are somewhat different and the professional nature of European products out of the gate where things work is different than in the USA. This was really apparent with System One, which is an incredible product. System One has all the web 2.0 buzzwords under the hood, but they focus on a simple to use tool that pulls together the best of the new components, but only where it makes sense to create a simple tool that addresses complex problems.
Culture of Understanding Complex to Make Simple
It seems the European approach is to understand and embrace the complex and make it simple through deep understanding of how things are built. It is very similar to Instructables as a culture. The approach in the USA seems to include the tools, but have lacked the understanding of the underlying components and in turn have left out elements that really embrace simplicity. Google is a perfect example of this approach. They talk simplicity, but nearly every tool is missing elements that make it fully usable (calendar not having sync, not being able to only have one or two Google tools on rather than everything on). This simplicity is well understood by the designers and they have wonderful solutions to the problems, but the corporate culture of churning things out gets in the way.
Breaking It Down for Use and Reuse
Information in simple forms that can be aggregated and viewed as people need in their lives is essential to us moving forward and taking the pain out of technology that most regular people experience on a daily basis. It is our jobs to understand the underlying complexity, create simple usable and reusable structures for that data and information, and allow simple solutions that are robust to be built around that simplicity.
System One Takes Information Workflow to a New Level
While at Microlearning Conference 2006 Bruno and Tom demonstrated their System One product. This has to be one of the best knowledge/information tools that I have seen in years. They completely understand simplicity and interaction design and have used it to create an information capture and social software tool for the enterprise. Bruno pointed me to a System One overview screen capture (you do not have to login to get started) that features some of the great elements in System One.
One of the brilliant aspects of System One is their marketing of the product. While it has easily usable wiki elements, heavy AJAX, live search, etc. they do not market these buzzwords, they market the ease of use to capture information (which can become knowledge) and the ease of finding information. The simplicity of the interface and interaction make it one of the best knowledge management tools available. Most knowledge management tools fall down on the information entry perspective. Building tools that are part of your workflow, inclusion of information from those that do not feed the KM tool, is essential and System One is the first tools that I have seen that understands this an delivers a product that proves they get it.
The enterprise social software market is one that is waiting to take off, as there is a very large latent need (that has been repressed by poor tools in the past). System One tool is quite smart as they have built e-mail search, file access, Google live file search (you type in the wiki (you do not need to know it is a wiki) and the terms used are searched in Google to deliver a rather nice contextual search. This built in search solves the Google complexity of building solid narrow search queries, but the person using the system just needs to have the capability to enter information into the screen.
Those of us that are geeks find Google queries a breeze, but regular people do not find it easy to tease out the deeply buried gems of information hidden in Google. Surfacing people who are considered experts, or atleast connectors to experts on subjects is part of the System One tool as well and this is an insanely difficult task in an enterprise.
My only wish was that I worked in an organization that would be large enough to use this tool, or there was a personal version I could use to capture and surface my own information when I am working.
You may recognize System One as the developer of retreivr, the Flickr interactive tool that allows you to draw a simple picture and their tool will find related photos in Flickr based on the drawing's pattern and colors. It is a brilliant tool, but not as smart as their main product.
To the Skies Again
I am off again, but this time I have clothes out of the cleaners and laundry done. The turn around from last trip to this trip was only a couple days. I am looking forward to being home for a bit, after this trip. I have about 18 hours of travel before I get where I am going.
I am quite looking forward to being with many people that are passionate about microcontent and microlearning. A conversation in early 2001 got me completely hooked on microcontent and its possibilities. We are finally beginning to see tangents of the microcontent world slip into use in the world of the general public. APIs, aggregation, tracking, metadata access, pushing to mobile, etc. are all components, when they work right. We are only a slice of the way there, but each little step gets better and better.
Is Mechanical Turk for Digital Human Filters?
The New York Times article, Software to Look for Experts Among Your Friends, brought to mind a couple conversations from my recent trip to Amsterdam. At XTech there were a couple presentations about Mechanical Turk (as well as other Amazon web services) and the big question was "what is a killer application for Mechanical Turk?" Many of us through out suggestions, many good, but I have kept thinking about one.
The one killer application is not really a problem a human solves, but the collective interests and amalgamation of information and trained services humans use and have trained. You see, for me Yahoo's MyWeb 2.0 is my killer search solution as it pays attention to what I and my "community" have interest in and mirrors our vocabulary (we used our own vocabulary to tag items of interest on the web in the MyWeb social bookmarking tool). These items of interest in MyWeb 2.0 are offered to me first in my web searches and more often than not what I am seeking is in this collection of mine or my contact's.
These searches that are based on human filters and collaborative filtering of information and structuring what we find of interest and bookmark bringing the information out from 5 to 40 pages deep in the web. It mirrors our interests and perspectives. But, what if we need information outside of our own interests? Now we need not our own corpus of items and possibly one outside those in our loose collective. Let us think I need information about transmission of disease between monkeys. This is not something I have interest in and I am not sure anybody I know (or would be couple of degrees of separation away) would have interest in. This is where I would love to turn to Mechanical Turk to use the filtering and aggregation capabilities of another person's life filtering. I may not want to add the monkey disease transmission to my own corpus, but I need to use a corpus of somebody who has this interest. In a sense I need to uses somebody's digital brain and information attraction mechanisms.
I am not sure if this is really a job for Mechanical Turk, but it requires another person's interests and permission. The Tacit company seems to have a similar product, but only works on Windows and most of the people I would want to tap are not Windows users, but most have web-based information resources.
An Overview of the Local InfoCloud is Available
I have finally posted a write-up on Exposing the Local InfoCloud, which explains the attributes and components that comprise the Local InfoCloud. This is a write-up of an explanation that has been bubbling for a couple years and I finally put into a presentation last Fall for Design Engaged. The Local InfoCloud include resources that are familiar to us and can often be the social software elements with which we interact as trusted resources.
During a recent symposium on social software (I attended virtually) there was much grumbling around the term "community" and when I stated I was going to try and redact that term from my usage, I was quickly asked what I was going to replace it with. Community is broadly used and for most of my uses the components in the Local InfoCloud are more distinct pointers to what people can mean when they discuss community. The components are also can help us describe the human and digital resources that bring data, information, and media objects closer to us.
When I have chatted with people about the attributes and components in the past it leads to more questions and wonderful discussions. The point of this piece is to clarify the framework for the Local InfoCloud and capture discussion. I have incorporated the feedback from chats I have had, where it was convincing, relevant, and I was happy with how it jelled. I am looking for more feedback and discussion, which is part of the reason it is posted at the Personal InfoCloud (comments and trackback capabilities are available, but moderated, there).
Developing the Web for Whom?
Google Web Developer Toolkit for the Closed Web
Andrew in his post "Reading user interface libraries" brings in elements of yesterday's discussion on The Battle to Build the Personal InfoCloud. Andrew brings up something in his post regarding Google and their Google Web Developer Toolkit (GWT. He points out it is in Java and most of the personal web (or new web) is built in PHP, Ruby [(including Ruby on Rails), Python, and even Perl].
When GWT was launched I was at XTech in Amsterdam and much of the response was confusion as to why it was in Java and not something more widely used. It seems that by choosing Java for developing GWT it is aiming at those behind the firewall. There is still much development on the Intranet done in Java (as well as .Net). This environment needs help integrating rich interaction into their applications. The odd part is many Intranets are also user-experience challenged as well, which is not one of Google's public fortés.
Two Tribes: Inter and Intra
This whole process made me come back to the two differing worlds of Internet and Intranet. On the Internet the web is built largely with Open Source tools for many of the big services (Yahoo, Google, EBay, etc.) and nearly all of the smaller services are Open Source (the cost for hosting is much much lower). The Open Source community is also iterating their solutions insanely fast to build frameworks (Ruby on Rails, etc.) to meet ease of development needs. These sites also build for all operating systems and aim to work in all modern browsers.
On the Intranet the solutions are many times more likely to be Java or .Net as their is "corporate" support for these tools and training is easy to find and there is a phone number to get help from. The development is often for a narrower set of operating systems and browsers, which can be relatively easy to define in a closed environment. The Google solution seems to work well for this environment, but it seems that early reaction to its release in the personal web it fell very flat.
13 Reasons
A posting about Top 13 reasons to CONSIDER the Microsoft platform for Web 2.0 development and its response, "Top 13 reasons NOT to consider the Microsoft platform for Web 2.0 development" [which is on a .Net created site] had me thinking about these institutional solutions (Java and .Net) in an openly developed personal web. The institutional solutions seem like they MUST embrace the open solutions or work seamlessly with them. Take any one of the technical solutions brought up in the Microsoft list (not including Ray Ozzie or Robert Scoble as technical solutions) and think about how it would fit into personal site development or a Web 2.0 developed site. I am not so sure that in the current state of the MS tools they could easily drop in with out converting to the whole suite. Would the Visual .Net include a Python, PHP, Ruby, Ruby On Rails, or Perl plug-in?The Atlas solution is one option in now hundreds of Ajax frameworks. To get use the tools must had more value (not more cost or effort) and embrace what is known (frogs are happy in warm water, but will not enter hot water). Does Atlas work on all browsers? Do I or any Internet facing website developer want to fail for some part of their audience that are using modern browsers?
The Web is Open
The web is about being browser agnostic and OS agnostic. The web makes the OS on the machine irrelevant. The web is about information, media, data, content, and digital objects. The tools that allow us to do things with these elements are increasingly open and web-based and/or personal machine-based.
Build Upon Open Data and Open Access
The web is moving to making the content elements (including the microconent elements) open for use beyond the site. Look at the Amazon Web Services (AWS) and the open APIs in the Yahoo Developer Network. Both of these companies openly ease community access and use of their content and services. This draws people into Amazon and Yahoo media and properties. What programming and scripting languages are required to use these services? Any that the developer wants.. That is right, unlike Google pushing Java to use their solution, Amazon and Yahoo get it, it is up to the developer to use what is best for them. What browsers do the Amazon and Yahoo solutions work in? All browsers.
I have been watching Microsoft Live since I went to Search Champs as they were making sounds that they got it too. The Live Clipboard [TechCrunch review] that Ray Ozzie gave at O'Reilly ETech is being developed in an open community (including Microsoft) for the whole of the web to use. This is being done for use in all browsers, on all operating systems, for all applications, etc. It is open. This seems to show some understanding of the web that Microsoft has not exhibited before. For Microsoft to become relevant, get in the open web game, and stay in the game they must embrace this approach. I am never sure that Google gets this and there are times where I am not sure Yahoo fully gets it either (a "media company" that does not support Mac, which the Mac is comprised of a heavily media-centric community and use and consume media at a much higher rate than the supported community and the Mac community is where many of the trend setters are in the blogging community - just take a look around at SXSW Interactive or most any other web conference these days (even XTech had one third of the users on Mac).
Still an Open Playing Field
There is an open playing field for the company that truly gets it and focusses on the person and their needs. This playing field is behind firewalls on Intranet and out in the open Internet. It is increasingly all one space and it continues to be increasingly open.
Expansive Web 2.0 Conference List
Eric Weaver has put together a killer list of upcoming conferences for Web 2.0 related subjects. Now to get this list in Upcoming.
The Battle to Build the Personal InfoCloud
Over at Personal InfoCloud I posted The Future is Now for Information Access, which was triggered by an interview with Steve Ballmer (Microsoft) about the future of technology and information. I do not see his future 10 years out, but today. I see the technology in the pockets of people today. People are frustrated with the information not being easily accessed and use and reuse not being as simple as it should. Much of this is happening because of the structure of the information.
Personal InfoCloud is the Focus
One thing that struck me from the article, which I did not write about, was the focus on Google. Personally I find it odd as Yahoo is sitting on the content and the structure for more than 90 percent of what is needed to pull off the Personal InfoCloud. Yahoo is beginning to execute and open access to their data in proper structures. Ballmer lays out a nearly exact scenario for aggregating one's own information and putting it in our lives to the one I have been presenting the last few years. Yahoo has the components in place today to build on top of and make it happen. Google is not only lacking the structure, but they are not executing well on their products they produce. Google does the technically cool beta, but does not iterate and fix the beta nor are they connecting the dots. Yahoo on the other hand is iterating and connecting (they need to focus on this with more interest, passion, and coordinated direction).
The Real Battle
I really do not see the battle as being between Google and the others. The real battle is between Yahoo and Microsoft. Why? Both focus on the person and that person's use and need for information in their life and with their context. Information needs to be aggregated (My Yahoo is a great start, but it goes deeper and broader) and filtered based on interest and need. We are living in a flood of information that has crossed into information pollution territory. We need to remove the wretched stench of information to get back the sweet smell of information. We need to pull together our own creations across all of the places we create content. We need to attract information from others whom have similar interests, frameworks, and values (intellectual, social, political, technological, etc.). The only foundation piece Yahoo is missing is deep storage for each person's own information, files, and media.
Microsoft Live Gets It
Microsoft has the same focus on the person. I have become intrigued with the Microsoft Live properties (although still have a large disconnect with their operating systems and much of their software). Live is aiming where Yahoo is sitting and beyond. Microsoft has the cash and the interest to assemble the pieces and people to get there. Live could get there quickly. Looking at the Live products I saw in January at Search Champs with some in relatively early states and what was launched a few months later, the are iterating quickly and solidly based on what real people want and need in their lives (not the alpha geeks, which Google seems to target). Live products are not done and the teams are intact and the features and connections between the components are growing. They are leaving Google in the dust.
Can Yahoo Stay Ahead of Microsoft?
The question for Yahoo is can they keep up and keep ahead of Microsoft? Google has the focus in search as of today (not for me as the combination of Yahoo! MyWeb 2.0 and Yahoo! Search combined blow away anything that Google has done or seemingly can do. Yahoo! does need to greatly improve the simplicity, ease of use, and payoff (it takes a while for the insanely great value of MyWeb 2.0 to kick in and that needs to come much earlier in the use phase for regular people).
I am seeing Microsoft assembling teams of smart passionate people who want to build a killer web for regular people. It seems Ray Ozzie was the turn around for this and is part of the draw for many heading to work on Live products. The competition for minds of people who get it puts Live in competition with Google, Yahoo, EBay, Amazon, and even Apple. I am seeing Live getting the people in that they need. Recently (last week) Microsoft even started changing their benefits and employee review practices to better compete and keep people. It seems that they are quite serious and want to make it happen now.
Yahoo Under Valued
Recent comments about Yahoo being under valued in the long term are dead on in my view. A recent Economist article about Google pointed out how poorly they execute on everything but their core service (search). This waking up starts to bring a proper focus on what those of us who look at regular people and their needs from information and media in their lives have been seeing, Yahoo gets it and is sitting on a gold mine. Yahoo has to realize that Microsoft sees the same thing and is pushing hard with a proper focus and passion to get there as well.
Google Overvalued
What does this mean for Google? I am not sure. Google is a technology company that is focussed on some hard problems, but it has to focus on solutions that people can use. Google aims for simple interfaces, but does not provide simple solutions or leaves out part of the solutions to keep it simple. They need a person-centered approach to their products. The addition of Jeff Veen and his Measure Map team should help, if they listen. Google has some excellent designers who are focussed on usable design for the people, but it seems that the technology is still king. That needs to change for Google to stay in the game.
More XTech 2006
I have had a little time to sit back and think about XTech I am quite impressed with the conference. The caliber of presenter and the quality of their presentations was some of the best of any I have been to in a while. The presentations got beneath the surface level of the subjects and provided insight that I had not run across elsewhere.
The conference focus on browser, open data (XML), and high level presentations was a great mix. There was much cross-over in the presentations and once I got the hang that this was not a conference of stuff I already knew (or presented at a level that is more introductory), but things I wanted to dig deeper into. I began to realize late into the conference (or after in many cases) that the people presenting were people whose writting and contributions I had followed regularly when I was doing deep development (not managing web development) of web applications. I changed my focus last Fall to get back to developing innovative applications, working on projects that are built around open data, and that filled some of the many gaps in the Personal InfoCloud (I also left to write, but that did get side tracked).
As I mentioned before, XTech had the right amount of geek mindset in the presentations. The one that really brought this to the forefront of my mind was on XForms, an Alternative to Ajax by Erik Bruchez. It focussed on using XForms as a means to interact with structured data with Ajax.
Once it dawned on me that this conference was rather killer and I sould be paying attention to the content and not just those in the floating island of friends the event was nearly two-thirds the way through. This huge mistake on my part was the busy nature of things that lead up to XTech, as well as not getting there a day or two earlier to adjust to the time, and attend the pre-conference sessions and tutorials on Ajax.
I was thrilled ot see the Platial presentation and meet the makers of the service. When I went to attend Simon Willison's presentation rather than attending the GeoRSS session, I realized there was much good content at XTech and it is now one on my must attend list.
As the conference was progressing I was thinking of all of the people that would have really benefitted and enjoyed XTech as well. A conference about open data and systems to build applications with that meet real people's needs is essential for most developers working out on the live web these days.
If XTech sounded good this year in Amsterdam, you may want to note that it will be in Paris next year.
Light Overview of XTech and Amsterdam (including BarCamp Amsterdam)
This trip to Amsterdam for XTech 2006 (and now bits of BarCamp Amsterdam II has been quite different from previous trips, in that Amsterdam is now getting to be very familiar. I also did not spend a day on the front end of the trip walking around adjusting to the time change and spent it inside at XTech, where I saw many friends, which really made it feel more like a floating island comprised of geographically distributed friends that I see when I travel.
It has been great seeing good friends that I really wish I could see more and/or work with on projects as I believe some killer things could get done. I also met people and got to hang out with many new people, which is always great. I was pleased to spend time with people I have only partially spent time with in the past.
I quite enjoyed XTech as it was a good amount of geekery, which provided sparks of inspiration, very good feedback on the "Come to Me Web" and "Personal InfoCloud" stuff I presented. The session had Paul Hammond, Tom Coates, and then myself presenting ideas that focussed on open data, using open data, and building for personal use and reuse of information in our three presentations. It was a fantastic set up.
There were many Mozilla folks around, which was fantastic to hear where the Mozilla/Firefox development is going. This was a very good cross pollination of people, ideas, and interests.
I also realized I need to through out my presentation on Personal InfoCloud and Come to Me Web and rebuild it from scratch. I was finding that my presentation that I have been iterating on for the past year or so is something that needs restructuring and refocussing. I get very positive comments on the presentation, but in delivering the presentation I have made many minor tweaks that have disrupted my flow of delivery. I believe that starting from scratch will help me focus on what gets delivered when. I am really do not write out the presentation in long form as I think that would make it stale for me.
I am heading home tomorrow, but I have not quite felt like I was in Amsterdam as it is really no longer a foreign place. It is still one of my favorite places to be. I spent much time exploring thoughts, spending time with people, playing with digital things, but not deeply finding the new bits of Amsterdam (outside of a few hours this morning). Ah well, I am back in a few short weeks.
Live Data Could Solve the Social Bookmarking Problem with Information Volatility
Alex brings up something in his Go and microformat stuff! covering what is in the works with Microformats at Microsoft. Scroll down to where Alex talks about "mRc = Live data wiring", now this live data access is incredibly important.
One of the elements that has been bugging me with social bookmarking it the volatility of the information is not taken into account when the bookmark is made. No, I am not talking about the information blowing up, but the blood pressure of the person bookmarking may rise if the data changes in some way. I look at social bookmarking, or bookmarking in general as a means to mark a place, but it fails as an indicator of state or status change of the information we are pointing to. The expressing of bookmarking and/or tagging is an expression of our explicit interest in that object we bookmarked and/or tagged. The problem is our systems so far are saying, "yes, you have interest, but so what".
What the live data approach does is makes our Personal InfoCloud active. If we could bookmark information and/or tag chunks of information as important we should be able to find out when that information changes, or better get an alert before the information changes. One area where this is essential and will add huge value is shopping. What happens with products in the real world? The prices change, they go out of stock, the product is modified, production of the product stopped, etc. The permeations are many, but those expressing interest should be alerted and have their information updated.
One of the things I have been including in my "Come to Me Web" presentations is the ability to think about what a person needs when they use and want to reuse information. We read about a product we desire, we read the price, but we may think about the product or put it on a wish list that is related to an event in the future. When we go to act on the purchase the information we have gathered and bookmarked may be out of date.
One solution I have been talking about in my presentations is providing an RSS/ATOM feed for the page as it is bookmarked so the person gets the ability to get updated information. I have built similar functionality into past products years ago that let people using data know when the data changed (e-mail) but also provided the means to show what the data was prior and what it had changed to. It was functionality that was deeply helpful to the users of the system. Live data seems a more elegant solution, if it provides the means to see what information had changed should the person relying on or desiring the information want it.
Upcoming Conferences I am Presenting at and Attending
Okay, things have been quite busy here. But, here will be changing as I am hitting the skies a bit in the short term. This means I may be near you so reach out and we can hang out and chat. I am completely looking forward to all the places on my schedule and seeing all of the people.
XTech
I am off to Amsterdam, Netherlands (no not that other one) this week to speak at XTech. I will be presenting Developing for the Personal InfoCloud on Thursday at 11:45 in the morning.
BarCamp Amsterdam
On Saturday I will be attending BarCamp Amsterdam for part of the time.
Seattle Area
Following the Amsterdam trip I should be in the Seattle area for work. I don't have dates as of yet, but if you shoot an e-mail I will be sure and connect.
Microlearning 2006 Conference
I will be heading to Innsbruck, Austria for the Microlearning Conference and preconference (June 7). I will be talking about microcontent in the Personal InfoCloud and our ability and desire to manage it (one means of doing this is folksonomy, but will be discussing much more).
Following Innsbruck I may be in Europe a bit longer and a little farther north. I will be in Amsterdam just following the conference, but beyond that my schedule has not yet fully jelled.
WebVisions 2006
I will be heading to WebVisions 2006 in Portland, Oregon July 20th and 21st. I will be speaking on Friday the 21st about Tagging in the Real World. This will look at how people are making use of tagging (particularly tagging services) and looking at the best practices.
The Fall
In September it looks like I will be in Brighton, UK for a wonderful event. I should also be in Australia later in September for another conference.
As these events get closer, I will be letting you know.
Yes, I know I need to be publishing this information in hCal, but I have been quite busy of late. But, I am moving in that direction very soon. You can also follow what I am watching and attending in Upcoming for vanderwal.
Congressional Race Reaches MySpace
I got a new and different request today in a social site, a request to "add Allan Lichtman as a MySpace friend". While the campaign has an official Allan Lichtman for Congress site, I thought the MySpace page was a new and interesting approach, particularly for a candidate of the people.
First, I do not know of any other campaigns that have a presence on MySpace or other social networking sites. I am sure they exist, but the MySpace entry interests me as it is a place where young voters (potential voters atleast) hang out. Getting people within collision space of running into a candidate in a web-based community is interesting. A friend of a friend in virtual space is a good play in my book.
While, I know very little about Allan Lictman I added him as a friend as a marker to spend a little more time finding out about him and his campaign. The request and the failure of social networking systems lack of ability to more granularly define relationships or components of relationships that "friends" have defined is really made clear here. I can not define how I know Allan Lichtman, nor can I define the elements that I find of interest in his stances.
Do we need a political social network? I really do not think that more verticals to social spaces are needed, but the social spaces we have need to stop being relative failures for the millions that use them and find them still frustrating.
So, for now I am a "friend" of Allan Lichtman on MySpace.
Mr. Thackara Provides Fodder for Two Loose Thoughts
Things have been busy of late after a return from Vancouver, British Columbia from the IA Summit. It took a week to get through taxes stuff for a four month old company (as of January 1) and a large stack of e-mail (in which responding meant more e-mail).
Yesterday morning I had the pleasurable fortune to have breakfast with John Tackara of Doors of Perception fame. I really need to put two things out there that popped up in conversation that I need to think about more deeply.
Children today are not into the web tied to computers, but focus on their friends through the mobile voice and text messaging. I have been running into this comment quite a bit from Europeans, but increasingly from parents of teen and pre-teens in America. Having a computer is not a large interest for them, but they live in their mobiles as a means to connect, filter, share, convene, and stay tethered to those that matter to them, their friends. The quick 10 to 20 year scenario for this could mean the web is dead and is a technology that had an immense impact, but was a technology that was relatively short lived. Are our communication technologies trending through ever shorter life spans? This could bolster my thinking that the web is increasingly a temporary terminus for information to be shared and picked up and used in context in other media that is better situated within people's lives.
This leads us to context. I keep looking at much of the information that is on the web as being out of the proper use context for most people. We read information on the web, but the web is not the context in which we will make use of this information we find. The web as it has been traditionally built is marginally better than television, in that an address for a car lot flashed on the television screen is as usable as it is plastered on a web page. The address does us more good in our pocket, in driving (or mass transit) routes, in our mobile that is in our pocket, etc. than it does on the web page, but few web pages today get that clue.
I have also been thinking about tagging and particularly tagging in the folksonomy subset as the tags providing mental context to external information. We use the tags to pull back these bits of information or to aggregate this information when we pull on the digital threads that draw what is at the ends of these tag tethers closer to ourselves. A chunk of information or media out on the web is lacking context to our lives with out these tags. When we have needs, most always framed in the context of a need related to a subject we use tags related to that subject to draw back in that which we found or other people with similar vocabularies have found. These tags provide context for the few thousand chunks of information out of the billions that we have explicitly expressed interest in and have placed the context upon based on its relationship between the information or object and ourself.