Inline Messaging

by Thomas Vander Wal in , , , , , ,


Many of the social web services (Facebook, Pownce, MySpace, Twitter, etc.) have messaging services so you can communication with your "friends". Most of the services will only ping you on communication channels outside their website (e-mail, SMS/text messaging, feeds (RSS), etc.) and require the person to go back to the website to see the message, with the exception of Twitter which does this properly.

Inline Messaging

Here is where things are horribly broken. The closed services (except Twitter) will let you know you have a message on their service on your choice of communication channel (e-mail, SMS, or RSS), but not all offer all options. When a message arrives for you in the service the service pings you in the communication channel to let you know you have a message. But, rather than give you the message it points you back to the website to the message (Facebook does provide SMS chunked messages, but not e-mail). This means they are sending a message to a platform that works really well for messaging, just to let you know you have a message, but not deliver that message. This adds extra steps for the people using the service, rather than making a simple streamlined service that truly connects people.

Part of this broken interaction is driven by Americans building these services and having desktop-centric and web views and forgetting mobile is not only a viable platform for messaging, but the most widely used platform around the globe. I do not think the iPhone, which have been purchased by the owners and developers of these services, will help as the iPhone is an elite tool, that is not like the messaging experience for the hundreds of millions of mobile users around the globe. Developers not building or considering services for people to use on the devices or application of their choice is rather broken development these days. Google gets it with Google Gears and their mobile efforts as does Yahoo with its Yahoo Mobile services and other cross platform efforts.

Broken Interaction Means More Money?

I understand the reasoning behind the services adding steps and making the experience painful, it is seen as money in their pockets through pushing ads. The web is a relatively means of tracking and delivering ads, which translates into money. But, inflicting unneeded pain on their customers can not be driven by money. Pain on customers will only push them away and leave them with fewer people to look at the ads. I am not advocating giving up advertising, but moving ads into the other channels or building solutions that deliver the messages to people who want the messages and not just notification they have a message.

These services were somewhat annoying, but they have value in the services to keep somebody going back. When Pownce arrived on the scene a month or so ago, it included the broken messaging, but did not include mobile or RSS feeds. Pownce only provides e-mail notifications, but they only point you back to the site. That is about as broken as it gets for a messaging and status service. Pownce is a beautiful interface, with some lightweight sharing options and the ability to build groups, and it has a lightweight desktop applications built on Adobe AIR. The AIR version of Pownce is not robust enough with messaging to be fully useful. Pownce is still relatively early in its development, but they have a lot of fixing of things that are made much harder than they should be for consuming information. They include Microfomats on their pages, where they make sense, but they are missing the step of ease of use for regular people of dropping that content into their related applications (putting a small button on the item with the microformat that converts the content is drastically needed for ease of use). Pownce has some of the checkboxes checked and some good ideas, but the execution of far from there at the moment. They really need to focus on ease of use. If this is done maybe people will comeback and use it.

Good Examples

So who does this well? Twitter has been doing this really well and Jaiku does this really well on Nokia Series60 phones (after the first version Series60). Real cross platform and cross channel communication is the wave of right now for those thinking of developing tools with great adoption. The great adoption is viable as this starts solving technology pain points that real people are experiencing and more will be experiencing in the near future. (Providing a solution to refindability is the technology pain point that del.icio.us solved.) The telecoms really need to be paying attention to this as do the players in all messaging services. From work conversations and attendees to the Personal InfoCloud presentation, they are beginning to get the person wants and needs to be in control of their information across devices and services.

Twitter is a great bridge between web and mobile messaging. It also has some killer features that add to this ease of use and adoption like favorites, friends only, direct messaging, and feeds. Twitter gets messaging more than any other service at the moment. There are things Twitter needs, such as groups (selective messaging) and an easier means of finding friends, or as they are now appropriately calling it, people to follow.

Can we not all catch up to today's messaging needs?


Understanding Taxonomy and Folksonmy Together

by Thomas Vander Wal in , , , ,


I deeply appreciate Joshua Porter's link to from his Taxonomies and Tags blog post. This is a discussion I have quite regularly as to the relation and it is in my presentations and workshops and much of my tagging (and social web) training, consulting, and advising focusses on getting smart on understanding the value and downfalls of folksonomy tagging (as well as traditional tagging - remember tagging has been around in commercial products since at least the 1980s). The following is my response in the comments to Josh' post...

Response to Taxonomy and Tags

Josh, thanks for the link. If the world of language were only this simple that this worked consistently. The folksonomy is a killer resource, but it lacks structure, which it crucial to disambiguating terms. There are algorithmic ways of getting close to this end, but they are insanely processor intensive (think days or weeks to churn out this structure). Working from a simple flat taxonomy or faceted system structure can be enabled for a folksonomy to adhere to.
This approach can help augment tags to objects, but it is not great at finding objects by tags as Apple would surface thousands of results and they would need to be narrowed greatly to find what one is seeking.
There was an insanely brilliant tool, RawSugar [(now gone thanks to venture capitalists pulling the plug on a one of a kind product that would be killer in the enterprise market)], that married taxonomy and folksonomy to help derive disambiguation (take appleseed as a tag, to you mean Johnny Appleseed, appleseed as it relates to gardening/farming, cooking, or the anime movie. The folksonomy can help decipher this through co-occurrence of terms, but a smart interface and system is needed to do this. Fortunately the type of system that is needed to do this is something we have, it is a taxonomy. Using a taxonomy will save processor time, and human time through creating an efficient structure.
Recently I have been approached by a small number of companies who implemented social bookmarking tools to develop a folksonomy and found the folksonomy was [initially] far more helpful than they had ever imagined and out paced their taxonomy-based tools by leaps and bounds (mostly because they did not have time or resources to implement an exhaustive taxonomy (I have yet to find an organization that has an exhaustive and emergent taxonomy)). The organizations either let their taxonomist go or did not replace them when they left as they seemed to think they did not need them with the folksonomy running. All was well and good for a while, but as the folksonomy grew the ability to find specific items decreased (it still worked fantastically for people refinding information they had personally tagged). These companies asked, "what tools they would need to start clearing this up?" The answer a person who understands information structure for ease of finding, which is often a taxonomist, and a tool that can aid in information structure, which is often a taxonomy tool.
The folksonomy does many things that are difficult and very costly to do in taxonomies. But taxonomies do things that folksonomies are rather poor at doing. Both need each other.

Complexity Increases as Folksonomies Grow

I am continually finding organizations are thinking the social bookmarking tools and folksonomy are going to be simple and a cure all, but it is much more complicated than that. The social bookmarking tools will really sing for a while, but then things need help and most of the tools out there are not to the point of providing that assistance yet. There are whole toolsets missing for monitoring and analyzing the collective folksonomy. There is also a need for a really good disambiguation tool and approach (particularly now that RawSugar is gone as a viable approach).


Stitching Conversation Threads Fractured Across Channels

by Thomas Vander Wal in , , , , , , , , , , , ,


Communicating is simple. Well it is simple at its core of one person talking with another person face-to-face. When we communicate and add technology into the mix (phone, video-chat, text message, etc.) it becomes more difficult. Technology becomes noise in the pure flow of communication.

Now With More Complexity

But, what we have today is even more complex and difficult as we are often holding conversation across many of these technologies. The communication streams (the back and forth communication between two or more people) are now often not contained in on communication channel (channel is the flavor or medium used to communicate, such as AIM, SMS, Twitter, e-mail, mobile phone, etc.).

We are seeing our communications move across channels, which can be good as this is fluid and keeping with our digital presence. More often than not we are seeing our communication streams fracture across channels. This fracturing becomes really apparent when we are trying to reconstruct our communication stream. I am finding this fracturing and attempting to stitch the stream back together becoming more and more common as for those who are moving into and across many applications and devices with their own messaging systems.

The communication streams fracture as we pick-up an idea or need from Twitter, then direct respond in Twitter that moves it to SMS, the SMS text message is responded back to in regular SMS outside of Twitter, a few volleys back and forth in SMS text, then one person leaves a voicemail, it is responded to in an e-mail, there are two responses back and forth in e-mail, an hour later both people are on Skype and chat there, in Skype chat they decide to meet in person.

Why Do We Want to Stitch the Communication Stream Together?

When they meet there is a little confusion over there being no written overview and guide. Both parties are sure they talked about it, but have different understandings of what was agreed upon. Having the communication fractured across channels makes reconstruction of the conversation problematic today. The conversation needs to be stitched back together using time stamps to reconstruct everything [the misunderstanding revolved around recommendations as one person understands that to mean a written document and the other it does not mean that].

Increasingly the reality of our personal and professional lives is this cross channel communication stream. Some want to limit the problem by keeping to just one channel through the process. While this is well intentioned it does not meet reality of today. Increasingly, the informal networking leads to meaningful conversations, but the conversations drifts across channels and mediums. Pushing a natural flow, as it currently stands, does not seem to be the best solution in the long run.

Why Does Conversation Drift Across Channels?

There are a few reasons conversations drift across channels and mediums. One reason is presence as when two people notice proximity on a channel they will use that channel to communicate. When a person is seen as present, by availability or recently posting a message in the service, it can be a prompt to communicate. Many times when the conversation starts in a presence channel it will move to another channel or medium. This shift can be driven by personal preference or putting the conversation in a medium or channel that is more conducive for the conversation style between people involved. Some people have a preferred medium for all their conversations, such as text messaging (SMS), e-mail, voice on phone, video chat, IM, etc.. While other people have a preferred medium for certain types of conversation, like quick and short questions on SMS, long single responses in e-mail, and extended conversations in IM. Some people prefer to keep their short messages in the channel where they begin, such as conversations that start in Facebook may stay there. While other people do not pay attention to message or conversation length and prefer conversations in one channel over others.

Solving the Fractured Communication Across Channels

Since there are more than a few reasons for the fractured communications to occur it is something that needs resolution. One solution is making all conversations open and use public APIs for the tools to pull the conversations together. This may be the quickest means to get to capturing and stitching the conversation thread back together today. While viable there are many conversations in our lives that we do not want public for one reason or many.

Another solution is to try to keep your conversations in channels that we can capture for our own use (optimally this should be easily sharable with the person we had the conversation with, while still remaining private). This may be where we should be heading in the near future. Tools like Twitter have become a bridge between web and SMS, which allows us to capture SMS conversations in an interface that can be easily pointed to and stitched back together with other parts of a conversation. E-mail is relatively easy to thread, if done in a web interface and/or with some tagging to pull pieces in from across different e-mail addresses. Skype chat also allows for SMS interactions and allows for them to be captured, searched, and pulled back together. IM conversations can easily be saved out and often each item is time stamped for easy stitching. VoIP conversations are often easily recorded (we are asking permission first, right?) and can be transcribed by hand accurately or be transcribed relatively accurately via speech-to-text tools. Voice-mail can now be captured and threaded using speech-to-text services or even is pushed as an attachment into e-mail in services as (and similar to) JConnect.

Who Will Make This Effortless?

There are three types of service that are or should be building this stitching together the fractured communications across channels into one threaded stream. I see tools that are already stitching out public (or partially public) lifestreams into one flow as one player in this pre-emergent market (Facebook, Jaiku, etc.). The other public player would be telecoms (or network provider) companies providing this as a service as they currently are providing some of these services, but as their markets get lost to VoIP, e-mail, on-line community messaging, Second Life, etc., they need to provide a service that keeps them viable (regulation is not a viable solution in the long run). Lastly, for those that do not trust or want their conversation streams in others hands the personally controlled application will become a solutions, it seems that Skype could be on its way to providing this.

Is There Demand Yet?

I am regularly fielding questions along these lines from enterprise as they are trying to deal with these issues for employees who have lost or can not put their hands on vital customer conversations or essential bits of information that can make the difference in delivering what their customers expect from them. Many have been using Cisco networking solutions that have some of these capabilities, but still not providing a catch all. I am getting queries from various telecom companies as they see reflections of where they would like to be providing tools in a Come to Me Web or facilitating bits of the Personal InfoCloud. I am getting requests from many professionals that want this type of solution for their lives. I am also getting queries from many who are considering building these tools, or pieces of them.

Some of us need these solutions now. Nearly all of us will need these solutions in the very near future.


Folksonomy Provides 70 Percent More Terms Than Taxonomy

by Thomas Vander Wal in , , ,


While at the WWW Conference in Banff for the Tagging and Metadata for Social Information Organization Workshop and was chatting with Jennifer Trant about folksonomies validating and identifying gaps in taxonomy. She pointed out that at least 70% of the tags terms people submitted in Steve Museum were not in the taxonomy after cleaning-up the contributions for misspellings and errant terms. The formal paper indicates (linked to in her blog post on the research more steve ... tagger prototype preliminary analysis) the percentage may even be higher, but 70% is a comfortable and conservative number.

Is 70% New Terms from Folksonomy Tagging Normal?

In my discussion with enterprise organizations and other clients that are looking to evaluate their existing tagging services, have been finding 30 percent to nearly 70 percent of the terms used in tagging are not in their taxonomy. One chat with a firm who had just completed updating their taxonomy (second round) for their intranet found the social bookmarking tool on their intranet turned up nearly 45 percent new or unaccounted for terms. This firm knew they were not capturing all possibilities with their taxonomy update, but did not realize their was that large of a gap. In building their taxonomy they had harvested the search terms and had used tools that analyzed all the content on their intranet and offered the terms up. What they found in the folksonomy were common synonyms that were not used in search nor were in their content. They found vernacular, terms that were not official for their organization (sometimes competitors trademarked brand names), emergent terms, and some misunderstandings of what documents were.

In other informal talks these stories are not uncommon. It is not that the taxonomies are poorly done, but vast resources are needed to capture all the variants in traditional ways. A line needs to be drawn somewhere.

Comfort in Not Finding Information

The difference in the taxonomy or other formal categorization structure and what people actually call things (as expressed in bookmarking the item to make it easy to refind the item) is normally above 30 percent. But, what organization is comfortable with that level of inefficiency at the low end? What about 70 percent of an organizations information, documents, and media not being easily found by how people think of it?

I have yet to find any organization, be it enterprise or non-profit that is comfortable with that type of inefficiency on their intranet or internet. The good part is the cost is relatively low for capturing what people actually call things by using a social bookmarking tool or other folksonomy related tool. The analysis and making use of what is found in a folksonomy is the same cost of as building a taxonomy, but a large part of the resource intensive work is done in the folksonomy through data capture. The skills needed to build understanding from a folksonomy will lean a little more on the analytical and quantitative skills side than the traditional taxonomy development. This is due to the volume of information supplied can be orders of magnitude higher than the volume of research using traditional methods.


Copy, Point & Tether

by Thomas Vander Wal in , , ,


In the Personal InfoCloud people have various means of keeping information close to them. But, these actions come down to two actions, which are non-exclusive (we can do both). These actions are copying the information to a device or location we are comfortable storing the information and pointing to where the information is located.

Copy, Point: What are We Missing?

Each action (copy & point) has its own benefits and detractors, but both methods are missing a central component that causes us grief at some point or another. We are missing a tether to that information. In our action of copy & point we are making an explicit action declaring our interest in the item. But, when that item changes we should have the option of finding that out. Sometimes it is a minor inconvenience that the information changed, such as spelling errors have been corrected or a more detailed explanation offered. When the modifications are not minor, we most often would benefit from having notification, such as the price of an object changes, the definition is changed, errors corrected, the date or time of an event changed, new article/post on the subject has been made, etc.

We live in a world of ever increasing updates and modifications, mostly because it is so easy to make the modification and we are increasingly becoming an iterative society. Digital information is making providing a platform for this iterative society. In a limited interaction we hit refresh to see the news updates on CNN or ESPN. This iterative society also builds tools to track systems that have a lot of volume and volatility, such as EBay.

Information volatility is a given, but when the flood of information we have interest in or have made explicit action statements about is more than we can refresh, how do we cope? How do those developing and designing these information systems make it easier for us?

We are tethering to the inflow of information. We subscribe to e-mail for product updates from services like Amazon and we have RSS/ATOM feeds for inflow of news services. These inflow tethers are adding to our information overload. We are finding new tools that allow us to copy & point to the bits that are of interest to us, such as social bookmarking tools like del.icio.us, Shadows, etc., but these tools do not truly tether the information in a manner that is helpful when the information changes.

Similarly, when we copy a chunk of information (microcontent) from a report, article, or on-line source we have no means to get an alert when that body of work we copied from was altered. We may be using, or more appropriately reusing the information. We should have an easy means of providing attribution for the item we found of interest (allowing credit where credit is dues is not something that most people are against, but it is not currently easy). There needs to be a solution that would allow easy attribution, in that action we are tethering the information chunk to our source. This attribution mechanism is similar to social bookmarking for larger addressable works (some services like Yahoo's MyWeb 2.0 save a copy of the work we are bookmarking), but neither provides an active tether that alerts us to changes in something we have made an explicit action show our interest in.


Exposing the Local InfoCloud

by Thomas Vander Wal in , , , , , , ,


I have spent a lot of time and effort focussing on the Personal InfoCloud, but the past year or two I have been seeing that the interaction between the person and their information resources that are closest to them (the Local InfoCloud) is extremely important. I have gone around the Local InfoCloud looking at ways to best explain it and bring it to life in a more understandable manner. This past November at Design Engaged 2005 my presentation needed me to dig into the Local InfoCloud and it various components. Since Design Engaged I have been using the slide and ideas around it to explain its relationship to the Personal InfoCloud and the "Come to Me Web". I have iterated on the idea and received some good feedback (particularly from Liz Lawley. Are you ready to dig in?

Overview of the Local InfoCloud

The Local InfoCloud started as an idea of information that was physically close. What is stored or accessed by physical location (information that is physically close) as in an Intranet or location-based information accessed on your mobile device. The more I thought about it and chatted with others it became clear it was more than physical location, it is information resources that are familiar and easier to access than the whole of the web (Global InfoCloud) as a framing concept.

As the my understanding began to lean toward familiarity as a core component of the definition of Local InfoCloud, the term began to embrace the social and community aspects (I am working on shying away from the term community as it is a broadly used term and I am trying to be a little more precise). Interactions with people, services, networks, applications, etc. that are familiar are means of bringing information closer to us as people with data, information, and media needs. The Local InfoCloud eases access. It eases the ability to find and refind information. It is information that is closer to us, not necessarily in physical proximity, but in the ability to access, in which familiarity is bread.

I spent much time considering changing the label from local to community or social, but there were elements that did not perfectly fit that either. Location-based services may be created by a service, but understanding the mindset, terminology, dialect, and cognitive frameworks that are germane to that physical location the information can be structured to resemble or mirror the social elements of understanding in that place. I will get to a better understanding of this when I talk about the Location aspect of the Local InfoCloud. As well, thinking in the Model of Attraction framework the Local InfoCloud is that which is attracted closer to us than the Global InfoCloud.

Important Attributes

There are some attributes that are important to the Local InfoCloud and separate it from the Global InfoCloud and ease the ability to integrate or draw the information and/or media in to the Personal InfoCloud.

Familiarity

As mentioned above familiarity is an essential attribute. Familiarity can be through vocabulary and terminology used to describe or discuss information and objects that people are trying to find and use. The taxonomy or germane ontologies are important to understand as they help ease the connection between the person seeking the information and objects and those providing it.

Access

Access to a resource is very important as it is in the ease of access that we rely on the Local InfoCloud. There is information that is in systems or in locations that others can not get to (that would make it in other's Eternal InfoCloud), but ability to get to the information is important. The ability to get back to the information (through password locked systems, access only by location, etc.) that dictates access is a key attribute.

Structure

Structure is a key attribute in the seeking, finding, and refinding information and objects. In a physical neighborhood we know that a corner store is on the corner, but in a portal we know that movie reviews have a certain URL structure and/or that we can click on a Entertainment button/link to get to the page that links to the movie reviews. Reading one movie review in a familiar site we know how to get to other movie reviews. These browsing structures allow the person to interact and attract information to their screen easily.

Known Actions

Known actions are the element in peoples lives that provide patterns that can be repeated to get to what the person desires. Many times people know how to get to, or more appropriately get back to what they are interested in through indirect connections. A favorite resource may be on a friend's link page as they have not set a direct means to connect to that source or to even draw that information to them to cut down the effort expended. Applications and location-based information are other environments that depend upon known actions to connect people to that which they desire.

Consistency

Consistency is a main attribute driver to our use and reuse of a component in the Local InfoCloud. Consistency breeds familiarity as people learn the terminology, can bookmark, use the known actions to get back to information, or guess how to get access to other items of interest. Having URL structures that are consistent provides a means to get at open information as well as permits the person to restructure means to keep that information closer to them (external social bookmarking as an example).

Copy, Point & Tether

Copy, Point & Tether are actions that a person can take to move information from a Local InfoCloud to the Personal InfoCloud. The attributes are germane to the Personal InfoCloud, but also have importance in the Local InfoCloud often the Local InfoCloud embraces these concepts to ease these actions.

When a person finds data, information, or media objects of interest they most often do one of three things: Copy the item to keep it close (hard drive, flash drive, scan to a drive, scrape to a drive, etc.); Point to the location where the information is located (bookmark, link, blog, wiki, etc.); or Tether the item which is desired by copying or pointing, but then setting a means to get notified when that item has been changed, updated, moved, etc. through tools like RSS/ATOM, e-mail, a pinging service, etc. The tethering is insanely important for items that are anything but completely static over the very long-term (think years not shorter) and it will be getting its own long write-up in the future (subscribe to the RSS here to tether your interest to the future content).

Components

Localinfocloud_overviewNow we can look at the components that can comprise the Local InfoCloud. Each of these have one or more of the attributes. The components are digital and physical in nature. Components may or may not be exclusive, as some Local InfoCloud resources may be comprised of more than one component.

Location

Location was the first component of the Local InfoCloud I considered. Location is important as the physical place has characteristics that draw various attributes together. Location often has a familiarity with terms and language that frame the items within it. The structure of the physical surroundings play an important part in how and where things are located in that location. Tools that are implemented by location are kiosks, GPS/location-based information systems, games that use physical space to provide rewards or clues, language translation tools that are needed in a location, physical location can provide, ease, hinder or censor access to information, and access points to get information can be germane to location (mobile devices need local permissions to access services, etc.).

Friends (and Family)

There is one area that is often over looked as friends and family are not always digital resources, but can provide incredible means of information. Knowing a friend (or she has a good friend) who is an expert in the subject that we need understanding of is very helpful. We can call or visit that person, but we can also e-mail, chat, or have a video conversation with the person to get access to the information or knowledge. In social networks it is common that people will use those whom they are most familiar as a resource to get access to stored knowledge or use the person as a ready pointer to how to get the items they need. Access and familiarity are very strong attributes with friends and family. Often we do not have to tap the person to get the information, but the friend will e-mail us a pointer that they believe we have an interest in consuming. We can save e-mail (as that pointer or container of information is structured by a face or name we have known connections and have put them in context, much in the way they do with us.

A person's preferred method becomes a known action for us. We know the times we can tap somebody with a question or what tools they prefer to communicate using. We know friends who love to talk and their best means of interaction is the phone or an audio chat, while others are more apt to respond to e-mail, text chat, text messaging on their mobile device, or respond to a blog post. Over time we learn not only what is easy to get from whom, but the best means to interact to get the what we desire.

Near in Thought

We have resources that we rely upon because we have similar taste, interest, and/or perspectives on the genre or facet of life the resource covers, more directly these resources are near in thought to us. Politics is an easy example as the terminology used in and around the items we are seeking is known to us and we have expectations that we will like or agree with what is provided from that resource. Beyond politics we have resources with similar interests, perspectives and taste that help filter and provide easier access to items we desire. These resources are not only familiar but they often are structured in a manner that we understand the naming conventions for categories and other resource are easy for use to use and predict what will be brought closer to us through actions. These resources may be whole web sites, journals, writers, blogs, periodicals, etc.

Affiliations

In our life we belong to many groups. These groups have their own terminology and structures for things. Some of these affiliations will be easy to grasp how to access the resources at first opportunity, while others will come through enculturation of learning the structures and terminology. Through consistancy of the affiliations we increase our ability to use these resources to our benefit.

Organizations

Organizations are things we can belong to or join, like a knitting group, local chapter of a national affinity society, etc. These memberships in the organizations allow greater interaction with others with similar interests and/or needs. Organizations can have gated resources that are only access through membership or affiliation with that group.

Work

Work was the initial driver behind the Local InfoCloud as it was information and resources on an Intranet that was the initial understanding of local. But, work also has its own terminology, known actions, and structure. Over time we learn the resources, both digital, physical, and human that provide us access to information and knowledge.

Social Software

Social software can be device-based or network-based (web, internet, intranet, etc.) and the software builds consistency, structure, and known actions over time. If the software is built well the hurdles will be low to understanding how to get at the items we want and need. The software connects people and provides individuals the ability to contribute content and connect with others with similar interest and needs. Social software may connect people over time in an asynchronous manner as a person can leave an answer to a question at one point in time, but everybody with that same question or interest will have the capability to get to the same answer and potentially connect with that person as a known expert/resource through time.

The software becomes the conduit for connecting people and the data, information, and media the people share and/or discuss and augment. It also provides the means to connect people who are near in thought. It is one means for us to share things we would like feedback on. Social software mitigates distance for connecting people around common interests and can mitigate time as we do not need to be on at the same time to interact. Some examples are online discussion groups, listserves, social bookmarking, social networking, blogs, chat software, etc.

Portals

Portals in this meaning are the large aggregation sites that collect information and media into a familiar interface. Tools like AOL, Yahoo, news sites, aggregated shopping sites, etc. are portals with familiar structures that are consistent. Portals make a learnable interface to a variety of data, information, and media objects. Some are interest-based, while others are extremely broad. Similar to a newspaper or magazine the portal has one set of structures to grasp and access remains constant over time. We will easily know where to find movie reviews, car sales, discussion lists, various genre of news, etc.

Wrap-up

This is still a work in progress to some degree. Feedback on these attributes and components is always welcome. There may be some editing to this page, but more than likely the modifications will be in pages and posts that follow-on under this Local InfoCloud category.


The Future is Now for Information Access

by Thomas Vander Wal in , , , , , , , , , , , , , , ,


An interview with Microsoft's Steve Ballmer in the in the San Francisco Chronicle regarding Steve's thoughts about the future of technology, information, and Microsoft (including their competition) sparked a few things regarding the Personal InfoCloud and Local InfoCloud. It could be the people I hang out with and the stay-at-home parents I run across during the day, but the future Ballmer talks about is happening now! The future will more widely distributed in 10 years, but the desire and devices are in place now. The thing holding everything back is content management systems that are built for the "I Go Get Web" and people implementing those systems that see technology and not a web of data.

Let's begin with Ballmer's response to the question, "Ten years from now, what is the digital world going to look like? To which Ballmer responds: A: People are going to have access to intelligence in multiple ways. I'm going to want to have intelligence in my pocket. I'm going to want to have intelligence in my TV. I'm going to want to have intelligence in my den and in my office. And what I may want in terms of size, of screen size, of input techniques, keyboard, handwriting, voice, may vary.

I think what we'll see is, we have intelligence everywhere. We have multiple input techniques, meaning in some sense you may have some bit of storage which travels with you everywhere, effectively. Today, people carry around these USB storage devices, but you'll carry around some mobile device.

The problem is people have the devices in their pockets today in the form of Blackberries, Treos, Nokia 770s, and just regular mobile phones with browsing and syncing. The access to the information is in people's pockets. The software to make it simple with few clicks is where the battle lies. My Palm OS-based Treo 650 is decent, but it has few clicks to get me to my information. My friends with the Windows version of the same device have six or more clicks for basic things like calendar and address book. Going through menus is not simplicity. Going directly to information that is desired is simplicity. A mobile devices needs simplicity as it is putting information in our hands with new contexts and other tasks we are trying to solve (driving, walking, meeting, getting in a taxi, getting on a bus, etc.).

The Information

Not only does the software have to be simple to access information in our Personal InfoCloud (the information that we have stated we want and need near us, but have structured in our personal framework of understanding). We also interact with the Local InfoCloud with is information sources that is familiar to us to which we have set a means of easing interaction (cognitively, physically, or mechanically).

This "intelligence" that Ballmer refers to is information in the form of data. It needs to be structured to make solid use of that information in our lives. This structure needs to ascend below the page level to at least the object level. The object level can be a photo with the associated metadata (caption, photographer, rights, permanent source, size, etc.), event information (event name, location, date and time, permanent location of the information, organizer, etc.), full-text and partial-text access (title, author, contact info, version, date published, rights, headers, paragraphs, etc.).

These objects may comprise a page or document on the web, but they not only have value as a whole, they have value as discrete objects. The web is a transient information store for data and media, it is a place to rest this information and object on its journey of use and reuse. People use and want (if not need) to use these objects in their lives. Their lives are comprised of various devices with various pieces of software that work best in their life. They want to track events, dates, people, ideas, media, memes, experts, friends, industries, finances, workspaces, competition, collaborators, entertainment, etc. as part of their regular lives. This gets very difficult when there is an ever growing flood of information and data bombarding us daily, hourly, consistently.

This is not a future problem. This is a problem right now! The information pollution is getting worse every moment we sit here. How do we dig through the information? How do we make sense of the information? How do we hold on to the information?

The solutions is using the resources we have at our finger tips. We need access to the object level data and the means to attach hooks to this data. One solution that is rising up is Microformats, which Ray Ozzie of Microsoft embraces and has been extending with his Live Clipboard, which is open for all (yes all operating systems and all applications) to use, develop, and extend. The web, as a transient information store, must be open to all comers (not walled off for those with a certain operating system, media player, browser, certain paid software, etc.) if the information is intended for free usage (I am seeing Microsoft actually understand this and seemingly embrace this).

Once we have the information and media we can use it and reuse it as we need. But, as we all know information and media is volatile, as it changes (for corrections, updates, expanding, etc.) and we need to know that what we are using and reusing is the best and more accurate information. We need the means to aggregate the information and sync the information when it changes. In our daily lives if we are doing research on something we want to buy and we bookmark it, should we not have the capability to get updates on the prices of the item? We made an explicit connection to that item, which at least conveys interest. Is it not in the interest of those selling the information to make sure we have the last price, if not changes to that product? People want and need this. It needs to be made simple. Those that get this right will win in the marketplace.

What is Standing in the Way?

So, the big question is, "what is standing in the way"? To some degree it is the tools with which we create the information and some of it is people not caring about the information, data, and media they expose.

The tools many of the large information providers are using are not up to the task. Many of the large content management systems (CMS) do not provide simple data structures. The CMS focusses on the end points (the devices, software, tools, etc.) not the simple data structures that permit simple efficient use and reuse of the objects. I have witnessed far too many times a simple web page that is well structured that is relatively small (under 40KB) get turned into an utter mess that is unstructured and large (over 200KB). Usable, parseable, and grabable information is broken by the tools. The tools focus on what looks good and not what is good. Not only is the structure of the data and objects broken, but they are no longer addressable. There are very few CMS that get it right, or let the developers get it right (one that gets it right is Axiom [open disclosure: I have done work with Siteworx the developer of Axiom]).

The other part of the problem is the people problem, which is often driven by not understanding the medium they are working within. They are focus on the tools, which are far from perfect and don't care enough to extend the tools to do what they should. Knowing the proper format for information, data, media, etc. on the web is a requirement for working on the web, not something that would be nice to learn someday. Implementing, building, and/or creating tools or content for the web requires understanding the medium and the structures that are inherent to building that well. I have had far too many discussions with people who do not understand the basics of the web nor the browser, which makes it nearly impossible to explain why their implementation fails. Content on the web has requirements to be structured well and the pages efficiently built. The pages need to degrade (not with an $80,000 plug-in) by default. Media on the web that is for open consumption must work across all modern systems (this should be a 3 year window if not longer for the "modern" definition).

Summary

So what is the take away from this? Content needs to be built with proper structure to the sub-object level (objects need the metadata attached and in standard formats). The content needs to be open and easily accessed. Portability of the information into the tools people use that put information in our pockets and lives must be done now. We have the technology now to do this, but often it is the poorly structured or formatted information, data, media, etc. that stands in the way. We know better and for those that don't know yet the hurdle is quite low and easy to cross.


The Come To Me Web

by Thomas Vander Wal in , , , , , , , , , , ,


Until May of 2005 I had trouble with one element in my work around the Model of Attraction and Personal InfoCloud (including the Local and Global InfoClouds as well) to build a framework for cross-platform design and development of information and media systems and services. This problem was lack of an easy of explaination of what changes have taken place in the last few years on the web and other means of accessing digital information. In preparing for a presentation I realized this change is manifest in how people get and interact with the digital information and media.

This change is easily framed as the "Come to Me" web. The "Come to Me" web, which is not interchangeable with the push/pull ideas and terms used in the late 90s (I will get to this distinction shortly). It is a little closer to the idea of the current, "beyond the page" examinations, which most of us that were working with digital information pre-web have always had in mind in our metaphors and ideologies, like the Model of Attraction and InfoClouds.

The I Go Get Web

Before we look at the "Come to Me" web we should look at what preceded it. The "I Go Get" metaphor for the web was the precursor. In this incarnation we sought their information. The focus was on the providers of the content and the people consuming the information (or users) were targeted and lured in, in the extreme people were drawn in regardless of a person's interest in the information or topic covered. The content was that of the the organization or site that provided that information.

This incarnation focussed on people accessing the information on one device, usually the desktop computer. Early on the information was developed for proprietary formats. Each browser variant had their own proprietary way of doing things, based around a few central markup tags. People had to put up with the "best view with on X browser" messages. Information was also distributed in various other proprietary formats that required software on the device just so the person could get the information.

The focus providing information was to serve one goal (or use) reading. Some of this was driven by software limitations. But it was also an extension of information distribution in the analog physical space (as opposed to the digital space). In the physical space the written word was distributed on paper and it was consumed by reading (reuse of it meant copying it for reading) and it took physical effort to reconstruct those words to repurpose that information (quoting sources, showing examples, etc.).

The focus was on information creation and the struggle was making it findable. On the web there were only limited central resources used to find information, as many of the search engines were not robust enough, did not have friendly interfaces. Findability was a huge undertaking, either to get people what they desired/needed or to "get eyeballs".

Just as the use of the information was an extension of the physical realm that predated the digital information environment, the dominant metaphor in the "I Go Get" web was based in the physical realm. We all designed and developed for findability around the navigation/wayfinding metaphor. This directly correlates to going somewhere. Cues we use to get us to information were patterned and developed from practices in the physical world.

Physical? Digital? Does it Matter?

You ask, "So what we used ideas from the physical world to develop our metaphors and methodologies for web design and development?" We know that metaphors guide our practices. This is a very good thing. But, metaphors also constrain our practices and can limit our exploration for solutions to those that fit within the boundaries of that metaphor. In the physical realm we have many constraints that do not exist in the digital realm. Objects are not constrained by the resources they are made from (other than the energy to drive digital realm - no power no digital realm). Once an object exists in the digital realm replicating them is relatively insignificant (just copy it).

Paths and connections between information and objects is not constrained by much, other than humans choosing to block its free flow (firewalls, filtering, limiting access to devices, etc.). Much like Peter Merholz desire lines where people wear the path between two places in a manner that works best for them (the shortest distance between two points is a straight line). Now, don't think of the physical limitation between two points, I need to go from my classroom on the fourth floor of building "X" to across campus, up the hill to the sixth floor office of my professor. Draw a straight line and walk directly. This does not work in physical space because of gravity and physical impediments.

Now we are ready to understand what really happens on the web. We go from the classroom to our professors office, but we don't move. The connection brings what we desire to us and our screen. In this case we may just chat (text or video - it does not matter) with the professor from our seat in the classroom (if we even need to be in the classroom). Connections draw objects to our screens through the manifestation of links. As differently as people's minds work to connect ideas together, there can be as many paths between two objects. Use of physical space is limited by limitations outlined in physics, but the limitations are vastly different in digital space, use of the same information and media has vastly different limitations also.

It is through breaking the constraints of old metaphors and letting the digital realm exist that we get to a new understanding of digital information on the networks of the digital realm, which include the web.

The Come to Me Web

The improved understanding of the digital realm and its possibilities beyond our metaphors of the physical environment allows us to focus on a "Come to Me" web. What many people are doing today with current technologies is quite different than was done four or five years ago. This is today for some and will be the future for many.

When you talk to people about information and media today they frame it is terms of, "my information", "my media", and "my collection". This label is applied to not only information they created, but information they have found and read/used. The information is with them in their mind and more often than not it is on one or more of their devices drives, either explicitly saved or in cache.

Many of us as designers and developers have embraced "user-centered" or "user experience" design as part of our practice. These mantras place the focus on the people using our tools and information as we have moved to making what we produce "usable". The "use" in "usable" goes beyond the person just reading the information and to meeting peoples desires and needs for reusing information. Microformats and Structured Blogging are two recent projects (among many) that focus on and provide for reuse of information. People can not only read the information, but can easily drop the information into their appropriate application (date related information gets put in the person's calendar, names and contact information are easily dropped into the address book, etc.). These tools also ease the finding and aggregating of the content types.

As people get more accustom to reusing information and media as they want and need, they find they are not focussed on just one device (the desktop/laptop), but many devices across their life. They have devices at work, at home, mobile, in their living space and they want to have the information that they desire to remain attracted to them no matter where they are. We see the proliferation of web-based bookmarking sites providing people access their bookmarks/favorites from any web browser on any capable device. We see people working to sync their address books and calendars between devices and using web-based tools to help ensure the information is on the devices near them. People send e-mail and other text/media messages to their various devices and services so information and files are near them. We are seeing people using their web-based or web-connected calendars to program settings on their personal digital video recorders in their living room (or wherever it is located).

Keeping information attracted to one's self or within easy reach, not only requires the information and media be available across devices, but to be in common or open formats. We have moved away from a world where all of our information and media distribution required developing for a proprietary format to one where standards and open formats prevail. Even most current proprietary formats have non-proprietary means of accessing the content or creating the content. We can do this because application protocols interfaces (APIs) are made available for developers or tools based on the APIs can be used to quickly and easily create, recreate, or consume the information or media.

People have moved from finding information and media as being their biggest hurdle, to refinding things in "my collection" being the biggest problem. Managing what people come across and have access to (or had access to) again when they want it and need it is a large problem. In the "come to me" web there is a lot of filtering of information, as we have more avenues to receive information and media.

The metaphor and model in the "I go get" web was navigation and wayfinding. In the "come to me" web a model based on attraction. This is not the push and pull metaphor from the late 1990s (as that was mostly focussed on single devices and applications). Today's usage is truly focussed on the person and how they set their personal information workflow for digital information. The focus is slightly different. Push and pull focussed on technology, today the focus is on person and technology is just the conduit, which could (and should) fade into the background. The conduits can be used to filter information that is not desired so what is of interest is more easily identified.


Location? Location? Where am I?

by Thomas Vander Wal in , , , , , , , , , , ,


I have been traveling more than usual this year to places in the United States and Europe. Some I have been to before and others I have not. Many of the trips are to places for only a few days and are set around meetings, conferences, or speaking engagements. I am often making plans at the last minute or having to make arrangements on the fly as ancillary meetings (not the prime reason I am there) get moved or cancelled. I am often looking for food, coffee, wifi, electronic stores, hardware stores, etc. in a location I am not completely familiar with. I am needing services of the local businessman, but I am not local.

The "Local Services"

You say, "there are many local services". Yes, there are Yahoo Local, Google Local, A9 Yellow Pages search, and other more local guides. But, none of them work on a mobile. There are Google SMS search and Mobile Yahoo, which has search that can tie to your local info, but if I am traveling I most likely have not save where I am looking for options.

Most modern phones know your location, they have to by law in the United States for emergency service calls. The phones do not provide easy access to that location software because the carriers providing the service do not want you to have it for free, they want somebody to pay for that information. If I call information they are not going to tell me where I am, nor the type of service or store I am seeking.

A Hack Finds "Where"

My current hack is to stand in front of a store, which I know the street name and I send the request for information about the place to Google SMS (ritual coffee. san francisco, ca) and I get one important piece of information back, the zip code. The zip code in the United States is the key to getting location information. There is nothing when driving (or actually riding as a passenger, because one never text messages while driving) or walking around that tells you the zip code (I have given up asking strangers on the street the zip code as it is more often than not incorrect). Once I have the zip code I can ask the mobile services for "coffee 94110" and get another place to get coffee and sit down because Ritual Coffee Roasters is utterly packed and already has seat vultures hovering.

Ministry of Silly Steps

Doing this little dance I get options, but it is a few steps that I should never have to take. The information most needed in a local search when mobile is location

Zip It, Zip, Z..

With the zip code I can dump that into my Mobile Yahoo! "new location" and get results. But, even because Yahoo! Mobile knows it is me (they offered me my stored locations (such as Home and Work)) it does not use that information to give me things I have reviewed and stored in Yahoo! Local. In the online version of Yahoo! Local I get reviews from people in my "community" (that really really needs to get a firm understanding of the granular social network), which is often helpful (if I know the person and can adjust my perception because I know how close that person's preferences are to mine on that subject). Sometimes I need an extension cord or an Apple Store (or a good substitution).

Elsewhere: Missing Even Partial Solutions

Additionally, this only works in the United States. The global local versions of Yahoo don't have fleshed out local services that are anything close to what is available in the United States and my "community" (as imperfect of an approach as it is at the moment) is still more helpful at filtering than nothing and I know I have many people in my "community" that have not only been to the same locations I am in, but have reviewed restaurants, local stores, etc. on the web and I want to be able to pull that information back in. Yes, this means the services need to grasp and embrace digital identity to make this work (or just build a social network capable address book that knows who my friend's identities are on various other services and social networking tools where this information may be sitting - not rocket science by any means). I heard some native language services were around, but those would not be fully helpful to me (I think I could get through it however), but if I tried a service that did not work it is not pointing me to one that does (now that would be insanely helpful and I would likely go to the kind service people for everything first as they would point me to just the right place every time).

Ya Beats Goo

Well at least Yahoo! understands there are places outside the United States. Google's services are not there, or any where on the mobile front it seems. In my last trip to Europe nobody knew that Google offered these services, which it seems they do not, in one of the most mobile use intensive cultures in the Western Hemisphere.

Enough

I know, enough. I agree. We need mobile information that works. WiFi is not here everywhere. Even if it were I am not foolish enough to pull out my laptop to try and get a signal and then get the information I need. I have a mobile device with the perfect capability to do just this. Actually there are more than double (if not triple - can not put my fingers on this info) the users with this capability on their mobile than laptop users in the United States (foolishly most laptops do not have locative hardware in them to ease this possibility if it was your last possibility). The technologies are here. Why are we not using them?


Europe Presentations from October

by Thomas Vander Wal in , , , , , , , ,


I am late in posting the links to my two presentations given in Europe. I presented the Personal Digital Convergence as the opening keynote to the SIGCHI.NL - HCI Close to You conference. I have also posted the final presentation, IA for the Personal InfoCloud, at the Euro IA Summit 2005.


Interview on Digital Identity and the Personal InfoCloud

by Thomas Vander Wal in


Today Thomas Vander Wal appears in Under the Radar: That cloud kinda looks like you! in an interview with Scott Hirsch who is a partner at MIG5. The interview focusses on digital identity, which is integral to the Personal InfoCloud and interacting with other people and services in the digital world. Increasingly it seems digital identity is tethered to physical world identity for access to buildings, accessing our computers, medical services, etc. This has some problems around privacy that must be addressed and there must be trust in the services that interact with out digital identity.


Mash-ups and the Model of Attraction

by Thomas Vander Wal in , , , , , , , , ,


I have been thinking a lot about web2.0 mash-ups like Housing Maps since I was on a panel with Paul Rademacher. Particularly I have been trying to make sense of mash-ups in the context of the Model of Attraction. It was not difficult to use these models as a lens to better understand what is going on in these mash-ups. The irony is I needed to do a tiny mash-up of my own to better understand what is going on.

Let us use the Housing Maps as our sole example. Housing Maps takes the housing listing information from Craigslist and displays them by location as a layer in the Google Maps. Paul had built the tool in his spare time as the result of showing up at the same location to rent twice. The visual representation of the listings on a map helped him keep from doing this again. The visual representation also helps others better discern proximity and location (next to a freeway is why it is cheap, or near playground for junior, etc.).

The interpretation of this mash-up and other web2.0 developments require using a slight mash-up of the Model of Attractions's receptors (the receptors are intellectual (cognitive), perceptual (sensory), mechanical, and physical). One uses the receptors as a whole to design and develop information/media access for people in different contexts, with different devices, varying needs, and in different contextual needs. In the case of understanding the Housing Maps we know what the mechanical receptor is, it is a desktop/laptop computer as that is what the interface requires to use the tool. Housing Maps implicitly requires full visual capabilities, and the means to control a pointing device (mouse, etc) for the physical receptors.

The two receptors we will look at are the Intellectual Receptor and the Perceptual Receptors. The Intellectual Receptor is used in the design and development phases to understand how a person thinks about the information/media by understanding vocabulary, information structures, complexity of conveyance (what level and style of writing are used to convey the ideas), level of detail used, the amount of explanation given, use of metaphors, etc. The Perceptual Receptors are used to understand what sensory elements are understood by the people using the information/media. The sensory elements are comprised of visual, auditory, motion/animation, touch (haptic), etc.

The Housing Maps requires understanding the limitations of the resources being used prior to Paul's remixing. The information that Paul was using was Craigslist to find a new place to live. Craigslist is a rich information source that has a large variety of things for sale or giving as well as social connective communities (personals pages). Paul was using the housing section in the San Francisco Bay Area as his information source. The housing entries have descriptions of the properties for rent/let/buy, much like the old classified real estate ads in the newspapers (remember those) but with a little more detail and often including photos of the property. One element that many of the properties include is a location variable (address).

While the Craigslist information is rich and robust and a fantastic resource, Craigslist has a simple interface. This interface, much like that of a classified ad is about providing the information and using the space efficiently. The reality is no mater what is done to the visual appearance of Craigslist the information in text form and the photos are just those simple elements. A map included in each of the entries would be a little more helpful, but it is still rather limiting as it does not give an idea of what is really on the market and where all of the properties of interest are located (in the given parameters of the person's query). We have the Intellectual Receptors largely sated. The Perceptual Receptors (what does the page look like how does a person interact with the information (passively/actively)) could use a little more tweaking, but within the context of the static HTML page the information interface offers little opportunity for improvement.

The missing element in the Craigslist information is not data that is missing (except where locative data is not included in the Craigslist entry). The missing element is in the Perceptual Receptor which then augments the Intellectual Receptor. The contextual framework for locative information is missing from the interface. The array of information provided in the Craigslist interface needs another vector to view the information (Craigslist limits by price, rough geographical area, type of property arrangement (rent/lease/sublet/buy/share/own), animals, and keywords). This vector is a more fine grained view of the location information and put into a context that helps make sense of the information easily. The context is a map, which works well for displaying location-based information.

The Google map is used for the visual representation layer, which provides the context to the location information. The Google map is an open interface that is available to use for the display of location relevant information from external data sources. The interface if very helpful for this type of information and it is freely available for those with the skill sets needed to parse and feed the information into the Google maps interface.

The web2.0 mash-ups extract information from one source and display that information in a different interface. Tools like Bloglines do this with feeds and display the information in an interface separate from the website's interface from which the information was posted by the content creator/owner.

These mash-ups serve to provide the person consuming the information a tool that works for their needs. In a "come to me web" this is very important. The content provider/owner would have to invest many resources to provide a broad array of interfaces to search each person and each person's needs and desires for information. Additionally, as it is with nearly everything on the web the interface that aggregates information from a broad variety of information sources provides a richer set of information for the person to use and analyze for their own needs. Not only are the Intellectual Receptors augmented by the network effect of the information, but offering the personal consuming the information a means/lens (for their Perceptual Receptor needs) to view the information/media in means that adds value for their need is required for people to better embrace the web as a source of information that is a layer woven into their life rather than technology tools that augment their lives.