Click here to Skip to main content
65,938 articles
CodeProject is changing. Read more.
Articles
(untagged)

RSS Feed Aggregator and Blogging Smart Client

0.00/5 (No votes)
16 Aug 2005 2  
RSS Feed aggregator and blogging Smart Client which uses Enterprise Library, Updater Application Block, lots of XML hacks and desktop tricks. A comprehensive guide to real life hurdles of Smart Client development.

Contents

Introduction

RSS Feeder.NET is a free open source desktop RSS feed aggregator which downloads feeds from web sources and stores them locally for offline viewing, searching and processing. It is also a rich blogging tool which you can use to blog a variety of blog engines including WordPress, B2Evolution, .Text, Community Server etc. You can be fully MS Outlook® dependent or can run fully standalone. You can also use both at the same time whichever you find comfortable to work with. It does not increase the Outlook load time, nor does it make the Outlook slow or prevent it from closing properly. It is a Smart Client that makes best use of both Local Resource and Distributed Web Information sources.

Update History

  • August 9, 2005 - Source code updated, Web Log Manager has several fixes.
  • August 8, 2005 - When you first load the project and build, you will find an error in WebLogManager.resx. The solution is to simply load the form, do something to make it dirty, and save the form.

Features

  • Newspaper mode. You can read feeds in a more readable newspaper mode called “Blogpaper”.
  • Auto Discovery. Drag any hyper link and I will find out whether there is any RSS Feed in that page.
  • Outlook integration. You can store feeds in Outlook folders.
  • Blogging. It provides you a Outlook 2003 style convenient workspace to manage your blog accounts and write rich posts.
  • Blog from Outlook. You can specify an Outlook folder for a weblog account. All the posts from that folder is automatically posted to the weblog during synchronization. You can write posts as HTML using Word editor. Post content (HTML markup) is cleaned rigorously before posting to the weblog.
  • Outlook View. It uses a customized view to present a more readable list of feeds in Outlook Folders. The standard Post view is not easy to browse through quickly. The view puts the subject first in bold and an excerpt of the post under the subject.
  • Optimized Startup. You can safely put RSS Feeder at startup and it won’t make your Windows® start slower. A clever lazy loading process puts no effort on Windows® during startup instead starts the app when Windows® has finally regained its strength after the long boot up struggle.
  • Newsgator Import. Newsgator users can use RSS Feeder to import all the subscriptions and seamlessly replace Newsgator without modifying the Outlook folder locations.

Feature walkthrough

Let’s have a small walkthrough of all the features of RSS Feeder .NET:

Outlook style view

The three pane view of Outlook is a user interface engineering marvel. It makes browsing through information fast and easy. In RSS Feeder, I have followed the same concept. The left most pane is the channel list. When you click a channel you see the list of feeds in the middle list. You can instantly search through feeds using the search box at the bottom middle. When you click on a feed, the feed content is displayed on the right side viewer.

You can directly edit channel properties from the left bottom property grid. Click on the title “Properties” to show the property grid. .NET’s built in property grid makes it very convenient to expose objects and allow users to modify the objects conveniently. I have implemented several extensions on property grid which I will be explaining later on.

Daily Blogpaper – Newspaper style reading

The most convenient reading view for blogs is the newspaper style view which I call “The Daily Blogpapper”. You can just open the Blogpaper, read through all the latest posts collected from all the channels (optionally excluded) and click on “Mark as Read”. It’s as simple as this.

The previous Outlook style view is useful to work with feeds and channels. But it is not a convenient reading environment. The Blogpaper gives you a truly comfortable reading environment for everyday reading.

Rich blogging feature

RSS Feeder is not just an RSS Aggregator; it’s also an equally powerful feature rich blogging tool.

Outlook 2003 style view is created using the amazing free UI components which you can find at Divelements Ltd. On the left side, you can see the weblogs accounts and your posts in that account. The right side is the editing environment.

RSS Feeder currently supports MetaWebLogAPI supported blog engines like WordPress, B2Evolution, Drupal etc. and also some XML web service based blog engines like .Text and Community Server. We will see how we can implement these in detail later on.

Outlook integration

If you don’t like my app, you can remain a MS Outlook user. When you install the application, it will ask you whether you need Outlook integration. You just need to specify a base folder which will contain all the child channel folders and the RSS Feeder will feed Outlook with all the feeds.

The convenient view of feed list makes it very easy to skim through lots of post rapidly. The auto preview also gives a glimpse of the post content eliminating the need to open each post and we can decide whether to keep or throw it away.

Blogging from Outlook

This is my favorite, you can create posts in an Outlook folder and then create a web log account which maps to that folder. All the posts in that folder are delivered to the web log. You can drag posts from some other place or can write new posts in that folder. Whenever RSS Feeder runs its periodical synchronization (every 5 mins) it will read the posts from the folder and then send the posts to the corresponding blog site.

Informative Progress Meter

The Send/Receive window of RSS Feeder gives you detailed statistics about feed synchronization. You can view the errors received from the server, you can see the speed of your internet connection and the number of feeds sent to Outlook.

While posting blogs, it also gives you the post ID generated from the server or the detail error message received while posting.

Newsgator import

RSS Feeder will automatically import Newsgator settings including all the subscription and Outlook folders at first run. This gives you zero effort for migrating from Newsgator to RSS Feeder.

Smart Client – definition and requirements according to MSDN

In order to qualify an application as a smart client, the application needs to fulfill the following requirements according to the definition of Smart Client at MSDN.

Local resources and user experience

All smart client applications share an ability to exploit local resources such as hardware for storage, processing or data capture such as compact flash memory, CPUs and scanners for example. Smart client solutions offer hi-fidelity end-user experiences by taking full advantage of all that the Microsoft® Windows® platform has to offer. Examples of well known smart client applications are Word, Excel, MS Money, and even PC games such as Half-Life 2. Unlike "browser-based" applications such as Amazon.Com or eBay.com, smart client applications live on your PC, laptop, Tablet PC, or smart device.

Connected

Smart client applications are able to readily connect to and exchange data with systems across the enterprise or the internet. Web services allow smart client solutions to utilize industry standard protocols such as XML, HTTP and SOAP to exchange information with any type of remote system.

Offline capable

Smart client applications work whether they are connected to the Internet or not. Microsoft® Money and Microsoft® Outlook are two great examples. Smart clients can take advantage of local caching and processing to enable operation during periods of no network connectivity or intermittent network connectivity. Offline capabilities are not only used in mobile scenarios but also by desktop solutions where they can take advantage of offline architecture to update backend systems on background threads, thus keeping the user interface responsive and improving the overall end-user experience. This architecture can also provide cost and performance benefits since the user interface need not be shuttled to the smart client from a server. Since smart clients can exchange just the data needed with other systems in the background, reductions in the volume of data exchanged with other systems are realized (even on hard-wired client systems this bandwidth reduction can realize huge benefits). This in turn increases the responsiveness of the user interface (UI) since the UI is not rendered by a remote system.

Intelligent deployment and update

In the past traditional client applications were difficult to deploy and update. It was not uncommon to install one application only to have it break another. Issues such as "DLL Hell" made installing and maintaining client applications difficult and frustrating. The Updater Application Block for .NET from the patterns and practices team provides prescriptive guidance to those who wish to create self-updating .NET Framework-based applications that are deployed across multiple desktops. The release of Visual Studio 2005 and the .NET Framework 2.0 will beckon a new era of simplified smart client deployment and updating with the release of a new deploy and update technology known as ClickOnce.

(The above text is copied and shortened from MSDN site.)

How is RSS Feeder a smart client

Let's look at how the RSS Feeder .NET is a smart client application:

Local resources and user experience

After downloading the application, it runs locally, stores all the feeds and your personal blogs in an MS Access Database and XML store respectively. It also gives you a rich user interface like Outlook 2003 to work with. As a result, you get the full benefits of desktop convenience yet all the information you are working with are produced in some web source.

Connected

The application connects to RSS Feed sources using HTTP and downloads the feeds to local store. It uses XML RPC for XMLRPC enabled blog engines in order to post web logs. Some of the famous XMLRPC supported blog engines are WordPress, B2Evolution, Drupal etc. It also uses Web Service to communicate with Blog engines like .Text and CommunityServer.

Offline capable

The application is fully offline capable. When it is not connected, you read the feeds from local store which are already downloaded. But when it gets connected, it automatically downloads recent feeds in the background and updates the view seamlessly.

Intelligent deployment and update

The application uses Updater Application Block 2.0 to provide auto update feature. Whenever I release a new version or deploy some bug fixes to a central server, all the users of that application automatically get the update behind the scene. This saves each and every user from going to the website and downloading the new version every time I release something new. It also allows me to instantly deliver bug fixes to everyone within a very short time.

Error reporting

This is my personal requirement for smart client. As the smart client is running on a distant computer, unlike web sites, you do not know whether the users are facing any problems or not and whether the applications are generating any exceptions. We need to provide some kind of error reporting feature that automatically captures the error and transmits the error to some central server so that the error can be fixed. You have seen this error reporting feature in Windows XP or Office 2003 applications. Whenever there is an error, it sends the error report to Microsoft. Similarly, this application also traps errors behind the scene and sends the exception trace to a tracking system at Sourceforge.

Multithreaded

Another favorite requirement of mine for making a client really smart is to make the application fully multithreaded. Smart clients should always be responsive. It must not get stuck whenever it is downloading or uploading data. Users will keep on working without knowing that there is something big going on in the background. For example, while a user is writing a blog, the application should download the feeds in the background without hindering the user’s authoring environment at all.

Crash proof

A smart client becomes a dumb client when it crashes in front of a user showing the dreaded “Continue” or “Quit” dialog box. In order to make your app truly smart, you need to catch any unhandled error and publish the error safely. In this app, I will show you how this has been done.

Topics covered

RSS Feeder is a complete application which deals with XML, XSLT, HTML, HTTP GET/POST, Configuration management, Cryptography, Logging, Auto Update, Rich UI, RSS/RDF/ATOM feed processing, XML RPC, Web Service, Blogging, Outlook automation, Multithreading and a lot more. Covering all these in detail requires the volume of a book. So, in this article I will just cover the tips and tricks I have used in all these areas. By the time you reach the end of this article, you will be fully equipped with real life experience of making a rich connected Smart Client application implementing the best practices and neat tricks you can find all over the web.

Enterprise Library

Enterprise Library is a major new release of the Microsoft patterns and practices application blocks. Application blocks are reusable software components designed to assist developers with common enterprise development challenges. Enterprise Library brings together new releases of the most widely used application blocks into a single integrated download.

The overall goals of the Enterprise Library are the following:

  • Consistency: All Enterprise Library application blocks feature consistent design patterns and implementation approaches.
  • Extensibility: All application blocks include defined extensibility points that allow developers to customize the behavior of the application blocks by adding it in their own code.
  • Ease of use: Enterprise Library offers numerous usability improvements, including a graphical configuration tool, a simpler installation procedure, and a clearer and more complete documentation and samples.
  • Integration: Enterprise Library application blocks are designed to work together and are tested to make sure that they do. It is also possible to use the application blocks individually (except in cases where the blocks depend on each other, such as on the Configuration Application Block).

Application blocks help address the common problems that developers face in every project. They have been designed to encapsulate the Microsoft recommended best practices for .NET applications. They can be added into .NET applications quickly and easily. For example, the Data Access Application Block provides access to the most frequently used features of ADO.NET in simple-to-use classes, boosting developer productivity. It also addresses scenarios not directly supported by the underlying class libraries. (Different applications have different requirements and you will find that every application block is not useful in every application that you build. Before using an application block, you should have a good understanding of your application requirements and of the scenarios that this application block is designed to address.)

RSS Feeder uses the following application blocks:

  • Caching Application Block: Caching frequently accessed data like the XSLT file which renders HTML from RSS feeds every time you click on an RSS Feed.
  • Configuration Application Block. Both static and dynamic configuration is handled by this block. For example, global configurations like whether you want Outlook integration, feed download interval etc.
  • Cryptography Application Block. Your weblog account password is encrypted using TripleDES algorithm.
  • Exception Handling Application Block. The entire application uses the exception handling block to handle all handled or unhandled exceptions.
  • Logging and Instrumentation Application Block. Informative, warning, debugs level logging and also error loggings are sent to a text file using this block.
  • Updater Application Block. Provides auto update from a central web server.

Connectivity

RSS Feeder has three ways connectivity:

  • Download latest updates from update server.
  • Download feeds from feed source.
  • Post to blog engines.

Auto update using Updater Application Block 2.0

The Updater Application block is a real pain to implement. After trying for weeks, I have finally given up its default BITSDownloader which does not work at all in my case. I have started using HTTPDownloader which works without any problem. When I use BITSDownloader I always get an error that the response header does not contain Content-Length. But if I capture the traffic, I can clearly see that Content-Length is present in the HTTP Response header. So, I have moved to the HTTPDownloader created by Kent Boogaart, for Katersoft. The HTTPDowloader uses nothing but the built-in HttpWebRequest to download the files from the server. It works both synchronously and asynchronously which suites all update scenarios.

The updater works this way:

  • First it gets the URL of the manifest XML. A manifest is prepared and stored in a server which describes the files that are needed to be downloaded.
  • After downloading the manifest, first it finds out the files that are needed to be downloaded according to the manifest. The it starts downloading them.
  • Once the download is complete, it spawns a console application which waits until RSS Feeder closes.
  • When RSS Feeder closes, the console application performs the updates and quits.

One problem I faced while downloading manifest is that the Proxy caches manifests. Normally your internet service provider uses proxies to cache web content. As XML is a web content, proxy servers nowadays cache XML files. As a result, even if I frequently update the manifest XML on the server, the proxies do not update their cache accordingly and returns an old version of the manifest file. This prevents auto update.

In order to solve this, I have injected a tick count at the end of the URI of the manifest. This always produces a unique URL and thus prevents proxies from returning old content.

/// Get the default configuration

// provider for updater block


Microsoft.ApplicationBlocks.Updater.
  Configuration.UpdaterConfigurationView view =
          new Microsoft.ApplicationBlocks.Updater.
              Configuration.UpdaterConfigurationView();

// Get the default manifest uri in order to add a unique

// timestamp at its end

// in order to avoid manifest file caching at proxies


Uri manifestUri = view.DefaultManifestUriLocation;
string uri = manifestUri.ToString();
uri += "?" + DateTime.Now.Ticks.ToString();

manifests = updater.CheckForUpdates(new Uri( uri ));

Deciding whether to overwrite old files with newer files

When you implement auto update in your application, you will come across a dilemma, whether to overwrite an older file with a newer file or not. For example, my RSS Feeder has several XSLT files which render RSS feeds to HTML. Now, the user has the freedom to change those files. So, if I release a new version of the file, I cannot overwrite those files without knowing whether the user has changed them or not. But I do need to update it because I may have fixed some problems in some of the files. Now you may think, duh! It’s an easy solution. Just check if the file dates are equal to the EXE’s date. If it is then the user has not changed the file. This is not always true because I release the updated EXE frequently. So, I cannot compare with the EXE’s date. The only way we can ensure we do not overwrite files accidentally is to preserve MD5 hash of each file and before copying ensure if the hash matches. I am not sure whether it is built into Updater Application block, but it would be nice to have this feature that automatically checks whether MD5 of a particular file matches before it overwrites a file.

Using Enterprise Library the easy way

Enterprise Library is a great piece of work. It can really take off a lot of framework development load from any type of application you develop. Normally when we go for developing a new project, we start with a framework which provides configuration management, security, cryptography, database access etc. All these require time to integrate and test. Enterprise Library saves you from developing such framework because it provides the best practices collected from years of experience of successful projects.

However, using Enterprise Library classes directly from all layers of your application is problematic. Very soon, you will feel the need for a wrapper class. Also, as it is a generic library, you first need to customize it a bit before getting started. Here I will present some extra tweaking that you can use to get started with Enterprise Library.

Let’s look at a typical configuration file which contains Caching, Cryptography, Exception and Logging application blocks:

Now look at the encircled sections. These are the things that you have to remember while using EL classes. For example, if you want to use the logging block, you have to provide the Category name:

Logger.Write( "Log message", "Debug" );

Again, if you want to use the Cache Manager, you will have to remember the name:

CacheManager manager = CacheFactory.GetCacheManager("XSL Cache");

If you want to use Security Block classes, you need to specify the security provider name e.g. MD5CryptoServiceProvider:

Cryptographer.CreateHash( "MD5CryptoServiceProvider", “hash me” );

So, very soon, your code will be full of EL classes and when any class’s signature changes or new version is released, as it did from Pattern and Practices Application Blocks to the new Enterprise Library, you will have to perform Search and Replace throughout your project to upgrade your code. This is a real pain. So, what I have done here is, I have made an Enterprise Library Helper class named EntLibHelper which you can use in the following way:

EntLibHelper.Info( “Information loggin”); // Log block


//Caching block

string hashedValue = EntLibHelper.Hash( “Hash me” );

// Exception block

EntLibHelper.Exception( “Some error occurred”, x );

So, next time, when Microsoft releases something called Universal Library, all you need to do is change the code inside EntLibHelper. Moreover, EntLibHelper handles all the troubles of initializing Enterprise Library blocks, ensures proper usage and provides a convenient interface which frees developers from remembering EL interfaces.

Check the source code for EntLibHelper. Remember, it is compatible with the app.config for RSSFeeder. If you make changes in any of the names, for example, renaming Logging Category or changing MD5 hashing to SHA hashing, you will have to change the names specified in the constants at the top of the class.

private const string EXCEPTION_POLICY_NAME = "General Policy";
private const string SYMMETRIC_INSTANCE = "DPAPI";

private const string HASH_INSTANCE = "MD5CryptoServiceProvider";
public const string XSL_CACHE_MANAGER = "XSL Cache";

RSS Feeder object model

RSS Feeder has a tiny object model as shown below:

Channels is a collection of Channel which represents a source of RSS feed. Channel is a collection of RSSFeed objects. RSSFeed object contains some basic information about a post like title, publish date, and the GUID which uniquely identifies a particular item. The entire XML received for each item is stored in the XML field.

Some of the items are extracted from the original XML to the RSS object for faster access. For example, when we render the list of feeds in a Listview, we need to show the title and the publish date. Moreover we need to sort on the publish date. This is why some of the fields are duplicated in the RSSFeed object which is taken from the actual XML.

Generally both RSS and Atom XML contain the title of the post, the author’s name, a unique ID, link to the post, publish date and a detail body.

RSS 2.0 XML format

Let’s look at a sample XML of RSS 2.0 format:

<?xml version="1.0"?>

<rss version="2.0" xmlns:dc=http://purl.org/dc/elements/1.1/
xmlns:admin=http://webns.net/mvcb/
xmlns:rdf=http://www.w3.org/1999/02/22-rdf-syntax-ns#
xmlns:content="http://purl.org/rss/1.0/modules/content/">

<channel>
   <title>.NET Community Blog</title>
   <link>http://localhost/b2evolution/index.php</link>
   <description>.NET Community Blog</description>

   <language>en-US</language>
   <docs>http://backend.userland.com/rss</docs>
   <ttl>60</ttl>

   <item>
      <title>Important information</title>
      <link>http://localhost/b2evolution/index.php?...</link>
      <pubDate>Fri, 17 Jun 2005 10:05:52 +0000</pubDate>

      <category domain="external">Announcements [A]</category>
      <category domain="alt">Announcements </category>

      <category domain="main">b2evolution Tips</category>
      <guid isPermaLink="false">21@http://localhost/b2evolution</guid>

      <description>Blog B contains a few posts in the
      'b2evolution Tips' category. </description>
      <content:encoded>
      <![CDATA[ <p>Blog B contains a few posts in the 'b2evolution Tips'
      category.</p>]]>

      </content:encoded>
      <comments>http://localhost/b2evolution/...</comments>
   </item>

The root node is <rss> which contains one or more <channel> nodes. <item> node represents one post. The body of the post is available in the <description> node. Some RSS Feed generators write both HTML and plain text content inside the <description>. However, some advanced generators write a plain text version inside the <description> and the actual HTML version with all the formatting inside the <content:encoded> node.

Uniquely identifying an item is troublesome because not all sites generate the guid node. For example, CodeProject RSS feeds contain no guid node. As a result, the only option you have to uniquely identify a feed is either to generate an MD5/SHA hash of the entire content and use that hash value as identifier or use the link node. RSS Feeder first looks whether there is any guid node, if not it uses the link node as the unique identifier. (To CodeProject: If this is not correct, let me know.)

Not all sites follow the RSS 2.0 format. Some sites are still using the RSS 0.9 format. Even those who do follow, do not always generate all the nodes properly. An example is the CodeProject RSS feeds, where you can see that the guid node is missing. You need to be careful while parsing RSS.

Atom 0.3 format

<?xml version="1.0" encoding="utf-8"?>

<feed version="0.3" xml:lang="en-US"
    xmlns="http://purl.org/atom/ns#">
<title>.NET Community Blog</title>

<link rel="alternate" type="text/html"
    href="http://localhost/b2evolution/index.php" />

<tagline>.NET Community Blog</tagline>
<generator url="http://b2evolution.net/"
    version="0.9.0.10">b2evolution</generator>

<modified>1970-01-01T00:00:00Z</modified>
<entry>
  <title type="text/plain"
      mode="xml">Important information</title>

  <link rel="alternate" type="text/html"
  href=http://localhost/b2evolution/index.php... />

  <author>
  <name>admin</name>
  </author>
  <id>http://localhost/b2evolution/index.php?...</id>

  <issued>2005-06-17T10:05:52Z</issued>
  <modified>1970-01-01T00:00:00Z</modified>
  <content type="text/html" mode="escaped">

  <![CDATA[ <p>Blog B contains a few posts in the
  'b2evolution Tips' category.</p>
  <p>All these entries are designed to help you so,
  as EdB would say:
  "<em>read them all before you start
  hacking away!</em>" ...]]></content>

</entry>

Atom type is also similar but looks a bit better to me than the RSS 2.0 format. The better things that I have noticed are, one the content node which has a nice type attribute that defines the type of the content. The mode attribute that helps to identify whether HTML decoding is required or not. It also has a nice id attribute which uniquely identifies an entry. The nodes are obvious and best of all; everyone generates consistent output when they produce Atom feed whereas RSS has several versions that are still widely being used.

Date handling

When you make an RSS aggregator, you will soon realize that people do not follow a consistent date format. Some use .NET’s DateTime format, some use PHP’s date format, some use Java’s date format and some even use RFC822 or RFC1123 date formats. There are so many different date formats people are using nowadays that you cannot find any code that can parse them all. After much struggle, I have finally come down to this function which tries to digest several possible date formats:

private DateTime FormatDate( string date )
{
    string RFC822 = "ddd, dd MMM yyyy HH:mm:ss zzz";

    //string RFC1123 = "yyyyMMddTHHmmss";


    //string RFCUnknown = "yyyy-MM-ddTHH:mm:ssZ";


    int indexOfPlus = date.LastIndexOf('+');
    if( indexOfPlus > 0 )
        date = date.Substring( 0, indexOfPlus-1 );

    string [] formats = new string[] { "r", "S", "U" };
    try

    {
      // Parse the dates using the standard

      // universal date format

      return DateTime.Parse(date,
           CultureInfo.InvariantCulture,
           DateTimeStyles.AdjustToUniversal);
    }
    catch

    {
       try
       {
          // Standard formats failed, try the "r" "S"

          // and "U" formats


          return DateTime.ParseExact( date, formats,
                    DateTimeFormatInfo.InvariantInfo,
                    DateTimeStyles.AdjustToUniversal);
       }
       catch
       {
           try
           {
              // All the standards formats have failed,


              //try the dreaded RFC822 format

              return DateTime.ParseExact( date, RFC822,
                      DateTimeFormatInfo.InvariantInfo,
                      DateTimeStyles.AdjustToUniversal);
           }
           catch
           {
              // All failed! The RSS Feed source

              // should be sued


              return DateTime.Now;
           }
       }
    }
}

Creating a generic parser for Atom, RDF and RSS feeds

A variety of widely accepted formats make developers’ lives difficult because they have to support all of the widely used specifications in their application. A feed aggregator needs to support RSS, Atom and RDF all together because all of these are widely used. This results in design complexity because you need to make your object model and parsing process generic for parsing and storing feeds of three different formats.

After much searching, I have finally realized that everyone produces different object models for RSS and Atom feeds. However, having different object model means you need to create different table structures in the database which is difficult to maintain. You also need to write code in all places which first checks whether it is dealing with Atom or RSS feed. This makes your application complicated. Such a problem is solved by XML. In XML, you can store different types of data, yet fully structured. Using XSL, you can easily render different XML structures the same way. So, it does not matter whether you have Atom or RSS content inside your XML, an XSL can easily check the type and produce same HTML output for viewing.

First let’s see the generic feed parser that I have made. The FeedProcessor.cs is the generic feed parser which can parse RSS/Atom/RDF the same way and produce Channel and RSSFeed objects in one generic format.

While parsing, first it sees whether the XML contains RSS or Atom or RDF:

public IList Parse( XmlReader reader )
{
    IList channels = new ArrayList();

    while( reader.Read() )
    {
       if( reader.NodeType == XmlNodeType.Element )
       {
           string name = reader.Name.ToLower();

           switch( name )
           {
               case "atom:feed":      // We have Atom Feed


               case "feed":   // We have Atom Feed

                      channels.Add( this.ProcessAtomFeed(reader));
                      break;
               case "rdf:rdf":        // We have rdf feed


               case "rdf":            // We have rdf feed

               case "rss:rss":        // We have rss feed


               case "rss":            // We have rss feed

                      channels.Add( this.ProcessRssFeed(reader));
                      break;
           }
       }
    }

    return channels;
}

For Atom feed, it calls the ProcessAtomFeed which parses the channel properties only:

private RssChannel ProcessAtomFeed( XmlReader reader )
{
    RssChannel channel = new RssChannel();
    channel.Type = RssTypeEnum.Atom;

    channel.Feeds = new ArrayList();
    while( reader.Read() )
    {
       if( reader.NodeType == XmlNodeType.Element )
       {
           string name = reader.Name;

           switch( name )
           {
               case "title":  // title for channel


                   channel.Title = ReadString( reader );
                   break;
               case "link":   // link to website

                   reader.MoveToAttribute("href");
                   if( reader.ReadAttributeValue() )
                   {
                           channel.Link = reader.Value;
                   }
                   break;
               case "tagline": // description of the channel


                   channel.Description = ReadString( reader );
                   break;
               case "description": // Same

                   channel.Description = ReadString( reader );
                   break;
               case "entry": // Aha! an entry


                   channel.Feeds.Add(this.ProcessAtomEntry(reader));
                   break;
           }
       }
       else if( reader.NodeType == XmlNodeType.EndElement )
       {
               if( reader.Name == "feed" )
                       break;
       }

    }

    return channel;
}

Similarly ProcessRssFeed function does the same job.

The complex part is parsing the entry or item node which actually contains the XML content. We need to parse it in two ways:

  • We need to discover some essential properties like publish date, title and GUID according to what we are parsing (Arom/RSS).
  • We need to store the everything we are reading in a buffer because XmlReader is a one way reader and we cannot go back once we have forwarded.

So, we not only need to use an XmlReader to read, but also an XmlWriter to write the same thing as we read in a temporary buffer.

The next design complexity is writing a generic function which parses the entry and item node in the same way. Although we can make two different functions for parsing entry and item nodes, the functions will have 90% code duplicated; the only differences are the name of some nodes and some structural differences. So, here is the function which parses it all:

private RssFeed ProcessFeedNode( XmlReader reader,
    string itemNodeName, string titleNodeName,
    string guidNodeName, string linkNodeName,
    string pubDateNodeName )
{
        RssFeed feed = new RssFeed();

        // Build a buffer which stores the


        // entire XML content of the entry

        StringBuilder buffer = new StringBuilder(1024);
        XmlTextWriter writer =
            new XmlTextWriter(new StringWriter(buffer));
        writer.Namespaces = false;
        writer.Indentation = 1;
        writer.IndentChar = '\t';
        writer.Formatting = Formatting.Indented;

        writer.WriteStartElement(itemNodeName);

        string lastNode = reader.Name;
        while( (reader.NodeType == XmlNodeType.Element 
           && lastNode != reader.Name) || reader.Read() )
        {
               if( reader.NodeType == XmlNodeType.Element )
               {
                       lastNode = reader.Name;

                       writer.WriteStartElement( reader.Name );
                       writer.WriteAttributes( reader, true );
                       if( reader.Name == titleNodeName )
                       {
                               feed.Title = ReadString( reader );
                               writer.WriteString(feed.Title);
                       }
                       else if( reader.Name == guidNodeName )
                       {
                               feed.Guid = ReadString( reader );
                               writer.WriteString(feed.Guid);
                       }
                       else if( reader.Name == linkNodeName )
                       {
                               // Atom feed contains the link as "href" attribute


                               string link = reader.GetAttribute("href", "");
                               if( null == link )
                               {
                                      // but Rss feed has the link as value


                                      link = ReadString( reader );
                                      writer.WriteString( link );
                               }

                               if( feed.Guid == null )
                               {
                                      feed.Guid = link;
                               }

                       }
                       else if( reader.Name == pubDateNodeName )
                       {
                               string date = ReadString( reader );
                               feed.PublishDate = this.FormatDate( date );
                               writer.WriteString(date);
                       }
                       else

                       {
                               writer.WriteRaw( reader.ReadInnerXml() );
                       }

                       // Close the element started

                       writer.WriteEndElement();

                       // For empty elements, ReadEndElement fails

                       if( reader.NodeType == XmlNodeType.EndElement )
                       {
                               if( reader.Name == itemNodeName ) break;

                               reader.ReadEndElement();
                       }
               }

               if( reader.NodeType == XmlNodeType.EndElement )
               {
                       if( reader.Name == itemNodeName )
                               break;
               }
        }

        writer.WriteEndElement();
        writer.Close();

        feed.XML = buffer.ToString();

        return feed;
}

Although this function is not optimal, we can optimize it in many ways. But it does the job pretty well. It parses a 200 KB feed in a fraction of a second without even occupying 5% of the CPU.

Tips 1. ReadString( reader ) or reader.ReadString()

You will see in the above code that I have used a custom function called ReadString instead of using XmlReader’s ReadString method. The documentation says ReadString method is supposed to read the content of the string. It is not supposed to jump off the end tag. But in practice, it does go over the end tag and stops at the next begin tag. So, if you are reading the <title> node, and call ReadString, the next node you will get is the <pubDate> node, not the </title>. But we need to know when a tag is closed so that we can close the tag in the XmlWriter also. This is why I have made the custom ReadString method:

private string ReadString( XmlReader reader )
{
    /// Reuse existing buffer in order to prevent


    /// frequent StringBuffer allocation

    buffer.Length = 0;
    /// Empty elements have no content

    if( reader.IsEmptyElement ) return string.Empty;

    /// Skip the begin tag and all white spaces before


    /// the first character of content is found

    while(!reader.EOF
         && ( reader.NodeType == XmlNodeType.Element
         || reader.NodeType == XmlNodeType.Whitespace ) )
       reader.Read();

    /// Read and store in buffer when we are getting text

    /// and CDATA sections.


    /// But stop immediately

    /// whenever we read the end element.

    while( reader.NodeType == XmlNodeType.CDATA
        || reader.NodeType == XmlNodeType.Text
        && reader.NodeType != XmlNodeType.EndElement )
    {
        buffer.Append( reader.Value );
        reader.Read();
    }

    /// Now the read is poting to the EndElement. Return


    /// the content of the buffer

    /// we have prepared for this node

    return buffer.ToString();
}

Converting Atom 0.3 to RSS 2.0

The best way to make this application simple and think only about RSS is to convert Atom XML to RSS XML just after downloading the content from web source. This way, the whole application can deal with the RSS feed and need not worry about all other formats. So, in future, if another format gets popular, all I need to do is write another converter which converts that format to RSS format. The application receives no significant changes.

In the source code, you will find atomtorss2.xslt which converts Atom 0.3 XML to RSS 2.0 XML. Here is a little excerpt of the XSLT which does the conversion:

<xsl:template name="items">
<xsl:for-each select="atom:entry">

  <item>
        <title><xsl:value-of select="atom:title"/></title>
        <link>

          <xsl:value-of select="atom:link[@rel='alternate']/@href"/>
        </link>
        <guid><xsl:value-of select="atom:id" /></guid>

        <description>
               <xsl:value-of select="atom:content" />
        </description>

        <pubDate>
               <xsl:choose>
               <xsl:when test='atom:issued'>
               <xsl:value-of select="date:format-date(atom:issued,'EEE,
                                              dd MMM yyyy hh:mm:ss z')"/>

               </xsl:when>
               <xsl:when test='atom:modified'>
               <xsl:value-of select="date:format-date(atom:modified,'EEE,
                                               dd MMM yyyy hh:mm:ss z')"/>

               </xsl:when>
               </xsl:choose>
        </pubDate>
  </item>
</xsl:for-each>

</xsl:template>

Now, here you will notice that, there’s a function format-date inside the select attribute. This function is not available in XSLT processor. So, how do we do this?

Introducing EXSLT project

EXSLT is taking XSLT processing to the next level. It supports a rich collection of functions that you can use in XSLT scripts which makes XSL a truly powerful script for real world XML transformation. The beauty of XsltTransformer is that, it allows you to write pure .NET functions which are invoked whenever they are called from XSLT script. Using this feature, you can make complicated XML transformation which fully utilizes the full power of .NET platform. You can even write functions that can call a database and get dynamic values and put that value in the resulting XML.

Using EXSLT is very easy. The following code shows how an Atom 0.3 XML is converted to RSS 2.0 XML:

void EXSLT()
{
        ExsltTransform xslt = new ExsltTransform();
        xslt.SupportedFunctions = ExsltFunctionNamespace.All;
        xslt.MultiOutput = false;
        xslt.Load("atomtorss2.xslt");
        xslt.Transform("atom.xml", "rss.xml");
}

XML serialization/deserialization

Although we all know it, I always forget to do the right thing to serialize ArrayList which contains custom objects. So, I have made a convenient SerializationHelper class which exposes variants of Serialize or Deserialize functions which you can use in your applications:

public static XmlWriter Serialize( Stream stream, object o )
{
        XmlTextWriter writer = new XmlTextWriter( stream,
                               System.Text.Encoding.UTF8 );
        XmlSerializer serializer =
                          new XmlSerializer( o.GetType() );
        serializer.Serialize( writer, o );
        return writer;
}

public static XmlWriter Serialize( Stream stream,
                         ArrayList array, Type type )
{
        XmlTextWriter writer = new XmlTextWriter( stream,
                                System.Text.Encoding.UTF8 );
        XmlSerializer serializer = new
           XmlSerializer(typeof(ArrayList), new Type[] {type});
        serializer.Serialize( writer, array );
        return writer;
}


public static object Deserialize( Stream stream, Type t )
{
        XmlTextReader reader = new XmlTextReader( stream );
        XmlSerializer serializer = new XmlSerializer( t );
        object o =  serializer.Deserialize( reader );
        return o;
}


public static ArrayList DeserializeArraylist(Stream stream,
                                                    Type t)
{
        XmlTextReader reader = new XmlTextReader( stream );
        XmlSerializer serializer = new
            XmlSerializer( typeof( ArrayList ), new Type [] { t } );
        ArrayList list =
            (ArrayList) serializer.Deserialize( reader );
        return list;
}

In order to serialize an ArrayList which contains objects of type RSSFeed, you issue the following command:

Serialize( stream, arrayList, typeof( RSSFeed ) );

Tips 2. Serializing an object which has property of type array

If you have a property of type ArrayList, the serialized XML looks very ugly and does not look like what we call – strongly typed if that applies to XML. In order to customize how an ArrayList is serialized, you can try the two attributes XmlArray and XmlArrayItem:

private ArrayList _WebLogs = new ArrayList();

[XmlArray("weblogs"), XmlArrayItem("weblog", typeof(WebLog)) ]

public ArrayList WebLogs
{
        get { return _WebLogs; }
        set { _WebLogs = value; }
}

This way you will get a nice XML output.

Embedded resource

The easiest way to carry additional files with your projects is not to carry them as external files, instead putting them inside an assembly. Visual Studio has a nice property named Build Action for this purpose:

If you set the property to Embedded Resource, that file is linked inside the assembly. This means that the entire content of the file is embedded inside the assembly as inline resource. As a result, you can make one assembly which contains all the files you need. You can directly read the files from the assembly as a Stream without opening as a file. Each file is embedded inside the assembly by the following naming convention:

NameSpace.FileName.Extention

So, for the file atomtorss2.xslt, the full name of the embedded resource is:

RSSFeederResources.atomtorss2.xslt

The best thing about embedded resources is that they are not in the file system as separate physical files. So, you never need to worry about the path of the files. You can directly read the content of the file as a Stream. However, the embedded resource is read-only. If you want to modify the content, then there is no way known to me so far to do that. In that, case you will have to create a file from embedded resource and use that file always.

The SerializationHelper has some convenient functions to deal with embedded assemblies. For example:

public static Assembly GetResourceAssembly()
{
   return Assembly.LoadFrom("RSSFeederResources.dll");
}

public static Stream GetStream( string name )
{
   return GetResourceAssembly().GetManifestResourceStream(
                           RESOURCE_ASSEMBLY_PREFIX + name);
}

GetStream returns an embedded resource as a Stream.

public static System.Drawing.Icon GetIcon( string name )
{
        using( Stream stream = GetStream( name ) )
        {
               return new System.Drawing.Icon( stream );
        }
}

You get an Icon object directly from an embedded icon file. This is very convenient for carrying all the icons that your application uses. You don’t need to maintain separate icon files and worry about their path.

public static void WriteEmbeddedFile( string name, string fileName )
{
        using( Stream stream = GetStream( name ) )
        {
               FileInfo file = new FileInfo( fileName );
               using( FileStream fileStream = file.Create() )
               {
                       byte [] buf = new byte[ 1024 ];
                       int size;
                       while( (size = stream.Read( buf, 0, 1024 )) > 0 )
                       {
                               fileStream.Write( buf, 0, size );
                       }
               }
        }
}

The WriteEmbeddedFile function generates a file from the content of the embedded resource.

Storing data in application data folder

The best place to store application specific data is not in the folder where your program is installed but in the “Application Data” folder that Windows® creates for each user. There are actually two Application Data folders. One is directly under the user folder and another is inside the hidden folder "Local Settings". So, if you user name is Omar AL Zabir, the path to these folders will be:

C:\Documents and Settings\Omar Al Zabir\Application Data

And the secret one is:

C:\Documents and Settings\Omar Al Zabir\Local Settings\Application Data

The major difference between them is that the first one is visible. You can browse to that folder using Explorer. But the second one is hidden by default. You need to turn on “Show Hidden Files” from Explorer Options in order to see that folder.

Another difference is that, when you have a roaming profile in Windows®, the first visible folder is synchronized to the network store. So, whatever you store in that folder, you can easily access it when you login from another computer inside the domain. This location is better for storing files than the location where your application .exe is located because that does not roam. Remember, the second folder is computer specific and it is not synchronized when you move from one computer to another.

RSS Feeder stores all the data in the second folder.

You can get the path to these special folders using the Environment.GetFolderPath function. It takes an enumeration as shown in the following code snippet:

enum System.Environment.SpecialFolder
{
  ApplicationData
  CommonApplicationData
  LocalApplicationData
...
}

There are many other useful folders like the Desktop, Program Files, My Documents etc. All these are made available from this enum. Here’s how you use it:

// Prepare the path where all application specific


// settings will be stored

string appDataPath =  Environment.GetFolderPath(
         Environment.SpecialFolder.LocalApplicationData);

ApplicationSettings.ApplicationDataPath =
    Path.Combine( appDataPath, "RSS Feeder" );

// Check if all these paths exist

if( !Directory.Exists(ApplicationSettings.ApplicationDataPath))
{
   Directory.CreateDirectory(ApplicationSettings.ApplicationDataPath);
}

This way you can setup your own folder for storing all the application specific files.

OPML

OPML is an XML-based format that allows exchange of outline-structured information between applications running on different operating systems and environments. OPML is used to store information about RSS Feed sources. For example, a blog site uses OPML to store all the blog titles and feed locations that it contains. If you go to blogs.msdn.com you will get the OPML for all the Microsoft bloggers’ feed URL and title.

<opml>
  <body>
    <outline text="Microsoft Bloggers">

      <outline title="Alex Lowe's .NET Blog"
    htmlUrl=http://blogs.msdn.com/alowe/default.aspx
    xmlUrl="http://blogs.msdn.com/alowe/rss.aspx" />

      <outline title="Michał Cierniak"
    htmlUrl=http://blogs.msdn.com/michaljc/default.aspx
    xmlUrl="http://blogs.msdn.com/michaljc/rss.aspx" />

...
...

The above XML is an excerpt of OPML from blogs.msdn.com.

RSS aggregators use OPML to exchange subscription information among other aggregators. For example, you can export all your subscriptions as OPML from Newsgator and then import the OPML to my RSS Feeder. In fact you don’t need to do it at all. Whenever you run it, it will import the Newsgator settings. Beware, Newsgator's XML contains "xmlurl" but all other OPML use "xmlUrl". The difference is in the case of "U".

Newsgator stores its subscription information in an OPML. Its OPML is a bit different:

<?xml version="1.0" encoding="utf-8"?>

<opml xmlns:ng="http://newsgator.com/schema/opml">
  <body>
    <outline title="NewsGator News and Updates"
      description="NewsGator News and Updates"

      lastItemMD5="eFY8dWcOAPZBbpG7Ha1l5g==,XqCFJLx/DJigD7YRFnD3OA==,..."
      xmlurl=http://www.newsgator.com/news/rss.aspx
      htmlurl="http://www.newsgator.com"
      ng:folderName="" ng:folderType="auto"

      ng:useDefaultCredentials="false"
      ng:username="" ng:passwordenc=""
      ng:domain="" ng:useGuid="false"

      ng:interval="0" ng:nntpMostRecent="-1"
      ng:newsPage="true" ng:renderTransform=""

      ng:downloadAttachments="false" />
</body>
</opml>

It uses additional properties like ng:folderName and ng:folderType to describe Outlook Folder location where subscriptions are mapped. RSS Feeder provides Newsgator import feature. It uses this information to map to the same folder as Newsgator does.

The OpmlHelper class in the source code provides OPML parsing and generation functionality.

Database optimization

As RSS Feeder uses an MS Access database, it does not have much problem with multiple clients connecting to database or extra design issues that we need to consider while using SQL Server. However, we do need to optimize connection open and close as MS Access takes pretty long time to open and close connections.

So, what I have done here is open a static connection when the app starts and use that connection throughout the application. When the application closes, it closes the open connection. This gives significant performance boost than opening and closing connection whenever we access the database.

However, static connection object leads to multithreading issues. Sometimes two threads can try to execute commands on the same connection at the same time. For example, imagine Feed Downloader is downloading feeds in the background and you are reading the feeds. Now both you and the feed downloader are trying to read/write feeds from the database at the same time. At least one will fail as the connection will be in Executing state instead of Open state. In order to prevent this, I have implemented a Thread.Sleep whenever I see the connection is in Executing state.

private static OleDbConnection __Connection = null;

private static OleDbConnection _Connection
{
    get
    {
       if( null == __Connection )
       {
           string connectionString = string.Format
           ( "Provider=Microsoft.Jet.OLEDB.4.0;Data Source={0};",
                          ApplicationSettings.DatabaseFilePath );
           __Connection = new OleDbConnection( connectionString );

           __Connection.Open();
       }
       else

       {
           while( ConnectionState.Executing == __Connection.State
               || ConnectionState.Fetching == __Connection.State )
           {
                   System.Threading.Thread.Sleep( 50 );
           }

           if( ConnectionState.Open != __Connection.State )
                   __Connection.Open();
       }

       return __Connection;
    }
}

public static void Close()
{
    if( null != __Connection )
       if( ConnectionState.Closed != __Connection.State )
       {
           __Connection.Close();
           __Connection.Dispose();
       }
}

So, when a thread tries to get hold of a connection object while the connection is executing, the thread goes into a 50 ms sleep. This is pretty much enough time to let the execution complete.

In a normal single processor computer, you will rarely face this problem. But if you have a P4 Hyper Threading Processor which simulates two processors, or your own HP6000 with six Xeon Processors in your house, then you will see this problem frequently.

MS Outlook® integration

This is the most complicated part. The first problem you will face while automating Outlook is releasing all the object references used in your code. If you do not release the references properly, Outlook will never close even if it disappears from the screen. The second problem is making a version independent solution. If you “Add Reference...” to Outlook COM library from Visual Studio and set reference to a specific version of DLL, the interop assembly that is generated becomes a version specific assembly. As a result, it will not work properly with older versions of Outlook. The only way you can make it version independent is to develop against the oldest version of Outlook you want to support and make builds from that version.

Another solution to make a truly version independent Outlook automation is to use Late Binding. You can use the Activator class in .NET framework to instantiate any COM object using ProgID. So, you can launch Outlook using late binding in the following way:

public static void StartOutlook( ref object app, ref object name )
{
        try

        {
            // Try to get existing outlook instance

            app = Marshal.GetActiveObject("Outlook.Application");
            name = GetProperty( app, "Session" );
        }
        catch
        {
            // Create new instance of outlook


            app = GetObject("Outlook.Application");
            name = GetProperty( app, "Session" );
        }
}
public static object GetObject( string typeName )
{
    Type objectType = Type.GetTypeFromProgID(typeName);
    return Activator.CreateInstance( objectType );
}


public static object GetProperty( object obj,
                               string propertyName )
{
    return obj.GetType().InvokeMember( propertyName,
          BindingFlags.GetProperty, null, obj, null );
}

Although this makes coding a nightmare, but you can make helper functions to deal with the objects and properties. The OutlookHelper class provided with the source code gives you many helper functions to automate Outlook. However, you still have to memorize the object model of the Outlook first before writing code using Late Bound approach. In order to have a safe late bound approach to COM interop and a version independent yet strongly typed Outlook Automation Library, check my article SafeCOMWrapper. If you use this, you don’t have to write the cumbersome code and you don’t need to remember the object model.

Here’s a late bound way to show Folder Picker dialog box from the Outlook:

public static string SelectFolderPath()
{
    // Show outlook folder picker


    object app = null, name = null;

    StartOutlook( ref app, ref name );

    // Ensure at least one explorer is visible


    // otherwise we cannot show the folder

    // picker

    bool isOutlookInvisible = EnsureExplorer( app, name );

    object folder = CallMethod( name, "PickFolder" );

    string path = null;
    // Get the selected folder name


    if( null != folder )
      path = GetProperty( folder, "FolderPath" ) as string;

    if( null != folder )
           Marshal.ReleaseComObject( folder );
    folder = null;

    Marshal.ReleaseComObject( name );
    name = null;

    if( isOutlookInvisible )
           CloseOutlook(app);

    Marshal.ReleaseComObject( app );
    app = null;

    // Outlook still hangs around in the task


    // manager until we call this

    GC.Collect();

    return path;
}

Creating posts in Outlook from RSS Feeds

Whenever the application downloads RSS feeds, it also sends them to a mapped Outlook folder. First it transforms XML of an entry to HTML using XSLT transformation described in this article. Then it uses a UserProperty to store the actual XML against the post. This way, you can write macros later on to work on the XML of the original entries.

foreach( RssFeed item in rssItems )
{
    // Create outlook item


    object post = OutlookHelper.CallMethod( folderItems,
                                             "Add", 6 );
    //Outlook.OlItemType.olPostItem = 6


    //post.CreationTime = item.PublishDate;


    OutlookHelper.SetProperty( post, "Subject",
                                       item.Title );

    // Clear stream

    stream.Position = 0;
    stream.SetLength(0);

    // Transform RSS item XML to HTML


    XMLHelper.TransformXml( xsltFileName,
                           item.XML, stream );

    // Read the content from the stream

    stream.Position = 0;
    string html = reader.ReadToEnd();

    OutlookHelper.SetProperty( post,
                       "HTMLBody", html );
    OutlookHelper.SetProperty( post,
                       "BodyFormat", 2 );
    // Outlook.OlBodyFormat.olFormatHTML = 2



    // Store the actual item XML in an user property


    object userProperties =
       OutlookHelper.GetProperty( post, "UserProperties" );
    object missing = System.Reflection.Missing.Value;
    object xmlProperty =
       OutlookHelper.CallMethod( userProperties, "Add", "XML",
                                        1, missing, missing );
    // Outlook.OlUserPropertyType.olText = 1



    Marshal.ReleaseComObject( userProperties );

    OutlookHelper.SetProperty( xmlProperty,
                             "Value", item.XML );
    OutlookHelper.SetProperty( post, "UnRead", 1 );

    OutlookHelper.CallMethod( post, "Post" );

    Marshal.ReleaseComObject( post );

    itemsAdded ++;
}

Reading posts from Outlook folder and delivering to blog site

RSS Feeder provides blogging directly from Outlook. First you create a “Mail and Post” type folder. Then you create Posts in that folder. You can also drag posts from other folders to this folder. RSS Feeder picks up the posts available in the folder and then publishes it to the weblog site.

The following code shows how this is done the hard way which is the late bound way:

object folderItems = OutlookHelper.GetProperty( folder, "Items" );

int index = 1;
while(index <=
       (int)OutlookHelper.GetProperty(folderItems,
                                         "Count"))
{
    object item =
       OutlookHelper.GetItem( folderItems, index );

    string messageClass =
       (string)OutlookHelper.GetProperty( item,
                                 "MessageClass" );

    if( "IPM.Post" == messageClass )
    {
       string subject =
         (string)OutlookHelper.GetProperty(item,
                                      "Subject");
       string categories =
         (string)OutlookHelper.GetProperty( item,
                               "Categories", "" );
       string html;

       try

       {
         html =
           (string)OutlookHelper.GetProperty( item,
                                       "HTMLBody");
       }
       catch
       {
          ...

          MessageBox.Show( this,
          "You did not allow me to read" +
          "the post from Outlook." +
          "Please allow it next time.",
          "Outlook Error", MessageBoxButtons.OK,
          MessageBoxIcon.Exclamation );

          ...
          ...
          ...

          continue;
       }

       // Contruct a new post


       Post p = new Post();
       p.Title = subject;
       p.Text = html;
       p.Date = DateTime.Now;

       p.Categories =
          webLog.GetCategories( categories );

       // Post to web log, note this goes over web

       try
       {
         WebLogProvider.Instance.PostBlog(webLog, p);

         // After successful post, move


         // the post to "Sent" folder

         OutlookHelper.CallMethod( item,
                        "Move", sentFolder );
       }
       catch( Exception x )
       {
            ...
       }
    }

    Marshal.ReleaseComObject( item );
}

Downloading feeds using HTTP

Although there is nothing special in this area, but I can show you some handy tips. For example, I have made a handy function which downloads content from a URL and gives progress update as the download continues. As a result, you can make a progress dialog box which can show the downloaded content size and the connection speed while the download continues.

Tips 3. Downloading content in a MemoryStream with progress update

The following function handles the common download scenario with proxy support:

private static void DownloadContent( Uri uri,
        string proxyName, int port,
        string proxyUserName, string proxyPassword,
        ProgressEventHandler progressHandler,
        MemoryStream memoryStream )
{
   WebRequest webRequest = HttpWebRequest.Create(uri);

   // Set proxy settings to the web request


   if( null != proxyName && proxyName.Length > 0 )
   {
       webRequest.Proxy = new System.Net.WebProxy( proxyName, port );
       if( proxyUserName.Length > 0 )
       {
           // Decrypt the encrypted password


           string password = RSSCommon.PropertyEditor.PasswordEditor.
                              PasswordProvider.Decrypt(proxyPassword);
           // Set proxy credentials

           webRequest.Proxy.Credentials =
           new System.Net.NetworkCredential( proxyUserName, password );
       }
   }
   else
       webRequest.Proxy = System.Net.WebProxy.GetDefaultProxy();

   DateTime startTime = DateTime.Now;

   using(System.Net.WebResponse webResponse =
                                  webRequest.GetResponse())
   {
       using(System.IO.Stream stream =
                           webResponse.GetResponseStream())
       {
           byte [] buffer = new byte[ 1024 * 2 ]; // 2 KB buffer


           int size;

           while( (size = stream.Read( buffer, 0,
                           buffer.Length) ) > 0 )
           {
               memoryStream.Write( buffer, 0, size );

               // Perform some speed calculation



               // Total time elapsed

               TimeSpan duration = DateTime.Now - startTime;
               // Downloaded kilobytes

               double kb = (memoryStream.Length / 1024.0);
               double speed = kb / duration.TotalSeconds;
               string message = string.Format( "{0} kb, {1} kbps",
                               kb.ToString("n2"), speed.ToString("n2") );

               // notify download progress


               progressHandler( null,
                  new ProgressEventArgs( message , 0 ) );
           }
       }
   }

}

The above function takes a delegate of type ProgressEventHandler which is defined as follows:

public delegate void ProgressEventHandler( object sender, ProgressEventArgs e );

The ProgressEventArgs extends .NET framework’s EventArgs:

[Serializable]
public class ProgressEventArgs : EventArgs
{
        public string Message;
        public int Value;
        public ProgressEventArgs(string message, int value)
        {
               this.Message = message;
               this.Value = value;
        }
}

This is the standard way to make Events. I have seen people making custom delegates and taking all the required information as parameters. That’s not the prescribed way. The suggested way is to make delegates similar to EventHandler delegate and make your own class extend the EventArgs which contains all the event parameters. This makes the design consistent and other developers can pick up the concept easily as it is something closer to what they already use.

Tips 4. Updating UI from non-UI threads the easy and safe way

One of the common mistakes we all do is that we try to update the UI from background threads or self created threads than using the main UI thread and thus make our apps crash frequently. While designing multithreaded applications, you need to ensure that you are in no way using any resource on the UI from any other thread other than the main thread. Normally the common mistake we all make is, we breakup the code of another thread in different functions and on those functions we forget about the fact that it is running in another thread. As a result, the app crashes frequently whenever we try to access the UI elements.

In order to update UI from a function called from or running in another thread, you need to use the Invoke or BeginInvoke method of the Control or Form. Here is my favorite way to do this:

private void UpdateWebLogProgress( object sender,
                                 ProgressEventArgs e )
{
    if( base.Disposing ) return;

    if( base.InvokeRequired )
    {
       this.Invoke( new ProgressEventHandler(
                   this.UpdateWebLogProgress ),
                       new object[] { sender, e } );
       return

    }

    // Access UI elements here, no problemo.

}

The function which accesses UI element checks itself whether it requires invoke. If it requires invoke, it calls itself using a delegate via the Invoke method. If it does not require invoke, it executes the code which handles the UI elements. Thus it guarantees to prevent the problem of accessing UI elements from another thread. It also saves code because you need not Invoke it explicitly from all the places.

this.Invoke( new ProgressEventHandler(this.UpdateWebLogProgress),
    new object[] {this, new ProgressEventArgs(“Complete”, 100)});

Instead you need to do the following from any thread:

UpdateWebLogProgress( this, new ProgressEventArgs( “Complete”, 100 ) );

Thus it frees the caller from remembering the function that works with UI elements.

RSS Auto Discovery

An interesting feature of RSS Aggregators is to detect the presence of RSS feed or RSS Feed location from any URL. There are different scenarios you have to consider while implementing auto discovery:

  • The URL can itself be an RSS Feed, not an HTML.
  • The URL can be a HTML page having <link rel="alternate" type="application/rss+xml"> which refers to a RSS source.
  • The HTML page may contain hyperlinks to RSS Feed sources.

RSS Feeder can intelligently identify what a URL contains and take actions accordingly. For example if you provide the URL of an HTML page, it will find all the RSS feed sources specified inside the HTML. So, if you provide the URL msdn.microsoft.com it will automatically detect all the RSS Channels specified in that page.

On the other hand, if you give it a URL containing some type of feed, it will automatically detect the type of feed.

The HTML processing is done by the famous SgmlReader. Here’s is the code which discovers the type of content that is found after downloading data from a URL:

private static RssTypeEnum Discover( MemoryStream memoryStream,
        ref IList channelSources, ProgressEventHandler progressHandler )
{
    // Parse the content and find out whether the content is HTML or XML


    Sgml.SgmlReader reader = new Sgml.SgmlReader();
    bool isHtml = false;
    RssTypeEnum feedType = RssTypeEnum.Unknown;

    memoryStream.Position = 0;

    /// Parse the response using a SGML parser in order to identify whether

    /// the response is HTML or XML.


    reader.InputStream = new StreamReader( memoryStream );

    try
    {
       while( reader.Read() )
       {
           if( null != reader.Name )
           {
               string name = reader.Name.Trim().ToLower();

               /// If we get the html tag in the document,


               /// then it is definitely a HTML page

               if( name == "html" || name == "body" )
               {
                  isHtml = true;
                  feedType = RssTypeEnum.Unknown;
               }
               // RSS XML starts from <rss> node


               else if( (name == "rss" || name == "rdf:rdf")
                                          && !isHtml )
               {
                  feedType = RssTypeEnum.RSS;

                  // We need to parse in a different


                  // way for RSS stream

                  break;
               }
               // Atom XML starts from <feed> node

               else if( name == "feed" && !isHtml )
               {
                  feedType = RssTypeEnum.Atom;

                  // We need to parse in a different


                  // way for RSS stream

                  break;
               }
               else if( isHtml && name == "link" )
               {
                  // Let's see if this link indicats a RSS feed


                  string rel = reader.GetAttribute("rel");
                  string type = reader.GetAttribute("type");
                  string title = reader.GetAttribute("title");
                  string href = reader.GetAttribute("href");

                  if( rel.ToLower() == "alternate"

                     && type.ToLower() == "application/rss+xml")
                  {
                     channelSources.Add(new string [] { title, href });
                  }
                  if( rel.ToLower() == "alternate"
                    && type.ToLower() == "application/atom+xml")
                  {
                     channelSources.Add(new string [] { title, href });
                  }
               }
           }
       } // while( reader.Read() )


    }
    catch( Exception x )
    {
       EntLibHelper.Exception( "RSS Discovery", x );

       // RSS discovery failed. SgmlReader can't parse it.

       // May be this really is an RSS feed which .NET's framework

       // can parse better.


       feedType = RssTypeEnum.RSS;
    }

    return feedType;
}

The code reads the node one by one until it discovers either html, rss, rdf or feed node. If it reaches the html node, it assumes that this is an HTML document and looks for the link node which specifies the reference to RSS or Atom feed source. If it finds such sources, it makes a collection of those sources and returns the collection.

XSL transformation of RSS feeds

There are two types of XSL transformation of RSS XML in this program. One is to render a single post and another one is to render a newspaper which contains multiple channels and multiple feeds.

Rendering a single post

Believe it or not, you cannot transform an XML to HTML properly using XSL if it contains embedded HTML content inside the XML nodes. .NET framework’s built in XSL transformer cannot decode the HTML content inside the nodes. Although the XSL specification says you can do this:

<xsl:value-of disable-output-escaping="yes" select="description" />

This is used to prevent XSL transformer from escaping the content of the description node and to render as it is, but it does not work. The XSL transformer is going to encode the HTML content no matter what happens. As a result, instead of getting a nicely formatted HTML output, you see the HTML codes.

Due to this problem, I created the HtmlReader and HtmlWriter classes. Details about this can be found in my HTML Cleaner article.

HtmlWriter extends XmlWriter and prevents escaping of text content inside the tags. It overrides the WriteString method and invokes WriteRaw from it.

public override void WriteString(string text)
{
    // Change all non-breaking space to normal space


    text = text.Replace( " ", " " );
    /// When you are reading RSS feed and writing

    /// Html, this line helps remove


    /// those CDATA tags

    text = text.Replace("<![CDATA[","");
    text = text.Replace("]]>", "");

    if( this.FilterOutput )
    {
       // We want to replace consecutive spaces to


       // one space in order to save horizontal

       // width

       if(this.ReduceConsecutiveSpace) text =
             text.Replace("   ", " ");
       if(this.RemoveNewlines) text =
           text.Replace(Environment.NewLine, " ");
       base.WriteRaw( text );
    }
    else

    {
           base.WriteRaw( text );
    }
}

This class eliminates the need for having the disable-output-escaping attribute set.

Rendering the newspaper – multiple post from multiple channels

This is trickier than rendering a single post as you have to render multiple channels in one HTML. The process is as follows:

  • Combine all the channels into one XML which contains one RSS node but multiple channel nodes.
  • Transform the combined XML using XSLT in order to render HTML.

The first problem I encountered was combining multiple XML using XmlWriter. XmlWriter’s WriteRaw method is supposed to write everything I give it without doing any encoding. But it does not do that, instead it encodes any XML given as a node’s value. As a result, I cannot write XML of an RSS entry using XmlWriter. In order to do this, I need to first create a StreamWriter and then wrap that using XmlTextWriter. Regular XML construction is done by the XmlTextWriter, but whenever raw XML needs to be written, I call the StreamWriter’s WriteLine method.

So, the code for generating one channel is like this:

// Write the channel header

writer.WriteStartElement("channel");
writer.WriteElementString( "title",
                      channel.Title );
writer.WriteElementString( "link",
         channel.FeedURL.ToString() );

IList items =
  DatabaseHelper.GetTopRssItems(channel.Id,
            (int)itemCountUpDown.Value, false);

foreach( RssFeed feed in items )
{
        streamWriter.WriteLine( feed.XML );
}

writer.WriteEndElement();

Single instance of an application

This is a common requirement for many desktop applications where you need only one instance of your app to be allowed to run.

Doing this requires a bit of Win32 level knowledge. When an app is created, it registers a Mutex. But before creating it first checks whether there is any Mutex already registered with the same name. If there is, then another instance of application is already running and so we quit.

public static bool IsAlreadyRunning()
{
    string strLoc =
         Assembly.GetExecutingAssembly().Location;
    FileSystemInfo fileInfo = new FileInfo(strLoc);
    string sExeName = fileInfo.Name;
    bool bCreatedNew;

    mutex = new Mutex(true,
        "Global\\"+sExeName, out bCreatedNew);
    if (bCreatedNew)
           mutex.ReleaseMutex();

    return !bCreatedNew;
}

You will find the solution in this article:

Making responsive UI using timers

Suppose you have a list of items. Every time a user selects an item, you need to perform some database operation and generate HTML which takes time. You have written this code in the SelectedIndexChanged event of the list. So if the user selects an item by clicking it and then holds down the DOWN arrow for a while, the event is going to be fired repeatedly which results in frequent database calls and HTML rendering. As those operations take time, the selection pointer does not smoothly go down, instead it gets stuck on every item for a while making the whole UI unresponsive until all the method executions are finished.

Another scenario is, if the user first clicks an item and then scrolls down using a mouse and holds SHIFT and clicks another item to make multiple selections, the SelectedIndexChanged event of the list is going to be fired for each and every item that falls within the selection range. As a result, the user is going to be stuck for a while until all the events are fired and the code execution for all the events is finished.

In order to solve these problems, we will use a timer to invoke the method which does the actual work instead of calling it directly. The process is as follows:

  • On SelectedIndexChanged event of the list, we first stop the timer in order to initialize its state. Then we start the timer with an interval.
  • We set a delegate to the method which needs to be called.
  • On the Elapsed event, we call the delegate, stop the timer and clear the delegate.

As a result, no matter how many times the SelectedIndexChanged event of the list is fired, it just sets a timer which in turn queues a method invoke. As we use a delegate to point to a method instead of an event, the delegate is called only once which in turn calls the actual method just once.

Here’s how we set the timer:

this._CallbackMethod =
    new MethodInvoker( this.AutoSelectChannel );
this.callbackTimer.Interval =
    STARTUP_SHOW_CHANNEL_DURATION;
this.callbackTimer.Start();

On the Elapsed event, we have this code:

private void callbackTimer_Elapsed(object sender,
                     System.Timers.ElapsedEventArgs e)
{
    callbackTimer.Stop();

    if( null != this._CallbackMethod )
           this._CallbackMethod();

    this._CallbackMethod = null;

}

Reducing memory footprint of .NET applications

.NET Winforms applications are memory hog apps. An innocent one button, one form application takes about 15 to 20 MB memory when it loads, whereas a VB 6 counter part takes just 15 to 20 KB.

However, if you open Task Manager while RSS Feeder is in the tray and no window is visible, you will notice that it takes only around 3 to 5 MB memory.

This is done by reducing the WorkingSet of the Process of the app. WorkingSet is a property of the Process class which contains how much memory the process is consuming. The interesting thing is that, it is not a read-only variable; you can increase or decrease it. If you try to decrease it, you will notice the memory usage goes down significantly. Although it does not work always, we can give it a try whenever we want it.

public static void ReduceMemory()
{
    try

    {
       System.Diagnostics.Process loProcess =
           System.Diagnostics.Process.GetCurrentProcess();
       if(m_TipAction == true)
       {
           loProcess.MaxWorkingSet =
               (IntPtr)((int)loProcess.MaxWorkingSet - 1);
           loProcess.MinWorkingSet =
               (IntPtr)((int)loProcess.MinWorkingSet - 1);
       }
       else

       {
           loProcess.MaxWorkingSet =
               (IntPtr)((int)loProcess.MaxWorkingSet + 1);
           loProcess.MinWorkingSet =
               (IntPtr)((int)loProcess.MinWorkingSet + 1);
       }
       m_TipAction = !m_TipAction;
    }
    catch( Exception x )
    {
       System.Diagnostics.Debug.WriteLine( x );
    }
}

In the source code, you will find a MemoryHelper class which does this work. You can call the ReduceMemory method whenever your app is deactivated or minimized leaving some memory for others to use.

Register application on startup

We all know how to do it, add a key in the registry and that’s all. But putting an application at startup, specially a .NET application at startup is a serious pain for Windows. Normally if you try to load a Winforms application for the first time, you will see how slow it is and how much hard disk activity it consumes. Putting such applications at startup when Windows is doing a lot of other tasks is going to make the boot-up much slower and make the user lose his patience.

That’s why, I have made an application loader in VB 6 which takes only 20 KB and loads in the blink of an eye. When the RSS Feeder is set to load at startup by the user, it actually puts the loader at startup with a 5 minutes delay. The loader immediately goes to sleep for 5 minutes after loading, allowing Windows to breathe freely for a while and wipe out sweats after heavy boot-up pushups. When everything is nice and quiet, it starts the jumbo .NET app.

Here’s the VB 6 code of the loader:

Private Declare Sub Sleep Lib "kernel32" (_
                    ByVal dwMilliseconds As Long)


Sub Main()

    On Error GoTo ExitDoor

    ' The first command line argument

    ' specifies the delay and the


    ' second argument specifies the exe to launch

    Dim strArguments() As String
    strArguments = Split(Command$, " ")

    Dim lngDelay As Long
    lngDelay = Val(strArguments(0))

    Dim strFile As String

    strFile = Trim(strArguments(1))

    ' Pause for the amount of time

    Call Sleep(lngDelay * 1000)

    ' Run the program


    Shell strFile

ExitDoor:


End Sub

Notify tray icon

RSS Feeder normally stays at the system tray as an icon. So, we need to use Window 2000/XP/2003’s Balloon notification feature in order to provide information about its activities.

This is done by the wonderful article in CodeProject.

Error reporting

As the RSS Feeder runs all over the world without letting me know anything about it, I need a way to know whether my users are happily using it or not. So, I have implemented a silent crash report module which traps any unhandled exceptions and reports back to me via Source Forge tracking system.

Error reporting works in the following way:

  • Whenever an unhandled exception occurs, the error report is queued to an Error Queue.
  • The queue is frequently monitored by an Error Reporting Engine. Whenever it picks up an error from the queue, it launches a background thread.
  • The background thread prepares an error report collecting some context information like the current login name of the user which helps to identify duplicate reports from the same user.
  • The message is posted as HTTP POST to Source Forge Tracker System.

The error reporting engine uses a timer to periodically check the queue.

private void errorReportTimer_Tick(object sender,
                               System.EventArgs e)
{
    if( base.Disposing ) return;

    // If no reports in queue, close


    if( 0 == _ErrorReports.Count )
    {
           this.errorReportTimer.Stop();
           _Instance = null;
           base.Close();
    }
    else
    {
       if( !this._IsErrorBeingSent )
       {
           this._IsErrorBeingSent = true;

           base.Show();
           base.Refresh();

           // Get the first error report item


           ErrorReportItem item =
               _ErrorReports[0] as ErrorReportItem;
           _Instance.errorReportTextBox.Text = item.Details;

           ThreadPool.QueueUserWorkItem( new
                  WaitCallback( SendQueuedReport ), item );
       }
    }
}

The error report is sent from a background thread so that the UI remains responsive.

private void SendQueuedReport(object state)
{
    if( base.Disposing ) return;

    ErrorReportItem item = state as ErrorReportItem;
    try

    {
       // Send the report

       ErrorReportHelper.PostTrackerItem(item.Summary,
                                        item.Details);
    }
    catch
    {
           // Who cares

    }

    lock( _ErrorReports )
    {
           _ErrorReports.Remove( state );
    }

    this._IsErrorBeingSent = false;
}

The actual marvel is in the ErrorReportHelper class which performs the task of making an HTTP POST to the Source forge Tracker System.

public static void HttpPost( string url,
  string referer, params string [] formVariables )
{
    HttpWebRequest req = null;
    try

    {
       req = (HttpWebRequest)
               HttpWebRequest.Create(new Uri(url));
    }
    catch
    {
           return;
    }

    CookieContainer CookieCont= new CookieContainer();

    req.AllowAutoRedirect = true;
    req.UserAgent = "Mozilla/4.0 (compatible; " +
        "MSIE 6.0; Windows NT 5.0; .NET CLR 1.1.4322)";
    req.CookieContainer = CookieCont;

    req.Referer = referer;

    req.Accept = "image/gif, image/x-xbitmap,
              image/jpeg, image/pjpeg,
              application/msword,
              application/vnd.ms-powerpoint,
              application/vnd.ms-excel, */*";

    req.Method = "POST";
    req.ContentType =
        "application/x-www-form-urlencoded";

Now we have prepared a HttpWebRequest object to perform an HTTP POST. Next we need to prepare the body which contains the actual content that needs to be posted.

StringBuilder postData = new StringBuilder();

for( int i = 0; i < formVariables.Length; i += 2 )
{
  postData.AppendFormat("{0}={1}&",
        HttpUtility.UrlEncode(formVariables[i]),
        HttpUtility.UrlEncode(formVariables[i+1]));
}
postData.Remove(postData.Length - 1, 1);


byte[] postDataBytes =
    Encoding.UTF8.GetBytes(postData.ToString());

req.ContentLength = postDataBytes.Length;

Our next task is to connect to the actual tracker system index.php and transmit the data.

Stream postDataStream = req.GetRequestStream();
postDataStream.Write(postDataBytes, 0,
                      postDataBytes.Length);
postDataStream.Close();

In order to make a successful HTTP dialog, we do need to get the response a bit, but not the whole response. Source forge’s pages are pretty heavy. Do not think of downloading a full page. Besides the page that is returned contains all the posted items and there may be thousands of error reports!

    // Get the response stream in order


    // to complete a full HTTP POST

    HttpWebResponse resp = null;
    try
    {
       resp = (HttpWebResponse) req.GetResponse();
    }
    catch
    {
       return;
    }

    // We are not interested to read anything


    Stream rcvStream = resp.GetResponseStream();

    resp.Close();
    rcvStream.Close();
}

Trick 5. Centralized error handling

There is a very convenient way to catch any exception that someone mistakenly skipped putting inside the try catch block. The Application class provides an event ThreadException which traps any unhandled exception that was not caught by any catch block. When you forget to catch any exception, you see the shameful “Continue” or “Quit” dialog box with detailed exception trace exposing all your mistakes. If you catch the ThreadException you can capture any leaked exception, and can remain quiet or silently write to a log file in order to pick it up later on.

RSS Feeder hides its shame by subscribing to two events:

AppDomain.CurrentDomain.UnhandledException +=
    new UnhandledExceptionEventHandler(
           CurrentDomain_UnhandledException);
Application.ThreadException += new
    System.Threading.ThreadExceptionEventHandler(
                       Application_ThreadException);

These exceptions are quietly handled by my EntLibHelper which exposes a nice method to handle unhandled exceptions.

private static void Application_ThreadException(object sender,
    System.Threading.ThreadExceptionEventArgs e)
{
    EntLibHelper.UnhandledException(e.Exception);
}


private static void CurrentDomain_UnhandledException(object sender,
                                     UnhandledExceptionEventArgs e)
{
  if( e.ExceptionObject is Exception )
    EntLibHelper.UnhandledException(e.ExceptionObject as Exception);
}

EntLibHelper consults the configuration file to decide what to do with the unhandled exception. You can specify in the configuration file what to do with exceptions in the unhandled exception category. If you specify to throw it again, it shuts down the application. But if you specify to trap it, it sends the exception to the Error Reporting engine for sending it to a tracking system. As a result, you will see the error reporting window pops up which notifies me about the error. This is a zero hassle for both you and me.

internal static void UnhandledException(Exception x)
{
    // An unhandled exception occured

    // somewhere in our application. Let


    // the 'Global Policy' handler have

    // a try at handling it.

    try
    {
       bool rethrow = ExceptionPolicy.HandleException(x,
                                 "Unhandled Exception");
       if (rethrow)
       {
         // Something has gone very


         // wrong - exit the application.

         System.Windows.Forms.Application.Exit();
       }
       else
       {
         ErrorReportForm.QueueErrorReport(
           new ErrorReportItem( x.Message, x.ToString() ) );
       }
    }
...
...

Blogging features

RSS Feeder is not only an RSS aggregator but also an equally powerful blogging tool. It gives you a nice Outlook 2003 style UI to work with multiple weblogs.

You can create one or more weblog accounts, write posts and save for publishing it later on. When you send a post for publishing, it goes to a send queue which is collected by the send/receive module. If there is any problem in posting, you can come back and open the post to see the error report.

XMLRPC

Most of the widely used blog engines use XML RPC specially those which are developed using PHP. Some of them are WordPress, B2Evolution, Drupal etc. XML RPC is similar to SOAP where a method call and parameters are serialized to an XML. You can learn more about XML RPC by Googling on XMLRPC and also from this site.

Most of the Blog engines support a well defined API called MetaWeblogAPI. This API uses XMLRPC to call a fixed set of web methods. Although the API is fixed, the method names differ from one blog engine to another. As a result, you cannot really write a generic code for MetaWeblogAPI that works on all blog engines which support this API. I believe there is an urgent need to standardize a universally accepted list of methods and parameters or in short a fixed interface for all blog engines. For example, if you are blogging to B2Evolution, you need to use the following methods:

  • b2.getCategories: Get the categories for a blog.
  • b2.newPost: Make a new post.

On the other hand, for WordPress which supports the same MetaWeblogAPI, the method names are:

  • metaWeblog.getCategories
  • metaWeblog.newPost

This makes life difficult to implement a blogging library that works universally for all. However, this has been done pretty well in RSS Feeder. Currently it supports almost all PHP based blog engines and famous .NET based blog engines like .Text and Community Server.

Editing environment

The editing environment has been made as rich as possible while maintaining the restrictions the blog engines apply to formatting. Not all HTML formatting are supported by blog engines. Most of them strip of style attribute and style sheets. Some don't allow the <font> tag. Some prevent <div> inside the <p> tag. So, the editor that I have made more or less accommodates sufficient formatting tools that suites all.

Tips 6. Convenient Toolbar handling

When you have a big toolbar containing 30 to 40 buttons, it becomes difficult to write code for handling click on each of these buttons. You end up either writing click handlers for each button or one toolbar level handler with a big if else or switch case to determine which button was clicked.

Instead of this, I have tried a convenient approach. Each button contains a method name in the tag of a button. For example, the Save button has a tag value set to $SavePost. The user control has a public method named SavePost. Inside the click handler of the toolbar, I have this code:

private void DesignSandBar_ButtonClick(object sender,
                     TD.SandBar.ToolBarItemEventArgs e)
{
    // Get the tag which contains the


    // function name to be invoked

    string name = e.Item.Tag as string;

    if( null == name || 0 == name.Length ) return;

    // if self call, call the function on self


    if( name.StartsWith("$") )
    {
      MethodInfo method =
         this.GetType().GetMethod(name.TrimStart('$'));
      method.Invoke( this, null );
    }
    else

    {
       // The function is on the editor component

       MethodInfo [] methods =
             this._Editor.GetType().GetMethods( );
       // Call the method which has no parameter

       foreach( MethodInfo method in methods )
           if( method.Name == name &&

                   0 == method.GetParameters().Length )
           {
               try
               {
                   method.Invoke( this._Editor, null );
               }
               catch( Exception x )
               {
                   Debug.WriteLine( x );
               }
           }
    }
}

It uses reflection to find the method on the user control which matches the name specified in the tag and invokes the method. So, you don’t need to write the click handler or switch blocks to decide which button is clicked and for that which method needs to be called.

Moreover, if the Tag does not start with a $ sign, it means the method needs to be invoked on the DHTML editor control where the user writes the body of the post. The DHTML editor control exposes lots of formatting methods which are thus directly mapped to the toolbar buttons.

DHTML editor control

The editor you see is the DHTML editor control that comes with Internet Explorer. This is the control that Microsoft uses in most of their products including the HTML Designer Window of Visual Studio. You will find the control in this path:

C:\Program Files\Common Files\Microsoft Shared\Triedit\dhtmled.ocx

Working with this control is pretty difficult because almost all the functionality that it provides is via a method named ExecCommand. You can find a reference to this method at MSDN. This function lets you control the editor’s behavior, change formatting style, show common dialog boxes like Find, Link, Image etc. For example, you can use the following command to set the editor to Bold mode:

Object nullArg = null;
editorControl.ExecCommand( DHTMLEDITCMDID.DECMD_BOLD,
   OLECMDEXECOPT.OLECMDEXECOPT_DODEFAULT, ref nullArg );

As working with this control requires you to memorize all the function names and parameters, a convenient wrapper is necessary. That’s what you have in the HtmlEditor control provided in the source code. It exposed almost all the functionalities of DHTML editor as convenient public methods and properties like SetBold, SelectedText, OrderList etc. As the wrapper is a .NET class which extends the DHTML editor control, you can easily create a new instance of it and put it inside a container. Here’s the code sample which shows you how to create a new editor and put it inside a Panel:

private void CreateEditor()
{
    _Editor = new HtmlEditor();
    _Editor.Dock = DockStyle.Fill;
    _Editor.HandleCreated +=
         new EventHandler(_Editor_HandleCreated);
    editorPanel.Controls.Add( _Editor );
}

One important issue you need to notice is that, you can only work with this component when the handle is created, which means that IsHandleCreated property is true. If you want to work with this component on the Load event of a Control or Form then it will fail. The editor does not load properly when the Load event is fired. It for some reason takes a long time to get initialized properly. So, any task that you need to perform on the load needs to be performed on the HandleCreated event of the control. Also keep in mind to always check the IsHandleCreated property before using this control. For example:

if( this._Editor.IsHandleCreated )
{
        this._Editor.NewDocument();
}

UI

You must be wondering how I made the UI. It’s developed using the marvelous free controls found at Divelements Ltd. I am a great fan of their tools and I have also used it in my other project Smart UML.

Conclusion

I have tried to make RSS Feeder an ideal Smart Client application by implementing all the best practices that I have been able to collect from the web. This application is a good example of the use of Enterprise Library in a real application. It also uses the Application Updater Block 2.0 which was a real pain to implement. Hope you will find this article a good source of essentials of desktop application development. Although it’s been only a month since this app was developed and released, I hope it’s pretty much stable and feature-rich to be a part of your daily life. I will look forward to get feedback from you which will guide me to improve this in future and make it the best tool for feed aggregation and blogging.

License

This article has no explicit license attached to it but may contain usage terms in the article text or the download files themselves. If in doubt please contact the author via the discussion board below.

A list of licenses authors might use can be found here